1 Introduction

Image analysis is a universal need in image processing and computer vision. However, lack of high-frequency details in an image leads to incorrect image analysis. To overcome such issues, super resolution (SR) technique has been used to generate a high resolution (HR) image from the given low resolution (LR) image(s) observations [23]. Prior to SR algorithm, another possible approach was adopted to increase the number of pixels per unit area to combat the loss of high-frequency details. Such process considers the decrease in pixel size and increases the number of sensors [36,37,38]. However, due to diffraction effect decreasing the pixel size prompts to high capacitance which in turn slows down the system performance and results in increase of noise. SR algorithm is devoid of such limitations and found to be suitable in applications like satellite and areal imaging, medical image processing, textured recognition, face recognition, video surveillance, etc.

Various research articles have been reported on SR techniques over the past two decades. Generally, the SR can be achieved on both frequency as well as spatial domain [2, 18, 30, 35, 45]. Furthermore, the SR methods can be broadly categorized into two types such as (i) Multi frame SR and (ii) Single frame SR based on the number of LR reference images considered. Various multi image based SR approaches [21, 28, 38, 41] require accurate registration among images which lead to ringing effects and artifacts in the reconstructed image. In addition, they do not work well for unprocessed data. On the other hand, single image LR image based SR methods do not suffer from these problems. Further, existing single image based SR approaches can be classified into three general classes, i.e., interpolation based, reconstruction based and learning based. In the interpolation based methods [29, 51], it is hard to restore the high-frequency details. Therefore, blurred edge and complex texture are found in interpolation method. In reconstruction based method [30, 54], prior knowledge like edge, gradient, etc. are required to reconstruct the HR image. This method can suppress the aliasing artifacts and preserve sharper edges. However, it cannot generate fine details in larger magnification. To tackle this problem, learning based SR methods [5, 7, 9, 20, 52] have been introduced by inferring the mapping relationship between LR-HR image pairs.

Learning based SR approach can be further categorized into three categories such as belief network [3, 19, 20], manifold learning [6, 7, 9, 13] and compressive sensing [46, 47, 50]. Freeman et al. [19] are the forerunners who introduced learning based single image super resolution scheme in 1999. Later on, they have used Markov random field (MRF) model and loopy belief propagation to find the relationship between LR-HR image patch pairs. However, boundary effect and incompatibility between reconstructed HR patches and its surrounding patches have been seen in texture images [19, 20]. Based on these works, Sun et al. [40] proposed primal sketch apriori to overcome the blur present in edges, ridges and corners. To overcome the incompatibility between HR patches and its neighboring patches, manifold learning based super resolution techniques have been introduced in [4, 5, 7, 9]. Chang et al. [9] introduced single image SR using manifold learning where local geometry based features are extended from LR-HR patch pairs inferred to reconstruct the expected HR patches using optimal reconstruction weight. Thereafter, a novel feature selection scheme for neighbor embedding SR has been proposed by Chan et al. [7]. Guo et al. [24] have used maximum a posteriori (MAP) estimation to combine the global parametric constraint with a patch-based local non-parametric constraint. In another work, Gao et al. [22] have proposed a method to project the original HR and LR patch onto the jointly learning unified feature subspace. Later on, compressive sensing techniques have drawn attention from many researchers for single image super resolution [16, 26]. Mishra et al. [33] have introduced a robust scheme to generate a HR image from the registered LR image. In this work,they remove the outlier generated during registration using robust locally linear embedding (RLLE). Further, they used Zernike moment as feature selection for neighborhood preservation during HR image reconstruction [34].

In addition to these, few other studies have also proposed for the enhancement of texture image. Wei et al. [44] have discussed the extensions and applications of texture synthesis. HaCohen et al. [25] introduced consistent fine-scale detail to textured regions and produced edges with proper sharpness. To enhance the quality of complicated texture structure in the image, an example based SR using clustering and partially supervised neighbor embedding has been proposed by Zhang et al. [52]. Due to overlapping in the frequency range with that of the noise in the texture images, de-noising or in-painting is used for single image super resolution [49]. Yoo et al. [48] proposed a novel texture enhancement strategy to improve the single image super resolution performance. In recent years, many neural network (NN) based approaches has addressed for single image super resolution with excellent performance. Ahn et al. [1] proposed a texture enhancement framework by the combining interpolation technique and customized style transfer technique. Here they used repetitive tiling concept for the style transfer technique. Sajjadi et al. [39] introduced an automated texture synthesis technique by using feed forward fully convolutional neural networks. This method gives an efficient result due to the feed-forward architecture which inference time since the LR image only needs to be passed through the network once to get the result. Liu et al. [31] used total variation (TV) regularization approach for SR image reconstruction for synthetic aperture radar (SAR) image. In this work, they used gradient profile prior or other texture image prior in the maximum a posteriori framework. Cruz et al. [15] used Wiener filter in similarity domain to generate a super resolution image from a single reference image. Here they introduced filtering of the patches in 1D similarity domain and coupled it with the iterative back propagation (IBP) frame work. Chang et al. [11] presented a bi-level semantic representation analysing framework for image/ video which is work for weight learning approach for multimedia representation. Later, they introduced a semi supervised feature selection technique. For maintain the correlation between various feature in the training data they adopted manifold learning to labeled and unlabelled training data [8]. In addition, to maintain the correlation between the linear effect and the nonlinear effect for embedding feature interaction, an optimization algorithm has been proposed by them [10]. An isotonic regularizer that is able to exploit the constructed semantic ordering information has introduced by Chang et al. [12] for the video shots.

The literature review reveals that most of the existing schemes fail to reproduce the fine details during image reconstruction. Different applications on SR need suitable algorithms in order to achieve an optimal reconstruction image. Enhancement in texture images still points as an outstanding problem. In this work, we propose an improved learning based SR algorithm for texture image to generate a HR image from a single LR image. The scheme utilizes neighbor embedding technique (manifold learning) and is suitably named as single image super resolution using manifold learning (SSRM). In this approach, HR patches are reconstructed from the input test LR patches by the prior information that fetched from the training LR-HR pairs. The major contribution of this paper includes

  1. 1.

    Feature representation using pseudo Zernike moments (PZM) to classify the patches

  2. 2.

    Finding the k most similar patches from the test patch using similarity measure, i.e. Feature similarity index (FSIM)

  3. 3.

    Computation of optimal reconstruction weight using least square (LS) and non-negative matrix factorization (NMF) for HR patch reconstruction

Experimental results and comparisons on textural images show that proposed method offers better performance than other existing methods. Additionally, the proposed scheme works well on standard images.

The rest of the paper is organized as follows. Section 2 discusses the proposed algorithm. The experimental results and discussion are presented in Section 3. Finally, Section 4 deals with the concluding remarks and the scope for future work.

2 Proposed SSRM algorithm

In this work, an improved single image SR using manifold learning is proposed. The scheme follows three main steps

  1. (i)

    feature extraction using pseudo Zernike moments (PZM)

  2. (ii)

    finding k-nearest neighbor using similarity measure FSIM

  3. (iii)

    optimal reconstruction weight computation using collaborative method

Features are extracted from each LR-HR patch separately by considering the phase and magnitude values of PZM. In addition, FSIM has been used as a similarity measure for searching the nearest neighbor patches for HR patch reconstruction. Thereafter, a collaborative technique is used to find out the optimal reconstruction weight to generate the HR patch from a given LR input patch. Figure 1 shows the overall frame work of the proposed scheme.

Fig. 1
figure 1

Overall block diagram of proposed scheme

The overall steps to generate the reconstructed HR image is listed in Algorithm 1.

figure f

2.1 Patch generation

In this approach, each image is first converted from RGB to YCbCr color model and only the luminance component of the image is considered. Each LR image is decomposed into a set of t × t patches with two overlapping pixels. The corresponding HR image patch size is γt × γt with 2γ overlapping pixels, where γ is the magnification factor. Let the training LR image, HR image and testing LR image are denoted by \(L_{r}=\left \{{l_{r}^{p}} \right \}_{p = 1}^{a}, H_{r}=\left \{{h_{r}^{p}} \right \}_{p = 1}^{a}\), and \(L_{t}= \left \{{l_{t}^{q}} \right \}_{q = 1}^{b}\), respectively. The number of patches for the LR image of size t × t is t × t × [(rowstepsize) × (colstepsize)], where stepsize = patchsizeoverlappingpixels. The number of patches for the HR image of size γt × γt is γt × γt × [(rowstepsize) × (colstepsize)], where stepsize = γ × patchsizeγ × overlappingpixels.

2.2 Feature selection by PZM

In this work, we use pseudo Zernike moment (PZM) for feature selection. It has higher description ability for same order as compared to Zernike moment. Moreover, it provides meaningful information for image reconstruction. It is a widely used orthogonal moment. It has several applications in various areas of image processing and pattern recognition. Because of its orthogonality and invariance properties, it is robust to noise and have better ability for image representation [14]. Figure 2 shows the linear transformation on a single patch through normalization.

Fig. 2
figure 2

Normalization of the patch in unit circle

The PZM of a continous image intensity f(x, y) is defined as

$$ PZM_{n,m} \left( f\left( {x,y} \right)\right) = \frac{{n + 1}}{\pi }\int {\int}_{{{x^{2}} + {y^{2}}} \le 1} f\left( {x,y} \right) V_{n,m}^{\ast} \left( {x,y} \right)dxdy, $$
(1)

where n, m is the order and repetition respectively. Here, \(n = 0,1,2,...,\infty \) and |m|≤ n such that n −|m| = even. The discrete form of PZM for digital image intensity f(xq,yp) is represented as

$$ PZM_{n,m}\left( f\left( {x_{q},y_{p}} \right)\right) = \frac{{n + 1}}{\pi \lambda(N) }\sum\limits_{x}\sum\limits_{y} V_{n,m}^ \ast \left( {x,y} \right) \left( f\left( {x_{q},y_{p}} \right)\right), $$
(2)

where \({x_{q}}=-\frac {\sqrt {2}}{2}+\frac {\sqrt {2}}{N-1}q, q = 0,1,..,(N-1)\) and \({y_{p}}=\frac {\sqrt {2}}{2}-\frac {\sqrt {2}}{N-1}p, p = 0,1,..,(N-1)\). Here \(\frac {{n + 1}}{\pi }\) is a normalization factor, \(\lambda (N)=\frac {N^{2}}{2}\). Vn, m is the complex polynomial and Vn, m∗ is the complex conjugate of Vn, m.

The pseudo Zernike polynomials is represented as

$$ V_{n,m} \left( {x,y} \right) = R_{n,m} \left( r \right)e^{jm\theta}, $$
(3)

where radial polynomials \(R_{n,m} \left (r \right )\) is defined as

$$ R_{n,m} \left( r \right) = \sum\limits_{s = 0}^{{n - \left| m \right|}} (-1)^{s}{\frac{(2n + 1-s)!}{s!(n-\left| m \right|-s)!(n+\left| m \right|+ 1-s)!}}r^{(n-s)}, $$
(4)

here the PZM are defined in terms of polar coordinates (r, 𝜃) with \(0 \le \left | m \right |\le n \) and \(\left | r \right | \le 1\).

It is reported that PZM is unsuitable to extract the global information from the whole image. Hence, in this scheme, we compute the feature vector for each patch with respect to magnitude and phase of PZM [27].

2.2.1 Generation of feature vector considering the magnitude of the PZM

Generally, the 2-D PZM feature is represented by a complex value, which is the combination of magnitude and phase coefficients. The magnitude based PZM feature vector of order n = 0to nmax, with m ≥ 0 is represented in Table 1. For nmax = 5, the total number of feature vector is 21. Since, PZM with higher order moment is sensitive to noise and distortion, we consider 5 as the maximum order. So the feature vector for each patch f(xq,yp)is defined as

$$ \stackrel{n_{max}}{PZM}\left[f(x_{q},y_{p})\right]=\left\{|PZM_{u,v}(f(x_{q}, y_{q}))|\vert, u = 0,1,...,{n_{\max },v = o,i,...,u} \right\}, $$
(5)

where |PZMu, v(f(xq,yq))| is the magnitude of PZMn, m(f(xq,yp)).

Table 1 PZM with n = 0 to 5

2.2.2 Generation of feature vector considering the phase of the PZM

The wide range of edges are represented by high phase values. The gradient based edge detection technique are sensitive to blurring and magnification. Hence, in this work, PZM is used to preserve the high-frequency details during image reconstruction. By using various heuristic criteria, the patch with edge and without edge can be calculated from the training patches. Hence, for a input test patches, it will be easy to classify the nearest neighbor patches from the whole training patch. Then each patch having edge is counted according to Table 2. It will speed-up the searching process. In order to compute the phase feature, rotation invariant PZM is defined as

$$ \left|PZM^{R}_{n,m}(f({x_{q}},{y_{p}}))\right| = \left| {PZM_{n,m}(f({x_{q}},{y_{p}}))} \right|{\exp^{-jm{\theta}}} , $$
(6)

where \(PZM^{R}_{n,m}(f({x_{q}},{y_{p}}))\) is the rotated version of PZM with the rotation angle 𝜃. To remove the rotational effect on phase, a complex valued rotation invariant is created by combining phase coefficients of different orders with repetitions. This is represented as

where \(\varphi ^{R}_{n,m}=\varphi _{n,m}-m\theta \). Here, φn, m the phase coefficient and \(\varphi ^{R}_{n,m}\) is the phase value of rotated image. When we set n0 = 1 the modified phase angle of PZM is

where φ1,1 = 0. Therefore, the phase feature vector of each patch f(xq,yp) is represented as

Table 2 Patch labeling by edge detection

After the phase feature vector generation, several heuristic criteria have been used for identify the patches with edge and without edge. If ϕ < 𝜃, then its corresponding patch is labelled as non edge else labelled the patch as edged. This makes the searching of the patches having more information faster.

2.3 Patch classification

The similar patches of an input patch are directly searched from the training database. However, this scheme is inefficient and similar optimal patches may not be obtained. Hence, to find out the similar optimal patches, we use feature similarity index (FSIM) [53] instead of the Euclidean distance, since it fails to search the appropriate nearest neighbor patches in the manifold. FSIM is computed by phase congruency (SPC(x)) and gradient (SG(x)) based similarity.

Feature similarity index is computed by combining SPC(x) and SG(x)as

$$ S_{L}(x)=[S_{PC}(x)]^{\alpha}.[S_{G}(x)]^{\beta}, $$
(10)

where α and β are parameters for adjusting relative importance of both the features. Range of the FSIM is 0 to 1. The FSIM of two similar patches is nearer to 1. The value of FSIM between two patches lies between 0 to 1 and higher the value of FSIM, the similar the patches are. Thus, optimal patches are computed by top k-nearest neighbor patches with respect to FSIM.

2.4 Collaborative optimal weight reconstruction

In this section, a collaborative weight reconstruction is used during the HR patch generation. Firstly, the weight is computed using least squarer technique. Further, another weight vector is generated using non-negative matrix factorization method. Thereafter, final weight vector is collaboratively constructed using LS and NMF technique.

2.4.1 Weight vector generated by LS method

After generating the patches from images, each testing patch is taken into account to generate an expected HR patch. For each input patch \({l_{t}^{q}}\) of \(L_{t}= \left \{{l_{t}^{q}}, q = 1,2,3,...m\right \}, k\)-nearest neighbor training patches are searched. Through least square (LS) method the weights related to k-nearest neighbor training patches are found by minimizing the local reconstruction error as

$$ {\varepsilon^{q}} = \min {\left\| {{l_{t}^{q}} - \sum\limits_{{l_{r}^{p}} \in {\aleph_{q}}} {{w_{qp}}{l_{r}^{p}}} } \right\|^{2}} \,, $$
(11)

where q is the neighbor of \({l_{t}^{q}}\) in training set Lr and element wqp of Wq is the weight for \({l_{r}^{p}}\), subject to two constraints \(\sum \nolimits _{{l_{r}^{p}} \in {\aleph _{q}}} {{w_{qp}}} = 1\) and wqp = 0 for any \({l_{r}^{p}} \notin {\aleph _{q}}\).

Let us define a local Gram matrix Gqfor ltq as

$$ {G_{q}} = {\left( {{l_{t}^{q}}{1^{T}} - \textbf{L}} \right)^{T}}\left( {{l_{t}^{q}}{1^{T}} - \textbf{L}} \right) \,, $$
(12)

where 1 is a column vector of ones and L is a D × K matrix with its columns being the neighbors of \({x_{t}^{q}}\). To form a K-dimensional weight vector Wq, each weight wqpis reordered by the subscript p.

To solve an eigenvalue problem using the Lagrange theorem, the formulation is given by

$$ \varepsilon = {w^{T}}Gw \,, $$
(13)

where

$$ w = \left( {\frac{1}{2}} \right)\lambda {G^{- 1}}1 \,, $$
(14)

and

$$ \lambda = \frac{2}{{{1^{T}}{G^{- 1}}1}} \,. $$
(15)

When G is singular, the result of the least squares problem for finding wq does not have the unique solution. The constrained least squares problem has the following closed-form solution as:

$$ {w_{q}} = \frac{{G_{q}^{- 1}1}}{{{1^{T}}G_{q}^{- 1}1}}\,. $$
(16)

Hence, the least square optimal weight vector is represented by wl where wq is the element.

2.4.2 Weight vector generated by NMF method

Apart from that another weight vector is generated by non-negative matrix factorization (NMF). According to non-negative matrix factorization the weight is generated by

$$ X \approx F{G^{T}} $$
(17)
$$ {x_{t}^{i}} \approx {x_{d}^{i}}w $$
(18)

The iterative solution for (17) is proved to converge to a local minimum of the Euclidean distance \( {\left \| {X - F{w_{n}}T} \right \|^{2}}\). The iterative solution consists of two multiplicative update rules, both for F and wij. In this case matrix F, formed by the actual LR patches in the dictionary, is fixed and no update rule is needed for it. But the weight of each patch can be updated by

$$ {w_{ij}} \leftarrow {w_{ij}}\sqrt{\frac{{({x_{d}^{i}}^{T}{x_{d}^{i}})_{j}^{+} + {{\left[ {{w^{iT}}{{({x_{d}^{i}}^{T}{x_{d}^{i}})}^{-} }} \right]}_{j}}}}{{({x_{d}^{i}}^{T}{x_{d}^{i}})_{j}^{-} + {{\left[ {{w^{iT}}{{({x_{d}^{i}}^{T}{x_{d}^{i}})}^{+} }} \right]}_{j}}}}}\,. $$
(19)

The final weight vector is represented as wn, where wijis the element.

Thus, a collaborative algorithm is developed by combine weight generated by LS and the NMF to compute the optimal weight vector. The collaborative optimal reconstruction weight for the HR patch reconstruction is defined as

$$ \left| {w = \frac{1}{2}{w_{q}} \cdot {w_{n}} + \frac{1}{4}({w_{q}} + {w_{n}})} \right|\,, $$
(20)

where wqlis the element of final optimal weight vector W.

2.5 HR image reconstruction

As the manifold between LR-HR space is similar, the high-resolution output image patch htq is computed as follows

$$ {h_{t}^{q}} = \sum\limits_{{l_{r}^{p}} \in {\aleph_{q}}} {{w_{ql}}{h_{r}^{p}}}+ \overline{{h_{r}^{p}}} \,, $$
(21)

where \(\overline {{h_{r}^{p}}}\) is the mean of the luminance and wqlis the final reconstruction weight. The HR patch generation is shown in Fig. 3. To generate the output image, the overlapping patches \(H_{t}=\left \{{h_{t}^{q}} \right \}_{q = 1}^{n}\) is merged. The major factor that improves the performance of the reconstruction image are neighborhood preservation between patches and optimal reconstruction weight. Hence, this paper mainly focuses on the feature selection, neighborhood patch search by similarity measures and the collaborative reconstruction weight. Finally, the reconstructed image in luminance channel has been combined with the interpolated chromatic channel. Thereafter, the YCbCr image is converted to RGB channel for final HR image reconstruction.

Fig. 3
figure 3

HR patch generation

3 Experimental results

3.1 Experimental set up

In this paper, we validate our algorithm with various texture images collected from https://www.robots.ox.ac.uk/vgg/data/dtd/ and http://sipi.usc.edu/database/database.php?volume=textures [43]. In addition, some standard images used in Matlab tool have been considered for the experiment. To conduct the experiment, LR image is produced from the original HR image by 5 × 5Gaussian blur with a down-sampling factor 3. Both color and gray texture images are taken into account as shown in Figs. 4 and 5, respectively. Further, the scheme is validated on standard images as shown in Fig. 6. As human eyes has different sensitivity to change in illumination rather than chromatic component, the color image for the experiment is transformed from RGB to YCbCr channel. Moreover, RGB component is not very compatible for image representation. Hence, the proposed algorithm executed on the Y component of the YCbCr channel. The performance of the proposed scheme has been evaluated and a comparison has been made with different existing schemes. The existing competent schemes are Bicubic, EBSR [20], SRNE [9], NEVPM [17], NeedFS [7] and SCSR [46]. To access the quality of the image, two commonly used performance measures are employed i.e., Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) [42] to evaluate the sensitivity of the noise and structural similarity between the reconstructed image and ground truth image respectively.

Fig. 4
figure 4

Testing color texture images

Fig. 5
figure 5

Testing gray texture images

Fig. 6
figure 6

Testing standard images

3.2 Experimental results and discussion

3.2.1 Performance evaluation with varying magnification factor (γ)

In this section, we validate the proposed approach with respect to various magnification factor. In this approach, magnification factors are 2,3 and 4. Figures 7 and 8 show PSNR evaluation and SSIM evaluation with various magnification factor. From the given graph, it is observed that the quality is gradually decreasing according to the increase in magnification size. Hence, we consider 3 as the suitable factor for the experiment.

Fig. 7
figure 7

PSNR evaluation with respect to various magnification factor (γ)

Fig. 8
figure 8

SSIM evaluation with respect to various magnification factor (γ)

3.2.2 Performance evaluation with varying patch size (t)

The performance measures are computed with different patch sizes i.e., 5, 7, and 9. Tables 3 and 4 list the PSNR values of color and gray texture images respectively, whereas Tables 5 and 6 record the SSIM values for the color and gray respectively. Tables 7 and 8 show the PSNR and SSIM results for few standard images. It observed that the proposed method performs better than other existing methods on color texture, gray texture images and standard images. Further, it may be noticed that the larger the patch size, the better is the PSNR and SSIM values.

Table 3 PSNR of color texture image
Table 4 PSNR of gray texture image
Table 5 SSIM of color texture image
Table 6 SSIM of gray texture image
Table 7 PSNR of standard image
Table 8 SSIM of standard image

3.2.3 Performance evaluation by neighborhood preservation rate

In this section, neighborhood preservation rate is considered as another criterion to evaluate the efficacy of the proposed feature selection. Firstly, a group of LR patches have been selected randomly from the testing patch set. Then for each LR patch, the k-nearest neighbor patches have been considered from the test set. Prior to this, the corresponding k-nearest neighbor HR patches are identified from the HR patches. Comparing the number of corresponding HR patches of the k-nearest neighbor of LR patches, the neighborhood preservation rate is obtained.

In this section, we have performed the comparison between LR and HR patch with different features used in the state-of-the-art schemes. In SRNE [9] method, first gradient and second gradient are used as features, whereas in NeedFS [7] a combination of first-order gradient and norm luminance are considered. NESRRL [32] considered first gradient and residual luminance as the feature vector to generate the expected HR image. Moreover, to verify the advantage of PZM over ZM, we have validated with [34]. In this work, we have used PZM for feature generation. Here, we randomly consider 1800 LR patches, and out of which 300 patches are used as testing LR patches. Then, we have evaluated the neighborhood preservation for 3 magnification with different neighborhood size.

The neighborhood preservation rate obtained by the existing schemes with respect to different neighborhood size are shown in Fig. 9. It can be observed that the proposed scheme achieves better results as compared to the existing schemes by means of neighborhood preservation rate due to better feature vector.

Fig. 9
figure 9

Neighborhood preservation of test image

3.2.4 Performance evaluation varying with neighborhood size (k)

It has been observed that the neighborhood size has significant impact on the performance of the model. Hence, in this experiment the neighborhood size is varied between 1 to 15 and the corresponding PSNR and SSIM results for color, gray texture images and standard images are shown in Figs. 10 and 11 respectively. It may be seen that the proposed scheme achieves better PSNR and SSIM results for neighborhood size between 5 to 8. Further, the visual performance comparisons for three different category of images are shown in Figs. 1213141516 and 17. From these figures, it may be observed that the propose scheme yields better visualization results than other state-of-the-art schemes.

Fig. 10
figure 10

PSNR evaluation with respect to various neighborhood size (k)

Fig. 11
figure 11

SSIM evaluation with respect to various neighborhood size (k)

Fig. 12
figure 12

Comparison of SR results with (3 ×) magnification for c1image

Fig. 13
figure 13

Comparison of SR results with (3 ×) magnification for c3image

Fig. 14
figure 14

Comparison of SR results with (3 ×) magnification for g1image

Fig. 15
figure 15

Comparison of SR results with (3 ×) magnification for g2image

Fig. 16
figure 16

Comparison of SR results with (3 ×) magnification for bird image

Fig. 17
figure 17

Comparison of SR results with (3 ×) magnification for butterfly image

4 Conclusion

In this paper, a new framework for learning based super resolution has been proposed for texture image. Pseudo Zernike moments are utilized for the feature selection. Moreover, FSIM is used for patch classification during neighbor embedding. Subsequently, a collaborative optimal reconstruction weight is computed to HR patch generation. Optimal reconstruction weight is defined by the combination of least square error and non-negative factorization matrix. The experimental results on texture and standard images demonstrate that the proposed scheme achieves better results than other competent schemes. In future, we intent to explore our method on different images like areal image, medical image, video surveillance, etc. In addition, other feature generation schemes could be investigated as the potential alternatives to PZM.