During the process of image acquisition, camera shaking, defocusing, and rapid object movement cause blurred images. Blur distortion results in the loss of a large number of sharp details and local feature information in the whole or a part of a given image and thus seriously affects the subsequent interpretation and recognition of the image. Therefore, evaluating the quality of blurred images reasonably is of great theoretical significance. As the subjective judgment of the human eyes is time consuming, laborious, and unstable, researchers have focused on building an objective quality evaluation method in line with the judgment of the human eyes. At present, objective blurred image quality assessment (IQA) methods can be divided into three categories according to whether the original image is needed: full reference (FR)[1], partial reference (RR)[2], and no-reference (NR)[3]. The first two methods need the original image or the collection of some features of the original image. However, because most reference images are not easy to obtain or cannot be obtained at all in many practical applications, the development of NR-IQA has a wide range of practical significance, and a large number of researchers have aimed to study and explore this problem.
Mainstream NR blurred image evaluation algorithms are mainly divided into the following categories[4]: transform domain-based algorithms, reblur evaluation algorithms based on pixel statistical information, edge-based evaluation algorithms, and neural network-based algorithms. 1) Transform domain-based algorithms include the kurtosis method based on the DCT domain[5], Fourier transform algorithm[6], and global phase coherence and local phase coherence algorithm[7-8]. Using the properties of blurred images in the transform domain, this type of algorithm is characterized by definite physical meaning. However, the calculation of the model is relatively complex. 2) For reblur evaluation algorithms based on pixel statistical information, Crete et al.[9] proposed the measurement of the gray-level changes of neighboring pixels by using the reblur theory; the approaches include the adjacent gray-level difference variance method[10] and covariance matrix eigenvalue method[11]. This type of algorithm uses the statistical information of an image and yields good robustness. However, it ignores the location information of pixels, thereby greatly affecting estimation. 3) Edge-based evaluation algorithms include the just noticeable blur (JNB) algorithm[12] and cumulative probability of blur detection (CPBD) algorithm[13]. This type of algorithm is intuitive in terms of concept as it considers the characteristics of the human vision. However, such algorithm has a certain dependence on image content. The first two types of algorithms are mainly used to evaluate the quality of images according to the changes in the blur degree of the same images. They are mainly used in the field of automatic focusing of imaging systems, but they have limited usage in realistic scenarios. This type of algorithm can evaluate the blur degree of different images and achieve good performance. However, the calculation process is generally complex, and it cannot be widely used. 4) In terms of neural network-based algorithms, Yu et al.[14] proposed a measurement method on the basis of the shallow convolutional neural network (CNN). Hosseini et al.[15] proposed a measurement method called the human visual system (HVS) MaxPol. Yu et al.[16] used the general regression neural network (GRNN) instead of the multilayer perceptron in the original CNN architecture to generate the CNN-GRNN model. Although these algorithms have made valuable progress in sharpness measurement and they have high performance indicators, their high computing overhead may limit their application scope, e.g., running on low-cost mobile devices. In addition, these models based on neural networks need a large number of training samples, and existing image quality databases only contain hundreds of blurred images. This type of evaluation algorithm leads to overfitting.
The essence of blurred images is the loss of high-frequency information. Such characteristic is mainly manifested in the reduction of edge sharpness and the loss of texture details. Gradients are the constituent elements of edges, and they can reflect the changes in details and textures. The visual attention mechanism is the main factor of the HVS, which pays close attention to the regions of interest during subjective evaluation. These two visual characteristics have been widely used in IQA[17-19], and good evaluation performance has been achieved. With the above considerations, this study proposed a NR blur IQA algorithm on the basis of gradient features and saliency. The proposed algorithm uses reblur theory to construct a reference image to calculate the gradient distortion measure of blurred images. The experiments on natural image databases, namely, the image and video engineering (LIVE) database, categorial image quality (CSIQ) database, and camera image database (CID2013), show that the proposed algorithm achieves good subjective consistency and features low time complexity.
The input RGB blurred image is transformed into a gray image. Specifically, a Gaussian low-pass filter is used to blur the image and then construct the reference image. The reference image is then combined with the distorted one for evaluation. The gradient distortion measurement (GDM) can reflect tiny detail contrast, and the changes in texture features in an image are obtained by combining gradient similarities. At the same time, the saliency algorithm is introduced to detect the saliency of the input RGB blurred image. Moreover, an adaptive threshold is used to binarize the saliency map to obtain the saliency region map. The GDM map of the image and reblurred image is then weighted by the subblock saliency to form the final blur evaluation result. The principle block diagram is shown in Fig.1.
Fig.1 Flow chart of proposed algorithm
In NR-IQA, the reference image cannot be obtained. Hence, these methods have certain limitations, and the accuracy of the algorithm is greatly reduced. In this work, the concept of reblur[9], which is based on the subjective perception of the human eyes, is introduced. Given the subjective perception of the human eyes, the differences between sharp images and blurred images are easier to be perceived than the differences between blurred images and reblurred images. The method in Ref.[20] is used to construct a reference image by using the variation of high-frequency components. The image blur measurement map is then obtained by fusing the similarity of the local standard deviation of the images before and after reblurring with the significance model of blurred images. The final evaluation result is obtained by weighting the measure map with the local standard deviation value map.
A distortion RGB image is the first input in our algorithm; it is transformed to grayscale version to obtain the corresponding reference image[9].
As shown in Fig.2, the difference between the first two images before and after blurring can be easily perceived, whereas the difference between the second image and the third image is not obvious.
(a)
(b)
(c)
Fig.2 Schematic of image reblurring effect. (a) Sharp image; (b) Blurred image; (c) Reblurred image
As observed by the human eyes, a sharp image loses a considerable amount of image edge details in the process of image blurring. For an image that has been blurred, the details lost by the second blurring are more greatly reduced than those after the first blurring. This outcome is consistent with the fact that blurring affects the high-frequency components in images and that low-frequency components remain stable[21]; specifically, the blurring process mainly reduces the high-frequency components of the original image. After reblurring, the number of high-frequency components is reduced, and the image does not change greatly. Hence, the degree of change from a visual observation is minimal.
For the minimum standard deviation of effective reblurring[22], the standard deviation of the algorithm after experimental verification is set to 0.55, and the window size is 7×8.
1.2.1 Image gradient similarity calculation
HVS research[23] has shown that the human eyes are suitable for extracting structural information in visual areas. Wang et al.[24] proposed an evaluation algorithm based on structural similarity (SSIM) that can effectively extract structural information in visual areas. Given the advantages of the SSIM algorithm, image quality can be evaluated accurately and reliably by constructing the reference image and then using the structural similarity between the reference image and the distorted image to measure sharpness. However, the SSIM algorithm’s evaluation results are not reasonable for degraded images with serious blur[25], which indicates that the evaluation value of blurred images is too high and does not conform to people’s subjective feelings. The experimental results show that the human eye is very sensitive to image edges[26-28] and that gradients can effectively reflect the changes in the details and textures of images. Such a result proves that edges are an important part of the structural information of images. For seriously degraded blurred images, extracting gradient features can help effectively evaluate the degree of image blur and align the evaluation results with the characteristics of human visual perception.
As described in Section 1.1, the blur degree of a blurred image can be determined by measuring the changes in the high-frequency components of the blurred image and its reblurred image. In extracting the gradient information of an image, most of the low-frequency components are removed. Hence, the high-frequency components become dominant, and the sensitivity to the changes in blur degree increases. As shown in Fig.3, the gradient information extracted from images with different blur degrees can effectively reflect the changes in image blur degree. In this work, the blurred image is reblurred to construct the reference image, and the blur degree of an image is calculated by combining the gradient similarities.
(a)
(b)
(c)
(c)
(e)
(f)
(g)
(h)
Fig.3 Image gradient changes affected by reblurring. (a) Sharp image; (b) Gradient map of image (a); (c) Blurred image; (d) Gradient map of image C; (e) Reblurred version of image (a); (f) Gradient map of image (e); (g) Reblurred version of image (c); (h) Gradient map of image (g)
The gradient is usually calculated by linear filtering convolution. The typical filters include the Roberts, Sobel, Prewitt, and Scharr filters. This work uses the Scharr horizontal operator and Scharr vertical operator to calculate the horizontal and vertical gradients, respectively. These operators are defined as follows:
(1)
where H and V represent the Scharr horizontal operator and vertical operator, respectively.
The calculation methods are as follows:
Gx=Ii⊗H
(2)
Gy=Ii⊗V
(3)
where the operation ⊗ is the convolution operation of two matrices; Ii is the input blurred image in pixel space i; Gx and Gy respectively represent the gradient images in the horizontal and vertical directions through gradient filtering. Then,
(4)
The similarity between the blurred image (Gb) and the corresponding gradient amplitude image of the reference image (Gr) is used to measure the gradient feature changes of the image before and after reblurring. The calculation of the GDM map is defined as
(5)
where C1 is a positive constant matrix to avoid a zero denominator; Mi is based on pixel-by-pixel calculation, and the gradient amplitude calculation is based on a small area block.
1.2.2 Salient region calculation
Region detection for images is a basic research topic in neuroscience and psychology.The human eyes employ a visual attention mechanism and can thus select specific regions of interest from a large amount of information, including important regions of an image. These regions of interest are called salient regions. With the deepening of the research on salient region extraction, many algorithms have been proposed[29-31]. The salient region extraction algorithm (SDSP) proposed by Zhang et al.[29] has been cited by a large number of scholars because of its simple structure and fast calculation speed. The attention mechanism of the human eyes based on some edge and local details outside location features is also an important factor affecting the visual quality of images. On the basis of the SDSP saliency model, this algorithm combines two prior features, namely, frequency saliency and color saliency.
(7)
where SF and SC respectively represent the frequency significance and color significance. They are defined as
(8)
where fL, fb, fc are the three components of the color space; g represents the log Gabor filters.
(9)
where fA represents the degree of green-red in the pixel; fB represents the degree of blue-yellow; σc is a constant parameter.
In this work, the frequency saliency feature and color saliency feature are fused on the basis of the SDSP saliency model. Relative to the classical SDSP saliency model, which fuses three features, the fused model can improve the accuracy of calculating the saliency regions of blurred images.
According to the adaptive threshold selection algorithm[32], the threshold in this work is obtained for blurred images with different contents and scenes. The final salient region map is obtained by binarization. This map can accurately reflect the changes in the GDM map of the original blurred image in the significant region. In the LIVE database, images of parrots with different blur degrees are selected and shown in Fig.4. Fig.4(b) is more blurred than Fig.4(a). Moreover, the image structure information of the sharper image is more degraded than that of the blurred image. In addition, the saliency map corresponding to the blurred image presents fewer changes than the saliency map of the sharp image. In this way, the saliency area of the GDM map of the blurred image can be accurately calculated. The algorithm ignores the influence of the background region and evaluates the quality of blurred images by only measuring the gradient structure distortion of the significant regions of blurred images.
(a)
(b)
(c)
(d)
(e)
(f)
Fig.4 Gradient distortion measurement (GDM) and salient region maps corresponding to images with different blur degrees. (a) Sharper image; (b) GDM map of image (a); (c) Salient region map of image (a); (d) Blurred image; (e) GDM map of image (d); (f) Salient region map of image (d)
After calculating the GDM map of the test image and reblurred image as long as the salient region, the GDM map is weighted by the salient value of the salient region subblock to obtain the final evaluation result.
(10)
where represents the significant region map of the input blurred images.
To verify the effectiveness of the algorithm, this study uses quality evaluation image databases, namely, LIVE[33], CSIQ[34], and CID2013[35] databases, in the simulation experiments. According to the Video Quality Experts Group[36], a certain nonlinear relationship exists between objective and subjective evaluation results. The performance of the proposed algorithm is objectively reflected using the nonlinear regression function of five parameters[37].
(11)
where f(x) is the prediction score after nonlinear regression, representing the objective evaluation value; βi(i=1,2,3,4,5) denote the parameters of the model.
According to Section 2.1, the minimum standard deviation of effective reblurring is 0.25. Hence, the actual reblurring value in the proposed algorithm σ2 should be greater than 0.25. On the basis of the effective minimum standard deviation, the index values of SROCC based on different windows and different standard deviations in two image databases are compared to select the appropriate Gaussian filter window and standard deviation. Fig.5 indicates that given the highest SROCC values in the LIVE and CSIQ databases, the window size in the algorithm is set to 7×8, and the blur standard deviation is 0.55.
(a)
(b)Fig.5 Comparison of SROCC under different windows and standard deviations. (a) Experiment in the LIVE database; (b) Experiment in the CSIQ database
On the basis of the SDSP algorithm, this study proposes to fuse only two salient features, namely, frequency salient features and color salient features. A comparative experiment is performed to verify the superiority of the algorithm. Select images with varying degrees of blur in LIVE database. The results are shown in Fig.6.
(a)
(b)
(c)
(d)
(e)
(f)
Fig.6 Comparison of experiment results. (a) Monarch map; (b) Saliency map of image (a) based on classical SDSP saliency model; (c) Saliency maps fusing two saliency features of image (a); (d) House map; (e) Saliency map of image (d) based on classical SDSP saliency model; (f) Saliency maps fusing two saliency features of image (d)
The experiment results reveal that the branches in the blurred background in Fig.6(b) are not marked on the basis of the classical SDSP saliency model and that the branch area also includes edge information. Meanwhile, Fig.6(c) shows that the combination of two saliency feature models not only marks the visual saliency area but also includes some edge information outside the central area. As shown in Fig.6(d), this algorithm can mark the edges of the saliency area and some important edge information flowers. This capability differs from the saliency map based on the classical SDSP saliency model shown in Fig.6(e). Therefore, when extracting the saliency of blurred images, the saliency models of the proposed algorithm, which only fuses frequency saliency features and color saliency features, perform better than the classical SDSP saliency model.
2.4.1 Natural blurred image database
In the experiment, we set C=115. Before calculating the gradient similarity, we need to filter the 3×3 image with a Gaussian weighted average and then perform downsampling. To show the evaluation effect of the algorithm intuitively, we compare it with five NR blur image quality evaluation algorithms that use the reblur[9] algorithm of the image reblurring effect, JNB model[12] of the just perceptible blur concept, CPBD[13], global phase consistency (GPC-SI)[7], and local phase consistency (LPC-SI)[8]. CPBD, LPC-SI, and the algorithm which uses reblur theory[20] are recognized as excellent evaluation algorithms in the research of NR blur image quality evaluation algorithms. The results of the proposed algorithm and five representative algorithms for the LIVE and CSIQ databases are shown in Tab.1.
As shown in Tab.1, the performance index values of the proposed algorithm are better than those of the other algorithms. In the CSIQ database, these performance index values are slightly lower than those of LPC-SI and the algorithm in Ref.[20], but they show improvement relative to the results of the other four evaluation algorithms. Given its performance in the LIVE and CSIQ databases, the proposed algorithm achieves significant improvement in its correlation and accuracy relative to the other algorithms. In general, the proposed algorithm achieves the best prediction performance in the LIVE and CSIQ databases.
Tab.1 Performance comparison of different evaluation algorithms in LIVE and CSIQ databases
Database standardReblurJNBMCPBDGPC-SILPC-SIRef. 20 ProposedmethodBIDSROCC0.879 20.787 20.918 20.864 00.938 90.925 80.940 0KROCC0.701 90.607 10.763 20.687 40.778 50.774 10.797 5PLCC0.883 40.818 80.912 40.884 80.931 60.931 50.951 8RMSE8.654 910.602 77.553 88.617 86.771 66.716 75.667 4CSIQSROCC0.853 10.762 40.879 00.865 10.907 10.896 20.888 3KROCC0.644 40.597 60.690 50.687 50.720 50.724 80.712 2PLCC0.885 00.870 90.882 20.901 80.825 60.925 40.909 0RMSE0.133 40.140 90.134 90.123 90.108 50.108 60.116 8
2.4.2 Real blurred image database
In the real world, blur distortion is not regular and is more difficult to evaluate than blurred images in natural image databases. To verify the performance of the proposed algorithm in a real distorted blurred image database, the camera image database CID2013 is selected for the experiment. CID2013 contains 473 images and provides subjective scores for sharpness, graininess, brightness, and color saturation for each image. Only the sharpness score is used in the experiment (see Tab.2).
Tab.2 Performance comparison of real blur image databases
DatabaseStandardReblurJNBCPBDGPC-SILPC-SIProposed methodCID2013SROCC0.229 00.443 80.432 90.402 00.597 30.517 1
To further verify the superiority of the proposed algorithm in terms of time complexity, we conduct a comparison test between the proposed algorithm and the other aforementioned evaluation algorithms (see Tab.3). The evaluation results are obtained by conducting 10 runs and adopting the mean for the image size of 512×512. The processor of the hardware platform is Intel© CoreTM i5@1.70 GHz with 4 GB memory and Windows 7 64-bit operating system. The software platform is MATLAB R2017a. Tab.3 shows that the algorithm has a relatively low time complexity. The algorithm achieves the best performance evaluation in the LIVE database (see Tab.2). Although its performance index for the CSIQ and CID2013 databases is slightly lower than that of LPC-SI, its operation time is greatly improved. Hence, the proposed algorithm is a good candidate for time-critical applications.
Tab.3 Running time comparison of different evaluation algorithms
MethodReblurJNBMCPBDGPC-SILPC-SIProposedmethodTime/s0.256 10.743 10.541 40.305 12.404 30.246 5
1) Following the theory of reblurring, this study proposes a NR blur image quality evaluation algorithm that is based on GDM and salient region maps. The image to be evaluated is blurred by a Gaussian low-pass filter to construct a reference image. Then, the GDM map, which can reflect the contrast and texture changes of tiny image details, is obtained by combining the gradient similarities.
2) The saliency region map is calculated, and the final evaluation result of the blurred image is obtained by weighting the saliency value of the saliency region subblock.
3) The experimental results on the LIVE, CSIQ, and CID2013 databases show that the gradient and saliency features are integrated to make the evaluation results subjectively and objectively consistent with human visual characteristics.
[1]Hadizadeh H, Bajic I V. Full-reference objective quality assessment of tone-mapped images [J]. IEEE Transactions on Multimedia, 2017, 20(2): 392-404. DOI: 10.1109/TMM.2017.2740023
[2]Ma J, An P, Shen L Q, et al. Reduced-reference stereoscopic image quality assessment using natural scene statistics and structural degradation [J]. IEEE Access, 2018, 6: 2768-2780. DOI: 10.1109/ACCESS.2017.2785282.
[3]Fang Y M, Yan J B, Li L D, et al.No reference quality assessment for screen content images with both local and global feature representation[J]. IEEE Transactions on Image Processing, 2018, 27(4): 1600-1610. DOI:10.1109 /TIP.2017.2781307.
[4]Wang Z. M. Review of no-reference image quality assessment [J]. Acta Automatica Sinica, 2015, 41(6):1062-1079. DOI:10.16383/j.aas.2015.c140404. (in Chinese)
[5]Caviedes J, Oberti F. A new sharpness metric based on local kurtosis, edge and energy information [J]. Signal Processing: Image Communication, 2004, 19(2): 147-161. DOI:10.1016/j.image.2003.08.002.
[6]Liu H T, Heynderickx I. A perceptually relevant no-reference blockiness metric based on local image characteristics [J]. EURASIP Journal on Advances in Signal Processing, 2009, 2009: 1-14. DOI:10.1155/2009/263540.
[7]Blanchet G, Moisan L. An explicit sharpness index related to global phase coherence[C]// IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP). Yoto, Japan,2012:1065-1068. DOI: 10.1109/ICASSP.2012.6288070.
[8]Hassen R, Wang Z,Salama M. M. A. Image sharpness assessment based on local phase coherence [J]. IEEE Transactions on Image Processing, 2013, 22(7): 2798-2810. DOI:10.1109/TIP.2013.2251643.
[9]Crete F, Dolmiere T, Ladret P, et al. The blur effect: Perception and estimation with a new no-reference perceptual blur metric[C]//2017 Human Vision And Electronic Imaging Ⅻ. Sun Jose, CA, USA, 2007, 6492: 64920I. DOI:10.1117/12.702790.
[10]Tsomko E, Kim H J. Efficient method of detecting globally blurry or sharp images[C]// 2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services. Klagenfurt, Austria, 2008: 171-174. DOI: 10.1109/WIAMIS.2008.28.
[11]Wee C Y, Paramesran R. Image sharpness measure using eigenvalues[C]//2008 9th International Conference on Signal Processing. Beijing, China, 2008: 840-843. DOI: 10 .1109/ICOSP.2008.4
[12]Ferzli R, Karam L J. A no-reference objective image sharpness metric based on the notion of just noticeable blur (JNB)[J].IEEE Transactions on Image Processing, 2009, 18(4): 717-728. DOI:10.1109/TIP.2008.2011760.
[13]Narvekar N D, Karam L J. A no-reference image blur metric based on the cumulative probability of blur detection (CPBD)[J].IEEE Transactions on Image Processing, 2011, 20(9): 2678-2683. DOI:10.1109/TIP.2011.2131660.
[14]Yu S D, Wu S B, Wang L, et al. A shallow convolutional neural network for blind image sharpness assessment[J].PLoS One, 2017, 12(5): e0176632. DOI:10.1371/journal.pone.0176632.
[15]Hosseini M S, Zhang, Y, Plataniotis K N. Encoding visual sensitivity by MaxPol convolution filters for image sharpness assessment[J]. IEEE Transactions on Image Processing, 2019, 28(9): 4510-4525. DOI: 10.1109/TIP.2019.2906582.
[16]Yu S, Jiang F, Li L, Xie Y. CNN-GRNN for image sharpness assessment[C]//2016 Asian Conference on Computer Vision. Cham:Springer, 2016: 50-61. DOI:10.1007/978-3-319-54407-6_4.
[17]Zhang L, Zhang L, Mou X Q, et al. FSIM:A feature similarity index for image quality assessment[J]. IEEE Transactions on Image Processing, 2011, 20(8): 2378-2386. DOI:10.1109/TIP.2011.2109730.
[18]Zhang L, Shen Y, Li H Y. VSI:A visual saliency-induced index for perceptual image quality assessment[J]. IEEE Transactions on Image Processing, 2014, 23(10): 4270-4281. DOI:10.1109/TIP.2014.2346028.
[19]Xue W F, Zhang L, Mou X Q, et al. Gradient magnitude similarity deviation: A highly efficient perceptual image quality index[J].IEEE Transactions on Image Processing, 2014, 23(2): 684-695. DOI:10.1109/TIP.2013.2293423.
[20]Lu Y F, Zhang T, Zheng J,et al. No-reference blurring image quality assessment based on local standard deviation and saliency map[J]. Journal of Jilin University (Engineering and Technology Edition), 2016, 46(4): 1337-1343. (in Chinese)
[21]Zhao P, LI L, Cai H. Saliency guided gradient similarity for fast perceptual blur assessment[J]. IEICE Transactions on Information and Systems, 2015, 98(8): 1613-1616.DOI:10.3969/j.issn.0254-3087.2016.07.026.
[22]Wang H Y, Feng J, Niu W,et al. No-reference image quality assessment based on re-blur theory[J]. Chinese Journal of Scientific Instrument, 2016, 37(7):1647-1655.DOI:10.13229/j.cnki.jdxbgxb201604046.(in Chinese)
[23]Furht B, Marqure O. The handbook of video databases: design and applications [M]. Boca Raton, FL, USA: CRC Press, 2003: 1041-1078.
[24]Wang Z, Bovik A C, Sheikh H R, et al. Image quality assessment: From error visibility to structural similarity[J].IEEE Transactions on Image Processing, 2004, 13(4): 600-612. DOI:10.1109/tip.2003.819861.
[25]Wang Z, Bovik A C, Lu L G.Why is image quality assessment so difficult?[C]// 2011 IEEE International Conference on Acoustics. Orlando, FL, USA, 2002, 4: 3313-3316. DOI: 10.1109/ICASSP.2002.5745362.
[26]Liu A M, Lin W S, Narwaria M. Image quality assessment based on gradient similarity[J].IEEE Transactions on Image Processing, 2012, 21(4): 1500-1512. DOI:10.1109/TIP.2011.2175935.
[27]Wang T H, Jia H Z, Shu H Z. Full-reference image quality assessment algorithm based on gradient magnitude and histogram of oriented gradient [J]. Journal of Southeast University (Natural Science Edition), 2018, 48(2): 276-281. DOI:10.3969/j.issn.1001-0505.2018.02.014. (in Chinese)
[28]Yang C L, Kuang K Z, Chen G H, et al. Gradient-based structural similarity for image quality assessment [J]. Journal of South China University of Technology (Natural Science Edition), 2006(9): 22-25. DOI:10.3321/j.issn:1000-565X.2006.09.005.(in Chinese)
[29]Zhang L, Gu Z Y, Li H Y, et al. SDSP: A novel saliency detection method by combining simple priors[C]//International Conference on Image Processing. Melbourne, Australia, 2013: 171-175. DOI: 10.1109/ICIP.2013.6738036.
[30]Itti L, Koch C, Niebur E, et al. A model of saliency-based visual attention for rapid scene analysis [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1998, 20(11): 1254-1259. DOI: 10.1109/34.730558.
[31]Xia Z, Cheng C, Li X. Visual comfort enhancement study based on visual attention detection for stereoscopic displays [J]. Journal of the Society for Information Display, 2016, 24(10): 633-640. DOI:10.1002/jsid.508.
[32]Zhou Y, Wang K, Zhang H X, Xu W Q, Li L. Blur image quality assessment method based on blur detection probability variation [J]. Laser & Optoelectronics Progress, 2020, 57(10): 55- 60. DOI:10.3788/LOP57.101004.(in Chinese).
[33]Sheikh H,Wang Z,Cormack L,et al. LIVE image quality assessment database release2[EB/OL]. (2005)[2020-05-13]. http: / /live.Ece.utexas.edu/ research/ quality.
[34]Larson E C. Chandler D M. Categorical image quality (CSIQ) database [EB/OL]. (2009)[2020-07-20].http: // vision.okstata.edu/csiq/.
[35]Virtanen T, Nuutinen M, Vaahteranoksa M, et al. Cid2013: A database for evaluating no-reference image quality assessment algorithms[J]. IEEE Transactions on Image Processing,2015,24(1): 390-402. DOI: 10.1109/TIP.2014.2378061.
[36]Video Quality Experts Group. Validation of objective models of video quality assessment, phase Ⅱ VQEG [R/OL]. (2003-08-12)[2020-05-13]. http://www.vqeg.org.
[37]Sheikh H R, Sabir M, Bovik A C, et al. A statistical evaluation of recent full reference image quality assessment algorithms [J]. IEEE Transactions on Image Processing, 2006, 15(11): 3440-3451. DOI: 10.1109/TIP.2006.881959.