Open Access
16 December 2013 No-reference image quality assessment based on log-derivative statistics of natural scenes
Author Affiliations +
Abstract
We propose an efficient blind/no-reference image quality assessment algorithm using a log-derivative statistical model of natural scenes. Our method, called DErivative Statistics-based QUality Evaluator (DESIQUE), extracts image quality-related statistical features at two image scales in both the spatial and frequency domains. In the spatial domain, normalized pixel values of an image are modeled in two ways: pointwise-based statistics for single pixel values and pairwise-based log-derivative statistics for the relationship of pixel pairs. In the frequency domain, log-Gabor filters are used to extract the fine scales of the image, which are also modeled by the log-derivative statistics. All of these statistics can be fitted by a generalized Gaussian distribution model, and the estimated parameters are fed into combined frameworks to estimate image quality. We train our models on the LIVE database by using optimized support vector machine learning. Experiment results tested on other databases show that the proposed algorithm not only yields a substantial improvement in predictive performance as compared to other state-of-the-art no-reference image quality assessment methods, but also maintains a high computational efficiency.

1.

Introduction

Rapid advances in digital image processing technology have revolutionized the way in which images are captured, stored, transmitted, and accessed. A crucial task for any image processing system is the ability to assess image quality in a manner that is consistent with human subjective judgment. To address this need, numerous image quality assessment (IQA) algorithms have been proposed over the past several decades, using a wide variety of image analysis techniques. Based on the availability of a reference image, these algorithms can be divided into three main categories: full-reference (FR), reduced-reference (RR), and no-reference (NR) IQA.

In this paper, we address the task of NR IQA, in which no reference image is available to the IQA algorithm. Although humans can easily assess the quality of a distorted image in an NR setting, this task remains extremely challenging for an algorithm. The vast majority of NR IQA algorithms have been designed for specific distortion types. These distortion-specific methods often assume that a particular distortion type is known, such as additive white noise (WN) or Gaussian blur (Gblur) (e.g., Refs. 12.3. to 4), or JPEG/JPEG2000 compression (e.g., Refs. 56.7. to 8). Then, based on the assumed distortion type, specific distortion-related features are extracted to assess image quality.

Another class of NR IQA methods has been recently developed which does not require knowledge of the distortion type, but instead evaluates image quality assuming that the distorted image shares some properties with those in the training database. Machine-learning techniques are used to develop a mapping from the input features to the target quality scores in the training database, and then the quality of the distorted image is estimated based on the mapping. Such non-distortion-specific approaches usually follow (1) a learning-based approach; (2) a natural scene statistics-based (NSS-based) approach (e.g., Refs. 910. to 11) usually rely on a large number of quality-related features that are designed based on the physical properties of image distortions. Assuming that image quality can be measured by various features (e.g., wavelet coefficient magnitude and phase, texture statistics), and assuming that certain features that are effective for one particular distortion type may not be applicable to others, various regression models are learned to combine these features such that high performance is always achieved for a variety of distortions. In comparison, the NSS-based NR IQA approaches (e.g., Refs. 1213.14. to 15) also extract quality-related features, but these features are based on a statistical representation of the image’s coefficients in the original/transformed domain under certain distribution models. Quality is estimated based on the extent to which these statistics match those of natural images. In Sec. 2, we provide a more detailed review of existing NR IQA algorithms.

Despite the difference of the statistical models and image transforms employed, these NSS-based NR IQA algorithms share a common thread: they estimate image quality based on single-domain properties. However, we argue that information from both the spatial domain and the frequency domain can play important roles in representing distortions. For instance, it is well known that a blurred image will lose some sharp edges in the spatial domain, which means a reduction in the variance of edge pixel values throughout the image, while in the frequency domain, blurring results in a reduction of the high-frequency components. As we have previously demonstrated in Ref. 16, the quality of a blurred image can be estimated by utilizing both the spatial and spectral properties. We argue that this dual-domain approach can be useful for general NR IQA.

In this paper, we present an NSS-based non-distortion-specific approach called derivative statistics-based quality evaluator (DESIQUE), which builds upon BRISQUE15 by utilizing statistical features measured in both the spatial and spectral domains. Our work draws on the concept of log-derivative statistics proposed by Huang and Mumford,17 and on the concepts proposed by Ruderman.18 We employ combined frameworks consisting of both one- and two-stage frameworks that are commonly used in NSS-based NR IQA algorithms to predict image quality.

For the spatial domain analysis, images are first decorrelated through local mean subtraction and divisive normalization18 to obtain normalized pixel values [also called mean-subtracted contrast-normalized (MSCN) coefficients]. Then, these normalized pixel values are statistically modeled in two ways: (1) pointwise-based statistics for single pixel values (following Ref. 15) and (2) pairwise-based log-derivative statistics for the relationship of pixel pairs. Figure 1 gives an example of three blurred images [Fig. 1(a)] and their computed histogram profiles [Fig. 1(b)] based on the two above-mentioned statistics. As shown in the figure, different distortion levels give rise to different histogram shapes in the spatial domain. Image3 is the most blurred, and thus we see its spatial-domain histograms are most narrow at the base and have sharp peaks. In comparison, the histograms for image1 appear to be less peaked, a result of the much sharper content in edge areas. Image2 is somewhat in between: it is clearly sharper than image3, but more blurred than image1, and thus its histogram’s shape is somewhere in between the other two. This figure and other corresponding histograms for another distortion type shown later (see Sec. 3) demonstrate that statistics in the spatial domain can effectively indicate different amounts of distortions.

Fig. 1

Three blurred images from the LIVE database19 with different histogram profiles in the spatial domain. D1 through D7 denote the log-derivatives computed at six orientations with various spatial contexts (see Sec. 3.2.1 for details about computing the spatial-domain-based statistics).

JEI_22_4_043025_f001.png

However, although the spatial-domain-based statistics can be a potential measure of image quality, they do not take into account spectral properties, which can also be a good indicator of perceived distortion. For example, two images can have very similar statistics in the spatial domain but appear to be of different visual quality. We demonstrate this assertion in Fig. 2, which shows two images corrupted by different amounts of high-frequency noise. As shown in the figure, the computed histograms of the spatial-domain-based log-derivative statistics appear to be quite similar. However, when applying such statistics to the high-frequency components of the image in the frequency domain, the resulting histograms can be distinct.

Fig. 2

Two high-frequency noise-corrupted images from the TID database20 with different visual qualities. Notice that the spatial-domain-based log-derivative statistics have similar histogram profiles, but the histogram shape varies in the frequency domain. (These histograms were generated from measure D6; see Sec. 3.2.2 for details about computing the frequency-domain-based statistics.)

JEI_22_4_043025_f002.png

To obtain the spectral properties contained in an image, we utilize log-Gabor filters. Numerous studies have shown that the response properties of simple cells in primary visual cortex can be well-modeled by Gabor filters,21,22 which offer simultaneous localization of spatial and frequency information. Log-Gabor filters overcome the frequency distribution problem encountered by the classical Gabor filters by maintaining Gaussian passbands on a logarithmic frequency scale.23 In our implementation, we decompose an image into several spatial frequency bands at two orientations (horizontal and vertical) and compute the log-derivative statistics of the high-spatial frequency log-Gabor coefficients. As shown in Figs. 1 and 2, the histograms computed in both domains can be fitted by a two-parameter generalized Gaussian distribution (GGD), which has been demonstrated in Ref. 24 to be an effective model of the subband statistics of natural images for RR IQA. The parameters of these GGD fits are utilized as the DESIQUE features, which are then fed into combined frameworks to estimate the image’s quality.

Although DESIQUE is built upon BRISQUE, there are key differences between the two algorithms. First, the two algorithms use different types of analyses. As opposed to the paired-product-based analysis used in BRISQUE, the log-derivative-based analysis used in DESIQUE provides the latter with (1) greater sensitivity to local contrast changes and (2) distributions that are symmetric and more easily modeled. To demonstrate these facts, Fig. 3 shows three images [Fig. 3(a)] with different contrast, as well as their corresponding paired product and log-derivative coefficient histograms in four orientations [Fig. 3(b)]. As shown in this figure, the computed histograms of log-derivative values in the spatial domain yield profiles that can be used to distinguish between the three images, whereas the histograms of the paired-product values do not provide such discriminability. Also, as shown in Fig. 3, the MSCN coefficients have a symmetric distribution when modeled by log-derivative statistics, whereas the paired product statistics give rise to asymmetric distributions. Thus, when using log-derivatives, fewer parameters are required for the histogram fitting. This reduction in parameters (and thus in the number of features) allows for the computational resources to be devoted to the frequency-domain analysis to improve the predictive performance without penalizing the computational (runtime) performance.

Fig. 3

Three images with different contrast and their corresponding paired product/log-derivative coefficient histograms in four orientations. Note that image3 is from the CSIQ database;25 image1 and image2 are generated by adjusting the local contrast in image3 using Adobe Photoshop CS3 software.

JEI_22_4_043025_f003.png

Second, DESIQUE uses a two-domain approach, whereas BRISQUE uses a single-domain approach. As mentioned previously, BRISQUE operates in the spatial domain only, while we have shown in Figs. 1 and 2 that both the spatial and spectral properties can be important indicators of perceived distortion. It is well known that some distortions are more easily detected in the frequency domain rather than in the spatial domain and vice versa. For example, the fast-fading distortion in the LIVE database is more detectable in the spatial domain, whereas different types of noise are more detectable in the frequency domain. As we will demonstrate, by adopting a dual-domain approach, the proposed DESIQUE algorithm can achieve not only better performance on various distortions across different databases, but also relatively higher robustness to images distorted by different types of noise (see Sec. 4.4).

The main contributions of this work are as follows: (1) DESIQUE analyzes distorted images in both the spatial and frequency domains by extracting NSS-based quality-related features, and the ultimate quality score is predicted based on combined frameworks. We show in this work that distortions affect image statistics in both domains and both can play important complementary roles in assessing image quality. (2) We propose to use the seven types of log-derivative statistics to model the relationship between neighboring pixel/coefficient values, which are sensitive to distortion. All of these statistics have symmetrical distributions and can be characterized by using a two-parameter GGD model. (3) We also show that DESIQUE can be adaptive to images distorted by other types of noise on which it was not trained. Although DESIQUE was trained on additive white Gaussian noise, it still performs well on other types of noise such as additive color noise, spatially correlated noise, and high-frequency noise. (4) DESIQUE not only demonstrates good predictive performance, but also maintains significant computational efficiency; the algorithm can achieve near-real-time performance.

This paper is organized as follows. In Sec. 2, we give a brief review of the previous work on NR IQA. In Sec. 3, we provide details of the DESIQUE algorithm. In Sec. 4, we present results of the proposed algorithm on different image databases and evaluate the performance of different algorithms. General conclusions are provided in Sec. 5.

2.

Previous Work

Current NR IQA algorithms can generally be divided into two categories: (1) those which assume that the distortion type is known, only determining the severity and (2) those which do not consider the prior knowledge of distortion type, but assume that the distortion information of the tested image can be similar to the training database. In this section, we provide a brief review of these NR IQA methods.

2.1.

Distortion-Specific NR IQA Algorithms

2.1.1.

JPEG IQA

JPEG artifacts mainly contain blocking, blurring, and aliasing due to the quantization of discrete cosine transform (DCT) coefficients. Therefore, the general approach to JPEG IQA is to measure the blocking artifacts or edge strength at block boundaries to estimate image quality.

In Ref. 5, the blocky image is modeled as a non-blocky image distorted by a pure blocky signal, and then the the blocking artifact is measured by detecting and estimating the power of the blocky signal. In Ref. 6, blocking artifacts are measured in DCT domain by modeling the blocking artifacts as two-dimensional (2-D) step functions. This information is then used with a human-visual-system (HVS)-based measurement of blocking artifact severity to esitimate quality. In Ref. 26, image blockiness is measured by detecting the low-amplitude step edges and estimating the edge parameters using a Hermite transform. In Ref. 7, image blockiness is measured based on the average differences across block boundaries and the activity of the image signal. The activity is measured using two factors: (1) the average absolute difference between in-block image samples and (2) the zero-crossing rate. Other JPEG-based NR IQA algorithms have used the DCT (e.g., Refs. 27 and 28), discrete Fourier transform (DFT) (e.g., Ref. 29), Sobel operator (e.g., Refs. 30 and 31), selected gradient and uniformity measures (e.g., Ref. 32), HVS models (e.g., Ref. 33), and machine learning (e.g., Ref. 34). In summary, each algorithm measures blocking artifacts in JPEG compressed images by using either the transformed domain properties or the fundamental characteristics introduced by quantization to the signal.

2.1.2.

JPEG2000 IQA

For JPEG2000-compressed images, the predominant artifacts introduced are blurring and ringing. Algorithms for assessing quality of this type of image are mainly focusing on measuring the amount of blurring or edge spreading by using edge-detection techniques. For example, Li et al.35 proposed a principal component analysis–based method to assess the quality of JPEG2000 image. First, by viewing all edge points in an image as distorted or undistorted, local features are extracted at each of the edge points, which indicate both blurring and ringing. Then, a model is built to describe the relationship between these local features and a local distortion metric through the probability of the edge point being distorted versus undistorted. The image’s quality is estimated based on these edge points (a similar method can be found in Ref. 36). Sazzad et al.37,38 proposed to use pixel distortion and edge information for the quality evaluation of JPEG2000 images. Their algorithms assume that the HVS is very sensitive to edge information and any kinds of artifacts that create pixel distortion from neighborhood pixels.

Other methods for NR JPEG2000 IQA may utilize certain kinds of transform such as the wavelet transform or DCT. Statistical regularizations of the coefficients are then used to assess image quality. For instance, Sheikh et al.8 found that wavelet coefficient distributions of JPEG2000 images provide a good indication of the loss of visual quality. Thus, by computing features from all wavelet subbands and applying a nonlinear combination, a weighted average of the transformed features is used for quality prediction. Zhou et al.39 proposed to evaluate JPEG2000 images based on three steps: (1) image is divided into blocks, among which textured blocks are employed for quality prediction; (2) projections of wavelet coefficients between adjacent scales with the same orientation are utilized to measure the positional similarity; and (3) a general regression neural network is adopted to conduct quality prediction according to features from (1) and (2). Zhang et al.40 utilized kurtosis in the DCT domain to blindly appraise quality of JPEG2000 images based on three measurements: frequency band-based one-dimensional (1-D) kurtosis, basis function-based 1-D kurtosis, and 2-D kurtosis. The method is simple (without edges/features extraction), parameter free, and correlates well with subjective quality scores.

2.1.3.

Blur/noise IQA

Researchers have also developed various methods to measure image quality when other distortion types and combinations are present. Since the most common distortion types in images are noise and blur (due, e.g., to the transmission, storage, and reconstruction), numerous algorithms have been proposed to estimate noise degradation and blurring. For instance, Corner et al.41 developed a noise estimation technique using data masking. Li et al.1 proposed an NR IQA technique by joining three different NR measures for three different image distortion types (blur, noise, and block/ringing artifacts). Brandao and Queluz3 proposed an NR IQA algorithm for estimating quantization noise introduced by lossy encoding such as JPEG or MPEG based on the statistics of DCT coefficients. Cohen and Yitzhaky4 presented an NR method to identify noise and blur by assuming that common statistics of natural images can be obtained from their power spectra. Vu and Chandler16 proposed a spectral and spatial measure of local perceived sharpness in natural images by using the slope of the magnitude spectrum and the total spatial variation.

2.2.

Non-Distortion-Specific NR IQA Algorithms

2.2.1.

Learning-based IQA

Learning-based NR IQA methods estimate quality by learning a mapping from a high-dimensional feature space to a scalar quality value. The feature space contains a number of quality-related features that are designed to capture relevant factors that affect image quality. Different techniques on feature extraction and regression lead to various algorithms.

Tong et al.42 proposed a learning-based method that estimates quality by learning from training examples that contain both the high- and low-quality classes. A binary classifier is built and the quality metric of the tested image is denoted by the extent to which it belongs to these two classes. Tang et al.9 proposed a learning-based blind image quality measurement that first extracts low-level image features from natural image measure and texture statistics. Then, via a machine-learning framework, the algorithm learns a mapping from these features to the subjective quality score. Li et al.43 developed an NR IQA algorithm based on a general regression neural network (GRNN). The features extracted include the mean value of a phase congruency image, the entropy of the phase congruency image, the entropy of the distorted image, and the gradient of the distorted image. Quality is estimated by learning the functional relationship between these features and subjective mean opinion scores using the GRNN. In Refs. 10 and 11, Ye et al. used the visual codebook to quantize Gabor-filter-based features and then learned a mapping from the quantized feature space to image quality scores using either an example-based method or a support vector machine (SVM). This method extracts patch-level features and has the potential to be used in real-time applications with parallel implementation.

2.2.2.

NSS-based IQA

NSS-based NR IQA methods assume that some statistical properties of natural scenes will remain across different reference images, but will change significantly in the presence of distortions. Thus, by measuring how the statistical properties of a distorted image deviate from those of a typical natural image, the distortion types as well as the image quality can be estimated. Typical NSS-based NR algorithms include BIQI,12 BLIINDS-II,14 DIIVINE,13 BRISQUE,15 and NIQE.44

BIQI12 is a wavelet-based NR IQA method that extracts image features by modeling the wavelet subband coefficients based on a two-parameter GGD model. It utilizes a discrete wavelet transform with three scales and three orientations using Daubechies 9/7 wavelet basis. A total of 18 wavelet-based statistical features are extracted for each image.

BLIINDS-II14 measures image quality based on the statistics of DCT coefficients of local image patches via two stages. In the first stage, non-overlapping small image blocks are prepared and statistical features are extracted. In the second stage, a multivariate GGD model is used to perform quality prediction. In order to perform prediction in the second stage, BLIINDS-II requires training, during which a Bayesian probabilistic model is obtained containing all information in the training database.

DIIVINE13 operates based on a steerable pyramid transform, from which the coefficient statistics of the distorted image are used as the feature values for IQA. The algorithm employs a two-stage framework: distortion identification followed by distortion-specific quality assessment, and 88 features are extracted to perform the IQA task through SVM learning.

BRISQUE15 is another NSS-based NR IQA method that operates in the spatial domain. The underlying features derive from the empirical distribution of the local normalized luminance values, as well as their neighboring products under a spatial NSS model. No required transformation and a small feature number allow the algorithm to run almost in real-time. Despite its simplicity, the algorithm represents the currently best-performing NSS-based NR IQA method.

The recently developed NIQE44 algorithm is the first NSS-based approach that does not require training on the distorted images. The algorithm extracts local image features based on the work in Ref. 15, and the quality of a given test image is expressed as the distance between the multivariate Gaussian fit of the NSS features extracted from the test image and from the corpus of natural images. Although NIQE does not perform as well as DIIVINE or BRISQUE, it demonstrates the possibility to extend current NSS-based NR IQA to a no-distortion-trained approach.

3.

Algorithm

The proposed DESIQUE algorithm is based on the assumption that log-derivative-based statistical properties of natural images in both the spatial and frequency domains are generally consistent across reference images, but vary significantly in the presence of distortions. In this way, it is possible to measure deviations in these expected (natural) statistical features as proxy measures of quality degradation. DESIQUE extracts log-derivative-based statistical features in two domains using seven derivative types and employs combined frameworks consisting of one used in BRISQUE (one-stage framework) and the other one in DIIVINE (two-stage framework).

3.1.

Log-Derivative Statistics

The derivative statistics of natural images were first studied in Ref. 17, in which the differences of gray-level values between one pixel and its neighboring pixels were considered as the derivatives. Motivated by the work in Refs. 45 and 15, derivatives between pairs of pixels can have six orientations: horizontal (H), vertical (V), main-diagonal (MD), secondary-diagonal (SD), horizontal-vertical (HV), and combined-diagonal (CD). Based on the six orientations and the spatial location of pixel values, we propose seven types of derivatives defined as

Eq. (1)

H:xI(i,j)=I(i,j+1)I(i,j),

Eq. (2)

V:yI(i,j)=I(i+1,j)I(i,j),

Eq. (3)

MD:xyI(i,j)=I(i+1,j+1)I(i,j),

Eq. (4)

SD:yxI(i,j)=I(i+1,j1)I(i,j),

Eq. (5)

HV:xyI(i,j)=I(i1,j)+I(i+1,j)I(i,j1)I(i,j+1),

Eq. (6)

CD1:cxcyI(i,j)1=I(i,j)+I(i+1,j+1)I(i,j+1)I(i+1,j),

Eq. (7)

CD2:cxcyI(i,j)2=I(i1,j1)+I(i+1,j+1)I(i1,j+1)I(i+1,j1),
where I(i,j) represents the pixel value (or coefficient) corresponding to spatial location i, j. Note that Eqs. (1) to (4) and Eq. (6) are defined on neighboring pixels/pixel pairs and Eqs. (5) and (7) are defined on a patch size of 3×3.

In order to efficiently model natural images/image subbands by using the distribution of derivative statistics, we compute the logarithm of each pixel’s value to create new images/image subbands J by

Eq. (8)

J(i,j)=log[I(i,j)+K],
where K is a small constant that prevents I(i,j) from being zero. Thus, based on Eqs. (1) to (8), we define seven types of log-derivatives as

Eq. (9)

D1:xJ(i,j)=J(i,j+1)J(i,j),

Eq. (10)

D2:yJ(i,j)=J(i+1,j)J(i,j),

Eq. (11)

D3:xyJ(i,j)=J(i+1,j+1)J(i,j),

Eq. (12)

D4:yxJ(i,j)=J(i+1,j1)J(i,j),

Eq. (13)

D5:xy=J(i1,j)+J(i+1,j)J(i,j1)J(i,j+1),

Eq. (14)

D6:cxcyJ(i,j)1=J(i,j)+J(i+1,j+1)J(i,j+1)J(i+1,j),

Eq. (15)

D7:cxcyJ(i,j)2=J(i1,j1)+J(i+1,j+1)J(i1,j+1)J(i+1,j1).

In the following sections, we will show that histograms computed from these seven types of log-derivative statistics are effective in modeling natural images; they exhibit consistent profiles across different scenes but vary significantly in the presence of distortions. We, thus, extract statistical features that capture changes in these profiles.

3.2.

Log-Derivative Statistics–Based Features

As mentioned in Sec. 2, recent work has focused on modeling natural-scene statistics either in the spatial domain15 or in a transform domain (see, e.g., Refs. 12 to 14). However, we believe that the visual quality of an image can be influenced by the statistical information contained in both the spatial and frequency domains. Thus, the proposed DESIQUE features consist of two parts: the spatial domain features and the frequency domain features, and we will demonstrate that these statistical features are generally consistent across different images but change in the presence of distortion. Figure 4 shows a block diagram illustrating how DESIQUE features are extracted in these two domains.

Fig. 4

A block diagram of DESIQUE feature extraction. Note that the filter band coefficients will only contain the first scale of the log-Gabor filter subbands corresponding to the high-frequency components of the image. LP denotes low-pass filter.

JEI_22_4_043025_f004.png

3.2.1.

Modeling log-derivative statistics in the spatial domain

The features extracted in the spatial domain consist of two types: (1) pointwise-based statistics for single pixel values (following Ref. 15) and (2) pairwise-based log-derivative statistics for the relationship of pixel pairs. Specifically, given an image I(i,j), we first compute locally normalized pixel values via local mean subtraction and divisive normalization18 defined as

Eq. (16)

I^=I(i,j)μ(i,j)σ(i,j)+C,
where i1,2,,M, j1,2,,N are spatial indices; M, N are the image height and width, respectively; and C=1 is a constant that prevents division by zero. The quantities μ(i,j) and σ(i,j) are defined as

Eq. (17)

μ(i,j)=k=KKl=LLωk,lIk,l(i,j),

Eq. (18)

σ(i,j)=k=KKl=LLωk,l[Ik,l(i,j)μ(i,j)]2,
where ω={ωk,l|k=K,,K,l=L,,L} is a 2-D circularly symmetric Gaussian weighting function sampled out to three standard deviations and rescaled to unit volume. As in Ref. 15, we also define K=L=3.

According to Ref. 15, the MSCN coefficients I^(i,j) can be modeled by a zero-mean GGD given by

Eq. (19)

f(x;α,σ2)=α2βΓ(1/α)exp[(|x|/β)α],
where β=σΓ(1/α)Γ(3/α) and Γ(x)=0tx1etdt,x>0 is the gamma function. The parameter α controls the general shape of the distribution and σ controls the variance. We estimate the two-parameter GGD model by using the moment-matching-based approach proposed in Ref. 46, and these two values form the first set of features in the spatial domain that will be used to capture image distortion, denoted by α0 and σ0.

The other set of features we extract in the spatial domain are formed by modeling the relationship of neighboring MSCN coefficient pairs based on the log-derivative statistics. In Ref. 15, statistical relationships between neighboring pixels are modeled by using the empirical distributions of pairwise products of neighboring MSCN coefficients along four orientations. These paired products can be parameterized by an asymmetric GGD.

In our implementation, we model the relationship between two adjacent MSCN coefficients by using the seven types of log-derivatives previously defined by Eqs. (9) to (15). Here J(i,j)=ln(|I^(i,j)|+K) and K=0.1 is a constant that prevents I^(i,j) from being zero. Under the Gaussian coefficient model, and assuming that the MSCN coefficients are zero mean and unit variance, these log-derivative values also follow a generalized Gaussian distribution and thus their parameters (α, σ) can be estimated by using the method proposed in Ref. 46. These 14 parameters form the second part of DESIQUE features in the spatial domain. In all, we extract 16 spatial domain features at one image scale denoted by

fspatial=[α0,αD1,αD2,αD3,αD4,αD5,αD6,αD7,σ0,σD1,σD2,σD3,σD4,σD5,σD6,σD7].

Here, D1 through D7 denote seven types of log-derivatives. In order to visualize how the aforementioned two statistics of MSCN coefficients in the spatial domain vary as a function of distortions, Fig. 5 plots their corresponding histograms for five reference images (shown in Fig. 6), and Fig. 7 plots the histograms for five distorted versions of one reference image (image sailing2, shown in Fig. 8). Notice that both the pointwise-based statistics and the pairwise-based log-derivative statistics of MSCN coefficients exhibit almost consistent profiles; they are largely independent of the particular reference image but change significantly in the presence of distortions.

Fig. 5

Histogram of mean-subtracted contrast-normalized (MSCN) coefficients (a) and the seven types of log-derivative statistics [(b) to (h)] for each of the five reference images (shown in Fig. 6). Notice that the histograms exhibit consistent profiles that are largely independent of the particular reference image from which log-derivative statistics are computed.

JEI_22_4_043025_f005.png

Fig. 6

Five reference images from the LIVE database19 used to demonstrate the consistency of natural scene statistics in both the spatial and frequency domains. Top row: Images ocean, stream. Bottom row: Images sailing2, sailing3, and statue.

JEI_22_4_043025_f006.png

Fig. 7

Histogram of MSCN coefficients (a) and their seven types of log-derivative statistics [(b) to (h)] for each of the five distorted versions of image sailing2 (shown in Fig. 8). Distortions from the LIVE database—JPEG2000 (JP2K) and JPEG compression (JPEG), additive white noise (WN), Gaussian blur (Gblur), and a Rayleigh fast-fading channel simulation (FF). Notice that the distortions tend to affect the peakiness of the characteristic profile observed for the reference images.

JEI_22_4_043025_f007.png

Fig. 8

Distorted versions of image sailing2. From left to right: Fast-fading, Gaussian blur, JPEG2000 compression, JPEG compression, and Gaussian white noise.

JEI_22_4_043025_f008.png

3.2.2.

Modeling log-derivative statistics in the frequency domain

Most distortion types have a pronounced effect on the higher-frequency components, resulting in considerable magnitude differences in the high-frequency band. To capture this change, we first decompose an image by using a log-Gabor filter defined by

Eq. (20)

Gs,o(ω,θ)=exp{[log(ω/ωs)]22[log(σs/ωs)]2}×exp[(θμo)22σo2],
where Gs,o is the log-Gabor filter denoted by spatial scale index s and orientation index o. The parameter ω is the normalized radial frequency and θ is the orientation. The decomposition is computed at the highest frequency scale (s=1) and two orientations (o=1,2). We set the center frequency of that scale to be 1/3, corresponding to ω1=2/3. The value of σ1/ω1, which determines the bandwidth, is set to be 0.975. These values result in an approximately 1.5 octave bandwidth. Two orientations are computed using μo=(o1)π/2, resulting in orientations of 0 and 90 deg.

Next, we apply the log-derivative statistics to model the log-Gabor filter subband coefficients. Note that we only apply six types of them corresponding to neighboring pixels/pixel pairs previously defined in Eqs. (9) to (12) and (14) to (15). Here, J(i,j)=ln[|g(i,j)|+K], where |g(i,j)| represents the coefficient magnitude and K=0.1 is a constant that prevents g(i,j) from being zero. Again, we fit the histogram of log-derivative statistics for each of these subbands with a generalized Gaussian distribution model. The parameters (α,σ) are estimated and yield 24 DESIQUE features (2subbands×6derivative types×2parameters/derivativetype) extracted in the frequency domain at the original scale.

ffrequency=[αD1,0°,αD2,0°,αD3,0°,αD4,0°,αD6,0°,αD7,0°,σD1,0°,σD2,0°,σD3,0°,σD4,0°,σD6,0°,σD7,0°,αD1,90°,αD2,90°,αD3,90°,αD4,90°,αD6,90°,αD7,90°,σD1,90°,σD2,90°,σD3,90°,σD4,90°,σD6,90°,σD7,90°].

Here, D1 through D4, D6, and D7 denote six types of log-derivatives; 0 deg and 90 deg denote the horizontal and vertical log-Gabor filter subbands, respectively.

To illustrate how the log-derivative statistics of the filter responses behave as a function of distortions, Fig. 9 plots the corresponding histograms for five reference images (shown in Fig. 6) as well as the histograms for five distorted versions of one reference image (image sailing2, shown in Fig. 8) as a comparison. Notice that for both orientations of the log-Gabor filters, the log-derivative statistics of the subband coefficients exhibit consistent profiles that are largely independent of the particular reference image, but change significantly in the presence of distortions. As demonstrated in Fig. 9, these features extracted in the frequency domain are effective at identifying distortion and thus can be used for measuring image quality.

Fig. 9

Histograms of the log-derivative statistics of the log-Gabor filter subband coefficients for five reference images in the LIVE database,19 and the five distorted versions of one of the reference images (image sailing2 shown in Fig. 8). D1 to D4 and D6 to D7 correspond to the six log-derivative types defined by Eqs. (9) to (12) and (14) to (15). Distortions from the LIVE database: JPEG2000 (JP2K) and JPEG compression (JPEG), additive white noise (WN), Gaussian blur (Gblur), and a Rayleigh fast-fading channel simulation (FF). Notice that for the five reference images, the histograms exhibit consistent shapes, but they vary significantly when distortions are present.

JEI_22_4_043025_f009.png

Since images are naturally multiscale and distortions affect image structure across scales, following from Ref. 15, we extract all spatial features at two scales: the original image scale and a low-pass downsampled (by a factor of 2) scale. The frequency features are extracted based on the six types of log-derivatives at the original scale, and one type (D7) at the low-pass downsampled scale. Thus, a total of 60 features (32 spatial features, 24 frequency features at the original scale, and 4 at the downsampled scale) are used to identify distortions and to perform distortion-specific quality assessment.

3.3.

Quality Evaluation

The final step is to map the feature vectors to estimates of image quality. For this task, we employ the existing learning-based approaches commonly used in other NSS-based NR IQA algorithms. We have specifically tested and reported the results of the following approaches: (1) the one-stage framework used in BRISQUE; (2) the two-stage framework used in DIIVINE; (3) a winner-take-all-based variation of (2); and (4) a combination of (1) and (2).

The one-stage framework employs a support vector regression (SVR) to train a regression model, which directly maps the feature vector to an associated quality score, denoted as DESIQUE-I. For the two-stage framework, DESIQUE features are used to perform (1) distortion identification and (2) distortion-specific quality assessment. As in Ref. 13, the distortion identification stage employs support vector classification (SVC) to measure the probability that the distortion in the distorted image falls into one of n distortion classes, and the distortion-specific quality assessment stage employs an SVR to obtain n regression modules, each of which maps the feature vectors to an associated quality score. Note that because each module is trained specifically for each distortion, these regression modules function as distortion-specific quality estimators. Let p denote this n-dimensional vector of probabilities and q denote the n-dimensional vector of estimated qualities obtained from these n regression modules; then the overall estimated quality, denoted by DESIQUE-II, is computed as follows:

Eq. (21)

DESIQUE-II=i=1np(i)q(i),
where p(i) and q(i) denote elements of p and q, respectively.

We have also investigated a variation of the two-stage framework in which the overall quality estimate is based only on the distortion-specific quality estimate corresponding to the greatest probability. This approach, denoted by DESIQUE-hp, is computed as follows:

Eq. (22)

DESIQUE-hp=q(im),im=argmax{p(i)}.

Another variation that we have investigated, and which is the approach that generally yields the best results (as we will demonstrate shortly) and is thus the approach we recommend, is to use the minimum of the one-stage and two-stage outputs. This approach, denoted by DESIQUE, is computed as follows:

Eq. (23)

DESIQUE=min{DESIQUE-I,DESIQUE-II}.

Smaller values of DESIQUE denote predictions of greater image quality. Note that both the SVC and SVR require training. As we demonstrate next, training these stages on images from one database can yield excellent predictive performance on similarly distorted images from other databases.

4.

Results

In this section, the performance of DESIQUE is analyzed in terms of its ability to predict subject ratings of image quality.

4.1.

Training

We trained our models on the LIVE database,19 which contains 29 reference images and 779 distorted images that span various distortion categories: JPEG compression, JPEG2000 compression, white noise, Gblur, and a Rayleigh fading channel (fast fading). Each distorted image has an associated score that indicates a representative of the perceived quality of the image. We use the LIBSVM package47 to implement the training. To improve predictive performance, optimal radial basis function kernel parameters were used for both classification and regression.

4.2.

Testing

To assess its predictive performance, four databases of subjective image quality were used: (1) the LIVE database19 (note that the LIVE database was used for both training and cross-validation test), (2) the CSIQ database,25 (3) the TID database,20 and (4) the Toyama database. The CSIQ database consists of 30 original images distorted using six different types of distortions at four to five different levels of distortion, and the overall ratings are given by 25 different observers in the form of differential mean opinion scores (DMOS). The TID database consists of 25 reference images and 1700 distorted images over 17 distortion types, ratings of which are presented as mean opinion scores (MOS). Among these 25 reference images, only 24 are natural images, and we test our algorithm only on these 24 images. The Toyama database contains 14 original images and 168 distorted versions that cover only two distortion types: JPEG and JPEG2000 compression. The subjective ratings in the Toyama database are provided in the term of MOS values. We compared DESIQUE with various FR and NR quality assessment methods for which code is publicly available. The five FR methods were peak signal-to-noise ratio (PSNR),48 SSIM,49 multiscale structure similarity (MS-SSIM),50 visual information fidelity (VIF),51 and most apparent distortion (MAD).52 The three NR methods were DIIVINE,13 BLIINDS-II,14 and BRISQUE,15 all of which are NSS-based and trained on LIVE. Since DESIQUE is trained on the LIVE database containing five distortion types, we tested and compared these algorithms only on those distortion types on which the algorithms have been trained: JPEG compression, JPEG2000 compression, WN, and Gblur.

Before evaluating the performance of a particular quality assessment method on a particular database, we applied a logistic transform to bring the predicted values on the same scales as the DMOS/MOS values. The logistic transform recommended by Video Quality Experts Group53 is a four-parameter sigmoid given by

Eq. (24)

f(x)=τ1τ21+exp(xτ3|τ4|)+τ2,
where x denotes the raw predicted score and τ1, τ2, τ3, and τ4 are free parameters selected to provide the best fit of the predicted scores to the MOS/DMOS values. Note that the Spearman rank-order correlation coefficient (SROCC) relies only on the rank-ordering and is thus unaffected by the logistic transform due to the fact that f(x) is a monotonic function of x that does not change the rank-order.

Three criteria were used to measure the prediction monotonicity and prediction accuracy of each algorithm: (1) SROCC, (2) the Pearson linear correlation coefficient (CC), and (3) the root mean square error (RMSE) after nonlinear regression. As recommended in Ref. 53, the SROCC serves as a measure of prediction monotonicity, while the CC and RMSE serve as measures of prediction accuracy. A value close to 1 for SROCC and CC, and 0 for RMSE, indicates good predictive performance.

Two additional criteria were used to measure the prediction consistency of each algorithm: (1) outlier ratio and (2) outlier distance (OD).52

4.3.

Cross-Validation Test on LIVE

In this subsection, we performed a cross-validation test on LIVE by splitting the database into two non-overlapping sets—an 80% subset of the database for training and an extra 20% subset for testing. We compared with three NR and three FR IQA methods in terms of median SROCC and CC values computed over 1000 trials. For comparison, we also include the DESIQUE result of using only one-stage framework, two-stage framework, and two-stage framework with regression module corresponding to the higher probability, denoted by DESIQUE-I/II/hp, respectively. The results are shown in Table 1.

Table 1

Median Spearman rank-order correlation coefficient (SROCC) and correlation coefficient (CC) values across 1000 train-test combinations on the LIVE database. Italicized entries denote no-reference image quality assessment (NR IQA) algorithms; others are full-reference (FR) IQA algorithms.

JPEG2000JPEGAdditive white noise (WN)BlurRayleigh fast-fading channelALL
SROCCPSNR0.86460.88310.94100.75150.87360.8636
SSIM0.93890.94660.96350.90460.93930.9129
MS-SSIM0.96270.97850.97730.95420.93860.9535
BLIINDS-II0.93230.93310.94630.89120.85190.9124
DIIVINE0.91230.92080.98180.93730.86940.9250
BRISQUE0.91390.96470.97860.95110.87680.9395
DESIQUE-I0.93540.96670.98490.94480.86120.9406
DESIQUE-II0.92750.96830.98350.93820.86610.9406
DESIQUE-hp0.92360.96900.98220.93280.85870.9374
DESIQUE0.93590.96930.98580.94820.86720.9437
CCPSNR0.87620.90290.91730.78010.87950.8592
SSIM0.94050.94620.98240.90040.95140.9065
MS-SSIM0.97460.97930.98830.96450.94880.9511
BLIINDS-II0.93860.94260.96350.89940.87890.9164
DIIVINE0.92330.93480.98660.93700.89160.9270
BRISQUE0.92290.97350.98510.95060.90300.9424
DESIQUE-I0.94850.97820.99100.95730.87290.9429
DESIQUE-II0.94160.98010.98960.94960.88870.9430
DESIQUE-hp0.93740.98030.98860.94830.88350.9407
DESIQUE0.94800.98030.99150.95850.88190.9465

In order to evaluate statistical significance, we performed a one-sided t-test54 with 95% confidence level between SROCC values generated by these algorithms across 1000 train-test trials. The results are shown in Table 2, in which “1,” “0,” “1” indicate that the mean correlation of the algorithm in row is statistically superior, equivalent, or inferior to the mean correlation of the algorithm in colunm. Also included are DESIQUE-I, DESIQUE-II, and DESIQUE-hp for comparison.

Table 2

Results of the one-sided t-test performed between SROCC values generated by different measures. “1,” “0,” “−1” indicate that the algorithm in the row is statistically superior, equivalent, or inferior to the algorithm in the column.

PSNRSSIMMS-SSIMBLIINDS-IIDIIVINEBRISQUEDESIQUE-IDESIQUE-IIDESIQUE-hpDESIQUE
PSNR0111111111
SSIM1011111111
MS-SSIM1101111111
BLIINDS-II1110111111
DIIVINE1111011111
BRISQUE1111100011
DESIQUE-I1111100011
DESIQUE-II1111100011
DESIQUE-hp1111111101
DESIQUE1111111110

To demonstrate that DESIQUE features can be used for different distortion identification, Table 3 shows the median and mean classification accuracy of the classifier in the two-stage framework for each of the distortions in the LIVE database, as well as across all distortions. To show the performance consistency of each of the algorithms considered here, Fig. 10 plots the mean and standard deviation of SROCC values across these 1000 trials for each of them.

Table 3

Mean and median classification accuracy across 1000 train-test trials.

Classification accuracy (%)JP2KJPEGWNBLURFFALL
Mean86.197.699.995.682.292.3
Median86.1100.0100.096.783.392.6

Fig. 10

Mean SROCC and standard error bars for various algorithms across the 1000 train-test trials on the LIVE database.

JEI_22_4_043025_f010.png

According to the cross-validation results, DESIQUE performs better than DESIQUE-I and II, which demonstrates that combined frameworks can achieve better performance than using either of them alone. Also, notice that the performance of DESIQUE-I and II are equivalent and are better than DESIQUE-hp. Such a slight dip in the performance of DESIQUE-hp might be attributable to imperfect distortion identification in the first stage of the two-stage framework. Compared with FR IQA methods, DESIQUE is more consistent than PSNR and SSIM, but still remains slightly inferior to MS-SSIM, indicating that there is more space for further improvement. Compared with other three NR IQA algorithms, DESIQUE is statistically the best. However, it is important to note that this cross-validation is not necessarily the best way to gauge the performance of any given algorithm because the amounts of distortions (i.e., the severity levels) are similar in both training and testing sets. Thus, in the following sections, we will show more results of these NR IQA algorithms in assessing image quality on other databases.

4.4.

Performances on Other Databases

In this section, we evaluate the performance of DESIQUE on subsets of the CSIQ, TID, and Toyama databases corresponding to the individual distortion types of JPEG, JPEG2000, Gaussian blurring, and additive Gaussian white noise to demonstrate the capability of the proposed algorithm on assessing quality of images with any distortion types on which it was trained. We also tested DESIQUE on different noise-corrupted images in the TID database to show a degree of adaptiveness to other noise distortion types.

4.4.1.

Overall performance

The overall testing results on CSIQ, TID, and Toyama are shown in Table 4. Italicized entries denote FR algorithms. Notice that all these FR IQA algorithms compared are only applied on those four distortion types that have been trained. The results of the best-performing FR algorithm in each case are italicized and bolded, and the results of the best-performing NR algorithm are bolded.

Table 4

Overall performances of DESIQUE and other algorithms on subsets of the CSIQ, TID, and Toyama databases. Italicized entries denote FR algorithms. Results of the best-performing FR algorithm are italicized and bolded, and results of the best-performing NR algorithm are bolded.

PSNRSSIMMS-SSIMVIFMADDIIVINEBLIINDS-IIBRISQUEDESIQUE
CCCSIQ0.9080.8510.9500.9670.9740.8540.9010.9240.942
TID0.8480.7360.9120.9500.9470.8770.8640.9070.925
Toyama0.6350.7960.8930.9140.9410.6340.7540.8500.871
Average0.8480.8050.9290.9540.9600.8300.8670.9070.926
SROCCCSIQ0.9220.8760.9530.9590.9670.8280.8730.9000.928
TID0.8700.7670.8970.9400.9350.8910.8400.8980.919
Toyama0.6130.7860.8860.9090.9360.6420.7240.8480.872
Average0.8600.8270.9240.9450.9520.8220.8400.8920.917
RMSECSIQ0.1190.1480.0880.0720.0640.1470.1230.1080.095
TID0.8401.0720.6500.4950.5090.7600.7980.6680.600
Toyama0.9660.7570.5640.5070.4250.9680.8640.6600.615
Average0.4830.5450.3450.2760.2650.4680.4560.3750.339
Outlier ratioCSIQ0.2720.3320.2200.1870.1570.3650.3200.2620.243
TID0.7270.7840.7340.6300.6220.7060.7660.6930.659
Toyama0.1900.0950.0660.0300.0240.2140.1550.0660.060
Average0.4120.4480.3690.3120.2930.4570.4440.3770.355
Outlier distanceCSIQ11.33119.3096.4735.0133.24324.56116.31812.6979.812
TID176.350233.982120.95778.15979.194137.472162.692121.259101.156
Toyama20.9049.3963.1721.3971.67316.51712.6854.1565.671

From Table 4 it is clear that compared with other NR IQA methods, DESIQUE performs quite well in predicting quality. It improves upon BRISQUE and is superior to DIIVINE and BLIINDS-II. Further, it even challenges some of the FR IQA methods such as PSNR and SSIM. The last rows of the SROCC, CC, and RMSE results in Table 4 show the average values, where the averages were weighted by the number of distorted images tested in each database. On average, DESIQUE demonstrates the best NR IQA performance. Note that OD is dependent on the dynamic range of the database and, therefore, cannot be used to compare across databases.

Figure 11 shows scatter-plots of logistic-transformed DESIQUE quality predictions versus subjective ratings (MOS or DMOS) on different databases. Although for each database there are some images whose quality scores are predicted far from their true MOS/DMOS values, the proposed DESIQUE algorithm can predict quality for most of them.

Fig. 11

Scatter plots of objective scores predicted by DESIQUE algorithm after logistic transform versus subjective scores on different image databases. Note that the x axis across all three figures represents the predicted value transformed via Eq. (24); the y axis represents the true differential mean opinion score (DMOS) value for the CSIQ database, true MOS value for the TID and Toyama databases.

JEI_22_4_043025_f011.png

In summary, when looking at the overall performance across databases, DESIQUE has a better average performance than other NR IQA methods.

4.4.2.

Statistical significance

A statistical significance test allows one to quantify whether the numerical difference between IQA performance is statistically significant. To assess the statistical significance, an F-test was employed to compare the variances of two algorithms’ residuals (errors in predictions) if the residuals are Gaussian-distributed. The test statistic is the ratio of variance, denoted by F=σA2/σB2, where σA2 and σB2 denote the variance of the residuals from each IQA algorithm. A smaller residual variance indicates a better prediction. Values of F>Fcritical (or F<1/Fcritical) indicate that at a given confidence level, algorithm A has significantly larger (or smaller) residuals than algorithm B, where Fcritical is computed based on the number of residuals and the confidence level. Also, note that if residuals are not Gaussian, then the significance test is often inconclusive. In this paper, a formal test using Jarque-Bera (JB) statistic55 is used to measure the Gaussianity of the residuals. A smaller value of the JB statistic denotes less deviation from Gaussianity, and vice versa.

Table 5 shows the summary for the overall statistical performance of each NR IQA algorithm on subsets of the CSIQ, TID, and Toyama databases. Each entry is the ratio of the residual variance of the algorithm in the row to the algorithm in the column. Bold entries denote that the algorithm in the row has a statistically smaller residual variance than the algorithm in the column with confidence greater than 95%. Italicized entries denote that the algorithm in the row has statistically greater residual variance with the same confidence. Also contained in Table 5 are the JB statistic measures of Gaussianity. Larger values of the JB statistic denote larger deviations from Gaussianity.

Table 5

Statistical significance relationships (ratio of residual variance) between NR IQA algorithms on subsets of the CSIQ, TID, and Toyama databases. A value <1 denotes that the algorithm in the row has smaller residuals than the algorithm in the column; a value >1 denotes larger residuals. See text for details.

DIIVINEBLIINDS-IIBRISQUEDESIQUE
CSIQDIIVINE1.4291.8362.382
BLIINDS-II0.7001.2851.667
BRISQUE0.5450.7781.297
DESIQUE0.4200.6000.771
JBSTAT143.778.5958.6661.0
TIDDIIVINE1.6492.3512.912
BLIINDS-II0.6071.4261.766
BRISQUE0.4250.7011.239
DESIQUE0.3430.5660.807
JBSTAT120.72.35.7104.7
ToyamaDIIVINE1.2552.1532.478
BLIINDS-II0.7971.7161.975
BRISQUE0.4650.5831.151
DESIQUE0.4040.5060.869
JBSTAT2.615.04.096.2

Although DESIQUE has statistically the smallest residual variance on the subsets of the CSIQ and TID databases, note that none of the DESIQUE residuals on these databases can be deemed as Gaussian, which is attributable to several outliers (see Fig. 11). However, the residual variance of DESIQUE is inflated by the existence of these outliers, and thus the fact that it can achieve significantly lower residual variance with these outliers is noteworthy. For the subsets of the Toyama database, the statistical significance is less selective due to the small number of images tested. Nonetheless, DESIQUE still performs competitively well.

4.4.3.

Performance on individual distortion types

We also tested the performance of DESIQUE on subsets of the three testing databases corresponding to the individual distortion types of JPEG, JPEG2000, Gaussian blurring, and additive Gaussian white noise. For this evaluation, we used the same logistic transform computed for each full database and then extracted the transformed scores corresponding to each distortion type. Three other NR IQA methods and five FR IQA algorithms were included to make the comparison. The results are shown in Table 6 in terms of SROCC and CC values; highlighted entries indicate the best performance. All of these NR IQA methods were trained on the 779 distorted images in the LIVE database with the same distortion types.

Table 6

SROCC, CC, and RMSE of DESIQUE and other quality assessment algorithms on different types of distortion on the CSIQ, TID, and Toyama databases. Italicized entries denote NR algorithms. Results of the best-performing FR algorithm are italicized and bolded, and results of the best-performing NR algorithm are bolded.

PSNRSSIMMS-SSIMVIFMADDIIVINEBLIINDS-IIBRISQUEDESIQUE
CC
CSIQJPEG20000.9470.9230.9770.9780.9830.8930.9120.8960.925
JPEG0.8910.9400.9810.9880.9830.6970.9120.9460.973
Gblur0.9250.9000.9590.9740.9760.8980.8970.9280.922
WN0.9530.9260.9470.9610.9560.7860.8970.9380.942
TIDJPEG20000.8850.8750.9740.9710.9820.8790.9190.9060.928
JPEG0.8780.9370.9660.9730.9610.8990.8890.9500.971
Gblur0.9300.9380.9510.9420.8010.8400.8250.8730.885
WN0.9420.8070.8100.9070.8190.8100.7140.8100.873
ToyamaJPEG0.3770.6520.7860.9000.9190.7090.8260.8640.877
JPEG20000.8580.9180.9490.9620.9610.6030.6860.8690.880
SROCC
CSIQJPEG20000.9360.9210.9690.9670.9750.8300.8840.8670.913
JPEG0.8880.9220.9620.9700.9620.7040.8810.9090.944
Gblur0.9290.9240.9720.9750.9680.8710.8700.9030.901
WN0.9360.9250.9470.9570.9540.7970.8860.9250.930
TIDJPEG20000.8250.8780.9730.9700.9740.9070.9110.9040.928
JPEG0.8750.9250.9400.9310.9250.8710.8380.9100.932
Gblur0.9340.9450.9630.9580.8470.8590.8260.8740.894
WN0.9180.8120.8180.9130.8330.8340.7150.8230.882
ToyamaJPEG0.2850.6260.8350.9070.9160.7020.8200.8570.880
JPEG20000.8610.9140.9450.9560.9550.6120.6270.8670.882
RMSE
CSIQJPEG20000.1020.1220.0670.0660.0580.1300.1420.1400.120
JPEG0.1390.1040.0590.0470.0570.1250.2190.0990.070
Gblur0.1090.1250.0810.0650.0620.1270.1260.1070.111
WN0.0510.0640.0540.0470.0500.0740.1040.0580.056
TIDJPEG20000.8910.9280.4300.4590.3661.1780.7560.8100.712
JPEG0.8090.5920.4350.3880.4680.7410.7740.5280.403
Gblur0.4280.4020.3590.3900.6940.6210.6550.5660.540
WN0.2060.3620.3590.2580.3510.3430.4280.3590.298
ToyamaJPEG1.1450.9371.2360.5400.4880.8720.7080.6220.595
JPEG20000.6500.5020.3980.3440.3481.0070.9830.6250.601

As shown in Table 6, DESIQUE provides better predictions in comparison to BLIINDS-II, DIIVINE, and BRISQUE on most distortion types. Notice that for the white noise images in the TID database, DESIQUE demonstrates better predictive performance compared with the other three NSS-based NR IQA algorithms. For some distortion types, such as the JPEG compression, in both the CSIQ and TID databases, DESIQUE performs competitively with many FR IQA methods. In summary, when looking at the performance on individual distortion types, DESIQUE also demonstrates the best NR IQA performance.

4.4.4.

Performance on noise-corrupted images

Although DESIQUE was trained on WN images in the LIVE database, it can also yield valid predictions of quality when other types of noise are present. To demonstrate the adaptiveness of DESIQUE to different noise types, we tested the algorithm on three other noise distortion types in the TID database: additive noise in color components (ACN), spatially correlated noise (SCN), and high-frequency noise (HFN) (notice that in this test all NR algorithms were still trained on LIVE with its five standard distortion types). We compared with three other NR algorithms in terms of CC, SROCC, and RMSE. The results are listed in Table 7.

Table 7

Performances on three other noise distortion types in the TID database.

DIIVINEBLIINDS-IIBRISQUEDESIQUE
SROCCAdditive noise in color components (ACN)0.7760.7370.7470.852
Spatially correlated noise (SCN)0.3640.4620.5690.825
High-frequency noise (HFN)0.8940.8710.7550.925
CCACN0.7900.7540.7680.866
SCN0.4030.5330.5700.811
HFN0.9080.8950.7050.946
RMSEACN0.3130.3360.3280.256
SCN0.5690.5260.5110.363
HFN0.4000.4260.6790.310

As shown in Table 7, DESIQUE demonstrates better performance than DIIVINE, BLIINDS-II, and BRISQUE on these three noise distortion types, despite the fact that none of them were trained on these distortions. These results demonstrate that different noise categories can be detected and measured by using both spatial and spectral feature types. Figure 12 shows the scatter plots of BRISQUE and DESIQUE on these different noise-corrupted images before logistic transform. Notice that BRISQUE even fails on predicting qualities of the spatially correlated noise images, as it produces an increasing relationship with DMOS values in the view of all data points, even though the corresponding SROCC and CC values are computed as 0.569 and 0.570, respectively. Also notice that for the HFN images, DESIQUE provides even more competitive results than BRISQUE, again owing to DESIQUE’s dual-domain analysis. In summary, our proposed DESIQUE algorithm demonstrates not only a competitive performance for the distortion types on which it was trained, but also a degree of adaptiveness to images distorted by different types of noise.

Fig. 12

Scatter plots of objective quality scores (before logistic transform) predicted by BRISQUE and DESIQUE on different noise-corrupted images in the TID database. Note that the x axis across all three figures represents the algorithm predicted DMOS value; the y axis represents the true MOS value.

JEI_22_4_043025_f012.png

4.4.5.

Performance on reference images

In this section, we show the performance of DESIQUE on reference images from the three testing databases. Notice that there are no valid human subjective quality scores for the reference images in the CSIQ and TID databases. Although in the Toyama database, human subjective quality scores are provided in terms of MOS values, it still seems to be inappropriate to measure an algorithm’s performance using SROCC/CC because the MOS values of reference images are too close to provide any meaningful linear/nonlinear relationships. Therefore, to show the performance of our proposed algorithm, we first predicted the original DMOS value of each reference image using our algorithm and then applied the logistic transform defined in Eq. (24) to bring the predicted values on the same scales as the true DMOS/MOS values for the distorted images. Here, the four parameters of the sigmoid function were estimated by using the distorted images in the same database. Finally, we compared these linearized predicted DMOS values to the corresponding scatter plots in Fig. 11 for each database to see if they are all located in a region representing the highest quality.

Figure 13 shows results of DESIQUE tested on all reference images in CSIQ, TID, and Toyama. Note that we tested only 24 undistorted versions of the natural images in the TID database. By referring to Fig. 11, we conclude that DESIQUE works well for most reference images. Specifically, for the CSIQ database, 28 out of 30 reference images are given quality scores below 0.15 (note in Fig. 11 that the true DMOS values of the reference images in the CSIQ database should be close to zero). Thus, quality scores of only two reference images were predicted slightly higher than their true DMOS values. For the TID databases, all reference images seem to be evaluated quite well (note in Fig. 11 that the undistorted/reference images in the TID database should have true MOS values around/above six). For the Toyama database, only one reference image was predicted below four, a bit far away from its true MOS value (note again in Fig. 11 that the undistorted/reference images in the TID database should have true MOS values around/above 4.5). All three figures demonstrate that the proposed DESIQUE algorithm performs quite well on these reference images.

Fig. 13

Objective quality scores predicted by DESIQUE algorithm after logistic transform. Note that the x axis across all three figures represents the index of reference images in each database; the y axis represents the predicted value transformed via Eq. (24).

JEI_22_4_043025_f013.png

4.5.

Contribution Analysis

To analyze the contributions of the log-derivative statistics and each of the two domains (spatial domain and frequency domain) toward the final performance, we performed the prediction test on the aforementioned four databases by using only the spatial-domain-based features (denoted by DESIQUE-S) and only the frequency-domain-based features (denoted by DESIQUE-F), respectively. Three types of testing were performed: (1) the 1000 cross-validation test of DESIQUE-S and F were performed on the LIVE database; (2) the overall performance of DESIQUE-S and F were tested on subsets of the CSIQ, TID, and Toyama databases; and (3) the performance of DESIQUE-S and F were tested on noise-corrupted images from both the CSIQ and TID databases. Note that in (3) we tested on ACN, SCN, and HFN images from TID, and on WN images from both CSIQ and TID (denoted by WN-CSIQ/TID). Also note that models of DESIQUE-S and F for (2) and (3) were obtained via training on the whole LIVE database, and quality was predicted in all three tests using the combined frameworks as recommended in this paper. For comparison, we also listed the corresponding results of the BRISQUE algorithm and of DESIQUE with combined domains employed. The experiment results are shown in Tables 8 and 9.

Table 8

Median SROCC, CC, and RMSE values of BRISQUE, DESIQUE-S, DESIQUE-F, and DESIQUE algorithms across 1000 train-test combinations on the LIVE database.

JP2KJPEGWNBLURFFALL
SROCCBRISQUE0.91390.96470.97860.95110.87680.9395
DESIQUE-S0.93610.96360.98620.95190.84230.9305
DESIQUE-F0.90820.96230.98400.92610.80000.9188
DESIQUE0.93590.96930.98580.94820.86720.9437
CCBRISQUE0.92290.97350.98510.95060.90300.9424
DESIQUE-S0.94590.97250.99140.96310.85760.9354
DESIQUE-F0.91810.97320.99000.93890.85030.9238
DESIQUE0.94800.98030.99150.95850.88190.9465
RMSEBRISQUE9.69347.16414.78565.645012.10249.1315
DESIQUE-S8.04707.36003.63194.968014.13299.6542
DESIQUE-F9.98367.24403.91116.300014.394910.4679
DESIQUE7.96926.21233.61725.228513.31728.7985

Table 9

SROCC, CC, and RMSE values of BRISQUE, DESIQUE-S, DESIQUE-F, and DESIQUE algorithms on subsets of the CSIQ, TID, and Toyama databases, as well as on different noise-corrupted images in the TID and CSIQ databases.

CSIQTIDToyamaACNSCNHFNWN-TIDWN-CSIQ
SROCCBRISQUE0.9000.8980.8480.7470.5690.7550.8230.925
DESIQUE-S0.9210.9180.8940.8460.4490.9050.8490.928
DESIQUE-F0.9190.8850.8160.8050.6440.9030.8800.935
DESIQUE0.9280.9190.8720.8520.8250.9250.8820.930
CCBRISQUE0.9240.9070.8500.7680.5700.7050.8100.938
DESIQUE-S0.9330.9250.8930.8560.4690.9300.8410.939
DESIQUE-F0.9320.8770.8190.8170.6640.9350.8730.945
DESIQUE0.9420.9250.8710.8660.8110.9460.8730.942
RMSEBRISQUE0.1080.6680.6600.3280.5110.6790.3590.058
DESIQUE-S0.1020.6000.5630.2640.5490.3520.3310.058
DESIQUE-F0.1030.7600.7180.2950.4650.3400.2980.055
DESIQUE0.0950.6000.6150.2560.3630.3100.2980.056

According to Table 8, the log-derivative-based analysis in either of the two domains (spatial and frequency) seems to be inferior to the paired-product-based analysis in the spatial domain when performing the cross-validation test on the LIVE database. However, combining the two domain analyses together can improve the performance significantly. According to Table 9, the log-derivative statistics improve upon paired product statistics in predicting image quality when training on LIVE and testing on other databases (note that although DESIQUE-S only has SROCC and CC values of 0.449 and 0.469 on SCN, BRISQUE actually fails on SCN as illustrated in Sec. 4.4.4), and this fact also demonstrates that the two statistics are fundamentally different. Although the log-derivative statistics in the frequency domain analysis contributes less than that in the spatial domain when consulting with the overall performance on subsets of the three testing databases, it does help improve or maintain the predicting performance on CSIQ and TID, and especially on those different noise-corrupted images. On the Toyama database, however, the frequency-domain analysis seems to be ineffective. In summary, in regards to the performance, when all databases are evaluated together, the log-derivative statistics proposed in this paper are superior to the paired product statistics in predicting image quality, and both domains (spatial and frequency) are required to achieve better performance across different databases and distortions.

4.6.

Computational Analysis

DESIQUE also exhibits relatively low computational complexity. Although DESIQUE extracts statistical features in both the spatial and frequency domains to evaluate image quality, and its number of features (60) is larger than that of BRISQUE (which uses only 36 features), it is still quite efficient in computation. To demonstrate this fact, we compared the overall computational complexity of DESIQUE with three other NR IQA methods: BLIINDS-II, DIIVINE, and BRISQUE on different image sizes (256×256, 512×512, 1024×1024, and 1600×1600pixels). The test was performed on a modern desktop computer (AMD Phenom II X4 965 Processor at 3.39 GHz, 4.00 GB RAM, Windows 7 Pro 64-bit, Matlab 7.8.0). Table 10 shows the average runtime of each algorithm in seconds, where the average was taken over 100 trials.

Table 10

Runtime requirements (seconds/image) for four NR IQA methods on different image sizes.

256×256512×5121024×10241600×1600
BRISQUE0.1600.2330.5441.158
DIIVINE8.66122.42884.436225.328
BLIINDS-II9.84338.904156.365380
DESIQUE0.1280.3671.3733.258

As shown in Table 10, DESIQUE is only a little slower than BRISQUE when images have large sizes. This is due to the fact that DESIQUE has to compute both the spatial- and frequency-domain features and that the parameter estimation needs to be performed 30 times for an entire image (BRISQUE only has 10 estimations) in total. However, considering its predictive performance improvement, we believe that the time cost is justified, and it is still much faster than DIIVINE and BLIINDS-II.

5.

Conclusion

This paper presented an algorithm for NR image quality assessment (DESIQUE), which operates by using log-derivative statistics of natural scenes. DESIQUE extracts log-derivative-based statistical features at two image scales in both the spatial and frequency domains, upon which combined frameworks perform NR IQA. We demonstrated that DESIQUE can achieve better performance in predicting image quality than many other well-known NR IQA methods across various databases. We also showed that DESIQUE demonstrates a degree of adaptiveness to different types of noise and is well performed on reference images. We also showed that DESIQUE is computationally efficient: it is among the fastest NSS-based NR IQA methods currently available. Future work involves building a more comprehensive training database and testing DESIQUE on a wider range of distortion types. Future work will also include exploring other features/feature combinations, as well as other frameworks (e.g., Dempster-Shafer framework56,57) to further improve the algorithm performance.

Acknowledgments

This material is based upon work supported by the National Science Foundation Awards 0917014 and 1054612, and by the U.S. Army Research Laboratory and the U.S. Army Research Office under contract/grant number W911NF-10-1-0015.

References

1. 

X. Li, “Blind image quality assessment,” in Proc. of Int. Conf. on Image Processing, I-449 –I-452 (2002). Google Scholar

2. 

S. GabardaG. Cristobal, “No-reference image quality assessment through von mises distribution,” J. Opt. Soc. Am. A, 29 (10), 2058 –2066 (2012). http://dx.doi.org/10.1364/JOSAA.29.002058 JOAOD6 0740-3232 Google Scholar

3. 

T. BrandaoM. P. Queluz, “No-reference image quality assessment based on DCT domain statistics,” Signal Process., 88 (4), 822 –833 (2008). http://dx.doi.org/10.1016/j.sigpro.2007.09.017 SPRODR 0165-1684 Google Scholar

4. 

E. CohenY. Yitzhaky, “No-reference assessment of blur and noise impacts on image quality,” Signal, Image Video Process., 4 (3), 289 –302 (2010). http://dx.doi.org/10.1007/s11760-009-0117-4 1863-1703 Google Scholar

5. 

Z. WangA. C. BovikB. L. Evans, “Blind measurement of blocking artifacts in images,” in Proc. of Int. Conf. on Image Processing, 981 –984 (2000). Google Scholar

6. 

A. C. BovikS. Liu, “DCT-domain blind measurement of blocking artifacts in DCT-coded images,” in Proc. of IEEE Int. Acoustics, Speech, and Signal Processing, 1725 –1728 (2001). Google Scholar

7. 

Z. WangH. R. SheikhA. C. Bovik, “No-reference perceptual quality assessment of JPEG compressed images,” in Proc. of IEEE Int. Conf. on Image Processing, 477 –480 (2002). Google Scholar

8. 

H. R. SheikhA. C. BovikL. Cormack, “No-reference quality assessment using natural scene statistics: JPEG2000,” IEEE Trans. Image Process., 14 (11), 1918 –1927 (2005). http://dx.doi.org/10.1109/TIP.2005.854492 IIPRE4 1057-7149 Google Scholar

9. 

H. TangN. JoshiA. Kapoor, “Learning a blind measure of perceptual image quality,” in Int. Conf. on Computer Vision and Pattern Recognition, 305 –312 (2011). Google Scholar

10. 

P. YeD. Doermann, “No-reference image quality assessment based on visual codebook,” in 18th IEEE Int. Conf. on Image Processing, 3089 –3092 (2011). Google Scholar

11. 

P. YeD. Doermann, “No-reference image quality assessment using visual codebook,” IEEE Trans. Image Process., 21 (7), 3129 –3138 (2012). http://dx.doi.org/10.1109/TIP.2011.2162421 IIPRE4 1057-7149 Google Scholar

12. 

A. K. MoorthyA. C. Bovik, “A two-step framework for constructing blind image quality indices,” IEEE Signal Process. Lett., 17 (5), 513 –516 (2010). http://dx.doi.org/10.1109/LSP.2010.2043888 IESPEJ 1070-9908 Google Scholar

13. 

A. K. MoorthyA. C. Bovik, “Blind image quality assessment: from natural scene statistics to perceptual quality,” IEEE Trans. Image Process., 20 (12), 3350 –3364 (2011). http://dx.doi.org/10.1109/TIP.2011.2147325 IIPRE4 1057-7149 Google Scholar

14. 

M. A. SaadA. C. Bovik, “Blind image quality assessment: a natural scene statistics approach in the DCT domain,” IEEE Trans. Image Process., 21 (8), 3339 –3352 (2012). http://dx.doi.org/10.1109/TIP.2012.2191563 IIPRE4 1057-7149 Google Scholar

15. 

A. MittalA. K. MoorthyA. C. Bovik, “No-reference image quality assessment in the spatial domain,” IEEE Trans. Image Process., 21 (12), 4695 –4708 (2012). http://dx.doi.org/10.1109/TIP.2012.2214050 IIPRE4 1057-7149 Google Scholar

16. 

C. VuT. D. PhanD. M. Chandler, “S3: a spectral and spatial sharpness measure,” IEEE Trans. Image Process., 21 (3), 934 –945 (2012). Google Scholar

17. 

J. HuangD. Mumford, “Statistics of natural images and models,” in IEEE Computer Society Conf. on Computer Vision and Pattern Recognition, 541 –547 (1999). Google Scholar

18. 

H. R. Sheikh, “The statistics of natural images,” Netw. Comput. Neural Syst., 5 (4), 517 –548 (1994). http://dx.doi.org/10.1088/0954-898X/5/4/006 NEWKEB 0954-898X Google Scholar

19. 

D. L. Ruderman, H. R. Sheikhet al., “Image and video quality assessment research at LIVE,” http://live.ece.utexas.edu/research/quality/ Google Scholar

20. 

N. Ponomarenkoet al., “TID2008—a database for evaluation of full-reference visual quality assessment metrics,” Adv. Modern Radioelectron., 10 30 –45 (2009). Google Scholar

21. 

J. Daugman, “Uncertainty relation for resolution in space, sptial frequency, and orientation optimized by two-dimensional visual cortical filters,” J. Opt. Soc. Am., 2 (7), 1160 –1169 (1985). http://dx.doi.org/10.1364/JOSAA.2.001160 JOSAAH 0030-3941 Google Scholar

22. 

S. Marcelja, “Mathematical description of the response of simple cortical cells,” J. Opt. Soc. Am., 70 (11), 1297 –1300 (1980). http://dx.doi.org/10.1364/JOSA.70.001297 JOSAAH 0030-3941 Google Scholar

23. 

D. J. Field, “Relations between the statistics of natural images and the response properties of cortical cells,” J. Opt. Soc. Am. A, 4 (12), 2379 –2394 (1987). http://dx.doi.org/10.1364/JOSAA.4.002379 JOAOD6 0740-3232 Google Scholar

24. 

Q. LiZ. Wang, “Reduced-reference image quality assessment using divisive normalization-based image representation,” IEEE J. Sel. Topics Signal Process., 3 (2), 202 –211 (2009). http://dx.doi.org/10.1109/JSTSP.2009.2014497 1932-4553 Google Scholar

26. 

L. MeestersJ. B. Martens, “A single-ended blockiness measure for JPEG-coded images,” Signal Process., 82 (3), 369 –387 (2002). http://dx.doi.org/10.1016/S0165-1684(01)00177-3 SPRODR 0165-1684 Google Scholar

27. 

F. Panet al., “Measuring blocking artifacts using edge direction information,” in 2004 IEEE Int. Conf. on Multimedia and Expo, 1491 –1494 (2004). Google Scholar

28. 

C.-S. ParkJ.-H. KimS.-J. Ko, “Fast blind measurement of blocking artifacts in both pixel and DCT domains,” J. Math. Imaging Vis., 28 (3), 279 –284 (2007). http://dx.doi.org/10.1007/s10851-007-0019-4 JIMVEC 0924-9907 Google Scholar

29. 

C. ChenJ. A. Bloom, “A blind reference-free blockiness measure,” in Proc. of the 11th Pacific Rim Conf. on Advances in Multimedia Information Processing: Part I, 112 –123 (2010). Google Scholar

30. 

C. PerraF. MassiddaD. D. Giusto, “Image blockiness evaluation based on sobel operator,” in Int. Conf. on Image Processing, 389 –392 (2005). Google Scholar

31. 

H. ZhangY. ZhouX. Tian, “A weighted sobel operator-based no-reference blockiness metric,” in Pacific-Asia Workshop on Computational Intelligence and Industrial Application, 1002 –1006 (2008). Google Scholar

32. 

J. Chenet al., “A no-reference blocking artifacts metric using selective gradient and plainness measures,” in Proc. of the 9th Pacific Rim Conf. on Multimedia: Advances in Multimedia Information Processing, 894 –897 (2008). Google Scholar

33. 

S. Suthaharan, “No-reference visually significant blocking artifact metric for natural scene images,” Signal Process., 89 (8), 1647 –1652 (2009). http://dx.doi.org/10.1016/j.sigpro.2009.02.007 SPRODR 0165-1684 Google Scholar

34. 

S. SureshR. V. BabuH. J. Kim, “No-reference image quality assessment using modified extreme learning machine classifier,” Appl. Soft Comput., 9 (2), 541 –552 (2009). http://dx.doi.org/10.1016/j.asoc.2008.07.005 1568-4946 Google Scholar

35. 

M. LiH. ZhangC. Zhang, “No-reference quality assessment for JPEG2000 compressed images,” in Int. Conf. on Image Processing, 3539 –3542 (2004). Google Scholar

36. 

P. M. Fredericet al., “Perceptual blur and ringing metric: application to JPEG2000,” Signal Process.: Image Commun., 19 (2), 163 –172 (2004). http://dx.doi.org/10.1016/j.image.2003.08.003 SPICEF 0923-5965 Google Scholar

37. 

Z. M. P. SazzadY. KawayokeY. Horita, “Spatial features based no reference quality assessment for JPEG2000,” in IEEE Int. Conf. on Image Processing, III–517 –III–520 (2007). Google Scholar

38. 

Z. M. P. SazzadY. KawayokeY. Horita, “No reference image quality assessment for JPEG2000 based on spatial features,” Signal Process.: Image Commun., 23 (4), 257 –268 (2008). http://dx.doi.org/10.1016/j.image.2008.03.005 SPICEF 0923-5965 Google Scholar

39. 

J. ZhouB. XiaoQ. Li, “A no reference image quality assessment method for JPEG2000,” in IEEE Int. Conf. on Neural Networks, 863 –868 (2008). Google Scholar

40. 

J. ZhangS. H. OngT. M. Le, “Kurtosis-based no-reference quality assessment for JPEG2000 images,” Signal Process.: Image Commun., 26 (1), 13 –23 (2011). http://dx.doi.org/10.1016/j.image.2010.11.003 SPICEF 0923-5965 Google Scholar

41. 

B. R. CornerR. M. NarayananS. E. Reichenbach, “Noise estimation in remote sensing imagery using data masking,” Int. J. Remote Sens., 24 (4), 689 –702 (2003). http://dx.doi.org/10.1080/01431160210164271 IJSEDK 0143-1161 Google Scholar

42. 

H. Tonget al., “Learning no-reference quality metric by examples,” in Proc. 11th Int. Multimedia Modelling Conf., 247 –254 (2005). Google Scholar

43. 

C. LiA. C. BovikX. Wu, “Blind image quality assessment using a general regression neural network,” IEEE Trans. Neural Netw., 22 (5), 793 –799 (2011). http://dx.doi.org/10.1109/TNN.2011.2120620 ITNNEP 1045-9227 Google Scholar

44. 

A. MittalR. SoundararajanA. C. Bovik, “Making a complete blind image quality analyzer,” IEEE Signal Process. Lett., 20 (3), 209 –212 (2013). http://dx.doi.org/10.1109/LSP.2012.2227726 IESPEJ 1070-9908 Google Scholar

45. 

T.-L. Chen, “On the statistics of natural images,” Division of Applied Mathematics at Brown University, (2005). Google Scholar

46. 

K. SharifiA. Leon-Garcia, “Estimation of shape parameter for generalized Gaussian distribution in subband decompositions of video,” IEEE Trans. Circuits Syst. Video Technol., 5 (1), 52 –56 (1995). http://dx.doi.org/10.1109/76.350779 ITCTEM 1051-8215 Google Scholar

47. 

C.-C. ChangC.-J. Lin, “LIBSVM: a library for support vector machines,” ACM Transactions on Intelligent Syst. Technol., 2 (3), 27:1 –27:27 (2001). Google Scholar

48. 

“Objective video quality measurement using a peak-signal-to-noise-ratio (PSNR) full reference technique,” (2001). Google Scholar

49. 

Z. Wanget al., “Image quality assessment: from error visibility to structural similarity,” IEEE Trans. Image Process., 13 (4), 600 –612 (2004). http://dx.doi.org/10.1109/TIP.2003.819861 IIPRE4 1057-7149 Google Scholar

50. 

Z. WangE. P. SimoncelliA. C. Bovik, “Multiscale structural similarity for image quality assessment,” in Conference Record of the Thirty-Seventh Asilomar Conf. on Signals, Systems and Computers, 1398 –1402 (2003). Google Scholar

51. 

H. R. SheikhA. C. Bovik, “Image information and visual quality,” IEEE Trans. Image Process., 15 (2), 430 –444 (2006). http://dx.doi.org/10.1109/TIP.2005.859378 IIPRE4 1057-7149 Google Scholar

52. 

E. C. LarsonD. M. Chandler, “Most apparent distortion: full-reference image quality assessment and the role of strategy,” J. Electron. Imaging, 19 (1), 011006 (2010). http://dx.doi.org/10.1117/1.3267105 JEIME5 1017-9909 Google Scholar

53. 

VQEG, “Final report from the video quality experts group on the validation of objective models of video quality assessment, phase ii,” http://www.its.bldrdoc.gov/vqeg/projects/frtv-phase-ii/frtv-phase-ii.aspx Google Scholar

54. 

D. J. Sheskin, Handbook of Parametric and Nonparametric Statistical Procedures, Chapman and Hall/CRC, Boca Raton, Florida (2007). Google Scholar

55. 

A. K. BeraC. M. Jarque, “Efficient tests for normality, homoscedasticity and serial independence of regression residuals,” Econ. Lett., 6 (3), 255 –259 (1980). http://dx.doi.org/10.1016/0165-1765(80)90024-5 0165-1765 Google Scholar

56. 

G. Shafer, A Mathematical Theory of Evidence, Princeton University Press, Princeton (1976). Google Scholar

57. 

R. R. Yager, “On the Dempster-Shafer framework and new combination rules,” Inf. Sci., 41 (2), 93 –137 (1987). http://dx.doi.org/10.1016/0020-0255(87)90007-7 ISIJBC 0020-0255 Google Scholar

Biography

JEI_22_4_043025_d001.png

Yi Zhang received his BS and MS degrees in electrical engineering from Northwestern Polytechnical University, Xi’an, China, in 2008 and 2011, respectively. He is currently working toward his PhD degree in the School of Electrical and Computer Engineering at Oklahoma State University, Stillwater, Oklahoma. His research areas include image and video processing, machine learning, pattern recognition, and computer vision.

JEI_22_4_043025_d002.png

Damon M. Chandler received his BS and MS degrees in biomedical engineering from Johns Hopkins University, Baltimore, Maryland, in 1998, and his MEng, MS, and PhD degrees in electrical engineering from Cornell University, Ithaca, New York, in 2000, 2003, and 2005, respectively. From 2005 to 2006, he was a postdoctoral research associate in the Department of Psychology at Cornell. He is currently an associate professor in the School of Electrical and Computer Engineering at Oklahoma State University, where he heads the image coding and analysis laboratory. His research interests include image processing, data compression, computational vision, natural scene statistics, and visual perception.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 Unported License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Yi Zhang and Damon M. Chandler "No-reference image quality assessment based on log-derivative statistics of natural scenes," Journal of Electronic Imaging 22(4), 043025 (16 December 2013). https://doi.org/10.1117/1.JEI.22.4.043025
Published: 16 December 2013
Lens.org Logo
CITATIONS
Cited by 117 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Image quality

Databases

Distortion

JPEG2000

Statistical analysis

Statistical modeling

Image analysis


CHORUS Article. This article was made freely available starting 16 December 2014

Back to Top