+

CN107590804A - Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network - Google Patents

Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network Download PDF

Info

Publication number
CN107590804A
CN107590804A CN201710826241.5A CN201710826241A CN107590804A CN 107590804 A CN107590804 A CN 107590804A CN 201710826241 A CN201710826241 A CN 201710826241A CN 107590804 A CN107590804 A CN 107590804A
Authority
CN
China
Prior art keywords
coordinate position
pixel
pixel value
whose coordinate
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710826241.5A
Other languages
Chinese (zh)
Inventor
周武杰
张爽爽
郑飘飘
邱薇薇
周扬
赵颖
何成
葛丁飞
金国英
陈寿法
郑卫红
李鑫
吴洁雯
王昕峰
施祥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Lover Health Science and Technology Development Co Ltd
Original Assignee
Zhejiang Lover Health Science and Technology Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Lover Health Science and Technology Development Co Ltd filed Critical Zhejiang Lover Health Science and Technology Development Co Ltd
Priority to CN201710826241.5A priority Critical patent/CN107590804A/en
Publication of CN107590804A publication Critical patent/CN107590804A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a screen image quality evaluation method based on channel characteristics and a convolutional neural network, which extracts ten channel characteristic graphs of a distorted screen image to be evaluated and acquires respective normalized images; acquiring respective normalized images of ten channel characteristic diagrams of each distorted screen image in a training set in the same way, training respective subjective score values of all distorted screen images in the training set and the normalized images of the ten channel characteristic diagrams by using a convolutional neural network to obtain an optimal weight vector and an optimal bias term, further constructing to obtain a convolutional neural network regression training model, testing the normalized images corresponding to the distorted screen images to be evaluated according to the convolutional neural network regression training model, and obtaining objective quality evaluation predicted values of the distorted screen images to be evaluated; the method has the advantages that the influence of various characteristics of the screen image on the visual quality can be fully considered, so that the correlation between objective evaluation results and subjective perception can be improved.

Description

基于通道特征和卷积神经网络的屏幕图像质量评价方法Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network

技术领域technical field

本发明涉及一种无参考图像质量评价方法,尤其是涉及一种基于通道特征和卷积神经网络的屏幕图像质量评价方法。The invention relates to a no-reference image quality evaluation method, in particular to a screen image quality evaluation method based on channel features and a convolutional neural network.

背景技术Background technique

随着图像处理行业的快速发展,图像质量评价已成为越来越重要的组成部分,人们对图像质量的要求也日益增高。由于图像的采集、存储、传输和显示等过程中,往往会有不同程度的失真,如图像模糊、视频终端图像失真、系统中图像质量不达标等,因此,建立有效的图像质量评价机制非常重要,如在图像去噪、图像融合等处理过程中可用于各种算法的性能比较、参数选择;在图像编码与通信领域可用于指导整个图像的传输过程并评估系统性能;在视频监控领域可用于检测视频画面的质量,及时发现并调整视频监控质量。With the rapid development of the image processing industry, image quality evaluation has become an increasingly important component, and people's requirements for image quality are also increasing. Due to the process of image acquisition, storage, transmission and display, there are often different degrees of distortion, such as image blur, video terminal image distortion, image quality in the system is not up to standard, etc. Therefore, it is very important to establish an effective image quality evaluation mechanism For example, it can be used for performance comparison and parameter selection of various algorithms in the process of image denoising and image fusion; it can be used to guide the entire image transmission process and evaluate system performance in the field of image coding and communication; it can be used in the field of video surveillance. Detect the quality of video images, discover and adjust the quality of video surveillance in time.

图像质量评价方法大体可以划分为两类,即主观评价方法和客观评价方法。主观评价方法是通过人眼直接判定图像的质量,受主观意识的操控;其评价结果的特点是:图像的舒适度越高,得到的评分结果就越高;其是一种相对比较可靠的评价方法,但是费时费力。客观评价方法是由机器根据一定算法得出的图像质量指标,其主要分为三种评价方法,即全参考图像质量评价方法、半参考图像质量评价方法和无参考图像质量评价方法,当前研究最多的是全参考图像质量评价方法,但是由于多数应用中无法获得相应的原始图像,因此无参考图像质量评价方法的研究更具实用价值。Image quality evaluation methods can be roughly divided into two categories, namely subjective evaluation methods and objective evaluation methods. The subjective evaluation method is to directly judge the quality of the image through the human eye, which is controlled by subjective consciousness; the characteristics of the evaluation result are: the higher the comfort of the image, the higher the scoring result; it is a relatively reliable evaluation method, but time consuming. The objective evaluation method is the image quality index obtained by the machine according to a certain algorithm. It is mainly divided into three evaluation methods, namely, the full reference image quality evaluation method, the semi-reference image quality evaluation method and the no-reference image quality evaluation method. The current research is the most However, since the corresponding original images cannot be obtained in most applications, the research on no-reference image quality assessment methods is more practical.

无参考图像质量评价方法可分为特定失真评价方法和通用评价方法两种,特定失真评价方法只能对某种特定失真类型的图像进行评价,例如JPEG、JPEG2K及Gblur失真等,无法对其它失真类型的图像及多种处理技术处理后的图像进行质量评价;通用评价方法可以同时对多种失真类型的图像进行质量评价。No-reference image quality evaluation methods can be divided into specific distortion evaluation methods and general evaluation methods. Specific distortion evaluation methods can only evaluate images of a certain type of distortion, such as JPEG, JPEG2K and Gblur distortion, etc., and cannot evaluate other distortions. Types of images and images processed by multiple processing techniques for quality evaluation; the general evaluation method can simultaneously evaluate the quality of images with multiple types of distortion.

现有的通用无参考图像质量评价方法主要针对一般的图像,而针对特殊图像(例如,屏幕图像)的研究相对较少,由于屏幕图像含有文字、图形和图像等内容,因此对屏幕图像采用通用无参考图像质量评价方法进行质量评价更具有挑战性。The existing general-purpose no-reference image quality evaluation methods are mainly aimed at general images, but there are relatively few studies on special images (such as screen images). Since screen images contain text, graphics, and images, etc., the general It is more challenging to perform quality assessment without reference image quality assessment methods.

发明内容Contents of the invention

本发明所要解决的技术问题是提供一种基于通道特征和卷积神经网络的屏幕图像质量评价方法,其能够充分考虑到屏幕图像多种特性对视觉质量的影响,从而能够提高客观评价结果与主观感知之间的相关性。The technical problem to be solved by the present invention is to provide a screen image quality evaluation method based on channel features and convolutional neural network, which can fully consider the impact of various characteristics of screen images on visual quality, thereby improving the objective evaluation results and subjective evaluation results. Correlation between perceptions.

本发明解决上述技术问题所采用的技术方案为:一种基于通道特征和卷积神经网络的屏幕图像质量评价方法,其特征在于包括以下步骤:The technical solution adopted by the present invention to solve the above-mentioned technical problems is: a method for evaluating the quality of screen images based on channel features and convolutional neural networks, characterized in that it comprises the following steps:

步骤一:令{Id(i,j)}表示待评价的失真屏幕图像,其中,1≤i≤W,1≤j≤H,W表示{Id(i,j)}的宽度,H表示{Id(i,j)}的高度,Id(i,j)表示{Id(i,j)}中坐标位置为(i,j)的像素点的像素值;Step 1: Let {I d (i, j)} represent the distorted screen image to be evaluated, where, 1≤i≤W, 1≤j≤H, W represents the width of {I d (i,j)}, H Represents the height of {I d (i, j)}, I d (i, j) represents the pixel value of the pixel whose coordinate position is (i, j) in {I d (i, j)};

步骤二:利用集总特征通道方法对{Id(i,j)}进行特征提取,得到{Id(i,j)}的十个通道特征图,分别为L通道特征图、U通道特征图、V通道特征图、梯度幅值通道特征图、第1个方向梯度直方图通道特征图、第2个方向梯度直方图通道特征图、第3个方向梯度直方图通道特征图、第4个方向梯度直方图通道特征图、第5个方向梯度直方图通道特征图、第6个方向梯度直方图通道特征图,对应记为{Ld(m,n)}、{Ud(m,n)}、{Vd(m,n)}、{Gd,0(m,n)}、{Gd,1(m,n)}、{Gd,2(m,n)}、{Gd,3(m,n)}、{Gd,4(m,n)}、{Gd,5(m,n)}、{Gd,6(m,n)},其中,1≤m≤M,1≤n≤N,符号为向下取整操作符号,M表示{Id(i,j)}的每个通道特征图的宽度,N表示{Id(i,j)}的每个通道特征图的高度,Ld(m,n)表示{Ld(m,n)}中坐标位置为(m,n)的像素点的像素值,Ud(m,n)表示{Ud(m,n)}中坐标位置为(m,n)的像素点的像素值,Vd(m,n)表示{Vd(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,0(m,n)表示{Gd,0(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,1(m,n)表示{Gd,1(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,2(m,n)表示{Gd,2(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,3(m,n)表示{Gd,3(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,4(m,n)表示{Gd,4(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,5(m,n)表示{Gd,5(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,6(m,n)表示{Gd,6(m,n)}中坐标位置为(m,n)的像素点的像素值;Step 2: Use the lumped feature channel method to perform feature extraction on {I d (i, j)}, and obtain ten channel feature maps of {I d (i, j)}, which are L channel feature map and U channel feature Figure, V channel feature map, gradient amplitude channel feature map, first directional gradient histogram channel feature map, second directional gradient histogram channel feature map, third directional gradient histogram channel feature map, fourth The directional gradient histogram channel feature map, the 5th directional gradient histogram channel feature map, and the 6th directional gradient histogram channel feature map, correspondingly recorded as {L d (m,n)}, {U d (m,n) )}, {V d (m,n)}, {G d,0 (m,n)}, {G d,1 (m,n)}, {G d,2 (m,n)}, { G d,3 (m,n)}, {G d,4 (m,n)}, {G d,5 (m,n)}, {G d,6 (m,n)}, where 1 ≤m≤M, 1≤n≤N, symbol is the symbol of the rounding down operation, M represents the width of each channel feature map of {I d (i,j)}, N represents the height of each channel feature map of {I d (i,j)}, L d (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {L d (m,n)}, and U d (m,n) represents the coordinates in {U d (m,n)} The pixel value of the pixel point whose position is (m, n), V d (m, n) represents the pixel value of the pixel point whose coordinate position is (m, n) in {V d (m, n)}, G d, 0 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,0 (m,n)}, and G d,1 (m,n) represents {G d,1 ( The pixel value of the pixel whose coordinate position is (m,n) in m,n)}, G d,2 (m,n) means that the coordinate position in {G d,2 (m,n)} is (m,n) ), G d,3 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {G d,3 (m,n)}, G d,4 ( m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,4 (m,n)}, G d,5 (m,n) represents {G d,5 (m, The pixel value of the pixel whose coordinate position is (m,n) in n)}, G d,6 (m,n) represents the pixel value of the coordinate position (m,n) in {G d,6 (m,n)} The pixel value of the pixel point;

步骤三:获取{Ld(m,n)}、{Ud(m,n)}、{Vd(m,n)}、{Gd,0(m,n)}、{Gd,1(m,n)}、{Gd,2(m,n)}、{Gd,3(m,n)}、{Gd,4(m,n)}、{Gd,5(m,n)}、{Gd,6(m,n)}各自的归一化图像,对应记为 其中,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值;Step 3: Get {L d (m,n)}, {U d (m,n)}, {V d (m,n)}, {G d,0 (m,n)}, {G d, 1 (m,n)}, {G d,2 (m,n)}, {G d,3 (m,n)}, {G d,4 (m,n)}, {G d,5 ( The normalized images of m,n)} and {G d,6 (m,n)} are correspondingly denoted as in, express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose middle coordinate position is (m,n);

步骤四:采用P幅原始的无失真屏幕图像,建立其在不同失真类型不同失真程度下的失真屏幕图像集合,将该失真屏幕图像集合作为训练集,训练集包括多幅失真屏幕图像;然后利用主观质量评价方法评价出训练集中的每幅失真屏幕图像的主观评分值,将训练集中的第j幅失真屏幕图像的主观评分值记为DMOSj;并按照步骤一至步骤三的操作,以相同的方式获取训练集中的每幅失真屏幕图像的十个通道特征图各自的归一化图像,将训练集中的第j幅失真屏幕图像的十个通道特征图各自的归一化图像对应记为 其中,P>1,1≤j≤K',K'表示训练集中包含的失真屏幕图像的总幅数,0≤DMOSj≤100,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值;Step 4: Using P original undistorted screen images, set up a set of distorted screen images under different distortion types and different degrees of distortion, and use the set of distorted screen images as a training set. The training set includes multiple distorted screen images; then use The subjective quality evaluation method evaluates the subjective score value of each piece of distorted screen image in the training set, and records the subjective score value of the jth piece of distorted screen image in the training set as DMOS j ; Obtain the respective normalized images of the ten channel feature maps of each distorted screen image in the training set by means of , and record the respective normalized images of the ten channel feature maps of the jth distorted screen image in the training set as Among them, P>1, 1≤j≤K', K' indicates the total number of distorted screen images contained in the training set, 0≤DMOS j ≤100, express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose middle coordinate position is (m,n);

步骤五:利用卷积神经网络对训练集中的所有失真屏幕图像各自的主观评分值及各自的十个通道特征图的归一化图像进行训练,使得经过训练得到的回归函数值与主观评分值之间的误差最小,拟合得到最优的权值矢量Wopt和最优的偏置项bopt;接着利用Wopt构造得到卷积神经网络回归训练模型;再根据卷积神经网络回归训练模型,对 进行测试,预测得到{Id(i,j)}的客观质量评价预测值,记为Q,Q=f(x),其中,Q是x的函数,f()为函数表示形式,x为输入变量,x表示 为Wopt的转置矢量,为x的卷积函数。Step 5: Use the convolutional neural network to train the respective subjective score values of all the distorted screen images in the training set and the normalized images of the respective ten channel feature maps, so that the relationship between the regression function value obtained after training and the subjective score value The error between is the smallest, and the optimal weight vector W opt and the optimal bias item b opt are obtained by fitting; then the convolutional neural network regression training model is obtained by using the W opt construction; and then according to the convolutional neural network regression training model, right Carry out the test and predict the objective quality evaluation prediction value of {I d (i, j)}, denoted as Q, Q=f(x), Among them, Q is a function of x, f() is a function representation, x is an input variable, and x represents is the transpose vector of W opt , Convolution function for x.

所述的步骤三的具体过程为:The concrete process of described step three is:

求得{Ld(m,n)}的局部平均值图像,记为{μ1(m,n)};并求得{Ld(m,n)}的局部方差图像,记为{σ1(m,n)};然后根据{Ld(m,n)}、{μ1(m,n)}、{σ1(m,n)},获取{Ld(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ1(m,n)表示{μ1(m,n)}中坐标位置为(m,n)的像素点的像素值,σ1(m,n)表示{σ1(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {L d (m,n)}, denoted as {μ 1 (m,n)}; and obtain the local variance image of {L d (m,n)}, denoted as {σ 1 (m,n)}; then according to {L d (m,n)}, {μ 1 (m,n)}, {σ 1 (m,n)}, get {L d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 1 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 1 (m,n)}, and σ 1 (m,n) represents {σ 1 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Ud(m,n)}的局部平均值图像,记为{μ2(m,n)};并求得{Ud(m,n)}的局部方差图像,记为{σ2(m,n)};然后根据{Ud(m,n)}、{μ2(m,n)}、{σ2(m,n)},获取{Ud(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ2(m,n)表示{μ2(m,n)}中坐标位置为(m,n)的像素点的像素值,σ2(m,n)表示{σ2(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {U d (m,n)}, denoted as {μ 2 (m,n)}; and obtain the local variance image of {U d (m,n)}, denoted as {σ 2 (m,n)}; then according to {U d (m,n)}, {μ 2 (m,n)}, {σ 2 (m,n)}, get {U d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 2 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 2 (m,n)}, and σ 2 (m,n) represents {σ 2 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Vd(m,n)}的局部平均值图像,记为{μ3(m,n)};并求得{Vd(m,n)}的局部方差图像,记为{σ3(m,n)};然后根据{Vd(m,n)}、{μ3(m,n)}、{σ3(m,n)},获取{Vd(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ3(m,n)表示{μ3(m,n)}中坐标位置为(m,n)的像素点的像素值,σ3(m,n)表示{σ3(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {V d (m,n)}, denoted as {μ 3 (m,n)}; and obtain the local variance image of {V d (m,n)}, denoted as {σ 3 (m,n)}; then according to {V d (m,n)}, {μ 3 (m,n)}, {σ 3 (m,n)}, get {V d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 3 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 3 (m,n)}, and σ 3 (m,n) represents {σ 3 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,0(m,n)}的局部平均值图像,记为{μ4(m,n)};并求得{Gd,0(m,n)}的局部方差图像,记为{σ4(m,n)};然后根据{Gd,0(m,n)}、{μ4(m,n)}、{σ4(m,n)},获取{Gd,0(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ4(m,n)表示{μ4(m,n)}中坐标位置为(m,n)的像素点的像素值,σ4(m,n)表示{σ4(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,0 (m,n)}, denoted as {μ 4 (m,n)}; and obtain the local variance image of {G d,0 (m,n)}, Denote it as {σ 4 ( m,n) } ; then get {G d ,0 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 4 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 4 (m,n)}, and σ 4 (m,n) represents {σ 4 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,1(m,n)}的局部平均值图像,记为{μ5(m,n)};并求得{Gd,1(m,n)}的局部方差图像,记为{σ5(m,n)};然后根据{Gd,1(m,n)}、{μ5(m,n)}、{σ5(m,n)},获取{Gd,1(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ5(m,n)表示{μ5(m,n)}中坐标位置为(m,n)的像素点的像素值,σ5(m,n)表示{σ5(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,1 (m,n)}, denoted as {μ 5 (m,n)}; and obtain the local variance image of {G d,1 (m,n)}, Recorded as {σ 5 (m,n)}; then according to {G d,1 (m,n)}, {μ 5 (m,n)}, {σ 5 (m,n)}, get {G d ,1 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 5 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 5 (m,n)}, and σ 5 (m,n) represents {σ 5 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,2(m,n)}的局部平均值图像,记为{μ6(m,n)};并求得{Gd,2(m,n)}的局部方差图像,记为{σ6(m,n)};然后根据{Gd,2(m,n)}、{μ6(m,n)}、{σ6(m,n)},获取{Gd,2(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ6(m,n)表示{μ6(m,n)}中坐标位置为(m,n)的像素点的像素值,σ6(m,n)表示{σ6(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,2 (m,n)}, denoted as {μ 6 (m,n)}; and obtain the local variance image of {G d,2 (m,n)}, Denote it as {σ 6 (m,n)}; then according to {G d,2 (m,n)}, {μ 6 (m,n)}, {σ 6 (m,n) } , get ,2 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 6 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 6 (m,n)}, and σ 6 (m,n) represents {σ 6 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,3(m,n)}的局部平均值图像,记为{μ7(m,n)};并求得{Gd,3(m,n)}的局部方差图像,记为{σ7(m,n)};然后根据{Gd,3(m,n)}、{μ7(m,n)}、{σ7(m,n)},获取{Gd,3(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ7(m,n)表示{μ7(m,n)}中坐标位置为(m,n)的像素点的像素值,σ7(m,n)表示{σ7(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,3 (m,n)}, denoted as {μ 7 (m,n)}; and obtain the local variance image of {G d,3 (m,n)}, Denote it as {σ 7 (m,n) } ; then get {G d ,3 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 7 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 7 (m,n)}, and σ 7 (m,n) represents {σ 7 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,4(m,n)}的局部平均值图像,记为{μ8(m,n)};并求得{Gd,4(m,n)}的局部方差图像,记为{σ8(m,n)};然后根据{Gd,4(m,n)}、{μ8(m,n)}、{σ8(m,n)},获取{Gd,4(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ8(m,n)表示{μ8(m,n)}中坐标位置为(m,n)的像素点的像素值,σ8(m,n)表示{σ8(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,4 (m,n)}, denoted as {μ 8 (m,n)}; and obtain the local variance image of {G d,4 (m,n)}, Denote it as { σ 8 (m,n)} ; then get {G d ,4 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 8 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 8 (m,n)}, and σ 8 (m,n) represents {σ 8 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,5(m,n)}的局部平均值图像,记为{μ9(m,n)};并求得{Gd,5(m,n)}的局部方差图像,记为{σ9(m,n)};然后根据{Gd,5(m,n)}、{μ9(m,n)}、{σ9(m,n)},获取{Gd,5(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ9(m,n)表示{μ9(m,n)}中坐标位置为(m,n)的像素点的像素值,σ9(m,n)表示{σ9(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,5 (m,n)}, denoted as {μ 9 (m,n)}; and obtain the local variance image of {G d,5 (m,n)}, Denote it as { σ 9 (m,n)} ; then get {G d ,5 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 9 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 9 (m,n)}, and σ 9 (m,n) represents {σ 9 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n);

求得{Gd,6(m,n)}的局部平均值图像,记为{μ10(m,n)};并求得{Gd,6(m,n)}的局部方差图像,记为{σ10(m,n)};然后根据{Gd,6(m,n)}、{μ10(m,n)}、{σ10(m,n)},获取{Gd,6(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ10(m,n)表示{μ10(m,n)}中坐标位置为(m,n)的像素点的像素值,σ10(m,n)表示{σ10(m,n)}中坐标位置为(m,n)的像素点的像素值。Obtain the local average value image of {G d,6 (m,n)}, denoted as {μ 10 (m,n)}; and obtain the local variance image of {G d,6 (m,n)}, Denote it as {σ 10 (m,n) } ; then get {G d ,6 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 10 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 10 (m,n)}, and σ 10 (m,n) represents {σ 10 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

与现有技术相比,本发明的优点在于:Compared with the prior art, the present invention has the advantages of:

1)本发明方法对失真屏幕图像进行特征提取,得到失真屏幕图像的十个通道特征图,进而获取失真屏幕图像的十个通道特征图的归一化图像,将训练集中的所有失真屏幕图像各自的十个通道特征图的归一化图像输入到卷积神经网络进行训练,得到卷积神经网络回归训练模型;再将待评价的失真屏幕图像的十个通道特征图的归一化图像输入到卷积神经网络回归训练模型中,预测得到待评价的失真屏幕图像的客观质量评价预测值,由于结合了失真屏幕图像的多种特征,而这些特征能够比较准确地描述失真屏幕图像,即本发明方法充分考虑到了屏幕图像多种特性对视觉质量的影响,从而能够有效地提高客观评价结果与主观感知之间的相关性。1) The inventive method carries out feature extraction to the distorted screen image, obtains ten channel feature maps of the distorted screen image, and then obtains the normalized image of the ten channel feature maps of the distorted screen image, and separates all distorted screen images in the training set Input the normalized images of the ten channel feature maps of the convolutional neural network into the convolutional neural network for training, and obtain the convolutional neural network regression training model; then input the normalized images of the ten channel feature maps of the distorted screen image to be evaluated into the In the convolutional neural network regression training model, the objective quality evaluation prediction value of the distorted screen image to be evaluated is predicted, and since multiple features of the distorted screen image are combined, these features can describe the distorted screen image more accurately, that is, the present invention The method takes full account of the influence of various characteristics of the screen image on the visual quality, so that the correlation between the objective evaluation results and the subjective perception can be effectively improved.

2)本发明方法采用基于卷积神经网络的监督学习方式,即模拟人眼视觉特性处理的监督特征学习方法进行训练和测试,使得本发明方法能够充分考虑到视觉感知特性对图像失真的敏感性。2) The method of the present invention adopts a supervised learning method based on a convolutional neural network, that is, a supervised feature learning method that simulates the processing of human visual characteristics for training and testing, so that the method of the present invention can fully take into account the sensitivity of visual perception characteristics to image distortion .

3)本发明方法采用符合人脑机理特性的卷积神经网络预测得到待评价的失真屏幕图像的客观质量评价预测值,能有效地提高客观评价结果与主观感知之间的相关性。3) The method of the present invention adopts the convolutional neural network conforming to the mechanism characteristics of the human brain to predict the objective quality evaluation prediction value of the distorted screen image to be evaluated, which can effectively improve the correlation between the objective evaluation result and the subjective perception.

附图说明Description of drawings

图1为本发明方法的总体实现框图;Fig. 1 is the overall realization block diagram of the inventive method;

图2为一幅失真屏幕图像;Figure 2 is a distorted screen image;

图3为图2所示的失真屏幕图像的L通道特征图;Fig. 3 is the L channel feature map of the distorted screen image shown in Fig. 2;

图4为图2所示的失真屏幕图像的U通道特征图;Fig. 4 is the U channel feature map of the distorted screen image shown in Fig. 2;

图5为图2所示的失真屏幕图像的V通道特征图;Fig. 5 is the V channel feature map of the distorted screen image shown in Fig. 2;

图6为图2所示的失真屏幕图像的梯度幅值通道特征图;Fig. 6 is the gradient amplitude channel feature map of the distorted screen image shown in Fig. 2;

图7为图2所示的失真屏幕图像的第1个方向梯度直方图通道特征图;Fig. 7 is the first directional gradient histogram channel feature map of the distorted screen image shown in Fig. 2;

图8为图2所示的失真屏幕图像的第2个方向梯度直方图通道特征图;Fig. 8 is the second directional gradient histogram channel feature map of the distorted screen image shown in Fig. 2;

图9为图2所示的失真屏幕图像的第3个方向梯度直方图通道特征图;Fig. 9 is the 3rd directional gradient histogram channel feature map of the distorted screen image shown in Fig. 2;

图10为图2所示的失真屏幕图像的第4个方向梯度直方图通道特征图;Fig. 10 is the 4th directional gradient histogram channel feature map of the distorted screen image shown in Fig. 2;

图11为图2所示的失真屏幕图像的第5个方向梯度直方图通道特征图;Fig. 11 is the 5th directional gradient histogram channel feature map of the distorted screen image shown in Fig. 2;

图12为图2所示的失真屏幕图像的第6个方向梯度直方图通道特征图。FIG. 12 is a feature map of the sixth directional gradient histogram channel of the distorted screen image shown in FIG. 2 .

具体实施方式detailed description

以下结合附图实施例对本发明作进一步详细描述。The present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments.

本发明提出的一种基于通道特征和卷积神经网络的屏幕图像质量评价方法,其总体实现框图如图1所示,其包括以下步骤:A method for evaluating screen image quality based on channel features and convolutional neural networks proposed by the present invention, its overall implementation block diagram is shown in Figure 1, and it includes the following steps:

步骤一:令{Id(i,j)}表示待评价的失真屏幕图像,其中,1≤i≤W,1≤j≤H,W表示{Id(i,j)}的宽度,H表示{Id(i,j)}的高度,Id(i,j)表示{Id(i,j)}中坐标位置为(i,j)的像素点的像素值。Step 1: Let {I d (i, j)} represent the distorted screen image to be evaluated, where, 1≤i≤W, 1≤j≤H, W represents the width of {I d (i,j)}, H Indicates the height of {I d (i, j)}, and I d (i, j) indicates the pixel value of the pixel at the coordinate position (i, j) in {I d (i, j)}.

步骤二:利用现有的集总特征通道方法(Aggregate Channel Features,ACF)对{Id(i,j)}进行特征提取,得到{Id(i,j)}的十个通道特征图,分别为L通道特征图、U通道特征图、V通道特征图、梯度幅值通道特征图、第1个方向梯度直方图通道特征图、第2个方向梯度直方图通道特征图、第3个方向梯度直方图通道特征图、第4个方向梯度直方图通道特征图、第5个方向梯度直方图通道特征图、第6个方向梯度直方图通道特征图,对应记为{Ld(m,n)}、{Ud(m,n)}、{Vd(m,n)}、{Gd,0(m,n)}、{Gd,1(m,n)}、{Gd,2(m,n)}、{Gd,3(m,n)}、{Gd,4(m,n)}、{Gd,5(m,n)}、{Gd,6(m,n)},其中,1≤m≤M,1≤n≤N, 符号为向下取整操作符号,M表示{Id(i,j)}的每个通道特征图的宽度,N表示{Id(i,j)}的每个通道特征图的高度,Ld(m,n)表示{Ld(m,n)}中坐标位置为(m,n)的像素点的像素值,Ud(m,n)表示{Ud(m,n)}中坐标位置为(m,n)的像素点的像素值,Vd(m,n)表示{Vd(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,0(m,n)表示{Gd,0(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,1(m,n)表示{Gd,1(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,2(m,n)表示{Gd,2(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,3(m,n)表示{Gd,3(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,4(m,n)表示{Gd,4(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,5(m,n)表示{Gd,5(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,6(m,n)表示{Gd,6(m,n)}中坐标位置为(m,n)的像素点的像素值。Step 2: Use the existing aggregated feature channel method (Aggregate Channel Features, ACF) to perform feature extraction on {I d (i, j)}, and obtain ten channel feature maps of {I d (i, j)}, They are L channel feature map, U channel feature map, V channel feature map, gradient magnitude channel feature map, gradient histogram channel feature map in the first direction, gradient histogram channel feature map in the second direction, and third direction The gradient histogram channel feature map, the fourth directional gradient histogram channel feature map, the fifth directional gradient histogram channel feature map, and the sixth directional gradient histogram channel feature map, correspondingly denoted as {L d (m,n )}, {U d (m,n)}, {V d (m,n)}, {G d,0 (m,n)}, {G d,1 (m,n)}, {G d ,2 (m,n)}, {G d,3 (m,n)}, {G d,4 (m,n)}, {G d,5 (m,n)}, {G d,6 (m,n)}, where, 1≤m≤M, 1≤n≤N, symbol is the symbol of the rounding down operation, M represents the width of each channel feature map of {I d (i,j)}, N represents the height of each channel feature map of {I d (i,j)}, L d (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {L d (m,n)}, and U d (m,n) represents the coordinates in {U d (m,n)} The pixel value of the pixel point whose position is (m, n), V d (m, n) represents the pixel value of the pixel point whose coordinate position is (m, n) in {V d (m, n)}, G d, 0 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,0 (m,n)}, and G d,1 (m,n) represents {G d,1 ( The pixel value of the pixel whose coordinate position is (m,n) in m,n)}, G d,2 (m,n) means that the coordinate position in {G d,2 (m,n)} is (m,n) ), G d,3 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {G d,3 (m,n)}, G d,4 ( m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,4 (m,n)}, G d,5 (m,n) represents {G d,5 (m, The pixel value of the pixel whose coordinate position is (m,n) in n)}, G d,6 (m,n) represents the pixel value of the coordinate position (m,n) in {G d,6 (m,n)} The pixel value of the pixel point.

图2给出了一幅失真屏幕图像,图3至图12对应给出了图2所示的失真屏幕图像的L通道特征图、U通道特征图、V通道特征图、梯度幅值通道特征图、第1个方向梯度直方图通道特征图、第2个方向梯度直方图通道特征图、第3个方向梯度直方图通道特征图、第4个方向梯度直方图通道特征图、第5个方向梯度直方图通道特征图、第6个方向梯度直方图通道特征图。Figure 2 shows a distorted screen image, and Figures 3 to 12 correspond to the L channel feature map, U channel feature map, V channel feature map, and gradient magnitude channel feature map of the distorted screen image shown in Figure 2 , the first directional gradient histogram channel feature map, the second directional gradient histogram channel feature map, the third directional gradient histogram channel feature map, the fourth directional gradient histogram channel feature map, and the fifth directional gradient Histogram channel feature map, the sixth directional gradient histogram channel feature map.

步骤三:采用现有的平均值求取方法,求得{Ld(m,n)}的局部平均值图像,记为{μ1(m,n)};并采用现有的方差求取方法,求得{Ld(m,n)}的局部方差图像,记为{σ1(m,n)};然后根据{Ld(m,n)}、{μ1(m,n)}、{σ1(m,n)},获取{Ld(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ1(m,n)表示{μ1(m,n)}中坐标位置为(m,n)的像素点的像素值,σ1(m,n)表示{σ1(m,n)}中坐标位置为(m,n)的像素点的像素值。Step 3: Use the existing average calculation method to obtain the local average value image of {L d (m,n)}, denoted as {μ 1 (m,n)}; and use the existing variance to obtain method to obtain the local variance image of {L d (m,n)}, denoted as {σ 1 (m,n)}; then according to {L d (m,n)}, {μ 1 (m,n) }, {σ 1 (m,n)}, get the normalized image of {L d (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 1 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 1 (m,n)}, and σ 1 (m,n) represents {σ 1 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Ud(m,n)}的局部平均值图像,记为{μ2(m,n)};并采用现有的方差求取方法,求得{Ud(m,n)}的局部方差图像,记为{σ2(m,n)};然后根据{Ud(m,n)}、{μ2(m,n)}、{σ2(m,n)},获取{Ud(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ2(m,n)表示{μ2(m,n)}中坐标位置为(m,n)的像素点的像素值,σ2(m,n)表示{σ2(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing mean value calculation method, obtain the local mean value image of {U d (m,n)}, denoted as {μ 2 (m,n)}; and use the existing variance calculation method, find Get the local variance image of {U d (m,n)}, denoted as {σ 2 (m,n)}; then according to {U d (m,n)}, {μ 2 (m,n)}, { σ 2 (m,n)}, get the normalized image of {U d (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 2 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 2 (m,n)}, and σ 2 (m,n) represents {σ 2 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Vd(m,n)}的局部平均值图像,记为{μ3(m,n)};并采用现有的方差求取方法,求得{Vd(m,n)}的局部方差图像,记为{σ3(m,n)};然后根据{Vd(m,n)}、{μ3(m,n)}、{σ3(m,n)},获取{Vd(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ3(m,n)表示{μ3(m,n)}中坐标位置为(m,n)的像素点的像素值,σ3(m,n)表示{σ3(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {V d (m,n)}, denoted as {μ 3 (m,n)}; and use the existing variance calculation method, to obtain Get the local variance image of {V d (m,n)}, denoted as {σ 3 (m,n)}; then according to {V d (m,n)}, {μ 3 (m,n)}, { σ 3 (m,n)}, get the normalized image of {V d (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 3 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 3 (m,n)}, and σ 3 (m,n) represents {σ 3 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,0(m,n)}的局部平均值图像,记为{μ4(m,n)};并采用现有的方差求取方法,求得{Gd,0(m,n)}的局部方差图像,记为{σ4(m,n)};然后根据{Gd,0(m,n)}、{μ4(m,n)}、{σ4(m,n)},获取{Gd,0(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ4(m,n)表示{μ4(m,n)}中坐标位置为(m,n)的像素点的像素值,σ4(m,n)表示{σ4(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,0 (m,n)}, denoted as {μ 4 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,0 (m,n)}, denoted as {σ 4 (m,n)}; then according to {G d,0 (m,n)}, {μ 4 (m ,n)}, {σ 4 (m,n)}, get the normalized image of {G d,0 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 4 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 4 (m,n)}, and σ 4 (m,n) represents {σ 4 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,1(m,n)}的局部平均值图像,记为{μ5(m,n)};并采用现有的方差求取方法,求得{Gd,1(m,n)}的局部方差图像,记为{σ5(m,n)};然后根据{Gd,1(m,n)}、{μ5(m,n)}、{σ5(m,n)},获取{Gd,1(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ5(m,n)表示{μ5(m,n)}中坐标位置为(m,n)的像素点的像素值,σ5(m,n)表示{σ5(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average value calculation method, obtain the local average value image of {G d,1 (m,n)}, denoted as {μ 5 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,1 (m,n)}, denoted as {σ 5 (m,n)}; then according to {G d,1 (m,n)}, {μ 5 (m ,n)}, {σ 5 (m,n)}, get the normalized image of {G d,1 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 5 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 5 (m,n)}, and σ 5 (m,n) represents {σ 5 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,2(m,n)}的局部平均值图像,记为{μ6(m,n)};并采用现有的方差求取方法,求得{Gd,2(m,n)}的局部方差图像,记为{σ6(m,n)};然后根据{Gd,2(m,n)}、{μ6(m,n)}、{σ6(m,n)},获取{Gd,2(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ6(m,n)表示{μ6(m,n)}中坐标位置为(m,n)的像素点的像素值,σ6(m,n)表示{σ6(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,2 (m,n)}, denoted as {μ 6 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,2 (m,n)}, denoted as {σ 6 (m,n)}; then according to {G d,2 (m,n)}, {μ 6 (m ,n)}, {σ 6 (m,n)}, get the normalized image of {G d,2 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 6 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 6 (m,n)}, and σ 6 (m,n) represents {σ 6 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,3(m,n)}的局部平均值图像,记为{μ7(m,n)};并采用现有的方差求取方法,求得{Gd,3(m,n)}的局部方差图像,记为{σ7(m,n)};然后根据{Gd,3(m,n)}、{μ7(m,n)}、{σ7(m,n)},获取{Gd,3(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ7(m,n)表示{μ7(m,n)}中坐标位置为(m,n)的像素点的像素值,σ7(m,n)表示{σ7(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,3 (m,n)}, denoted as {μ 7 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,3 (m,n)}, denoted as {σ 7 (m,n)}; then according to {G d,3 (m,n)}, {μ 7 (m ,n)}, {σ 7 (m,n)}, get the normalized image of {G d,3 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 7 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 7 (m,n)}, and σ 7 (m,n) represents {σ 7 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,4(m,n)}的局部平均值图像,记为{μ8(m,n)};并采用现有的方差求取方法,求得{Gd,4(m,n)}的局部方差图像,记为{σ8(m,n)};然后根据{Gd,4(m,n)}、{μ8(m,n)}、{σ8(m,n)},获取{Gd,4(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ8(m,n)表示{μ8(m,n)}中坐标位置为(m,n)的像素点的像素值,σ8(m,n)表示{σ8(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,4 (m,n)}, denoted as {μ 8 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,4 (m,n)}, denoted as {σ 8 (m,n)}; then according to {G d,4 (m,n)}, {μ 8 (m ,n)}, {σ 8 (m,n)}, get the normalized image of {G d,4 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 8 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 8 (m,n)}, and σ 8 (m,n) represents {σ 8 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,5(m,n)}的局部平均值图像,记为{μ9(m,n)};并采用现有的方差求取方法,求得{Gd,5(m,n)}的局部方差图像,记为{σ9(m,n)};然后根据{Gd,5(m,n)}、{μ9(m,n)}、{σ9(m,n)},获取{Gd,5(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ9(m,n)表示{μ9(m,n)}中坐标位置为(m,n)的像素点的像素值,σ9(m,n)表示{σ9(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,5 (m,n)}, denoted as {μ 9 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,5 (m,n)}, denoted as {σ 9 (m,n)}; then according to {G d,5 (m,n)}, {μ 9 (m ,n)}, {σ 9 (m,n)}, get the normalized image of {G d,5 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 9 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 9 (m,n)}, and σ 9 (m,n) represents {σ 9 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

采用现有的平均值求取方法,求得{Gd,6(m,n)}的局部平均值图像,记为{μ10(m,n)};并采用现有的方差求取方法,求得{Gd,6(m,n)}的局部方差图像,记为{σ10(m,n)};然后根据{Gd,6(m,n)}、{μ10(m,n)}、{σ10(m,n)},获取{Gd,6(m,n)}的归一化图像,记为中坐标位置为(m,n)的像素点的像素值记为 其中,μ10(m,n)表示{μ10(m,n)}中坐标位置为(m,n)的像素点的像素值,σ10(m,n)表示{σ10(m,n)}中坐标位置为(m,n)的像素点的像素值。Using the existing average calculation method, obtain the local average value image of {G d,6 (m,n)}, denoted as {μ 10 (m,n)}; and use the existing variance calculation method , get the local variance image of {G d,6 (m,n)}, denoted as {σ 10 (m,n)}; then according to {G d,6 (m,n)}, {μ 10 (m ,n)}, {σ 10 (m,n)}, get the normalized image of {G d,6 (m,n)}, denoted as Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 10 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 10 (m,n)}, and σ 10 (m,n) represents {σ 10 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).

步骤四:采用P幅原始的无失真屏幕图像,建立其在不同失真类型不同失真程度下的失真屏幕图像集合,将该失真屏幕图像集合作为训练集,训练集包括多幅失真屏幕图像;然后利用现有的主观质量评价方法评价出训练集中的每幅失真屏幕图像的主观评分值,将训练集中的第j幅失真屏幕图像的主观评分值记为DMOSj;并按照步骤一至步骤三的操作,以相同的方式获取训练集中的每幅失真屏幕图像的十个通道特征图各自的归一化图像,将训练集中的第j幅失真屏幕图像的十个通道特征图各自的归一化图像对应记为 其中,P>1,如取P=100,1≤j≤K',K'表示训练集中包含的失真屏幕图像的总幅数,0≤DMOSj≤100,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值。Step 4: Using P original undistorted screen images, set up a set of distorted screen images under different distortion types and different degrees of distortion, and use the set of distorted screen images as a training set. The training set includes multiple distorted screen images; then use Existing subjective quality evaluation method evaluates the subjective scoring value of each distorted screen image in the training set, and the subjective scoring value of the jth distorted screen image in the training set is recorded as DMOS j ; and according to the operation of step 1 to step 3, Obtain the normalized images of the ten channel feature maps of each distorted screen image in the training set in the same way, and record the normalized images of the ten channel feature maps of the jth distorted screen image in the training set correspondingly for Among them, P>1, such as taking P=100, 1≤j≤K', K' indicates the total number of distorted screen images contained in the training set, 0≤DMOS j ≤100, express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel whose middle coordinate position is (m,n).

步骤五:卷积神经网络(Convolution Neural Network,CNN)是基于大脑神经网络设计的机器学习方法,其可以有效地抑制过拟合,因此本发明方法利用卷积神经网络对训练集中的所有失真屏幕图像各自的主观评分值及各自的十个通道特征图的归一化图像进行训练,使得经过训练得到的回归函数值与主观评分值之间的误差最小,拟合得到最优的权值矢量Wopt和最优的偏置项bopt;接着利用Wopt构造得到卷积神经网络回归训练模型;再根据卷积神经网络回归训练模型,对 进行测试,预测得到{Id(i,j)}的客观质量评价预测值,记为Q,Q=f(x),其中,Q是x的函数,f()为函数表示形式,x为输入变量,x表示 (Wopt)T为Wopt的转置矢量,为x的卷积函数。Step 5: Convolution Neural Network (CNN) is a machine learning method based on brain neural network design, which can effectively suppress overfitting, so the method of the present invention utilizes Convolution Neural Network for all distorted screens in the training set The respective subjective score values of the images and the normalized images of the respective ten channel feature maps are used for training, so that the error between the regression function value obtained after training and the subjective score value is the smallest, and the optimal weight vector W is obtained by fitting. opt and the optimal bias item b opt t; then use W opt to construct the convolutional neural network regression training model; then according to the convolutional neural network regression training model, for Carry out the test and predict the objective quality evaluation prediction value of {I d (i, j)}, denoted as Q, Q=f(x), Among them, Q is a function of x, f() is a function representation, x is an input variable, and x represents (W opt ) T is the transpose vector of W opt , Convolution function for x.

卷积神经网络包括两个卷积层、一个全连接层、一个输出层,卷积层使用的滤波器的内核个数都是64,大小为3×3,其中第一个卷积层后面跟一个最大池化层,第二个卷积层后面跟一个最大池化层和一个平均池化层,两个最大池化层和一个平均池化层的大小都是7×7,步幅为2。使用的激活函数都是修正线性单元函数(Rectified Linear Unit,ReLU)。ReLU可以用f(y)=max(0,y)来表示,y为输入变量,max()为取最大值函数。使用ReLU作为激活函数的原因是:ReLU可以通过简单地将零激活矩阵阈值化来实现;ReLU不会饱和,并且使用ReLU作为激活函数可以很大程度地加快随机梯度下降的收敛。The convolutional neural network includes two convolutional layers, a fully connected layer, and an output layer. The number of filter kernels used in the convolutional layer is 64, and the size is 3×3. The first convolutional layer is followed by A max pooling layer, a second convolutional layer followed by a max pooling layer and an average pooling layer, both max pooling layers and an average pooling layer are of size 7×7 with a stride of 2 . The activation functions used are Rectified Linear Unit (ReLU). ReLU can be represented by f(y)=max(0,y), y is the input variable, and max() is the maximum value function. The reason for using ReLU as the activation function is: ReLU can be implemented by simply thresholding the zero activation matrix; ReLU will not saturate, and using ReLU as the activation function can greatly speed up the convergence of stochastic gradient descent.

为了进一步验证本发明方法的可行性和有效性,进行实验。In order to further verify the feasibility and effectiveness of the method of the present invention, experiments were carried out.

为此,采用失真屏幕图像数据库来分析利用本发明方法得到的失真屏幕图像的客观质量评价预测值与主观评分值之间的相关性。这里,利用评估图像质量评价方法的3个常用客观参量作为评价指标,即非线性回归条件下的Pearson相关系数(Pearson linearcorrelation coefficient,PLCC)、Spearman相关系数(Spearman rank ordercorrelation coefficient,SROCC)、均方误差(root mean squared error,RMSE),PLCC和RMSE反映失真屏幕图像的客观质量评价预测值的准确性,SROCC反映其单调性。For this reason, the distorted screen image database is used to analyze the correlation between the objective quality evaluation prediction value and the subjective rating value of the distorted screen image obtained by the method of the present invention. Here, three commonly used objective parameters for evaluating image quality evaluation methods are used as evaluation indicators, namely Pearson correlation coefficient (Pearson linear correlation coefficient, PLCC) under nonlinear regression conditions, Spearman correlation coefficient (Spearman rank order correlation coefficient, SROCC), mean square Error (root mean squared error, RMSE), PLCC and RMSE reflect the accuracy of the objective quality evaluation prediction value of the distorted screen image, and SROCC reflects its monotonicity.

利用现有的主观质量评价方法获得失真屏幕图像数据库中的每幅失真屏幕图像的主观评分值,再利用本发明方法计算失真屏幕图像数据库中的每幅失真屏幕图像的客观质量评价预测值。将按本发明方法计算得到的失真屏幕图像的客观质量评价预测值做五参数Logistic函数非线性拟合,PLCC和SROCC值越高,RMSE值越低说明客观评价方法的客观评价结果与主观评分值之间的相关性越好。反映本发明方法的质量评价性能的PLCC、SROCC和RMSE相关系数如表1所列。从表1所列的数据可知,按本发明方法得到的失真屏幕图像的客观质量评价预测值与主观评分值之间的相关性是很好的,表明客观评价结果与人眼主观感知的结果较为一致,足以说明本发明方法的可行性和有效性。表1利用本发明方法得到的失真屏幕图像的客观质量评价预测值与主观评分值之间的Using the existing subjective quality evaluation method to obtain the subjective scoring value of each distorted screen image in the distorted screen image database, and then using the method of the present invention to calculate the objective quality evaluation prediction value of each distorted screen image in the distorted screen image database. The objective quality evaluation prediction value of the distorted screen image calculated by the method of the present invention is used as a five-parameter Logistic function nonlinear fitting, the higher the PLCC and SROCC values, the lower the RMSE value, indicating the objective evaluation results and subjective scoring values of the objective evaluation method The better the correlation between. The PLCC, SROCC and RMSE correlation coefficients reflecting the quality evaluation performance of the method of the present invention are listed in Table 1. As can be seen from the data listed in Table 1, the correlation between the objective quality evaluation prediction value and the subjective scoring value of the distorted screen image obtained by the method of the present invention is very good, showing that the objective evaluation result and the result of subjective perception of human eyes are relatively good. Consistent enough to illustrate the feasibility and effectiveness of the method of the present invention. Table 1 Utilizes the difference between the objective quality evaluation prediction value and the subjective rating value of the distorted screen image obtained by the method of the present invention

相关性Correlation

Claims (2)

1.一种基于通道特征和卷积神经网络的屏幕图像质量评价方法,其特征在于包括以下步骤:1. a screen image quality evaluation method based on channel feature and convolutional neural network, is characterized in that comprising the following steps: 步骤一:令{Id(i,j)}表示待评价的失真屏幕图像,其中,1≤i≤W,1≤j≤H,W表示{Id(i,j)}的宽度,H表示{Id(i,j)}的高度,Id(i,j)表示{Id(i,j)}中坐标位置为(i,j)的像素点的像素值;Step 1: Let {I d (i, j)} represent the distorted screen image to be evaluated, where, 1≤i≤W, 1≤j≤H, W represents the width of {I d (i,j)}, H Represents the height of {I d (i, j)}, I d (i, j) represents the pixel value of the pixel whose coordinate position is (i, j) in {I d (i, j)}; 步骤二:利用集总特征通道方法对{Id(i,j)}进行特征提取,得到{Id(i,j)}的十个通道特征图,分别为L通道特征图、U通道特征图、V通道特征图、梯度幅值通道特征图、第1个方向梯度直方图通道特征图、第2个方向梯度直方图通道特征图、第3个方向梯度直方图通道特征图、第4个方向梯度直方图通道特征图、第5个方向梯度直方图通道特征图、第6个方向梯度直方图通道特征图,对应记为{Ld(m,n)}、{Ud(m,n)}、{Vd(m,n)}、{Gd,0(m,n)}、{Gd,1(m,n)}、{Gd,2(m,n)}、{Gd,3(m,n)}、{Gd,4(m,n)}、{Gd,5(m,n)}、{Gd,6(m,n)},其中,1≤m≤M,1≤n≤N,符号为向下取整操作符号,M表示{Id(i,j)}的每个通道特征图的宽度,N表示{Id(i,j)}的每个通道特征图的高度,Ld(m,n)表示{Ld(m,n)}中坐标位置为(m,n)的像素点的像素值,Ud(m,n)表示{Ud(m,n)}中坐标位置为(m,n)的像素点的像素值,Vd(m,n)表示{Vd(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,0(m,n)表示{Gd,0(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,1(m,n)表示{Gd,1(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,2(m,n)表示{Gd,2(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,3(m,n)表示{Gd,3(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,4(m,n)表示{Gd,4(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,5(m,n)表示{Gd,5(m,n)}中坐标位置为(m,n)的像素点的像素值,Gd,6(m,n)表示{Gd,6(m,n)}中坐标位置为(m,n)的像素点的像素值;Step 2: Use the lumped feature channel method to perform feature extraction on {I d (i, j)}, and obtain ten channel feature maps of {I d (i, j)}, which are L channel feature map and U channel feature Figure, V channel feature map, gradient amplitude channel feature map, first directional gradient histogram channel feature map, second directional gradient histogram channel feature map, third directional gradient histogram channel feature map, fourth The directional gradient histogram channel feature map, the 5th directional gradient histogram channel feature map, and the 6th directional gradient histogram channel feature map, correspondingly recorded as {L d (m,n)}, {U d (m,n) )}, {V d (m,n)}, {G d,0 (m,n)}, {G d,1 (m,n)}, {G d,2 (m,n)}, { G d,3 (m,n)}, {G d,4 (m,n)}, {G d,5 (m,n)}, {G d,6 (m,n)}, where 1 ≤m≤M, 1≤n≤N, symbol is the symbol of the rounding down operation, M represents the width of each channel feature map of {I d (i,j)}, N represents the height of each channel feature map of {I d (i,j)}, L d (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {L d (m,n)}, and U d (m,n) represents the coordinates in {U d (m,n)} The pixel value of the pixel point whose position is (m, n), V d (m, n) represents the pixel value of the pixel point whose coordinate position is (m, n) in {V d (m, n)}, G d, 0 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,0 (m,n)}, and G d,1 (m,n) represents {G d,1 ( The pixel value of the pixel whose coordinate position is (m,n) in m,n)}, G d,2 (m,n) means that the coordinate position in {G d,2 (m,n)} is (m,n) ), G d,3 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {G d,3 (m,n)}, G d,4 ( m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {G d,4 (m,n)}, G d,5 (m,n) represents {G d,5 (m, The pixel value of the pixel whose coordinate position is (m,n) in n)}, G d,6 (m,n) represents the pixel value of the coordinate position (m,n) in {G d,6 (m,n)} The pixel value of the pixel point; 步骤三:获取{Ld(m,n)}、{Ud(m,n)}、{Vd(m,n)}、{Gd,0(m,n)}、{Gd,1(m,n)}、{Gd,2(m,n)}、{Gd,3(m,n)}、{Gd,4(m,n)}、{Gd,5(m,n)}、{Gd,6(m,n)}各自的归一化图像,对应记为 其中,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值;Step 3: Get {L d (m,n)}, {U d (m,n)}, {V d (m,n)}, {G d,0 (m,n)}, {G d, 1 (m,n)}, {G d,2 (m,n)}, {G d,3 (m,n)}, {G d,4 (m,n)}, {G d,5 ( The normalized images of m,n)} and {G d,6 (m,n)} are correspondingly denoted as in, express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose middle coordinate position is (m,n); 步骤四:采用P幅原始的无失真屏幕图像,建立其在不同失真类型不同失真程度下的失真屏幕图像集合,将该失真屏幕图像集合作为训练集,训练集包括多幅失真屏幕图像;然后利用主观质量评价方法评价出训练集中的每幅失真屏幕图像的主观评分值,将训练集中的第j幅失真屏幕图像的主观评分值记为DMOSj;并按照步骤一至步骤三的操作,以相同的方式获取训练集中的每幅失真屏幕图像的十个通道特征图各自的归一化图像,将训练集中的第j幅失真屏幕图像的十个通道特征图各自的归一化图像对应记为 其中,P>1,1≤j≤K',K'表示训练集中包含的失真屏幕图像的总幅数,0≤DMOSj≤100,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值,表示中坐标位置为(m,n)的像素点的像素值;Step 4: Using P original undistorted screen images, set up a set of distorted screen images under different distortion types and different degrees of distortion, and use the set of distorted screen images as a training set. The training set includes multiple distorted screen images; then use The subjective quality evaluation method evaluates the subjective scoring value of each piece of distorted screen image in the training set, and records the subjective scoring value of the jth piece of distorted screen image in the training set as DMOS j ; method to obtain the normalized images of the ten channel feature maps of each distorted screen image in the training set, and record the normalized images of the ten channel feature maps of the jth distorted screen image in the training set as Among them, P>1, 1≤j≤K', K' indicates the total number of distorted screen images contained in the training set, 0≤DMOS j ≤100, express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose coordinate position is (m,n), express The pixel value of the pixel point whose middle coordinate position is (m,n); 步骤五:利用卷积神经网络对训练集中的所有失真屏幕图像各自的主观评分值及各自的十个通道特征图的归一化图像进行训练,使得经过训练得到的回归函数值与主观评分值之间的误差最小,拟合得到最优的权值矢量Wopt和最优的偏置项bopt;接着利用Wopt构造得到卷积神经网络回归训练模型;再根据卷积神经网络回归训练模型,对 进行测试,预测得到{Id(i,j)}的客观质量评价预测值,记为Q,Q=f(x),其中,Q是x的函数,f()为函数表示形式,x为输入变量,x表示 为Wopt的转置矢量,为x的卷积函数。Step 5: Use the convolutional neural network to train the respective subjective score values of all the distorted screen images in the training set and the normalized images of the respective ten channel feature maps, so that the relationship between the regression function value obtained after training and the subjective score value The error between is the smallest, and the optimal weight vector W opt and the optimal bias item b opt are obtained by fitting; then the convolutional neural network regression training model is obtained by using the W opt construction; and then according to the convolutional neural network regression training model, right Carry out the test and predict the objective quality evaluation prediction value of {I d (i, j)}, denoted as Q, Q=f(x), Among them, Q is a function of x, f() is a function representation, x is an input variable, and x represents is the transpose vector of W opt , Convolution function for x. 2.根据权利要求1所述的基于通道特征和卷积神经网络的屏幕图像质量评价方法,其特征在于所述的步骤三的具体过程为:2. the screen image quality evaluation method based on channel feature and convolutional neural network according to claim 1, is characterized in that the concrete process of described step 3 is: 求得{Ld(m,n)}的局部平均值图像,记为{μ1(m,n)};并求得{Ld(m,n)}的局部方差图像,记为{σ1(m,n)};然后根据{Ld(m,n)}、{μ1(m,n)}、{σ1(m,n)},获取{Ld(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ1(m,n)表示{μ1(m,n)}中坐标位置为(m,n)的像素点的像素值,σ1(m,n)表示{σ1(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {L d (m,n)}, denoted as {μ 1 (m,n)}; and obtain the local variance image of {L d (m,n)}, denoted as {σ 1 (m,n)}; then according to {L d (m,n)}, {μ 1 (m,n)}, {σ 1 (m,n)}, get {L d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 1 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 1 (m,n)}, and σ 1 (m,n) represents {σ 1 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Ud(m,n)}的局部平均值图像,记为{μ2(m,n)};并求得{Ud(m,n)}的局部方差图像,记为{σ2(m,n)};然后根据{Ud(m,n)}、{μ2(m,n)}、{σ2(m,n)},获取{Ud(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ2(m,n)表示{μ2(m,n)}中坐标位置为(m,n)的像素点的像素值,σ2(m,n)表示{σ2(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {U d (m,n)}, denoted as {μ 2 (m,n)}; and obtain the local variance image of {U d (m,n)}, denoted as {σ 2 (m,n)}; then according to {U d (m,n)}, {μ 2 (m,n)}, {σ 2 (m,n)}, get {U d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 2 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 2 (m,n)}, and σ 2 (m,n) represents {σ 2 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Vd(m,n)}的局部平均值图像,记为{μ3(m,n)};并求得{Vd(m,n)}的局部方差图像,记为{σ3(m,n)};然后根据{Vd(m,n)}、{μ3(m,n)}、{σ3(m,n)},获取{Vd(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为 其中,μ3(m,n)表示{μ3(m,n)}中坐标位置为(m,n)的像素点的像素值,σ3(m,n)表示{σ3(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {V d (m,n)}, denoted as {μ 3 (m,n)}; and obtain the local variance image of {V d (m,n)}, denoted as {σ 3 (m,n)}; then according to {V d (m,n)}, {μ 3 (m,n)}, {σ 3 (m,n)}, get {V d (m,n)} The normalized image of Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 3 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 3 (m,n)}, and σ 3 (m,n) represents {σ 3 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,0(m,n)}的局部平均值图像,记为{μ4(m,n)};并求得{Gd,0(m,n)}的局部方差图像,记为{σ4(m,n)};然后根据{Gd,0(m,n)}、{μ4(m,n)}、{σ4(m,n)},获取{Gd,0(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ4(m,n)表示{μ4(m,n)}中坐标位置为(m,n)的像素点的像素值,σ4(m,n)表示{σ4(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,0 (m,n)}, denoted as {μ 4 (m,n)}; and obtain the local variance image of {G d,0 (m,n)}, Denote it as {σ 4 ( m,n) } ; then get {G d ,0 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 4 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 4 (m,n)}, and σ 4 (m,n) represents {σ 4 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,1(m,n)}的局部平均值图像,记为{μ5(m,n)};并求得{Gd,1(m,n)}的局部方差图像,记为{σ5(m,n)};然后根据{Gd,1(m,n)}、{μ5(m,n)}、{σ5(m,n)},获取{Gd,1(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ5(m,n)表示{μ5(m,n)}中坐标位置为(m,n)的像素点的像素值,σ5(m,n)表示{σ5(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,1 (m,n)}, denoted as {μ 5 (m,n)}; and obtain the local variance image of {G d,1 (m,n)}, Recorded as {σ 5 (m,n)}; then according to {G d,1 (m,n)}, {μ 5 (m,n)}, {σ 5 (m,n)}, get {G d ,1 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 5 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 5 (m,n)}, and σ 5 (m,n) represents {σ 5 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,2(m,n)}的局部平均值图像,记为{μ6(m,n)};并求得{Gd,2(m,n)}的局部方差图像,记为{σ6(m,n)};然后根据{Gd,2(m,n)}、{μ6(m,n)}、{σ6(m,n)},获取{Gd,2(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ6(m,n)表示{μ6(m,n)}中坐标位置为(m,n)的像素点的像素值,σ6(m,n)表示{σ6(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,2 (m,n)}, denoted as {μ 6 (m,n)}; and obtain the local variance image of {G d,2 (m,n)}, Denote it as {σ 6 (m,n)}; then according to {G d,2 (m,n)}, {μ 6 (m,n)}, {σ 6 (m,n) } , get ,2 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 6 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 6 (m,n)}, and σ 6 (m,n) represents {σ 6 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,3(m,n)}的局部平均值图像,记为{μ7(m,n)};并求得{Gd,3(m,n)}的局部方差图像,记为{σ7(m,n)};然后根据{Gd,3(m,n)}、{μ7(m,n)}、{σ7(m,n)},获取{Gd,3(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ7(m,n)表示{μ7(m,n)}中坐标位置为(m,n)的像素点的像素值,σ7(m,n)表示{σ7(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,3 (m,n)}, denoted as {μ 7 (m,n)}; and obtain the local variance image of {G d,3 (m,n)}, Denote it as {σ 7 (m,n) } ; then get {G d ,3 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 7 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 7 (m,n)}, and σ 7 (m,n) represents {σ 7 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,4(m,n)}的局部平均值图像,记为{μ8(m,n)};并求得{Gd,4(m,n)}的局部方差图像,记为{σ8(m,n)};然后根据{Gd,4(m,n)}、{μ8(m,n)}、{σ8(m,n)},获取{Gd,4(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ8(m,n)表示{μ8(m,n)}中坐标位置为(m,n)的像素点的像素值,σ8(m,n)表示{σ8(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average image of {G d,4 (m,n)}, denoted as {μ 8 (m,n)}; and obtain the local variance image of {G d,4 (m,n)}, Denote it as { σ 8 (m,n)} ; then get {G d ,4 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 8 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 8 (m,n)}, and σ 8 (m,n) represents {σ 8 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,5(m,n)}的局部平均值图像,记为{μ9(m,n)};并求得{Gd,5(m,n)}的局部方差图像,记为{σ9(m,n)};然后根据{Gd,5(m,n)}、{μ9(m,n)}、{σ9(m,n)},获取{Gd,5(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ9(m,n)表示{μ9(m,n)}中坐标位置为(m,n)的像素点的像素值,σ9(m,n)表示{σ9(m,n)}中坐标位置为(m,n)的像素点的像素值;Obtain the local average value image of {G d,5 (m,n)}, denoted as {μ 9 (m,n)}; and obtain the local variance image of {G d,5 (m,n)}, Denote it as { σ 9 (m,n)} ; then get {G d ,5 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 9 (m,n) represents the pixel value of the pixel whose coordinate position is (m,n) in {μ 9 (m,n)}, and σ 9 (m,n) represents {σ 9 (m,n) )} in the pixel value of the pixel whose coordinate position is (m, n); 求得{Gd,6(m,n)}的局部平均值图像,记为{μ10(m,n)};并求得{Gd,6(m,n)}的局部方差图像,记为{σ10(m,n)};然后根据{Gd,6(m,n)}、{μ10(m,n)}、{σ10(m,n)},获取{Gd,6(m,n)}的归一化图像中坐标位置为(m,n)的像素点的像素值记为其中,μ10(m,n)表示{μ10(m,n)}中坐标位置为(m,n)的像素点的像素值,σ10(m,n)表示{σ10(m,n)}中坐标位置为(m,n)的像素点的像素值。Obtain the local average value image of {G d,6 (m,n)}, denoted as {μ 10 (m,n)}; and obtain the local variance image of {G d,6 (m,n)}, Denote it as {σ 10 (m,n) } ; then get {G d ,6 (m,n)} normalized image Will The pixel value of the pixel point whose coordinate position is (m, n) is recorded as Among them, μ 10 (m,n) represents the pixel value of the pixel point whose coordinate position is (m,n) in {μ 10 (m,n)}, and σ 10 (m,n) represents {σ 10 (m,n) )} The pixel value of the pixel whose coordinate position is (m,n).
CN201710826241.5A 2017-09-14 2017-09-14 Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network Pending CN107590804A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710826241.5A CN107590804A (en) 2017-09-14 2017-09-14 Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710826241.5A CN107590804A (en) 2017-09-14 2017-09-14 Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network

Publications (1)

Publication Number Publication Date
CN107590804A true CN107590804A (en) 2018-01-16

Family

ID=61051468

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710826241.5A Pending CN107590804A (en) 2017-09-14 2017-09-14 Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network

Country Status (1)

Country Link
CN (1) CN107590804A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108428227A (en) * 2018-02-27 2018-08-21 浙江科技学院 No-reference image quality assessment method based on fully convolutional neural network
WO2019141255A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Image filtering method and device
CN110473148A (en) * 2019-01-17 2019-11-19 张红梅 Computer Quality estimation system
CN112653884A (en) * 2019-10-10 2021-04-13 北京滴普科技有限公司 Method for evaluating quality of screen

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105208374A (en) * 2015-08-24 2015-12-30 宁波大学 Non-reference image quality objective evaluation method based on deep learning
CN105979253A (en) * 2016-05-06 2016-09-28 浙江科技学院 Generalized regression neural network based non-reference stereoscopic image quality evaluation method
CN107146216A (en) * 2017-04-07 2017-09-08 浙江科技学院 A No-Reference Image Objective Quality Assessment Method Based on Gradient Self-similarity

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105208374A (en) * 2015-08-24 2015-12-30 宁波大学 Non-reference image quality objective evaluation method based on deep learning
CN105979253A (en) * 2016-05-06 2016-09-28 浙江科技学院 Generalized regression neural network based non-reference stereoscopic image quality evaluation method
CN107146216A (en) * 2017-04-07 2017-09-08 浙江科技学院 A No-Reference Image Objective Quality Assessment Method Based on Gradient Self-similarity

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
万维: "基于深度学习的目标检测算法研究及应用", 《中国优秀硕士论文全文数据库》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019141255A1 (en) * 2018-01-18 2019-07-25 杭州海康威视数字技术股份有限公司 Image filtering method and device
CN108428227A (en) * 2018-02-27 2018-08-21 浙江科技学院 No-reference image quality assessment method based on fully convolutional neural network
CN108428227B (en) * 2018-02-27 2020-06-26 浙江科技学院 No-reference image quality evaluation method based on full convolution neural network
CN110473148A (en) * 2019-01-17 2019-11-19 张红梅 Computer Quality estimation system
CN112653884A (en) * 2019-10-10 2021-04-13 北京滴普科技有限公司 Method for evaluating quality of screen
CN112653884B (en) * 2019-10-10 2022-12-23 北京滴普科技有限公司 A method for evaluating screen quality

Similar Documents

Publication Publication Date Title
CN108428227B (en) No-reference image quality evaluation method based on full convolution neural network
Zhang et al. Hierarchical feature fusion with mixed convolution attention for single image dehazing
Zhang et al. Edge strength similarity for image quality assessment
CN108665460B (en) Image Quality Evaluation Method Based on Combination Neural Network and Classification Neural Network
CN109167996B (en) It is a kind of based on convolutional neural networks without reference stereo image quality evaluation method
CN104023230B (en) A kind of non-reference picture quality appraisement method based on gradient relevance
CN114066812B (en) No-reference image quality evaluation method based on spatial attention mechanism
CN105574901B (en) A general no-reference image quality assessment method based on local contrast mode
CN105357519B (en) Non-reference stereo image quality objective evaluation method based on self-similarity characteristics
CN112767385B (en) No-reference image quality evaluation method based on significance strategy and feature fusion
CN107743225B (en) A Method for No-Reference Image Quality Prediction Using Multi-Layer Depth Representations
CN107590804A (en) Screen Image Quality Evaluation Method Based on Channel Feature and Convolutional Neural Network
CN105451016A (en) No-reference video quality evaluation method suitable for video monitoring system
CN104658001A (en) Non-reference asymmetric distorted stereo image objective quality assessment method
CN106127234B (en) A no-reference image quality assessment method based on feature dictionary
CN115965889A (en) A video quality assessment data processing method, device and equipment
CN107146216A (en) A No-Reference Image Objective Quality Assessment Method Based on Gradient Self-similarity
CN106791822A (en) It is a kind of based on single binocular feature learning without refer to stereo image quality evaluation method
CN115239647A (en) Full-reference video quality evaluation method based on two stages of self-adaptive sampling and multi-scale time sequence
CN104574363B (en) A kind of full reference image quality appraisement method for considering gradient direction difference
CN110400307A (en) A Screen Image Quality Evaluation Method Based on Regional Difference
CN106683079B (en) A kind of non-reference picture method for evaluating objective quality based on structure distortion
CN107274388A (en) A no-reference screen image quality assessment method based on global information
CN105469413B (en) It is a kind of based on normalization ring weighting without refer to smear restoration image synthesis method for evaluating quality
CN106023152A (en) Reference-free stereo image quality objective evaluation method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180116

RJ01 Rejection of invention patent application after publication
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载