US20090161984A1 - Apparatus and method for image processing, and program for use therein, and learning apparatus - Google Patents
Apparatus and method for image processing, and program for use therein, and learning apparatus Download PDFInfo
- Publication number
- US20090161984A1 US20090161984A1 US12/326,413 US32641308A US2009161984A1 US 20090161984 A1 US20090161984 A1 US 20090161984A1 US 32641308 A US32641308 A US 32641308A US 2009161984 A1 US2009161984 A1 US 2009161984A1
- Authority
- US
- United States
- Prior art keywords
- image
- pixel
- interest
- pixel values
- processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000012545 processing Methods 0.000 title claims abstract description 186
- 238000000034 method Methods 0.000 title description 7
- 239000011159 matrix material Substances 0.000 claims description 27
- 230000002093 peripheral effect Effects 0.000 claims description 27
- 238000003672 processing method Methods 0.000 claims description 7
- 239000000284 extract Substances 0.000 description 12
- 230000006978 adaptation Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 6
- 238000004519 manufacturing process Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/98—Adaptive-dynamic-range coding [ADRC]
Definitions
- the present invention contains subject matter related to Japanese Patent Application JP 2007-330452 filed in the Japan Patent Office on Dec. 21, 2007, the entire contents of which being incorporated herein by reference.
- the present invention relates to an image processing apparatus, an image processing method, and a program for use therein, and a learning apparatus, and more particularly to an image processing apparatus and an image processing method each of which is capable of removing a noise, which is generally generated and which has luminance dependency, from a video signal corresponding to an image, and a program for use therein, and a learning apparatus.
- the noise is treated as a white color, and pixel values of the peripheral pixels are added to one another, thereby removing the noise from a video signal corresponding to an image.
- the noise removing processing it has been devised in the noise removing processing that the noise is assumed to have a color because there is a deviation in color of the noise, and thus the noise is removed with a high degree of accuracy as compared with the case where the noise is treated as the white color.
- the present invention has been made in the light of such circumferences, and it is therefore desirable to provide an image processing apparatus and an image processing method each of which is capable of removing a noise, which is generally generated and which has luminance dependency, from a video signal corresponding to an image, and a program for use therein, and a learning apparatus.
- an image processing apparatus for converting a first image into a second image having higher image quality than that of the first image, including: a first pixel value extracting section for extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; an estimate noise amount arithmetically operating section for obtaining estimate noise amounts for the plurality of pixel values extracted by the first pixel value extracting section, respectively; a processing coefficient generating section for generating second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher image quality than that of a student image in accordance with an arithmetic operation for the
- an image processing method for use in an image processing apparatus for converting a first image into a second image having higher image quality than that of the first image, the image processing method including the steps of: extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; obtaining estimate noise amounts for the plurality of pixel values, respectively; obtaining second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher quality than that of a student image in accordance with an arithmetic operation for the second processing coefficients obtained in accordance with estimate noise amounts about the pixel values within the student image corresponding to the first image, and the first processing coefficients, and
- the plurality of pixel values, within the first image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest within the second image having the higher image quality than that of the first image are extracted.
- the estimate noise amounts are obtained with respect to the plurality of pixel values thus extracted, respectively.
- the second processing coefficients are obtained in accordance with the arithmetic operation for the first processing coefficients previously learned from the normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image.
- the normal equation is obtained based on the relational expression for generating the teacher image corresponding to the second image having the higher image quality than that of the student image in accordance with the arithmetic operation for the second processing coefficients obtained in accordance with the estimate noise amounts about the pixel values, within the student image, corresponding to the first image and the first processing coefficients, and the student image.
- the plurality of pixel values, within the first image, corresponding to the position of the noticed image, and the peripheral positions of the position of the noticed image within the second image are extracted.
- the pixel value of the pixel of interest within the second image is generated in accordance with the arithmetic operation for the plurality of pixel values thus extracted, and the second processing.
- a program in accordance with which a computer executes image processing for converting a first image into a second image having higher image quality than that of the first image including the steps of: extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; obtaining estimate noise amounts for the plurality of pixel values, respectively; obtaining second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher quality than that of a student image in accordance with an arithmetic operation for the second processing coefficients obtained in accordance with estimate noise amounts about the pixel values within the student image corresponding to the first image, and the first processing coefficient
- a learning apparatus including: a normal equation generating section for obtaining estimate noise amounts from pixel values within a student image containing therein a noise having noise dependency, obtaining second processing coefficients in accordance with an arithmetic operation for the estimate noise amounts, and first processing coefficients, and generating a normal equation by using the pixel values of the student image, and pixel values of the teacher image, in order to solve a relational expression for generating the teacher image having higher image quality than that of the student image, in accordance with an arithmetic operation for the second processing coefficients and the student image; and a coefficient generating section for generating the first processing coefficients by solving the normal equation.
- the estimate noise amounts are obtained from the pixel values within the student image containing therein the noise having the luminance dependency.
- the second processing coefficients are obtained in accordance with the arithmetic operation for the estimate noise amounts, and the first processing coefficients.
- the normal equation is generated by using the pixel values of the student image, and the pixel values of the student image, in order to solve the relational expression for generating the teacher image having the higher image quality than that of the student image, in accordance with the arithmetic operation for the second processing coefficients and the student image.
- the first processing coefficients are generated by solving the normal equation.
- the noise which is generally generated and which has the luminance dependency can be removed from the video signal corresponding to the image.
- FIG. 1 is a block diagram showing a configuration of an image processing apparatus according to an embodiment of the present invention
- FIG. 2 is a graph showing light quantity dependency of amounts of noises which two kinds of image sensors have, respectively;
- FIG. 3 is a diagram explaining a norm for learning for predictive coefficients
- FIG. 4 is a view showing a class tap composed of pixel values of pixels of 3 ⁇ 3;
- FIG. 5 is a graph showing predictive coefficients which are multiplied by pixel values composing a predictive tap
- FIG. 6 is a flow chart explaining noise removing processing executed by the image processing apparatus shown in FIG. 1 ;
- FIG. 7 is a diagram showing an addition matrix
- FIG. 8 is a block diagram showing a configuration of a learning apparatus for learning processing coefficients according to an embodiment of the present invention.
- FIG. 9 is a flow chart explaining the learning processing executed by the learning apparatus shown in FIG. 8 ;
- FIG. 10 is a block diagram showing a configuration of hardware of a computer.
- FIG. 1 is a block diagram showing a configuration of an image processing apparatus 10 according to an embodiment of the present invention.
- the image processing apparatus 10 shown in FIG. 1 is composed of a class tap extracting portion 11 , a class classifying portion 12 , a coefficient storing portion 13 , a noise amount tap extracting portion 14 , a noise amount arithmetically operating portion 15 , a predictive coefficient generating portion 16 , a predictive tap extracting portion 17 , and a predicting portion 18 .
- the image processing apparatus 10 executes class classifying adaptation processing.
- class classifying adaptation processing a class for a pixel of interest in an output image which will be generated after this is generated, and predictive coefficients of the output image having higher image quality than that of an input image are generated from the input image by using predictive coefficients obtained based on a processing coefficient of the class, and an estimate value for an amount of noise as a variance value of a luminance dependency noise of the input image (hereinafter referred to as “an estimate noise amount”), and the input image.
- the processing coefficients used in the class classifying adaptation processing executed by the image processing apparatus 10 is obtained from learning (its details will be described later) using an image containing therein the luminance dependency noise, and an image from which the luminance dependency noise is removed. Therefore, the image processing apparatus 10 can generate an image, having high image quality, from which the luminance dependency noise is removed by executing the class classifying adaptation processing. From this, the class classifying adaptation processing executed by the image processing apparatus 10 can be the as noise removing processing for removing the luminance dependency noise from the input image as it were.
- the class tap extracting portion 11 successively determines pixels composing an output image (this output image is one which will be generated after this and does not exist at the current stage, and thus is virtually supposed) as pixels of interest.
- the class tap extracting portion 11 extracts pixel values of a plurality of pixels, within the input image, corresponding to a position of the pixel of interest, and peripheral positions of the position of the pixel of interest as a class tap from the input image.
- the pixel values of the plurality of pixels within the input image are used to classify the pixels of interest into classes.
- the class tap extracting portion 11 supplies the class tap to the class classifying portion 12 .
- the class classifying portion 12 classifies the pixels of interest into classes in correspondence to the feature of the class tap supplied thereto from the class tap extracting portion 11 , thereby generating the classes for the pixels of interest.
- Adaptive dynamic range coding (ADRC) or the like, for example, can be adopted as a method of classifying the pixels of interest into the classes.
- ADRC Adaptive dynamic range coding
- the pixel values composing the class tap are ADRC-processed, and the classes of the pixels of interest are determined in accordance with ADRC codes obtained as the result of executing the ADRC processing.
- the pixels of interest are classified into the classes in accordance with correlation (waveform) of the class tap.
- K-bit ADRC for example, a maximum value MAX and a minimum value MIN of the pixel values composing the class tap are detected.
- each of the plurality of pixel values as the class tap is re-quantized into K-bits based on the dynamic range DR. That is to say, the minimum value MIN is subtracted from each of the pixel values as the class tap, and each of the resulting values obtained from the subtraction is divided by DR/2 K (quantized).
- a bit string which is obtained by arranging the K-bit data, as the class tap, obtained in the manner described above in the predetermined order is set as the ADRC code.
- the class tap for example, is subjected to 1-bit ADRC processing
- the minimum value MIN is subtracted from each of the pixel values as the class tap is divided by 1 ⁇ 2 of a difference between the maximum value MAX and the minimum value MIN (rounding-down of a fractional part).
- each data is set as 1 bit (binarized).
- a bit string obtained by arranging the data of 1 bit in the predetermined order is set as the ADRC code.
- the data on the classes for the pixels of interest generated by the class classifying portion are supplied to the coefficient storing portion 13 .
- the coefficient storing portion 13 stores therein the optimal processing coefficients for the classes obtained from the learning which will be described later.
- the coefficient storing portion 13 reads out the processing coefficients corresponding to the classes the data on which is supplied from the class classifying portion 12 , and supplies the processing coefficients thus read out to the predictive coefficient generating portion 16 .
- the noise amount tap extracting portion 14 successively determines the pixels composing the output image as the pixels of interest similarly to the case of the class tap extracting portion 11 .
- the noise amount tap extracting portion 14 extracts the pixel values of the plurality of pixels, within the input image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the noise amount tap from the input image.
- the pixel values of the plurality of pixels within the input image are used to arithmetically operate the estimate noise amount corresponding to the pixel of interest.
- the noise amount tap extracting portion 14 supplies the noise amount tap to the noise amount arithmetically operating portion 15 .
- the noise amount arithmetically operating portion 15 arithmetically operates the estimate noise amount corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amount tap extracting portion 14 , and supplies the resulting estimate noise amount to the predictive coefficient generating portion 16 .
- the predictive coefficient generating portion 16 carries out a predetermined matrix arithmetic operation by using the processing coefficients supplied thereto from the coefficient storing portion 13 , and the estimate noise amount supplied thereto from the noise amount arithmetically operating portion 15 , thereby generating the predictive coefficients.
- the predictive coefficient generating portion 16 supplies the predictive coefficients thus generated to the predicting portion 18 .
- the predictive tap extracting portion 17 successively determines the pixels composing the output image as the pixels of interest similarly to the case of each of the class tap extracting portion 11 and the noise amount tap extracting portion 14 .
- the predictive tap extracting portion 17 extracts the pixel values of the plurality of pixels, within the input image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the predictive tap from the input image. In this case, the pixel values of the plurality of pixels within the input image are used to predict the pixels of interest. Also, the predictive tap extracting portion 17 supplies the predictive tap thus extracted to the predicting portion 18 .
- the predicting portion 18 carries out the predictive arithmetic operation for predicting the pixel of interest by using the predictive coefficients supplied thereto from the predictive coefficient generating portion 16 , and the predictive tap supplied thereto from the predictive tap extracting portion 17 . As a result, the predicting portion 18 generates the predictive value of the pixel value of the pixel of interest as the pixel value of the pixel of interest composing the output image. Also, the predicting portion 18 outputs the output image composed of the pixel values obtained from the predictive arithmetic operation.
- a graph of FIG. 2 shows light quantity (luminance) dependency of amounts of noises which two kinds of image sensors have, respectively. It is noted that in FIG. 2 , an axis of abscissa represents a total light quantity, that is, a pixel value, and an axis of ordinate represents an amount of noise.
- the amount of noise a can be expressed in the form of a quadratic expression of a true value L of the pixel value given by Expression (1):
- the noise amount arithmetically operating portion 15 may obtain the amount of noise by carrying out the arithmetic operation for Expression (1).
- the estimate noise amount ⁇ ′ is arithmetically operated by carrying out the arithmetic operation for Expression (2) by using a pixel value L′ containing therein the noise:
- the predicting portion 18 When it is assumed that in the image processing apparatus 10 shown in FIG. 1 , the predicting portion 18 , for example, carries out a linear predictive arithmetic operation as a predetermined predictive arithmetic operation, the pixel value y composing the output image is obtained in accordance with Expression (3) as a linear expression:
- the optical predictive coefficient W can be obtained from the learning which will be described later.
- the following two items are required as the norm for the learning thereof.
- a first item is such that the predictive coefficient W is changed based on the correlation (waveform) of the predictive tap. That is to say, the first item is such that the predictive coefficient W which is multiplied by the pixel value is made large in that pixel value, nearer the pixel value of the pixel of interest, of the pixel values composing the predictive tap.
- the predictive tap is composed of a pixel value x 1 of a pixel of an input image corresponding to a position of a pixel of interest, a pixel value x 0 near the pixel value x 1 , and a pixel value x 2 far from the pixel value x 1 .
- the predictive coefficient W needs to be learned so that the predictive coefficient w 0 which is multiplied by the pixel value x 0 becomes larger than the predictive coefficient w 2 which is multiplied by the pixel value x 2 which is located farther from the pixel value x 1 than the pixel value x 0 is located.
- a second item is such that the predictive coefficient W is changed based on the estimate noise amount ⁇ ′ of the input image. That is to say, the predictive coefficient W which is multiplied by the pixel value is made large in that pixel value having the less estimate noise amount ⁇ ′.
- a predictive tap is composed of a pixel value x 1 of the pixel of the input image corresponding to the position of a pixel of interest, a pixel value x 0 having an estimate noise amount ⁇ 0 , and a pixel value x 2 having an estimate noise amount ⁇ 2 larger than the estimate noise amount ⁇ 0 .
- the predictive coefficient W needs to be learned so that the predictive coefficient w 0 which is multiplied by the pixel value x 0 becomes larger than the predictive coefficient w 2 which is multiplied by the pixel value x 2 having the larger estimate noise amount x 2 than the estimate noise ⁇ 0 of the pixel value x 0 .
- Cluster classification learning is known as a method of learning the optimal predictive coefficient in accordance with the norm as described above.
- the pixels of interest are classified into classes based on the correlation of the cluster tap, and the estimate noise amount ⁇ ′, and the predictive coefficient is obtained every class.
- the number of classes explosively increases, so that it becomes difficult to learn the predictive coefficients. For example, when the class tap, as shown in FIG.
- the optimal processing coefficient for each of the classes obtained by the classification based on the correlation of the class tap is obtained in the learning which will be described later. Also, the image processing apparatus 10 optimizes that processing coefficient in correspondence to the estimate noise amount ⁇ ′ in the class classification adaptation processing, and carries out the predictive arithmetic operation with the optimized processing coefficient as the predictive coefficient W.
- the maximum signification of the class classification learning is to change the value of the predictive coefficient every class.
- the predictive coefficients which are multiplied by the pixel values, respectively, within the predictive tap are different from each other in two different classes.
- an axis of abscissa represents the positions of the pixels corresponding to the pixel values composing the predictive tap
- an axis of ordinate represents the predictive coefficients which are multiplied by the pixel values of the pixels, respectively.
- the image processing apparatus 10 optimizes the predictive coefficients in accordance with Expression (4) similarly to the case where the predictive coefficient W is changed in correspondence to the estimate noise amount ⁇ ′ based on the optimal processing coefficient for each of the classes obtained by the classification based on the correlation of the class tap obtained from the learning, thereby classifying and learning the pixels of interest based on the correlation of the class tap, and the estimate noise amounts.
- W 0 and W ⁇ ′i are processing coefficients of an n-dimensional vector corresponding to the predictive coefficient W of the n-dimensional vector
- m is the number of pixel values composing the noise amount tap
- ⁇ ′ i is an estimate noise amount of an i-th pixel value within the noise amount tap. It is noted that the number, m, of pixel values composing the noise amount tap may be either identical to or different from the number, n, of pixel values composing the predictive tap.
- Expression (4) is expressed in the form of a determinant of matrix as follows:
- w 0,0 , w 0,1 , . . . , w 0,n represent n elements of the processing coefficients W 0 of the n-dimensional vector, respectively, and w ⁇ i′,0 , w ⁇ i′,1 , . . . , w ⁇ i′,n represent n elements of the processing coefficients W ⁇ i′ , of the n-dimensional vector, respectively.
- Step S 11 the class tap extracting portion 11 , the noise amount tap extracting portion 14 , and the predictive tap extracting portion 17 determine one, of a plurality of pixels composing the output image, which is not yet set as a pixel of interest as the pixel of interest.
- Step S 12 the class tap extracting portion 11 extracts the class tap corresponding to the pixel of interest from the input image, and supplies the class tap thus extracted to the class classifying portion 12 .
- Step S 13 the class classifying portion 12 classifies the pixel of interest into the corresponding one of the classes in correspondence to the feature of the class tap supplied thereto from the class extracting portion 11 , thereby generating the class for the pixel of interest. Also, the class classifying portion 12 supplies the data on the class thus generated for the pixel of interest to the coefficient storing portion 13 .
- Step S 14 the noise amount tap extracting portion 14 extracts the noise amount tap corresponding to the pixel of interest from the input image, and supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operating portion 15 .
- Step S 15 the noise amount arithmetically operating portion 15 arithmetically operates the estimate noise amount ⁇ ′ corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amount tap extracting portion 14 in accordance with Expression (2), and supplies the data on the resulting estimate noise amount ⁇ ′ to the predictive coefficient generating portion 16 .
- Step S 16 the coefficient storing portion 13 reads out the processing coefficient corresponding to the class supplied thereto from the class classifying portion 12 , and supplies the processing coefficient thus read out to the predictive coefficient generating portion 16 .
- Step S 17 the predictive coefficient generating portion 16 carries out the arithmetic operation for the matrix of Expression (5) by using the processing coefficients supplied thereto from the coefficient storing portion 13 , and the estimate noise amounts ⁇ ′ supplied thereto from the noise amount arithmetically operating portion 15 , thereby generating the predictive coefficient W.
- the predictive coefficient generating portion 16 supplies the predictive coefficient W to the predicting portion 18 .
- Step S 18 the predictive tap extracting portion 17 extracts the predictive tap corresponding to the pixel of interest from the input image, and supplies the data on the predictive tap thus extracted to the predicting portion 18 .
- Step S 19 the predicting portion 18 carries out the predictive arithmetic operation for Expression (3) by using the predictive coefficient W supplied thereto from the predictive coefficient generating portion 16 , and the predictive tap supplied thereto from the predictive tap extracting portion 17 . As a result, the predicting portion 18 generates the predictive value for the pixel value of the pixel of interest as the pixel value of the pixel of interest composing the output image.
- Step S 20 the class tap extracting portion 11 , the noise amount tap extracting portion 14 , and the predictive tap extracting portion 17 determines whether or not all the pixels composing the output image are already determined as the pixels of interest, respectively.
- Step S 20 It is determined in Step S 20 that all the pixels composing the output image are not yet determined as the pixels of interest, respectively, the operation returns back to the processing in Step S 11 , and the processing described above is repeatedly executed.
- Step S 20 it is determined in Step S 20 that all the pixels composing the output image are already determined as the pixels of interest, respectively, in Step S 21 , the predicting portion 18 outputs the output image composed of the pixel values generated based on the predictive arithmetic operation, thereby completing the operation.
- the image processing apparatus 10 generates the predictive coefficient W by using the processing coefficients for the respective classes obtained based on the correlation of the class tap obtained from the learning which will be described later, and the estimate noise amounts ⁇ ′. Therefore, the optimal predictive coefficient W corresponding to the correlation of the class tap, and the estimate noise amounts ⁇ ′ can be generated without explosively increasing the number of classes. As a result, the image processing apparatus 10 can execute the highly precise noise removing processing by using the predictive coefficient W, thereby generating the output image, having the high image quality, from which the luminance dependency noise is removed.
- an S/N ratio of the output image in the image processing apparatus 10 becomes larger than that of the output image obtained from the class classifying adaptation processing in the related art by using the predictive coefficients obtained from the learning made for the classes obtained through the classification by using the correlation of the class tap, and the estimate noise amounts ⁇ ′.
- the former is 36.5 and the latter is 35.6.
- the learning for the processing coefficients is carried out by utilizing a least-square method.
- a true value of a pixel value of a pixel of an output image in a k-th sample is represented by y k
- the least-square method for example, is adopted as the norm representing that the predictive coefficient W is the optimal one.
- a minimization function Q is expressed by Expression (6):
- N is the number of samples used in set learning of the pixel values y k of the output image, and an n-dimensional vector of the input image composing the predictive tap about the pixel values y k (the number of samples for the learning).
- a minimum value (local minimum value) of the minimization function Q expressed by Expression (6) is given by a processing coefficient in which a value obtained by partially differentiating the minimization function Q by all variable numbers is made zero. Therefore, a linear simultaneous equation of an addition matrix shown in FIG. 7 which is structured so that the minimization function Q is partially differentiated by all the variable numbers, and the resulting value becomes zero is solved, thereby obtaining the optimal processing coefficient.
- the addition matrix shown in FIG. 7 is divided into small blocks (small blocks each composed of a matrix of n ⁇ n in the case of a matrix of a left-hand side of a left member, and small blocks each composed of a matrix of n ⁇ 1 in the case of each of a matrix of a right-hand side of a left member, and a matrix of the right member).
- Each of the small blocks of the matrix of the left-hand side of the left member, and the matrix of the right member is structured by multiplying elements of the addition matrix corresponding to the linear predictive arithmetic operation in the class classifying adaptation processing in the related art by the values of the estimate noise amounts ⁇ ′ corresponding to the positions of the small blocks.
- suffixes i and j (0 ⁇ i, j ⁇ n) represent the positions of the pixels corresponding to the pixel values within the predictive tap.
- the addition matrix shown in FIG. 7 is generated every class and thus the optimal processing coefficient is obtained every class.
- the matrix arithmetic operation for Expression (5) is carried out by using the processing coefficients for the respective classes obtained in the manner described above, thereby generating the predictive coefficient W. Also, the predictive arithmetic operation for Expression (3) is carried out by using the resulting predictive coefficient W, thereby converting the input image into the output image.
- FIG. 8 is a block diagram showing an example of a configuration of a learning apparatus 30 for learning the processing coefficients used by the image processing apparatus 10 shown in FIG. 1 .
- the learning apparatus 30 shown in FIG. 8 is composed of a learning pair storing portion 31 , a class tap extracting portion 32 , a class classifying portion 33 , a noise amount tap extracting portion 34 , a noise amount arithmetically operating portion 35 , a predictive tap extracting portion 36 , a normal equation generating portion 37 , a coefficient generating portion 38 , and a coefficient storing portion 39 .
- the learning pair storing portion 31 stores therein a set of an image containing therein a luminance dependency noise and corresponding to an input image in the image processing apparatus 10 as a student in learning for the processing coefficients (hereinafter referred to as “a student image”), and an image corresponding to an ideal output image into which the input image is converted as a teacher (hereinafter referred to as “a teacher image”) in the form of a learning pair.
- the learning pair storing portion 31 outputs the data on the student image of the learning pair to each of the class tap extracting portion 32 , the noise amount tap extracting portion 34 , and the predictive tap extracting portion 36 . Also, the learning pair storing portion 31 outputs the data on the teacher image of the learning pair to the normal equation generating portion 37 .
- the class tap extracting portion 32 successively determines the pixels composing the teacher image as the pixels of interest similarly to the case of the class tap extracting portion 11 shown in FIG. 1 . Also, the class tap extracting portion 32 extracts the pixel values of a plurality of pixels, within the student image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest in the form of the class tap from the student image. In this case, the pixel values of a plurality of pixels within the student image are used to classify the pixels of interest into the classes. The class tap extracting portion 32 supplies the data on the resulting class tap to the class classifying portion 33 .
- the class classifying portion 33 classifies the pixels of interest into the classes in correspondence to the feature of the class tap supplied thereto from the class tap extracting portion 32 similarly to the case of the class classifying portion 12 shown in FIG. 1 , thereby generating the classes corresponding to the pixels of interest.
- the class classifying portion 33 supplies the data on the classes to the normal equation generating portion 37 .
- the noise amount tap extracting portion 34 successively determines the pixels composing the teacher image as the pixels of interest, respectively, the case of the class tap extracting portion 32 . Also, the noise amount tap extracting portion 34 extracts the pixel values of a plurality of pixels, within the student image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest in the form of the noise amount tap from the student image similarly to the case of the noise amount tap extracting portion 14 shown in FIG. 1 . In this case, the pixel values of a plurality of pixels within the student image are used to arithmetically operate the estimate noise amount ⁇ ′ corresponding to the pixel of interest. The noise amount tap extracting portion 34 supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operating portion 35 .
- the noise amount arithmetically operating portion 35 arithmetically operates the estimate noise amount ⁇ ′ corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amount tap extracting portion 34 in accordance with Expression (2) similarly to the case of the noise amount arithmetically operating portion 15 shown in FIG. 1 . Also, the noise amount arithmetically operating portion 35 supplies the data on the estimate noise amount ⁇ ′ corresponding to the pixel of interest to the normal equation generating portion 37 .
- the predictive tap extracting portion 36 successively determines the pixels composing the teacher image as the pixels of interest, respectively, similarly to the case of each of the class tap extracting portion 32 , and the noise amount tap extracting portion 34 . Also, the predictive tap extracting portion 36 extracts the pixel values of a plurality of pixels within the student image corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the predictive tap from the student image similarly to the case of the predictive tap extracting portion 17 shown in FIG. 1 . In this case, the pixel values of a plurality of pixels within the student image are used to predict the pixel of interest. Also, the predictive tap extracting portion 36 supplies the data on the predictive tap thus extracted to the normal equation generating portion 37 .
- the normal equation generating portion 37 generates the addition matrix (refer to FIG. 7 ) in the form of the normal equation.
- the addition matrix is obtained by performing the addition by using the estimate noise amount ⁇ ′ supplied from the noise amount arithmetically operating portion 35 , the predictive tap supplied from the predictive tap extracting portion 36 , and the teacher image inputted from the learning pair storing portion 31 every class supplied from the class classifying portion 33 .
- the normal equation generating portion 37 substitutes the pixel values of the pixel of interest of the teacher image in the k-th sample in the form of y k into the addition matrix shown in FIG. 7 .
- the normal equation generating portion 37 substitutes the estimate noise amounts ⁇ ′ corresponding to the pixels of interest, respectively, into the addition matrix shown in FIG. 7 .
- the normal equation generating portion 37 generates the normal equation.
- the normal equation generated by the normal equation generating portion 37 is an equation for solving an expression corresponding to Expression (3) for generation of the teacher image based on the multiplication of the predictive coefficient W obtained from both the estimate noise amounts ⁇ ′ and the processing coefficients, and the predictive tap of the student image, thereby obtaining the processing coefficients.
- the normal equation generated from the normal equation generating portion 37 is supplied to the coefficient generating portion 38 .
- the coefficient generating portion 38 solves the normal equation supplied thereto from the normal equation generating portion 37 , thereby generating the processing coefficients. Also, the coefficient generating portion 38 instructs the coefficient storing portion 39 to store therein the processing coefficients thus generated.
- the processing coefficients which are learned in the manner as described above, and are then stored in the coefficient storing portion 39 are stored in the coefficient storing portion 13 shown in FIG. 1 , and are used in the image processing apparatus 10 .
- the learning pair storing portion 31 outputs the learning pair which is not yet outputted of the learning pairs stored therein. Specifically, the learning pair storing portion 31 inputs the data on the student image of the learning pair to each of the class tap extracting portion 32 , the noise amount tap extracting portion 34 , and the predictive tap extracting portion 36 . On the other hand, the learning pair storing portion 31 inputs the data on the teacher image of the learning pair to the normal equation generating portion 37 .
- Step S 31 the class tap extracting portion 32 , the noise amount tap extracting portion 34 , and the predictive tap extracting portion 36 determine one of the pixels, of a plurality of pixels composing the teacher image, which are not yet determined as the pixels of interest as the pixel of interest similarly to the case of the class tap extracting portion 11 shown in FIG. 1 .
- Step S 32 the class tap extracting portion 32 extracts the class tap corresponding to the pixel of interest from the student image, and supplies the data on the class tap thus extracted to the class classifying portion 33 .
- Step S 33 the class classifying portion 33 classifies the pixel of interest into the corresponding one of the classes in correspondence to the feature of the class tap supplied thereto from the class tap extracting portion 32 similarly to the case of the class classifying portion 12 shown in FIG. 1 , thereby generating the class for the pixel of interest.
- the class classifying portion 33 supplies the data on the class thus generated to the normal equation generating portion 37 .
- Step S 34 the noise amount tap extracting portion 34 extracts the noise amount tap corresponding to the pixel of interest from the student image, and supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operating portion 35 similarly to the case of the noise amount tap extracting portion 14 shown in FIG. 1 .
- Step 35 the noise amount arithmetically operating portion 35 arithmetically operates the estimate noise amount ⁇ ′ corresponding to the noticed image from the noise amount tap supplied thereto from the noise amount tap extracting portion 34 in accordance with Expression (2) similarly to the case of the noise amount arithmetically operating portion 15 shown in FIG. 1 . Also, the noise amount arithmetically operating portion 35 supplies the data on the estimate noise amount ⁇ ′ to the normal equation generating portion 37 .
- Step S 36 the predictive tap extracting portion extracts the predictive tap corresponding to the pixel of interest from the student image similarly to the case of the predictive tap extracting portion 17 shown in FIG. 1 , and supplies the data on the predictive tap thus extracted to the normal equation generating portion 37 .
- Step S 37 the normal equation generating portion carries out the addition for the addition matrix every class supplied thereto from the class classifying portion (refer to FIG. 7 ) by using the estimate noise amounts ⁇ ′ supplied thereto from the noise amount arithmetically operating portion 35 , the predictive tap supplied thereto from the predictive tap extracting portion 36 , and the teacher image inputted thereto from the learning pair storing portion 31 .
- Step S 38 the class tap extracting portion 32 , the noise amount tap extracting portion 34 , and the predictive tap extracting portion 36 determine whether or not all the pixels composing the teacher image are already determined as the pixels of interest.
- the operation returns back to the processing in Step S 31 , and the processing described above is repeatedly executed.
- Step S 39 the learning pair storing portion 31 determines whether or not the processing from Step S 30 to Step S 38 is already executed for all the learning pairs, that is, whether or not all the learning pairs stored therein are already outputted.
- Step S 39 the processing from Step S 30 to Step S 38 is not yet executed for all the learning pairs, the operation returns back to the processing in Step S 30 , and the processing described above is repeatedly executed.
- Step S 39 when it is determined in Step S 39 that all the learning pairs stored therein are already outputted, the normal equation generating portion 37 supplies the data on the normal equation generated by carrying out the addition in Step S 37 to the coefficient generating portion 38 .
- Step S 40 the coefficient generating portion 38 generates the processing coefficients by solving the normal equation supplied thereto from the normal equation generating portion 37 , and instructs the coefficient storing portion 39 to store therein the processing coefficients.
- the learning apparatus 30 learns the optimal processing coefficient every class obtained by the classification based on the correlation of the class tap by using the student image containing therein the luminance dependency noise, and the teacher image, as the ideal image, from which the luminance dependency noise is removed. Therefore, the image processing apparatus 10 generates the predictive coefficient W by using the optimal processing coefficients and the estimate noise amounts ⁇ ′, thereby making it possible to generate the optimal predictive coefficient W, for removal of the luminance dependency noise, corresponding to the correlation of the class tap, and the estimate noise amounts ⁇ ′ without explosively increasing the number of classes. As a result, the image processing apparatus 10 can generate the output image, having the high image quality, from which the luminance dependency noise is removed.
- the noticed images are classified into the classes in correspondence to the correlation of the class tap, the noticed images may not be classified.
- a series of processing described above can be executed by either the hardware or the software.
- a program composing the software is installed from a program recording medium either in a computer incorporated in a dedicated hardware, or, for example, in a general purpose personal computer or the like which can execute various functions by installing therein various programs.
- FIG. 10 shows an example of a configuration of the hardware in a computer 300 for executing the series of processing described above in accordance with a program.
- a Central Processing Unit (CPU) 301 a Central Processing Unit (CPU) 301 , a Read Only Memory (ROM) 302 , and a Random Access Memory (RAM) 303 are connected to one another through a bus 304 .
- CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- an I/O interface 305 is connected to the bus 304 .
- An input portion 306 , an output portion 307 , a storing portion 308 , a communication portion 309 , and a drive 310 are also connected to the I/O interface 305 .
- the input portion 306 is composed of a receiving portion or the like for receiving an instruction transmitted thereto from a keyboard, a mouse, and a microphone, a remote controller, or the like.
- the output portion 307 is composed of a display device, a speaker or the like.
- the storing portion 308 is composed of a hard disc, a nonvolatile memory or the like.
- the drive 310 drives a removable media 311 as a package media.
- the removable media 311 is composed of a magnetic disc (including a flexible disc), an optical disc (such as a Compact Disc-Read Only Memory (CD-ROM), or a Digital Versatile Disc (DVD)), a magneto optical disc, or a semiconductor memory.
- a magnetic disc including a flexible disc
- an optical disc such as a Compact Disc-Read Only Memory (CD-ROM), or a Digital Versatile Disc (DVD)
- CD-ROM Compact Disc-Read Only Memory
- DVD Digital Versatile Disc
- magneto optical disc or a semiconductor memory.
- the CPU 301 loads the program, for example, stored in the storing portion 308 into the RAM 303 through the I/O interface 305 and the bus 304 , thereby executing the series of processing described above.
- the program which the CPU 301 of the computer 300 executes for example, is recorded in the removable media 311 , or is provided through a wired or wireless transmission medium such as a Local Area Network (LAN), the Internet or a digital satellite-based broadcasting.
- LAN Local Area Network
- the Internet or a digital satellite-based broadcasting.
- the program can be installed in the storing portion 308 through the I/O interface 305 by equipping the drive 310 with the removable media 311 .
- the program can be received at the communication portion 309 through the wired or wireless transmission media to be installed in the storing portion 308 .
- the program can be previously installed either in the ROM 302 or in the storing portion 308 .
- the program which the CPU 301 of the computer 300 executes may be either a program in accordance with the processing is executed in a time series manner so as to follow the order explained in this specification, or a program in accordance with which the processing is executed at a necessary timing when a call is made.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Facsimile Image Signal Circuits (AREA)
Abstract
Description
- The present invention contains subject matter related to Japanese Patent Application JP 2007-330452 filed in the Japan Patent Office on Dec. 21, 2007, the entire contents of which being incorporated herein by reference.
- 1. Field of the Invention
- The present invention relates to an image processing apparatus, an image processing method, and a program for use therein, and a learning apparatus, and more particularly to an image processing apparatus and an image processing method each of which is capable of removing a noise, which is generally generated and which has luminance dependency, from a video signal corresponding to an image, and a program for use therein, and a learning apparatus.
- 2. Description of the Related Art
- In related art, in noise removing processing for removing a noise from a video signal corresponding to an image, the noise is treated as a white color, and pixel values of the peripheral pixels are added to one another, thereby removing the noise from a video signal corresponding to an image.
- On the other hand, in recent years, it has been devised in the noise removing processing that the noise is assumed to have a color because there is a deviation in color of the noise, and thus the noise is removed with a high degree of accuracy as compared with the case where the noise is treated as the white color.
- For example, it has been devised in the noise removing processing that black pixel signals generated in black pixels obtained by light-shielding an opening portion are subtracted from pixel signals outputted from the effective pixels, respectively, thereby removing a fixed pattern noise (FPN) with which the pixel signals are mixed due to a manufacturing error. This technique, for example, is descried in Japanese Patent Laid-Open No. 2007-116292. In that noise removing processing, in order to remove the noise due to the manufacturing error, the black pixel signals need to be detected every product.
- On the other hand, it has not been devised in the noise removing processing to remove a noise which, for example, does not depend on any of the manufacturing processes, but has luminance dependency which an image sensor itself generally has (hereinafter referred to as “a luminance dependency noise”).
- The present invention has been made in the light of such circumferences, and it is therefore desirable to provide an image processing apparatus and an image processing method each of which is capable of removing a noise, which is generally generated and which has luminance dependency, from a video signal corresponding to an image, and a program for use therein, and a learning apparatus.
- In order to attain the desire described above, according to an embodiment of the present invention, there is provided an image processing apparatus for converting a first image into a second image having higher image quality than that of the first image, including: a first pixel value extracting section for extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; an estimate noise amount arithmetically operating section for obtaining estimate noise amounts for the plurality of pixel values extracted by the first pixel value extracting section, respectively; a processing coefficient generating section for generating second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher image quality than that of a student image in accordance with an arithmetic operation for the second processing coefficients obtained in accordance with estimate noise amounts about the pixel values within the student image corresponding to the first image, and the first processing coefficients, and the student image; a second pixel value extracting section for extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; and a predicting section for generating a pixel value of the pixel of interest within the second image in accordance with an arithmetic operation for the plurality of pixel values extracted by the second pixel value extracting section, and the second processing coefficients.
- According to another embodiment of the present invention, there is provided an image processing method for use in an image processing apparatus for converting a first image into a second image having higher image quality than that of the first image, the image processing method including the steps of: extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; obtaining estimate noise amounts for the plurality of pixel values, respectively; obtaining second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher quality than that of a student image in accordance with an arithmetic operation for the second processing coefficients obtained in accordance with estimate noise amounts about the pixel values within the student image corresponding to the first image, and the first processing coefficients, and the student image; extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; and generating a pixel value of the pixel of interest within the second image in accordance with an arithmetic operation for the plurality of pixel values thus extracted, and the second processing coefficients.
- According to the embodiments of the present invention described above, the plurality of pixel values, within the first image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest within the second image having the higher image quality than that of the first image are extracted. The estimate noise amounts are obtained with respect to the plurality of pixel values thus extracted, respectively. The second processing coefficients are obtained in accordance with the arithmetic operation for the first processing coefficients previously learned from the normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image. In this case, the normal equation is obtained based on the relational expression for generating the teacher image corresponding to the second image having the higher image quality than that of the student image in accordance with the arithmetic operation for the second processing coefficients obtained in accordance with the estimate noise amounts about the pixel values, within the student image, corresponding to the first image and the first processing coefficients, and the student image. In addition, the plurality of pixel values, within the first image, corresponding to the position of the noticed image, and the peripheral positions of the position of the noticed image within the second image are extracted. Also, the pixel value of the pixel of interest within the second image is generated in accordance with the arithmetic operation for the plurality of pixel values thus extracted, and the second processing.
- According to still another embodiment of the present invention, there is provided a program in accordance with which a computer executes image processing for converting a first image into a second image having higher image quality than that of the first image, the program including the steps of: extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; obtaining estimate noise amounts for the plurality of pixel values, respectively; obtaining second processing coefficients in accordance with an arithmetic operation for first processing coefficients previously learned from a normal equation, and the estimate noise amounts obtained for the plurality of pixel values, respectively, within the first image, the normal equation being obtained based on a relational expression for generating a teacher image corresponding to the second image having higher quality than that of a student image in accordance with an arithmetic operation for the second processing coefficients obtained in accordance with estimate noise amounts about the pixel values within the student image corresponding to the first image, and the first processing coefficients, and the student image; extracting a plurality of pixel values, within the first image, corresponding to a position of a pixel of interest, and peripheral positions of the position of the pixel of interest within the second image; and generating a pixel value of the pixel of interest within the second image in accordance with an arithmetic operation for the plurality of pixel values thus extracted, and the second processing coefficients.
- According to yet another embodiment of the present invention, there is provided a learning apparatus, including: a normal equation generating section for obtaining estimate noise amounts from pixel values within a student image containing therein a noise having noise dependency, obtaining second processing coefficients in accordance with an arithmetic operation for the estimate noise amounts, and first processing coefficients, and generating a normal equation by using the pixel values of the student image, and pixel values of the teacher image, in order to solve a relational expression for generating the teacher image having higher image quality than that of the student image, in accordance with an arithmetic operation for the second processing coefficients and the student image; and a coefficient generating section for generating the first processing coefficients by solving the normal equation.
- According to the embodiments of the present invention described above, the estimate noise amounts are obtained from the pixel values within the student image containing therein the noise having the luminance dependency. The second processing coefficients are obtained in accordance with the arithmetic operation for the estimate noise amounts, and the first processing coefficients. In addition, the normal equation is generated by using the pixel values of the student image, and the pixel values of the student image, in order to solve the relational expression for generating the teacher image having the higher image quality than that of the student image, in accordance with the arithmetic operation for the second processing coefficients and the student image. Also, the first processing coefficients are generated by solving the normal equation.
- As set forth hereinabove, according to the present invention, the noise which is generally generated and which has the luminance dependency can be removed from the video signal corresponding to the image.
- In addition, it is possible to generate the processing coefficients for removing the noise which is generally generated and which has the luminance dependency from the video signal corresponding to the image.
-
FIG. 1 is a block diagram showing a configuration of an image processing apparatus according to an embodiment of the present invention; -
FIG. 2 is a graph showing light quantity dependency of amounts of noises which two kinds of image sensors have, respectively; -
FIG. 3 is a diagram explaining a norm for learning for predictive coefficients; -
FIG. 4 is a view showing a class tap composed of pixel values of pixels of 3×3; -
FIG. 5 is a graph showing predictive coefficients which are multiplied by pixel values composing a predictive tap; -
FIG. 6 is a flow chart explaining noise removing processing executed by the image processing apparatus shown inFIG. 1 ; -
FIG. 7 is a diagram showing an addition matrix; -
FIG. 8 is a block diagram showing a configuration of a learning apparatus for learning processing coefficients according to an embodiment of the present invention; -
FIG. 9 is a flow chart explaining the learning processing executed by the learning apparatus shown inFIG. 8 ; and -
FIG. 10 is a block diagram showing a configuration of hardware of a computer. -
FIG. 1 is a block diagram showing a configuration of animage processing apparatus 10 according to an embodiment of the present invention. - The
image processing apparatus 10 shown inFIG. 1 is composed of a classtap extracting portion 11, aclass classifying portion 12, acoefficient storing portion 13, a noise amount tap extractingportion 14, a noise amount arithmetically operatingportion 15, a predictivecoefficient generating portion 16, a predictivetap extracting portion 17, and a predictingportion 18. - The
image processing apparatus 10 executes class classifying adaptation processing. In the class classifying adaptation processing, a class for a pixel of interest in an output image which will be generated after this is generated, and predictive coefficients of the output image having higher image quality than that of an input image are generated from the input image by using predictive coefficients obtained based on a processing coefficient of the class, and an estimate value for an amount of noise as a variance value of a luminance dependency noise of the input image (hereinafter referred to as “an estimate noise amount”), and the input image. - It is noted that the processing coefficients used in the class classifying adaptation processing executed by the
image processing apparatus 10, for example, is obtained from learning (its details will be described later) using an image containing therein the luminance dependency noise, and an image from which the luminance dependency noise is removed. Therefore, theimage processing apparatus 10 can generate an image, having high image quality, from which the luminance dependency noise is removed by executing the class classifying adaptation processing. From this, the class classifying adaptation processing executed by theimage processing apparatus 10 can be the as noise removing processing for removing the luminance dependency noise from the input image as it were. - In the
image processing apparatus 10 shown inFIG. 1 , the classtap extracting portion 11 successively determines pixels composing an output image (this output image is one which will be generated after this and does not exist at the current stage, and thus is virtually supposed) as pixels of interest. The class tap extractingportion 11 extracts pixel values of a plurality of pixels, within the input image, corresponding to a position of the pixel of interest, and peripheral positions of the position of the pixel of interest as a class tap from the input image. In this case, the pixel values of the plurality of pixels within the input image are used to classify the pixels of interest into classes. The classtap extracting portion 11 supplies the class tap to theclass classifying portion 12. - The
class classifying portion 12 classifies the pixels of interest into classes in correspondence to the feature of the class tap supplied thereto from the class tap extractingportion 11, thereby generating the classes for the pixels of interest. Adaptive dynamic range coding (ADRC) or the like, for example, can be adopted as a method of classifying the pixels of interest into the classes. With a method using the ADRC, the pixel values composing the class tap are ADRC-processed, and the classes of the pixels of interest are determined in accordance with ADRC codes obtained as the result of executing the ADRC processing. As a result, the pixels of interest are classified into the classes in accordance with correlation (waveform) of the class tap. - Note that, in K-bit ADRC, for example, a maximum value MAX and a minimum value MIN of the pixel values composing the class tap are detected. DR=MAX−MIN is set as a local dynamic range of a set of a plurality of pixel values composing the class tap. Also, each of the plurality of pixel values as the class tap is re-quantized into K-bits based on the dynamic range DR. That is to say, the minimum value MIN is subtracted from each of the pixel values as the class tap, and each of the resulting values obtained from the subtraction is divided by DR/2K (quantized). Also, a bit string which is obtained by arranging the K-bit data, as the class tap, obtained in the manner described above in the predetermined order is set as the ADRC code.
- Therefore, when the class tap, for example, is subjected to 1-bit ADRC processing, after the minimum value MIN is subtracted from each of the pixel values as the class tap is divided by ½ of a difference between the maximum value MAX and the minimum value MIN (rounding-down of a fractional part). As a result, each data is set as 1 bit (binarized). Also, a bit string obtained by arranging the data of 1 bit in the predetermined order is set as the ADRC code. The data on the classes for the pixels of interest generated by the class classifying portion are supplied to the
coefficient storing portion 13. - The
coefficient storing portion 13 stores therein the optimal processing coefficients for the classes obtained from the learning which will be described later. Thecoefficient storing portion 13 reads out the processing coefficients corresponding to the classes the data on which is supplied from theclass classifying portion 12, and supplies the processing coefficients thus read out to the predictivecoefficient generating portion 16. - The noise amount
tap extracting portion 14 successively determines the pixels composing the output image as the pixels of interest similarly to the case of the classtap extracting portion 11. The noise amounttap extracting portion 14 extracts the pixel values of the plurality of pixels, within the input image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the noise amount tap from the input image. In this case, the pixel values of the plurality of pixels within the input image are used to arithmetically operate the estimate noise amount corresponding to the pixel of interest. The noise amounttap extracting portion 14 supplies the noise amount tap to the noise amount arithmetically operatingportion 15. - The noise amount arithmetically operating
portion 15 arithmetically operates the estimate noise amount corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amounttap extracting portion 14, and supplies the resulting estimate noise amount to the predictivecoefficient generating portion 16. - The predictive
coefficient generating portion 16 carries out a predetermined matrix arithmetic operation by using the processing coefficients supplied thereto from thecoefficient storing portion 13, and the estimate noise amount supplied thereto from the noise amount arithmetically operatingportion 15, thereby generating the predictive coefficients. The predictivecoefficient generating portion 16 supplies the predictive coefficients thus generated to the predictingportion 18. - The predictive
tap extracting portion 17 successively determines the pixels composing the output image as the pixels of interest similarly to the case of each of the classtap extracting portion 11 and the noise amounttap extracting portion 14. The predictivetap extracting portion 17 extracts the pixel values of the plurality of pixels, within the input image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the predictive tap from the input image. In this case, the pixel values of the plurality of pixels within the input image are used to predict the pixels of interest. Also, the predictivetap extracting portion 17 supplies the predictive tap thus extracted to the predictingportion 18. - The predicting
portion 18 carries out the predictive arithmetic operation for predicting the pixel of interest by using the predictive coefficients supplied thereto from the predictivecoefficient generating portion 16, and the predictive tap supplied thereto from the predictivetap extracting portion 17. As a result, the predictingportion 18 generates the predictive value of the pixel value of the pixel of interest as the pixel value of the pixel of interest composing the output image. Also, the predictingportion 18 outputs the output image composed of the pixel values obtained from the predictive arithmetic operation. - Next, an arithmetic operation for the estimate noise amount by the noise amount arithmetically operating
portion 15 shown inFIG. 1 will be described with reference toFIG. 2 . - A graph of
FIG. 2 shows light quantity (luminance) dependency of amounts of noises which two kinds of image sensors have, respectively. It is noted that inFIG. 2 , an axis of abscissa represents a total light quantity, that is, a pixel value, and an axis of ordinate represents an amount of noise. - As shown in
FIG. 2 , the amounts of noises which the two kinds of image sensors have, respectively, have the luminance dependency. Thus, the amount of noise a can be expressed in the form of a quadratic expression of a true value L of the pixel value given by Expression (1): -
σ2 =aL 2 +bL+c (1) - where a, b and c are parameters inherent in the image sensor, respectively.
- According to Expression (1), the amount of noise of the pixel corresponding to the true value L of the pixel value can be arithmetically operated from the true value L of the pixel value. Therefore, the noise amount arithmetically operating
portion 15 may obtain the amount of noise by carrying out the arithmetic operation for Expression (1). Actually, however, it is impossible to acquire the pixel value in which the true value of the pixel value, that is, the luminance dependency noise is not contained. For this reason, the estimate noise amount σ′ is arithmetically operated by carrying out the arithmetic operation for Expression (2) by using a pixel value L′ containing therein the noise: -
σ′2 =aL′ 2 +bL′+c (2) - Next, the class classifying adaptation processing executed in the
image processing apparatus 10 will be desired with reference toFIGS. 3 to 5 . - When it is assumed that in the
image processing apparatus 10 shown inFIG. 1 , the predictingportion 18, for example, carries out a linear predictive arithmetic operation as a predetermined predictive arithmetic operation, the pixel value y composing the output image is obtained in accordance with Expression (3) as a linear expression: -
y=WX (3) - where X represents an n-dimensional vector (X=(x0, x1, . . . , xn)) consisting of the n pixel values, of the n pixels of the input image, composing the predictive tap about the pixel value y of the output image, and W represents an n-dimensional vector (W=(w0, w1, . . . , wn)) consisting of the predictive coefficients which are multiplied by the n pixel values of the n pixels, respectively. From the above, the number of pixels composing the predictive tap is equal to the number of predictive coefficients composing the predictive coefficient W.
- However, the optical predictive coefficient W can be obtained from the learning which will be described later. However, in order to execute highly precise noise removing processing, the following two items are required as the norm for the learning thereof.
- A first item is such that the predictive coefficient W is changed based on the correlation (waveform) of the predictive tap. That is to say, the first item is such that the predictive coefficient W which is multiplied by the pixel value is made large in that pixel value, nearer the pixel value of the pixel of interest, of the pixel values composing the predictive tap.
- For example, as shown on a left-hand side of
FIG. 3 , it is assumed that the predictive tap is composed of a pixel value x1 of a pixel of an input image corresponding to a position of a pixel of interest, a pixel value x0 near the pixel value x1, and a pixel value x2 far from the pixel value x1. In this case, the predictive coefficient W needs to be learned so that the predictive coefficient w0 which is multiplied by the pixel value x0 becomes larger than the predictive coefficient w2 which is multiplied by the pixel value x2 which is located farther from the pixel value x1 than the pixel value x0 is located. - In addition, a second item is such that the predictive coefficient W is changed based on the estimate noise amount σ′ of the input image. That is to say, the predictive coefficient W which is multiplied by the pixel value is made large in that pixel value having the less estimate noise amount σ′.
- For example, as shown on a right-hand side of
FIG. 3 , it is assumed that a predictive tap is composed of a pixel value x1 of the pixel of the input image corresponding to the position of a pixel of interest, a pixel value x0 having an estimate noise amount σ0, and a pixel value x2 having an estimate noise amount σ2 larger than the estimate noise amount σ0. In this case, the predictive coefficient W needs to be learned so that the predictive coefficient w0 which is multiplied by the pixel value x0 becomes larger than the predictive coefficient w2 which is multiplied by the pixel value x2 having the larger estimate noise amount x2 than the estimate noise σ0 of the pixel value x0. - Cluster classification learning is known as a method of learning the optimal predictive coefficient in accordance with the norm as described above. In this case, with the cluster classification learning, the pixels of interest are classified into classes based on the correlation of the cluster tap, and the estimate noise amount σ′, and the predictive coefficient is obtained every class. However, in this case, the number of classes explosively increases, so that it becomes difficult to learn the predictive coefficients. For example, when the class tap, as shown in
FIG. 4 , is composed of the pixel values of the pixels of 3×3, if the correlation of the class tap is represented by subjecting each of the pixel values of the pixels to the 1-bit ADRC processing, and the estimate noise amount σ′ of each of the pixels is represented by 3 bits, the total number of classes becomes about 6.9×1010 (=512×89). - Thus, the optimal processing coefficient for each of the classes obtained by the classification based on the correlation of the class tap is obtained in the learning which will be described later. Also, the
image processing apparatus 10 optimizes that processing coefficient in correspondence to the estimate noise amount σ′ in the class classification adaptation processing, and carries out the predictive arithmetic operation with the optimized processing coefficient as the predictive coefficient W. - That is to say, the maximum signification of the class classification learning is to change the value of the predictive coefficient every class. For example, as shown in
FIG. 5 , the predictive coefficients which are multiplied by the pixel values, respectively, within the predictive tap are different from each other in two different classes. It is noted that inFIG. 5 , an axis of abscissa represents the positions of the pixels corresponding to the pixel values composing the predictive tap, and an axis of ordinate represents the predictive coefficients which are multiplied by the pixel values of the pixels, respectively. - Thus, the
image processing apparatus 10 optimizes the predictive coefficients in accordance with Expression (4) similarly to the case where the predictive coefficient W is changed in correspondence to the estimate noise amount σ′ based on the optimal processing coefficient for each of the classes obtained by the classification based on the correlation of the class tap obtained from the learning, thereby classifying and learning the pixels of interest based on the correlation of the class tap, and the estimate noise amounts. -
- where W0 and Wσ′i are processing coefficients of an n-dimensional vector corresponding to the predictive coefficient W of the n-dimensional vector, m is the number of pixel values composing the noise amount tap, and σ′i is an estimate noise amount of an i-th pixel value within the noise amount tap. It is noted that the number, m, of pixel values composing the noise amount tap may be either identical to or different from the number, n, of pixel values composing the predictive tap.
- In addition, Expression (4) is expressed in the form of a determinant of matrix as follows:
-
- where w0,0, w0,1, . . . , w0,n represent n elements of the processing coefficients W0 of the n-dimensional vector, respectively, and wσi′,0, wσi′,1, . . . , wσi′,n represent n elements of the processing coefficients Wσi′, of the n-dimensional vector, respectively.
- Next, the noise removing processing executed by the
image processing apparatus 10 shown inFIG. 1 will be described with reference to a flow chart ofFIG. 6 . - Firstly, in Step S11, the class
tap extracting portion 11, the noise amounttap extracting portion 14, and the predictivetap extracting portion 17 determine one, of a plurality of pixels composing the output image, which is not yet set as a pixel of interest as the pixel of interest. Next, in Step S12, the classtap extracting portion 11 extracts the class tap corresponding to the pixel of interest from the input image, and supplies the class tap thus extracted to theclass classifying portion 12. - In Step S13, the
class classifying portion 12 classifies the pixel of interest into the corresponding one of the classes in correspondence to the feature of the class tap supplied thereto from theclass extracting portion 11, thereby generating the class for the pixel of interest. Also, theclass classifying portion 12 supplies the data on the class thus generated for the pixel of interest to thecoefficient storing portion 13. - In Step S14, the noise amount
tap extracting portion 14 extracts the noise amount tap corresponding to the pixel of interest from the input image, and supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operatingportion 15. In Step S15, the noise amount arithmetically operatingportion 15 arithmetically operates the estimate noise amount σ′ corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amounttap extracting portion 14 in accordance with Expression (2), and supplies the data on the resulting estimate noise amount σ′ to the predictivecoefficient generating portion 16. - In Step S16, the
coefficient storing portion 13 reads out the processing coefficient corresponding to the class supplied thereto from theclass classifying portion 12, and supplies the processing coefficient thus read out to the predictivecoefficient generating portion 16. In Step S17, the predictivecoefficient generating portion 16 carries out the arithmetic operation for the matrix of Expression (5) by using the processing coefficients supplied thereto from thecoefficient storing portion 13, and the estimate noise amounts σ′ supplied thereto from the noise amount arithmetically operatingportion 15, thereby generating the predictive coefficient W. The predictivecoefficient generating portion 16 supplies the predictive coefficient W to the predictingportion 18. - In Step S18, the predictive
tap extracting portion 17 extracts the predictive tap corresponding to the pixel of interest from the input image, and supplies the data on the predictive tap thus extracted to the predictingportion 18. In Step S19, the predictingportion 18 carries out the predictive arithmetic operation for Expression (3) by using the predictive coefficient W supplied thereto from the predictivecoefficient generating portion 16, and the predictive tap supplied thereto from the predictivetap extracting portion 17. As a result, the predictingportion 18 generates the predictive value for the pixel value of the pixel of interest as the pixel value of the pixel of interest composing the output image. In Step S20, the classtap extracting portion 11, the noise amounttap extracting portion 14, and the predictivetap extracting portion 17 determines whether or not all the pixels composing the output image are already determined as the pixels of interest, respectively. - It is determined in Step S20 that all the pixels composing the output image are not yet determined as the pixels of interest, respectively, the operation returns back to the processing in Step S11, and the processing described above is repeatedly executed.
- On the other hand, it is determined in Step S20 that all the pixels composing the output image are already determined as the pixels of interest, respectively, in Step S21, the predicting
portion 18 outputs the output image composed of the pixel values generated based on the predictive arithmetic operation, thereby completing the operation. - As described above, the
image processing apparatus 10 generates the predictive coefficient W by using the processing coefficients for the respective classes obtained based on the correlation of the class tap obtained from the learning which will be described later, and the estimate noise amounts σ′. Therefore, the optimal predictive coefficient W corresponding to the correlation of the class tap, and the estimate noise amounts σ′ can be generated without explosively increasing the number of classes. As a result, theimage processing apparatus 10 can execute the highly precise noise removing processing by using the predictive coefficient W, thereby generating the output image, having the high image quality, from which the luminance dependency noise is removed. - In addition, according to the experiments, an S/N ratio of the output image in the
image processing apparatus 10 becomes larger than that of the output image obtained from the class classifying adaptation processing in the related art by using the predictive coefficients obtained from the learning made for the classes obtained through the classification by using the correlation of the class tap, and the estimate noise amounts σ′. For example, when the number of pixel values composing the predictive coefficient and the noise amount tap is 9, according to the experiments, the former is 36.5 and the latter is 35.6. As a result, it is understood that theimage processing apparatus 10 can generate the output image having the higher image quality. - Next, a description will be given with respect to the learning for the processing coefficients used to generate the predictive coefficient W in the
image processing apparatus 10. The learning for the processing coefficients, for example, is carried out by utilizing a least-square method. - Specifically, a true value of a pixel value of a pixel of an output image in a k-th sample is represented by yk, an n-dimensional vector X of an input pixel composing a predictive tap about the pixels of the output image in the k-th sample is represented by Xk (Xk=(xk0, xk1, . . . , xkn)), and the least-square method, for example, is adopted as the norm representing that the predictive coefficient W is the optimal one. In this case, a minimization function Q is expressed by Expression (6):
-
- where N is the number of samples used in set learning of the pixel values yk of the output image, and an n-dimensional vector of the input image composing the predictive tap about the pixel values yk (the number of samples for the learning).
- A minimum value (local minimum value) of the minimization function Q expressed by Expression (6) is given by a processing coefficient in which a value obtained by partially differentiating the minimization function Q by all variable numbers is made zero. Therefore, a linear simultaneous equation of an addition matrix shown in
FIG. 7 which is structured so that the minimization function Q is partially differentiated by all the variable numbers, and the resulting value becomes zero is solved, thereby obtaining the optimal processing coefficient. - The addition matrix shown in
FIG. 7 is divided into small blocks (small blocks each composed of a matrix of n×n in the case of a matrix of a left-hand side of a left member, and small blocks each composed of a matrix of n×1 in the case of each of a matrix of a right-hand side of a left member, and a matrix of the right member). Each of the small blocks of the matrix of the left-hand side of the left member, and the matrix of the right member is structured by multiplying elements of the addition matrix corresponding to the linear predictive arithmetic operation in the class classifying adaptation processing in the related art by the values of the estimate noise amounts σ′ corresponding to the positions of the small blocks. It is noted that suffixes i and j (0≦i, j≦n) represent the positions of the pixels corresponding to the pixel values within the predictive tap. - The addition matrix shown in
FIG. 7 is generated every class and thus the optimal processing coefficient is obtained every class. - According to the class classifying adaptation processing executed by the
image processing apparatus 10, the matrix arithmetic operation for Expression (5) is carried out by using the processing coefficients for the respective classes obtained in the manner described above, thereby generating the predictive coefficient W. Also, the predictive arithmetic operation for Expression (3) is carried out by using the resulting predictive coefficient W, thereby converting the input image into the output image. -
FIG. 8 is a block diagram showing an example of a configuration of alearning apparatus 30 for learning the processing coefficients used by theimage processing apparatus 10 shown inFIG. 1 . - The
learning apparatus 30 shown inFIG. 8 is composed of a learningpair storing portion 31, a classtap extracting portion 32, aclass classifying portion 33, a noise amounttap extracting portion 34, a noise amount arithmetically operatingportion 35, a predictivetap extracting portion 36, a normalequation generating portion 37, acoefficient generating portion 38, and acoefficient storing portion 39. - In the
learning apparatus 30, the learningpair storing portion 31 stores therein a set of an image containing therein a luminance dependency noise and corresponding to an input image in theimage processing apparatus 10 as a student in learning for the processing coefficients (hereinafter referred to as “a student image”), and an image corresponding to an ideal output image into which the input image is converted as a teacher (hereinafter referred to as “a teacher image”) in the form of a learning pair. - In addition, the learning
pair storing portion 31 outputs the data on the student image of the learning pair to each of the classtap extracting portion 32, the noise amounttap extracting portion 34, and the predictivetap extracting portion 36. Also, the learningpair storing portion 31 outputs the data on the teacher image of the learning pair to the normalequation generating portion 37. - The class
tap extracting portion 32 successively determines the pixels composing the teacher image as the pixels of interest similarly to the case of the classtap extracting portion 11 shown inFIG. 1 . Also, the classtap extracting portion 32 extracts the pixel values of a plurality of pixels, within the student image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest in the form of the class tap from the student image. In this case, the pixel values of a plurality of pixels within the student image are used to classify the pixels of interest into the classes. The classtap extracting portion 32 supplies the data on the resulting class tap to theclass classifying portion 33. - The
class classifying portion 33 classifies the pixels of interest into the classes in correspondence to the feature of the class tap supplied thereto from the classtap extracting portion 32 similarly to the case of theclass classifying portion 12 shown inFIG. 1 , thereby generating the classes corresponding to the pixels of interest. Theclass classifying portion 33 supplies the data on the classes to the normalequation generating portion 37. - The noise amount
tap extracting portion 34 successively determines the pixels composing the teacher image as the pixels of interest, respectively, the case of the classtap extracting portion 32. Also, the noise amounttap extracting portion 34 extracts the pixel values of a plurality of pixels, within the student image, corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest in the form of the noise amount tap from the student image similarly to the case of the noise amounttap extracting portion 14 shown inFIG. 1 . In this case, the pixel values of a plurality of pixels within the student image are used to arithmetically operate the estimate noise amount σ′ corresponding to the pixel of interest. The noise amounttap extracting portion 34 supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operatingportion 35. - The noise amount arithmetically operating
portion 35 arithmetically operates the estimate noise amount σ′ corresponding to the pixel of interest from the noise amount tap supplied thereto from the noise amounttap extracting portion 34 in accordance with Expression (2) similarly to the case of the noise amount arithmetically operatingportion 15 shown inFIG. 1 . Also, the noise amount arithmetically operatingportion 35 supplies the data on the estimate noise amount σ′ corresponding to the pixel of interest to the normalequation generating portion 37. - The predictive
tap extracting portion 36 successively determines the pixels composing the teacher image as the pixels of interest, respectively, similarly to the case of each of the classtap extracting portion 32, and the noise amounttap extracting portion 34. Also, the predictivetap extracting portion 36 extracts the pixel values of a plurality of pixels within the student image corresponding to the position of the pixel of interest, and the peripheral positions of the position of the pixel of interest as the predictive tap from the student image similarly to the case of the predictivetap extracting portion 17 shown inFIG. 1 . In this case, the pixel values of a plurality of pixels within the student image are used to predict the pixel of interest. Also, the predictivetap extracting portion 36 supplies the data on the predictive tap thus extracted to the normalequation generating portion 37. - The normal
equation generating portion 37 generates the addition matrix (refer toFIG. 7 ) in the form of the normal equation. In this case, the addition matrix is obtained by performing the addition by using the estimate noise amount σ′ supplied from the noise amount arithmetically operatingportion 35, the predictive tap supplied from the predictivetap extracting portion 36, and the teacher image inputted from the learningpair storing portion 31 every class supplied from theclass classifying portion 33. - Specifically, the normal
equation generating portion 37 substitutes the pixel values of the student image composing the predictive tap in the k-th sample as Xk (Xk=(xk0, xk1, . . . , xkn)) into the addition matrix shown inFIG. 7 for the classes. In addition, the normalequation generating portion 37 substitutes the pixel values of the pixel of interest of the teacher image in the k-th sample in the form of yk into the addition matrix shown inFIG. 7 . Also, the normalequation generating portion 37 substitutes the estimate noise amounts σ′ corresponding to the pixels of interest, respectively, into the addition matrix shown inFIG. 7 . As a result, the normalequation generating portion 37 generates the normal equation. - It is noted that the addition matrix, as described above, is such that the minimization function Q expressed by Expression (6) is made zero. Therefore, it can be the that the normal equation generated by the normal
equation generating portion 37 is an equation for solving an expression corresponding to Expression (3) for generation of the teacher image based on the multiplication of the predictive coefficient W obtained from both the estimate noise amounts σ′ and the processing coefficients, and the predictive tap of the student image, thereby obtaining the processing coefficients. The normal equation generated from the normalequation generating portion 37 is supplied to thecoefficient generating portion 38. - The
coefficient generating portion 38 solves the normal equation supplied thereto from the normalequation generating portion 37, thereby generating the processing coefficients. Also, thecoefficient generating portion 38 instructs thecoefficient storing portion 39 to store therein the processing coefficients thus generated. The processing coefficients which are learned in the manner as described above, and are then stored in thecoefficient storing portion 39 are stored in thecoefficient storing portion 13 shown inFIG. 1 , and are used in theimage processing apparatus 10. - Next, learning processing executed by the
learning apparatus 30 shown inFIG. 8 will be described with reference to a flow chart ofFIG. 9 . - Firstly, in Step S30, the learning
pair storing portion 31 outputs the learning pair which is not yet outputted of the learning pairs stored therein. Specifically, the learningpair storing portion 31 inputs the data on the student image of the learning pair to each of the classtap extracting portion 32, the noise amounttap extracting portion 34, and the predictivetap extracting portion 36. On the other hand, the learningpair storing portion 31 inputs the data on the teacher image of the learning pair to the normalequation generating portion 37. - Next, in Step S31, the class
tap extracting portion 32, the noise amounttap extracting portion 34, and the predictivetap extracting portion 36 determine one of the pixels, of a plurality of pixels composing the teacher image, which are not yet determined as the pixels of interest as the pixel of interest similarly to the case of the classtap extracting portion 11 shown inFIG. 1 . - In Step S32, the class
tap extracting portion 32 extracts the class tap corresponding to the pixel of interest from the student image, and supplies the data on the class tap thus extracted to theclass classifying portion 33. In Step S33, theclass classifying portion 33 classifies the pixel of interest into the corresponding one of the classes in correspondence to the feature of the class tap supplied thereto from the classtap extracting portion 32 similarly to the case of theclass classifying portion 12 shown inFIG. 1 , thereby generating the class for the pixel of interest. Theclass classifying portion 33 supplies the data on the class thus generated to the normalequation generating portion 37. - In Step S34, the noise amount
tap extracting portion 34 extracts the noise amount tap corresponding to the pixel of interest from the student image, and supplies the data on the noise amount tap thus extracted to the noise amount arithmetically operatingportion 35 similarly to the case of the noise amounttap extracting portion 14 shown inFIG. 1 . - In
Step 35, the noise amount arithmetically operatingportion 35 arithmetically operates the estimate noise amount σ′ corresponding to the noticed image from the noise amount tap supplied thereto from the noise amounttap extracting portion 34 in accordance with Expression (2) similarly to the case of the noise amount arithmetically operatingportion 15 shown inFIG. 1 . Also, the noise amount arithmetically operatingportion 35 supplies the data on the estimate noise amount σ′ to the normalequation generating portion 37. - In Step S36, the predictive tap extracting portion extracts the predictive tap corresponding to the pixel of interest from the student image similarly to the case of the predictive
tap extracting portion 17 shown inFIG. 1 , and supplies the data on the predictive tap thus extracted to the normalequation generating portion 37. - In Step S37, the normal equation generating portion carries out the addition for the addition matrix every class supplied thereto from the class classifying portion (refer to
FIG. 7 ) by using the estimate noise amounts σ′ supplied thereto from the noise amount arithmetically operatingportion 35, the predictive tap supplied thereto from the predictivetap extracting portion 36, and the teacher image inputted thereto from the learningpair storing portion 31. - In Step S38, the class
tap extracting portion 32, the noise amounttap extracting portion 34, and the predictivetap extracting portion 36 determine whether or not all the pixels composing the teacher image are already determined as the pixels of interest. When it is determined in Step S38 that all the pixels composing the teacher image are not yet determined as the pixels of interest, the operation returns back to the processing in Step S31, and the processing described above is repeatedly executed. - On the other hand, when it is determined in Step S38 that all the pixels composing the teacher image are already determined as the pixels of interest, in Step S39, the learning
pair storing portion 31 determines whether or not the processing from Step S30 to Step S38 is already executed for all the learning pairs, that is, whether or not all the learning pairs stored therein are already outputted. When it is determined in Step S39 that the processing from Step S30 to Step S38 is not yet executed for all the learning pairs, the operation returns back to the processing in Step S30, and the processing described above is repeatedly executed. - On the other hand, when it is determined in Step S39 that all the learning pairs stored therein are already outputted, the normal
equation generating portion 37 supplies the data on the normal equation generated by carrying out the addition in Step S37 to thecoefficient generating portion 38. - Also, in Step S40, the
coefficient generating portion 38 generates the processing coefficients by solving the normal equation supplied thereto from the normalequation generating portion 37, and instructs thecoefficient storing portion 39 to store therein the processing coefficients. - As described above, the
learning apparatus 30 learns the optimal processing coefficient every class obtained by the classification based on the correlation of the class tap by using the student image containing therein the luminance dependency noise, and the teacher image, as the ideal image, from which the luminance dependency noise is removed. Therefore, theimage processing apparatus 10 generates the predictive coefficient W by using the optimal processing coefficients and the estimate noise amounts σ′, thereby making it possible to generate the optimal predictive coefficient W, for removal of the luminance dependency noise, corresponding to the correlation of the class tap, and the estimate noise amounts σ′ without explosively increasing the number of classes. As a result, theimage processing apparatus 10 can generate the output image, having the high image quality, from which the luminance dependency noise is removed. - It is noted that although in the above description, the noticed images are classified into the classes in correspondence to the correlation of the class tap, the noticed images may not be classified.
- Next, a series of processing described above can be executed by either the hardware or the software. When the series of processing described above is executed by the software, a program composing the software is installed from a program recording medium either in a computer incorporated in a dedicated hardware, or, for example, in a general purpose personal computer or the like which can execute various functions by installing therein various programs.
-
FIG. 10 shows an example of a configuration of the hardware in acomputer 300 for executing the series of processing described above in accordance with a program. - In the
computer 300, a Central Processing Unit (CPU) 301, a Read Only Memory (ROM) 302, and a Random Access Memory (RAM) 303 are connected to one another through abus 304. - In addition, an I/
O interface 305 is connected to thebus 304. Aninput portion 306, anoutput portion 307, a storingportion 308, acommunication portion 309, and adrive 310 are also connected to the I/O interface 305. In this case, theinput portion 306 is composed of a receiving portion or the like for receiving an instruction transmitted thereto from a keyboard, a mouse, and a microphone, a remote controller, or the like. Theoutput portion 307 is composed of a display device, a speaker or the like. The storingportion 308 is composed of a hard disc, a nonvolatile memory or the like. Also, thedrive 310 drives aremovable media 311 as a package media. In this case, theremovable media 311 is composed of a magnetic disc (including a flexible disc), an optical disc (such as a Compact Disc-Read Only Memory (CD-ROM), or a Digital Versatile Disc (DVD)), a magneto optical disc, or a semiconductor memory. - With the
computer 300 configured in the manner as described above, theCPU 301 loads the program, for example, stored in the storingportion 308 into theRAM 303 through the I/O interface 305 and thebus 304, thereby executing the series of processing described above. - The program which the
CPU 301 of thecomputer 300 executes, for example, is recorded in theremovable media 311, or is provided through a wired or wireless transmission medium such as a Local Area Network (LAN), the Internet or a digital satellite-based broadcasting. - Also, the program can be installed in the storing
portion 308 through the I/O interface 305 by equipping thedrive 310 with theremovable media 311. In addition, the program can be received at thecommunication portion 309 through the wired or wireless transmission media to be installed in the storingportion 308. In addition thereto, the program can be previously installed either in theROM 302 or in the storingportion 308. - It is noted that the program which the
CPU 301 of thecomputer 300 executes may be either a program in accordance with the processing is executed in a time series manner so as to follow the order explained in this specification, or a program in accordance with which the processing is executed at a necessary timing when a call is made. - In addition, the embodiment of the present invention is by no means limited to the embodiments described above, and various changes thereof can be made without departing from the gist of the present invention.
- It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Claims (12)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2007330452A JP5061882B2 (en) | 2007-12-21 | 2007-12-21 | Image processing apparatus, image processing method, program, and learning apparatus |
JP2007-330452 | 2007-12-21 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090161984A1 true US20090161984A1 (en) | 2009-06-25 |
US8363970B2 US8363970B2 (en) | 2013-01-29 |
Family
ID=40788733
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/326,413 Expired - Fee Related US8363970B2 (en) | 2007-12-21 | 2008-12-02 | Image processing apparatus and learning apparatus for converting a first image into a second image having higher image quality than the first image |
Country Status (2)
Country | Link |
---|---|
US (1) | US8363970B2 (en) |
JP (1) | JP5061882B2 (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20110065997A (en) * | 2009-12-10 | 2011-06-16 | 삼성전자주식회사 | Image processing apparatus and image processing method |
JP6319709B2 (en) * | 2014-03-21 | 2018-05-09 | 株式会社Ihi | Debris detection method |
US9262810B1 (en) * | 2014-09-03 | 2016-02-16 | Mitsubishi Electric Research Laboratories, Inc. | Image denoising using a library of functions |
EP3629240B1 (en) * | 2018-09-07 | 2023-08-23 | Panasonic Intellectual Property Corporation of America | Generative adversarial networks for local noise removal from an image |
JP7548827B2 (en) | 2021-01-18 | 2024-09-10 | 住友重機械工業株式会社 | Image inspection device, inspection image display device, image inspection method, and image inspection program |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7085318B2 (en) * | 2000-06-15 | 2006-08-01 | Sony Corporation | Image processing system, image processing method, program, and recording medium |
US20070110333A1 (en) * | 2003-06-27 | 2007-05-17 | Sony Corporation | Signal processing device, signal processing method, program, and recording medium |
US7450764B2 (en) * | 2003-02-27 | 2008-11-11 | Sony Corporation | Image processing device and method, learning device and method, recording medium, and program |
US7460178B2 (en) * | 2004-05-14 | 2008-12-02 | Sony Corporation | Image processing apparatus and image processing method |
US20090161948A1 (en) * | 2007-12-20 | 2009-06-25 | Sony Corporation | Coefficient learning apparatus, coefficient learning method, image processing apparatus, image processing method and programs |
US20090161977A1 (en) * | 2007-12-21 | 2009-06-25 | Sony Corporation | Image processing apparatus, image processing method, program and learning apparatus |
US7814039B2 (en) * | 2001-02-21 | 2010-10-12 | Sony Corporation | Signal processing device and method which learn a prediction coefficient by least-Nth-power error minimization of student teacher data error to detect a telop within image signals which are input signals by selectively outputting input signals following decision of the processing deciding means |
US7853099B2 (en) * | 2006-01-16 | 2010-12-14 | Sony Corporation | Information processing apparatus, information processing method, learning apparatus and learning method and program |
US7876323B2 (en) * | 2006-01-06 | 2011-01-25 | Sony Corporation | Display apparatus and display method, learning apparatus and learning method, and programs therefor |
US7903890B2 (en) * | 2004-12-21 | 2011-03-08 | Sony Corporation | Image processing device, learning device, and coefficient generating device and method |
US7940993B2 (en) * | 2004-12-21 | 2011-05-10 | Sony Corporation | Learning device, learning method, and learning program |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007116292A (en) | 2005-10-19 | 2007-05-10 | Pentax Corp | Fixed pattern noise removal device |
-
2007
- 2007-12-21 JP JP2007330452A patent/JP5061882B2/en not_active Expired - Fee Related
-
2008
- 2008-12-02 US US12/326,413 patent/US8363970B2/en not_active Expired - Fee Related
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7085318B2 (en) * | 2000-06-15 | 2006-08-01 | Sony Corporation | Image processing system, image processing method, program, and recording medium |
US7814039B2 (en) * | 2001-02-21 | 2010-10-12 | Sony Corporation | Signal processing device and method which learn a prediction coefficient by least-Nth-power error minimization of student teacher data error to detect a telop within image signals which are input signals by selectively outputting input signals following decision of the processing deciding means |
US7450764B2 (en) * | 2003-02-27 | 2008-11-11 | Sony Corporation | Image processing device and method, learning device and method, recording medium, and program |
US20070110333A1 (en) * | 2003-06-27 | 2007-05-17 | Sony Corporation | Signal processing device, signal processing method, program, and recording medium |
US7460178B2 (en) * | 2004-05-14 | 2008-12-02 | Sony Corporation | Image processing apparatus and image processing method |
US7903890B2 (en) * | 2004-12-21 | 2011-03-08 | Sony Corporation | Image processing device, learning device, and coefficient generating device and method |
US7940993B2 (en) * | 2004-12-21 | 2011-05-10 | Sony Corporation | Learning device, learning method, and learning program |
US7876323B2 (en) * | 2006-01-06 | 2011-01-25 | Sony Corporation | Display apparatus and display method, learning apparatus and learning method, and programs therefor |
US7853099B2 (en) * | 2006-01-16 | 2010-12-14 | Sony Corporation | Information processing apparatus, information processing method, learning apparatus and learning method and program |
US20090161948A1 (en) * | 2007-12-20 | 2009-06-25 | Sony Corporation | Coefficient learning apparatus, coefficient learning method, image processing apparatus, image processing method and programs |
US20090161977A1 (en) * | 2007-12-21 | 2009-06-25 | Sony Corporation | Image processing apparatus, image processing method, program and learning apparatus |
Also Published As
Publication number | Publication date |
---|---|
US8363970B2 (en) | 2013-01-29 |
JP5061882B2 (en) | 2012-10-31 |
JP2009153003A (en) | 2009-07-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108228703B (en) | Image question-answering method, device, system and storage medium | |
US8346003B2 (en) | Image processing for reducing blur | |
US8363970B2 (en) | Image processing apparatus and learning apparatus for converting a first image into a second image having higher image quality than the first image | |
US20130016920A1 (en) | Image processing device, image processing method, program and recording medium | |
CN101465943A (en) | Image processing apparatus, image processing method, program and learning apparatus | |
EP0449516A2 (en) | Image processing using neural network | |
US20220076077A1 (en) | Quality estimation model trained on training signals exhibiting diverse impairments | |
US20230274758A1 (en) | Method and electronic device | |
US20100080451A1 (en) | Image processing apparatus and coefficient learning apparatus | |
Giboulot et al. | Multivariate side-informed Gaussian embedding minimizing statistical detectability | |
TW200715854A (en) | Image processing apparatus, image processing method, and program | |
CN118866012B (en) | Tone quality evaluation method for multidimensional fusion analysis | |
US7679675B2 (en) | Data converting apparatus, data converting method, learning apparatus, leaning method, program, and recording medium | |
US7925019B2 (en) | Method and apparatus for converting data, method and apparatus for inverse converting data, and recording medium | |
US20110243462A1 (en) | Coefficient learning apparatus and method, image processing apparatus and method, program, and recording medium | |
EP1501295A4 (en) | Data converting apparatus, data converting method, learning apparatus, leaning method, and recording medium | |
JP4310697B2 (en) | Signal processing apparatus and method, recording medium, and program | |
JP2009060481A (en) | Image processing device and method, learning device and method, and program | |
Pujol et al. | Color quantization for image processing using self information | |
US8311963B2 (en) | Signal processing apparatus, signal processing method, signal processing program and learning apparatus | |
JP4093114B2 (en) | Information signal processing apparatus and processing method, codebook generation apparatus and generation method, and program for executing each method | |
JP2003303330A5 (en) | ||
US8208735B2 (en) | Image processing device, image processing method, learning device, learning method, and program | |
EP4517647A1 (en) | Method and apparatus for scaling and super-resolution based on machine learning | |
EP4369251A1 (en) | Systems and methods for tensorizing convolutional neural networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION,JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SUTO, YASUHIRO;KONDO, TETSUJIRO;SHIRAKI, HISAKAZU;AND OTHERS;SIGNING DATES FROM 20081112 TO 20081119;REEL/FRAME:021919/0204 Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SUTO, YASUHIRO;KONDO, TETSUJIRO;SHIRAKI, HISAKAZU;AND OTHERS;SIGNING DATES FROM 20081112 TO 20081119;REEL/FRAME:021919/0204 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20170129 |