WO2006030519A1 - 顔認証装置及び顔認証方法 - Google Patents
顔認証装置及び顔認証方法 Download PDFInfo
- Publication number
- WO2006030519A1 WO2006030519A1 PCT/JP2004/013666 JP2004013666W WO2006030519A1 WO 2006030519 A1 WO2006030519 A1 WO 2006030519A1 JP 2004013666 W JP2004013666 W JP 2004013666W WO 2006030519 A1 WO2006030519 A1 WO 2006030519A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- face
- image
- feature
- feature quantity
- eyes
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 29
- 238000000605 extraction Methods 0.000 claims abstract description 66
- 239000000284 extract Substances 0.000 claims abstract description 4
- 238000001514 detection method Methods 0.000 claims description 57
- 238000004364 calculation method Methods 0.000 claims description 13
- 230000001815 facial effect Effects 0.000 claims description 11
- 238000010586 diagram Methods 0.000 description 22
- 230000004044 response Effects 0.000 description 21
- 238000001914 filtration Methods 0.000 description 4
- 238000010606 normalization Methods 0.000 description 4
- 210000004709 eyebrow Anatomy 0.000 description 3
- YBJHBAHKTGYVGT-ZKWXMUAHSA-N (+)-Biotin Chemical compound N1C(=O)N[C@@H]2[C@H](CCCCC(=O)O)SC[C@@H]21 YBJHBAHKTGYVGT-ZKWXMUAHSA-N 0.000 description 2
- FEPMHVLSLDOMQC-UHFFFAOYSA-N virginiamycin-S1 Natural products CC1OC(=O)C(C=2C=CC=CC=2)NC(=O)C2CC(=O)CCN2C(=O)C(CC=2C=CC=CC=2)N(C)C(=O)C2CCCN2C(=O)C(CC)NC(=O)C1NC(=O)C1=NC=CC=C1O FEPMHVLSLDOMQC-UHFFFAOYSA-N 0.000 description 2
- PCTMTFRHKVHKIS-BMFZQQSSSA-N (1s,3r,4e,6e,8e,10e,12e,14e,16e,18s,19r,20r,21s,25r,27r,30r,31r,33s,35r,37s,38r)-3-[(2r,3s,4s,5s,6r)-4-amino-3,5-dihydroxy-6-methyloxan-2-yl]oxy-19,25,27,30,31,33,35,37-octahydroxy-18,20,21-trimethyl-23-oxo-22,39-dioxabicyclo[33.3.1]nonatriaconta-4,6,8,10 Chemical compound C1C=C2C[C@@H](OS(O)(=O)=O)CC[C@]2(C)[C@@H]2[C@@H]1[C@@H]1CC[C@H]([C@H](C)CCCC(C)C)[C@@]1(C)CC2.O[C@H]1[C@@H](N)[C@H](O)[C@@H](C)O[C@H]1O[C@H]1/C=C/C=C/C=C/C=C/C=C/C=C/C=C/[C@H](C)[C@@H](O)[C@@H](C)[C@H](C)OC(=O)C[C@H](O)C[C@H](O)CC[C@@H](O)[C@H](O)C[C@H](O)C[C@](O)(C[C@H](O)[C@H]2C(O)=O)O[C@H]2C1 PCTMTFRHKVHKIS-BMFZQQSSSA-N 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 210000004209 hair Anatomy 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/74—Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
Definitions
- the present invention relates to a face authentication apparatus and a face authentication method for extracting a face area from an image of a face and performing authentication by comparing the image of the face area with previously registered data.
- Patent Document 1 Japanese Patent Application Laid-Open No. 2002-342760
- the present invention has been made in order to solve the above-described problems, and can accurately extract a face region even in various face images and can reduce the amount of calculation.
- An object is to obtain a face authentication device and a face authentication method.
- the face authentication device includes a feature quantity extraction image generating means for generating a feature quantity extraction image obtained by performing a predetermined calculation on each pixel value for an input image, and a feature quantity extraction device.
- the position of both eyes is detected from the face detection means for detecting the face area from the image and the feature amount extraction image.
- the image processing apparatus includes a face authentication unit that performs face authentication by comparing with the feature amount acquired by the means.
- FIG. 1 is a block diagram showing a face authentication apparatus according to Embodiment 1 of the present invention.
- FIG. 2 is a flowchart showing the operation of the face authentication apparatus according to Embodiment 1 of the present invention.
- FIG. 3 is an explanatory diagram showing a relationship between an original image and an integral image of the face authentication apparatus according to Embodiment 1 of the present invention.
- FIG. 4 is an explanatory diagram showing a method for dividing and processing an image of the face authentication apparatus according to the first embodiment of the present invention.
- FIG. 5 is an explanatory diagram of a rectangular filter of the face authentication device according to Embodiment 1 of the present invention.
- FIG. 6 is an explanatory diagram of a process for obtaining a total pixel value of the face authentication device according to the first embodiment of the present invention.
- FIG. 7 is an explanatory diagram of a process for obtaining the sum of pixel values in a rectangle when the integral image of the face authentication apparatus according to Embodiment 1 of the present invention is divided and obtained.
- FIG. 8 is an explanatory diagram of a search block to be detected when detecting a face area of the face authentication device according to the first embodiment of the present invention.
- FIG. 9 is a flowchart showing face area detection processing of the face authentication apparatus according to embodiment 1 of the present invention.
- FIG. 10 is an explanatory diagram showing a face area detection result of the face authentication device according to the first embodiment of the present invention.
- FIG. 11 is an explanatory diagram of a binocular search performed by the face authentication device according to the first embodiment of the present invention.
- FIG. 12 is an explanatory diagram of an eye region search operation of the face authentication device according to the first embodiment of the present invention.
- FIG. 13 is an explanatory diagram of a regular image process of the face authentication device according to the first embodiment of the present invention.
- FIG. 14 is an explanatory diagram of a feature amount database of the face authentication device according to the first embodiment of the present invention.
- FIG. 1 is a block diagram showing a face authentication apparatus according to Embodiment 1 of the present invention.
- the face authentication device includes an image input unit 1, a feature amount extraction image generation unit 2, a face detection unit 3, a binocular detection unit 4, a face image normalization unit 5, a feature amount acquisition unit 6, A feature quantity storage means 7, a feature quantity extraction image storage means 8, a feature quantity database 9, and a face authentication means 10 are provided.
- the image input unit 1 is a functional unit for inputting an image.
- the feature quantity extraction image generation means 2 is a means for acquiring a feature quantity extraction image obtained by performing a predetermined operation on each pixel value for the image input by the image input means 1.
- the feature amount extraction image is, for example, an integral image, and details thereof will be described later.
- the face detection unit 3 is a functional unit that detects a face area by a predetermined method based on the feature amount extraction image acquired by the feature amount extraction image generation unit 2.
- the binocular detection means 4 is a functional unit that detects the binocular area from the face area by the same method as the face detection means 3.
- the face image normalization means 5 is a functional unit that enlarges or reduces the face area to the image size to be face-authenticated based on the position of both eyes detected by the both-eye detection means 4.
- the feature quantity acquisition unit 6 is a functional unit that acquires a feature quantity for face recognition with normal facial recognition.
- the feature quantity storage unit 7 stores the feature quantity in the feature quantity database 9 and the face authentication unit 10. This is the function part to send.
- the feature quantity extraction image storage means 8 is a functional unit that stores the feature quantity extraction image acquired by the feature quantity extraction image generation means 2.
- the face detection means 3 -feature quantity acquisition means 6 This The feature amount extraction image storage means 8 is configured to perform various processes based on the feature extraction image stored in the feature amount extraction image storage means 8.
- the feature quantity database 9 includes the facial feature quantity used by the face detection means 3, the eye feature quantity used by the both-eye detection means 4, and the individual characteristics used by the face recognition means 10. It is a database that stores quantities.
- the face authentication means 10 compares the feature quantity to be authenticated acquired by the feature quantity acquisition means 6 with the feature quantity data of each person's face registered in the feature quantity database 9 in advance for face authentication. It is a functional part that performs
- FIG. 2 is a flowchart showing the operation.
- an image is input by the image input means 1 (step ST101).
- images taken with a digital camera equipped in a mobile phone or PDA, images input from an external memory, etc., images acquired using a means of internet communication, etc. are input to a mobile phone or PDA.
- images taken with a digital camera equipped in a mobile phone or PDA, images input from an external memory, etc., images acquired using a means of internet communication, etc. are input to a mobile phone or PDA.
- the feature quantity extraction image generation means 2 obtains a feature quantity extraction image (step ST102).
- the feature amount extraction image is an image used when filtering an image with a filter called a Rectangle Filter (rectangle filter) used to extract each feature in face detection, both-eye detection, and face authentication.
- a filter called a Rectangle Filter (rectangle filter) used to extract each feature in face detection, both-eye detection, and face authentication.
- FIG. 3 it is an integrated image obtained by calculating the total of pixel values in the direction of the coordinate axes (horizontal and vertical directions) of the X and Y coordinates.
- FIG. 3 is an explanatory diagram showing the result of converting the original image into an integral image by the feature quantity extraction image generation means 2.
- the integral image 12 is obtained. That is
- the calculated value of the integral image 12 corresponding to each pixel value of the original image 11 is a value obtained by adding each pixel value of the original image 11 in the pixel value force horizontal and vertical directions at the upper left of the drawing.
- the gray scale I can be obtained using the following equation, for example.
- the average value of each RGB component may be obtained.
- I (x, y) 0.2988I (x, y) + 0.5868I (x, y) +0.11441 (x, y)
- the image input means 1 when the input image size is a large size such as 3 million pixels, it cannot be expressed by integer type data used to express each pixel value of the integral image. There is. In other words, the integral value may overflow the data size of the integer type. Therefore, in the present embodiment, in consideration of such a case, the image is divided as follows within a range where overflow does not occur, and an integral image of each divided partial image is obtained.
- the integral image 12 is a value obtained by accumulating the pixel values of the original image 11 as they are, but the same applies to an integral image having a value obtained by squaring each pixel value of the original image 11. It is applicable to. However, in this case, since the integral value does not overflow the integer type data size, the division is further reduced (the divided image is small).
- FIG. 4 is an explanatory diagram showing a method for dividing and processing an image.
- 13-16 shows the divided images
- 17-19 shows the case where the search window overlaps the divided images.
- an integral image is obtained from the divided partial images 13, 14, 15, and 16.
- the rectangle for which the total value is calculated may extend over a plurality of divided images.
- three different cases are possible: 18 if they are different in the vertical direction, 17 if they are different in the horizontal direction, and 19 if they are different in the four divided images. There are two possible cases. The processing method in each of these cases will be described later.
- the face detection means 3 After obtaining the integrated image as described above, the face detection means 3 detects the image strength / face area (step ST104).
- characteristics of human face characteristics, eye characteristics, and individual differences of faces are represented by a combination of response values after filtering the image using multiple Rectangle Filters 20 shown in Fig. 5.
- the Rectangle Filter 20 shown in Fig. 5 is a value obtained by subtracting the sum of pixel values in a hatched rectangle from the sum of pixel values in a white rectangle in a fixed size search block, for example, a block of 24 X 24 pixels. Is what you want.
- the pixel value total is shown.
- Rectangle Filter 20 shown in FIG. 5 is a basic one, and actually there are a plurality of Rectangle Filters 20 having different positions and sizes in the search block.
- weighting is performed according to a plurality of filtering response values filtered using a plurality of Rectangle Filters suitable for detecting a human face, and the linear sum of the weighted values is greater than a threshold value. Whether or not the search block has a facial area power is determined by whether or not it is. In other words, the weight given according to the filtering response value represents the feature of the face, and this weight is acquired in advance using a learning algorithm or the like.
- the face detection means 3 is based on the total pixel value of each rectangle in the search block. And face detection. At this time, the integrated image obtained by the feature quantity extraction image generating means 2 is used as a means for efficiently performing the pixel value summation calculation.
- the total pixel value in the rectangle can be calculated by the following equation.
- the total pixel value in the rectangle can be obtained by only four computations, and the total pixel value in an arbitrary rectangle can be obtained efficiently.
- the integral pixel value of the integral image 12 is also represented by an integer, all the face authentication processing of the present embodiment in which various processes are performed using such an integral image 12 is an integer computation. Is possible.
- the overlapping pattern can be divided into 18 when overlapping in the vertical direction, 17 when overlapping in the horizontal direction, and 19 when overlapping with the four divided images.
- FIG. 7 is an explanatory diagram showing a case of three overlapping patterns.
- the sum of the pixel values of the portion overlapping each divided image may be added.
- the total pixel value of the rectangle AGEI can be calculated using the following equation.
- the search block used for extracting the facial feature value is fixed to, for example, 24 X 24 pixels, and the face image of the search block size is used when learning the facial feature value. Learning. Image area with an image size Cannot be detected using a search block with a fixed size. In order to solve this problem, there are methods of scaling the image to create multiple resolution images, or scaling the search block, and either method can be used. .
- the search block is enlarged or reduced.
- a face region of an arbitrary size can be detected by enlarging the search block at a constant enlargement / reduction ratio as follows.
- FIG. 8 is an explanatory diagram of a search block to be detected when detecting a face area.
- FIG. 9 is a flowchart showing face area detection processing.
- the enlargement / reduction ratio S is set to 1.0, and the process starts from an equal-size search block (step ST2 01).
- step ST202 it is determined whether the image in the search block is a face area while moving the search block one pixel at a time in the vertical and horizontal directions, and if it is a face area, the coordinates are stored (step ST202—step ST209).
- a new rectangular coordinate (coordinates of vertices constituting the rectangle) when the scaling factor S is applied to the rectangular coordinates in the Rectangle Filter is obtained (step ST204).
- each rectangular coordinate when the search block is enlarged or reduced is obtained by the following equation.
- top is the upper left Y coordinate of the rectangle
- left is the upper left X coordinate of the rectangle
- height is the height of the rectangle
- width is the width of the rectangle
- S is the scaling factor
- rc and cc are the rectangle
- the original The vertex coordinates, rn, cn are the vertex coordinates after conversion.
- a filter response is obtained based on the integral image stored in the feature quantity extraction image storage means 8 (step ST205). Since this filter response has an enlarged rectangle, it is larger by the enlargement / reduction ratio than the value of the search block size used during learning.
- the value when the filter response is obtained with the same search block size as that during learning is obtained by dividing the filter response by the enlargement / reduction ratio (step ST206).
- F is the response
- R is the response obtained from the enlarged rectangle
- S is the magnification.
- a weight corresponding to the response is obtained from the value obtained above, a linear sum of all weights is obtained, and whether or not the facial power is determined is determined by comparing the obtained value with a threshold value (step ST207). If it is a face, the coordinates of the search block at that time are stored.
- step ST210 After scanning the entire image, multiply the enlargement / reduction ratio S by a fixed value, for example, 1.25 (step ST210), and repeat the processing of step ST202 to step ST209 with the new enlargement / reduction ratio. Then, when the enlarged search block size exceeds the image size, the process ends (step ST211).
- the face area detected above is stored by determining a plurality of search blocks as face areas in the vicinity of the face in order to perform face area determination while moving the search block one pixel at a time as described above. Face area rectangles may overlap.
- FIG. 10 is an explanatory diagram showing this, and shows the detection result of the face area.
- the search blocks 25 in the figure are originally one area, so the rectangles overlap. In this case, the rectangles are integrated according to the overlapping ratio.
- the overlapping ratio can be obtained by the following equation, for example, when rectangle 1 and rectangle 2 overlap.
- Overlap rate Area of overlap area Area of Z rectangle 1
- Overlap ratio Area of overlap area Area of Z rectangle 2
- the two rectangles are integrated into one rectangle.
- the force to find the average value of the coordinates of each of the four points or the magnitude relationship force of the coordinate values can be obtained.
- both eyes are detected by the both eyes detecting means 4 from the face area obtained as described above (step ST105).
- the face detection means 3 If the features of the human face are taken into account from the face area detected by the face detection means 3, it is possible to predict in advance where the left eye and the right eye exist.
- the both-eye detection means 4 identifies each eye's search area from the coordinates of the face area, and detects eyes by paying attention to the search area.
- FIG. 11 is an explanatory diagram of the binocular search, in which 26 indicates the left eye search area and 27 indicates the right eye search area.
- Both eyes can be detected by the same process as the face detection in step ST104.
- the features of the left eye and the right eye are learned using the Rectangle Filter so that the center of the eye becomes the center of the search block, for example.
- eyes are detected while enlarging the search block in the same manner as in face detection step ST201—step ST211.
- an eye When an eye is detected, it may be set to end when the search block size after enlargement exceeds the search area size of each eye.
- searching for eyes it is very inefficient to scan from the upper left of the search area like the face detection means 3. This is because the eye position often exists near the center of the set search area.
- FIG. 12 is an explanatory diagram of the eye region search operation.
- the both-eye detection means 4 performs eye search processing from the center of the search range of both eyes in the detected face region to the periphery, and detects the positions of both eyes.
- the central force of the search area is also directed toward the periphery to search in a spiral shape.
- step ST106 the face image is normalized based on the positions of both eyes detected in step ST105 (step ST106).
- FIG. 13 is an explanatory diagram of normalization processing.
- the face image normalization means 5 also requires image power when the face area is enlarged / reduced from the positions 28 and 29 of both eyes detected by the both-eye detection means 4 so that the angle of view is required for face authentication. Extract facial features.
- the size of the normalized image 30 is, for example, the width and height are nwX nh pixels, and the position of the left eye and the position of the right eye are the coordinates L (xl, yl), R (xr, yr) in the normal image 30 Is set, the following processing is performed in order to make the detected face area conform to the set regular image.
- the enlargement / reduction ratio NS can be calculated by the following equation when the detected positions of both eyes are DL (xdl.ydl) and DR (xdr, ydr).
- NS ((xr-xl + l) 2 + (yr-yl + l) 2 ) / ((xdr-xdl + l) 2 + (ydr-ydl + 1) 2 )
- the position of the normalized image in the original image that is, the rectangular position to be authenticated is obtained using the obtained enlargement / reduction ratio and information on the positions of the left eye and right eye set on the normalized image.
- OrgNrImgTopLeft (x, y) (xdl-xl / NS, ydl-yl / NS)
- OrgNrmImgBtmRight (x, y) (xdl + (nw-xl) / NS, ydl + (nh-yl) / NS)
- a feature amount necessary for face authentication is extracted from the authentication target area obtained as described above using a Rectangle Filter for face authentication.
- the Rectangle Filter for face authentication is designed assuming a normal image size
- the rectangular coordinates in the Rectangle Filter are converted to the coordinates in the original image as in face detection, and the total pixel value is based on the integrated image. Therefore, the filter response at the normal image size can be obtained by multiplying the obtained filter response by the enlargement / reduction ratio NS obtained above.
- OrgRgn (x, y) (xdl + rx * NS, ydl + ry * NS)
- rx and ry are rectangular coordinates on the regular image 30.
- the pixel value of the integral image is referred to from the rectangular coordinates obtained here, and the pixel value in the rectangle is obtained.
- the response of the plurality of Rectangle Filters is obtained (step ST107).
- responses of a plurality of Rectangle Filters are stored in the feature value database 9 by the feature value storage means 7 (steps ST 108 and ST 109).
- FIG. 14 is an explanatory diagram of the feature quantity database 9.
- the feature quantity database 9 has a table structure of registration ID and feature quantity data as shown in the figure. That is, the response 31 of the plurality of Rectangle Filters 20 is obtained for the normalized image 30, and these responses 31 are associated with the registration ID corresponding to the individual.
- step ST110 and step ST111 in FIG. 2 processing for performing face authentication by the face authentication means 10 (step ST110 and step ST111 in FIG. 2) will be described.
- Face authentication is performed by comparing the feature quantity extracted by the feature quantity acquisition means 6 from the input image with the feature quantity stored in the feature quantity database 9. Specifically, when the feature value of the input image is RFc and the registered feature value is RFr, the weight is given as shown in the following equation (5) according to the difference between the feature values.
- feature amount storage registration processing
- face authentication authentication processing
- real-time processing can be realized even with, for example, a mobile phone or a PDA.
- I (x, y) is the total pixel value in the white rectangle
- I (x, y) is the image wwbb in the hatched rectangle
- an integrated image when used as the feature amount extraction image, it can be applied in the same manner as the above-described integrated image by corresponding to the integrated image as a feature amount expression. it can.
- an integrated image for obtaining the total obtained by subtracting the pixel values in the horizontal and vertical directions may be used as the feature quantity extraction image! ,.
- a feature amount extraction image for generating a feature amount extraction image obtained by performing a predetermined operation on each pixel value for an input image.
- a face detection unit for detecting a face area using learning data obtained by previously learning a facial feature from a feature quantity extraction image generated by an image generation unit and a feature quantity extraction image generation unit, and a detected face area Using the learning data that has learned the eye features in advance from the feature amount extraction image, the both eye detection means that detects the position of both eyes, and the image that normalizes the face area based on the position of both eyes!
- a feature amount acquisition means for extracting feature amounts
- a face authentication means for performing face authentication by comparing the feature amounts of individuals registered in advance with the feature amounts acquired by the feature amount acquisition means.
- the face detection means obtains a feature amount from a pixel value total difference of a specific rectangle in a predetermined search window in the feature amount extraction image, and the result
- the binocular detection means obtains a feature amount from a pixel value total difference of a specific rectangle within a predetermined search window in the feature amount extraction image, performs binocular detection based on the result, and detects the face.
- the authentication means performs face authentication using the result of obtaining the feature value from the pixel value total difference of the specific rectangle in the predetermined search window in the feature value extraction image. Therefore, the feature value can be accurately calculated with a small amount of calculation. Can be requested.
- face detection, both-eye detection, and face authentication processing are performed based on the feature amount extraction image obtained once, the processing efficiency can be improved.
- the feature quantity extraction image generation means is configured to extract an image having a value obtained by adding or multiplying the pixel value of each pixel in the direction of the coordinate axis. Picture Since the image is generated as an image, for example, the sum of pixel values in an arbitrary rectangle can be obtained only by calculation of four points, and the feature amount can be obtained efficiently with a small amount of computation.
- the face detection unit enlarges or reduces the search window, normalizes the feature amount according to the enlargement / reduction ratio, and detects the face area. Therefore, it is possible to increase memory efficiency without having to obtain a multi-resolution image and a feature amount extraction image corresponding to each resolution.
- the feature quantity extraction image generation means applies to each divided image divided within a range in which the calculated value of the feature quantity extraction image can be expressed. Since the feature amount extraction image is obtained, even when the image size becomes large, when the feature amount extraction image is obtained, it is not possible to cause overflow by dividing the image. There is an effect that can cope with various input image sizes.
- the feature amount extraction image for generating the feature amount extraction image data obtained by performing a predetermined operation on each pixel value with respect to the input image data.
- An acquisition step, a face area detection step for detecting a face area using learning data obtained by previously learning a facial feature from feature quantity extraction image data, and feature amount extraction image data for the detected face area From the two-eye detection step for detecting the position of both eyes using the learning data obtained by learning the eye characteristics in advance, and the feature amount data from the image data normalized based on the positions of both eyes! Since there is an authentication step that performs facial authentication by comparing the feature amount acquisition step to be extracted, the feature amount data of each individual registered in advance, and the feature amount data acquired in the feature amount acquisition step. Even if the input image Authentication processing can be performed, and face authentication processing can be performed with a small amount of computation.
- the face detection unit detects a face area from the input image, and the center of the search range of both eyes in the detected face area moves from the center to the periphery.
- a binocular detection means for performing a search and detecting the position of both eyes; a feature quantity acquisition means for extracting a feature quantity from an image obtained by normalizing a face area based on the positions of both eyes; and a personal feature quantity registered in advance.
- the face authentication unit since it is equipped with a face authentication unit that compares the feature amount acquired by the feature amount acquisition unit and performs face authentication, the amount of calculation in the binocular search process can be reduced, and as a result, the face authentication process is efficiently performed.
- a face area detection step of detecting a face area from input image data, and the periphery from the center of the search range of both eyes in the detected face area A eye detection process for detecting the position of both eyes, a feature amount acquisition step for extracting feature data from image data obtained by normalizing the face area based on the positions of both eyes, Since it includes a face authentication step that performs face authentication by comparing pre-registered individual feature data and feature data acquired in the feature acquisition step, perform binocular search processing with a small amount of computation As a result, the face recognition process can be efficiently performed.
- the face authentication device and the face authentication method according to the present invention perform face authentication by comparing an input image with a pre-registered image. Suitable for use in security systems.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Geometry (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Collating Specific Patterns (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/659,665 US20080080744A1 (en) | 2004-09-17 | 2004-09-17 | Face Identification Apparatus and Face Identification Method |
PCT/JP2004/013666 WO2006030519A1 (ja) | 2004-09-17 | 2004-09-17 | 顔認証装置及び顔認証方法 |
JP2006535003A JPWO2006030519A1 (ja) | 2004-09-17 | 2004-09-17 | 顔認証装置及び顔認証方法 |
CN2004800440129A CN101023446B (zh) | 2004-09-17 | 2004-09-17 | 脸部认证装置和脸部认证方法 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2004/013666 WO2006030519A1 (ja) | 2004-09-17 | 2004-09-17 | 顔認証装置及び顔認証方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006030519A1 true WO2006030519A1 (ja) | 2006-03-23 |
Family
ID=36059786
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2004/013666 WO2006030519A1 (ja) | 2004-09-17 | 2004-09-17 | 顔認証装置及び顔認証方法 |
Country Status (4)
Country | Link |
---|---|
US (1) | US20080080744A1 (ja) |
JP (1) | JPWO2006030519A1 (ja) |
CN (1) | CN101023446B (ja) |
WO (1) | WO2006030519A1 (ja) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008027275A (ja) * | 2006-07-24 | 2008-02-07 | Seiko Epson Corp | オブジェクト検出装置、オブジェクト検出方法および制御プログラム |
JP2009237634A (ja) * | 2008-03-25 | 2009-10-15 | Seiko Epson Corp | オブジェクト検出方法、オブジェクト検出装置、オブジェクト検出プログラムおよび印刷装置 |
JP2010500687A (ja) * | 2006-08-11 | 2010-01-07 | フォトネーション ビジョン リミテッド | デジタル画像取得装置におけるリアルタイムの顔探知 |
JP2011013732A (ja) * | 2009-06-30 | 2011-01-20 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
US8422739B2 (en) | 2006-08-11 | 2013-04-16 | DigitalOptics Corporation Europe Limited | Real-time face tracking in a digital image acquisition device |
US8463049B2 (en) * | 2007-07-05 | 2013-06-11 | Sony Corporation | Image processing apparatus and image processing method |
JP2015036123A (ja) * | 2013-08-09 | 2015-02-23 | 株式会社東芝 | 医用画像処理装置、医用画像処理方法及び分類器トレーニング方法 |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7953253B2 (en) * | 2005-12-31 | 2011-05-31 | Arcsoft, Inc. | Face detection on mobile devices |
US7643659B2 (en) * | 2005-12-31 | 2010-01-05 | Arcsoft, Inc. | Facial feature detection on mobile devices |
KR100771244B1 (ko) * | 2006-06-12 | 2007-10-29 | 삼성전자주식회사 | 동영상 데이터 처리 방법 및 장치 |
US9042606B2 (en) * | 2006-06-16 | 2015-05-26 | Board Of Regents Of The Nevada System Of Higher Education | Hand-based biometric analysis |
FI20075453A0 (sv) * | 2007-06-15 | 2007-06-15 | Virtual Air Guitar Company Oy | Bildsampling i en stokastisk modellbaserad datorvision |
JP5390943B2 (ja) | 2008-07-16 | 2014-01-15 | キヤノン株式会社 | 画像処理装置及び画像処理方法 |
JP5239625B2 (ja) * | 2008-08-22 | 2013-07-17 | セイコーエプソン株式会社 | 画像処理装置、画像処理方法および画像処理プログラム |
US20110199499A1 (en) * | 2008-10-14 | 2011-08-18 | Hiroto Tomita | Face recognition apparatus and face recognition method |
KR101522985B1 (ko) * | 2008-10-31 | 2015-05-27 | 삼성전자주식회사 | 영상처리 장치 및 방법 |
KR101179497B1 (ko) * | 2008-12-22 | 2012-09-07 | 한국전자통신연구원 | 얼굴 검출 방법 및 장치 |
US8339506B2 (en) * | 2009-04-24 | 2012-12-25 | Qualcomm Incorporated | Image capture parameter adjustment using face brightness information |
TWI413936B (zh) * | 2009-05-08 | 2013-11-01 | Novatek Microelectronics Corp | 人臉偵測裝置及其人臉偵測方法 |
JP2011128990A (ja) * | 2009-12-18 | 2011-06-30 | Canon Inc | 画像処理装置とその方法 |
JP5417368B2 (ja) * | 2011-03-25 | 2014-02-12 | 株式会社東芝 | 画像識別装置及び画像識別方法 |
KR101494874B1 (ko) * | 2014-05-12 | 2015-02-23 | 김호 | 사용자 인증 방법, 이를 실행하는 장치 및 이를 저장한 기록 매체 |
EP3174007A1 (en) | 2015-11-30 | 2017-05-31 | Delphi Technologies, Inc. | Method for calibrating the orientation of a camera mounted to a vehicle |
EP3173979A1 (en) | 2015-11-30 | 2017-05-31 | Delphi Technologies, Inc. | Method for identification of characteristic points of a calibration pattern within a set of candidate points in an image of the calibration pattern |
EP3534334B1 (en) | 2018-02-28 | 2022-04-13 | Aptiv Technologies Limited | Method for identification of characteristic points of a calibration pattern within a set of candidate points derived from an image of the calibration pattern |
EP3534333A1 (en) * | 2018-02-28 | 2019-09-04 | Aptiv Technologies Limited | Method for calibrating the position and orientation of a camera relative to a calibration pattern |
CN111144265A (zh) * | 2019-12-20 | 2020-05-12 | 河南铭视科技股份有限公司 | 一种人脸算法脸部图像提取方法和装置 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH04101280A (ja) * | 1990-08-20 | 1992-04-02 | Nippon Telegr & Teleph Corp <Ntt> | 顔画像照合装置 |
JPH05225342A (ja) * | 1992-02-17 | 1993-09-03 | Nippon Telegr & Teleph Corp <Ntt> | 動物体追跡処理方法 |
JP2000331158A (ja) * | 1999-05-18 | 2000-11-30 | Mitsubishi Electric Corp | 顔画像処理装置 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3426060B2 (ja) * | 1995-07-28 | 2003-07-14 | 三菱電機株式会社 | 顔画像処理装置 |
JP3350296B2 (ja) * | 1995-07-28 | 2002-11-25 | 三菱電機株式会社 | 顔画像処理装置 |
US6735566B1 (en) * | 1998-10-09 | 2004-05-11 | Mitsubishi Electric Research Laboratories, Inc. | Generating realistic facial animation from speech |
JP3600755B2 (ja) * | 1999-05-13 | 2004-12-15 | 三菱電機株式会社 | 顔画像処理装置 |
JP3969894B2 (ja) * | 1999-05-24 | 2007-09-05 | 三菱電機株式会社 | 顔画像処理装置 |
JP3695990B2 (ja) * | 1999-05-25 | 2005-09-14 | 三菱電機株式会社 | 顔画像処理装置 |
JP3768735B2 (ja) * | 1999-07-07 | 2006-04-19 | 三菱電機株式会社 | 顔画像処理装置 |
JP2001351104A (ja) * | 2000-06-06 | 2001-12-21 | Matsushita Electric Ind Co Ltd | パターン認識方法及びパターン認識装置、並びにパターン照合方法及びパターン照合装置 |
US7099510B2 (en) * | 2000-11-29 | 2006-08-29 | Hewlett-Packard Development Company, L.P. | Method and system for object detection in digital images |
US6895103B2 (en) * | 2001-06-19 | 2005-05-17 | Eastman Kodak Company | Method for automatically locating eyes in an image |
JP4161659B2 (ja) * | 2002-02-27 | 2008-10-08 | 日本電気株式会社 | 画像認識システム及びその認識方法並びにプログラム |
KR100438841B1 (ko) * | 2002-04-23 | 2004-07-05 | 삼성전자주식회사 | 이용자 검증 및 데이터 베이스 자동 갱신 방법, 및 이를이용한 얼굴 인식 시스템 |
US7369687B2 (en) * | 2002-11-21 | 2008-05-06 | Advanced Telecommunications Research Institute International | Method for extracting face position, program for causing computer to execute the method for extracting face position and apparatus for extracting face position |
KR100455294B1 (ko) * | 2002-12-06 | 2004-11-06 | 삼성전자주식회사 | 감시 시스템에서의 사용자 검출 방법, 움직임 검출 방법및 사용자 검출 장치 |
US7508961B2 (en) * | 2003-03-12 | 2009-03-24 | Eastman Kodak Company | Method and system for face detection in digital images |
JP2005044330A (ja) * | 2003-07-24 | 2005-02-17 | Univ Of California San Diego | 弱仮説生成装置及び方法、学習装置及び方法、検出装置及び方法、表情学習装置及び方法、表情認識装置及び方法、並びにロボット装置 |
US7274832B2 (en) * | 2003-11-13 | 2007-09-25 | Eastman Kodak Company | In-plane rotation invariant object detection in digitized images |
-
2004
- 2004-09-17 JP JP2006535003A patent/JPWO2006030519A1/ja active Pending
- 2004-09-17 WO PCT/JP2004/013666 patent/WO2006030519A1/ja active Application Filing
- 2004-09-17 CN CN2004800440129A patent/CN101023446B/zh not_active Expired - Fee Related
- 2004-09-17 US US11/659,665 patent/US20080080744A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH04101280A (ja) * | 1990-08-20 | 1992-04-02 | Nippon Telegr & Teleph Corp <Ntt> | 顔画像照合装置 |
JPH05225342A (ja) * | 1992-02-17 | 1993-09-03 | Nippon Telegr & Teleph Corp <Ntt> | 動物体追跡処理方法 |
JP2000331158A (ja) * | 1999-05-18 | 2000-11-30 | Mitsubishi Electric Corp | 顔画像処理装置 |
Non-Patent Citations (1)
Title |
---|
MIWA S. ET AL: "Rectangle Filter to AdaBoost o Mochiita Kao Ninsho Algorithm", THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS, 8 March 2004 (2004-03-08), pages 220 (D-12-54), XP002998200 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008027275A (ja) * | 2006-07-24 | 2008-02-07 | Seiko Epson Corp | オブジェクト検出装置、オブジェクト検出方法および制御プログラム |
JP2010500687A (ja) * | 2006-08-11 | 2010-01-07 | フォトネーション ビジョン リミテッド | デジタル画像取得装置におけるリアルタイムの顔探知 |
JP2010500836A (ja) * | 2006-08-11 | 2010-01-07 | フォトネーション ビジョン リミテッド | デジタル画像収集装置におけるリアルタイム顔追跡 |
US8422739B2 (en) | 2006-08-11 | 2013-04-16 | DigitalOptics Corporation Europe Limited | Real-time face tracking in a digital image acquisition device |
US8509498B2 (en) | 2006-08-11 | 2013-08-13 | DigitalOptics Corporation Europe Limited | Real-time face tracking in a digital image acquisition device |
US8666124B2 (en) | 2006-08-11 | 2014-03-04 | DigitalOptics Corporation Europe Limited | Real-time face tracking in a digital image acquisition device |
US8463049B2 (en) * | 2007-07-05 | 2013-06-11 | Sony Corporation | Image processing apparatus and image processing method |
JP2009237634A (ja) * | 2008-03-25 | 2009-10-15 | Seiko Epson Corp | オブジェクト検出方法、オブジェクト検出装置、オブジェクト検出プログラムおよび印刷装置 |
JP2011013732A (ja) * | 2009-06-30 | 2011-01-20 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
JP2015036123A (ja) * | 2013-08-09 | 2015-02-23 | 株式会社東芝 | 医用画像処理装置、医用画像処理方法及び分類器トレーニング方法 |
Also Published As
Publication number | Publication date |
---|---|
CN101023446A (zh) | 2007-08-22 |
CN101023446B (zh) | 2010-06-16 |
US20080080744A1 (en) | 2008-04-03 |
JPWO2006030519A1 (ja) | 2008-05-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2006030519A1 (ja) | 顔認証装置及び顔認証方法 | |
KR101923263B1 (ko) | 등록 및 인증을 위한 바이오메트릭 방법 및 시스템 | |
CN102663444B (zh) | 防止账号被盗用的方法和系统 | |
US7970185B2 (en) | Apparatus and methods for capturing a fingerprint | |
US7301564B2 (en) | Systems and methods for processing a digital captured image | |
US8908934B2 (en) | Fingerprint recognition for low computing power applications | |
US7853052B2 (en) | Face identification device | |
JP2007128480A (ja) | 画像認識装置 | |
JP2010049379A (ja) | 指紋画像取得装置、指紋認証装置、指紋画像取得方法及び指紋認証方法 | |
WO2004025565A1 (ja) | 虹彩コード化方法、個人認証方法、虹彩コード登録装置、虹彩認証装置、および虹彩認証プログラム | |
CN1820283A (zh) | 虹彩码生成方法、个人认证方法、虹彩码注册装置、个人认证装置 | |
CN103119623A (zh) | 瞳孔检测装置及瞳孔检测方法 | |
US20210073518A1 (en) | Facial liveness detection | |
CN111339897A (zh) | 活体识别方法、装置、计算机设备和存储介质 | |
JP2009237669A (ja) | 顔認識装置 | |
JP2009211490A (ja) | 画像認識方法および装置 | |
JP5393072B2 (ja) | 掌位置検出装置、掌紋認証装置、携帯電話端末、プログラム、および掌位置検出方法 | |
JP2017138674A (ja) | ナンバープレート認識装置及びナンバープレート認識システム並びにナンバープレート認識方法 | |
JP6229352B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
JP2009282925A (ja) | 虹彩認証支援装置及び虹彩認証支援方法 | |
JP4970385B2 (ja) | 2次元コード読取装置とそのプログラム | |
KR100880073B1 (ko) | 얼굴 인증 장치 및 얼굴 인증 방법 | |
JP2001243465A (ja) | 指紋画像照合方法および指紋画像照合装置 | |
US20120327486A1 (en) | Method and Device of Document Scanning and Portable Electronic Device | |
JP3567260B2 (ja) | 画像データ照合装置、画像データ照合方法、及び画像データ照合処理プログラムを記憶した記憶媒体 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006535003 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11659665 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020077006062 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200480044012.9 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 04822290 Country of ref document: EP Kind code of ref document: A1 |
|
WWP | Wipo information: published in national office |
Ref document number: 11659665 Country of ref document: US |