US20080267458A1 - Face image log creation - Google Patents
Face image log creation Download PDFInfo
- Publication number
- US20080267458A1 US20080267458A1 US11/741,235 US74123507A US2008267458A1 US 20080267458 A1 US20080267458 A1 US 20080267458A1 US 74123507 A US74123507 A US 74123507A US 2008267458 A1 US2008267458 A1 US 2008267458A1
- Authority
- US
- United States
- Prior art keywords
- face
- image
- quality
- criteria
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
Definitions
- the present invention relates to video surveillance systems and in particular to the creation of face image logs from captured video.
- face image logs In video surveillance systems it is often useful to summarize the collected surveillance video of people in the monitored scene by images of the faces visible in the original video sequence.
- the sequence of faces images is referred to as face image logs. Whether reviewed by security personnel, or an automated system; processed in real-time, or upon request; these logs allow investigators to determine who was in the vicinity of the surveillance camera at any particular moment in time without having to view the video sequence itself.
- face image logs need to be complete in the sense that they should contain, at the very least, one high quality image for each individual whose face appeared unobstructed in the original video.
- High quality images are important because they maximize the probability that an individual will be correctly identified.
- the most direct approach to constructing complete face image logs involves using existing face detection technologies to extract face images directly from video sequences, and immediately appending each of these detections to the face log.
- one face may be detected per person per frame.
- Surveillance footage captured at 15 frames per second, could potentially capture 900 face images per person per minute.
- the high rate of detections could easily overwhelm any human operator or automated biometric face recognition system that might be trying to process the face image log in real time. Real-time or not, much of this processing is wasteful since each individual may appear in the log numerous times.
- Face image validation systems have been used for analysis of images of faces to determine if they are suitable for use in identification documents such as passports but have been limited in their applicability to real-time applications. While the technique does provide a numeric quality score for input images, it is mainly for detecting images that do not meet the criteria established by the International Civil Aviation Organization; and their decisions are inherently binary. In addition the face images are acquired in a relatively controlled manner with on a limited range of variation in subject conditions. In contrast, a more continuous appraisal system is required when selecting high quality face images from video sequences.
- FIG. 1 is a system for face tracking and face log creation
- FIG. 2 is an illustrative example for determining pose estimation criteria
- FIG. 3 is an illustrative example for determining illumination criteria
- FIG. 4 is an illustrative example for determining sharpness criteria
- FIG. 5 is an illustrative example for determining skin content criteria
- FIG. 6 is an illustrative example for determining image resolution criteria
- FIG. 7 is a method of face log creation
- FIG. 8 is a method of face detection and tracking
- FIG. 9 is a method of face image quality analysis
- FIG. 10 is a method of face log purging
- FIG. 11 is an illustration of a face log comprising multiple face images of a single subject.
- Embodiments of the present invention are described below, by way of example only, with reference to FIGS. 1-11 .
- a method of generating face logs that is concise and complete from a video sequence, the method comprising the steps of: acquiring frames from the video sequence; determining one or more face images in each frames associated with one or more subjects; determining a quality score for each of the one or more face images based upon a plurality of criteria associated with the face image; and purging face images from the face log which are below a minimum quality score.
- a system for generating concise and complete face logs from a video sequence comprising a face detection module for determining face images of one or more subjects from candidate regions within each frame of the video sequence; a quality analysis module for receiving face images from the face detection module and determining a quality score based a plurality of criteria of the face images; and a storage means for storing the analyzed face images in a face log with an associated quality score, wherein the face log is periodically purged based upon the score of the acquired images.
- a computer readable medium containing instructions for generating concise and complete face logs from a video sequence, the instructions which when executed by a processor perform the steps of acquiring frames from the video sequence; determining one or more face images in each frames associated with one or more subjects; determining a quality score for each of the one or more face images based upon a plurality of criteria associated with the face image; and purging face images from the face log which are below a minimum quality score.
- a concise face log need not contain every instance of every face as it only takes one good quality image to identify an individual. With each image captured of the same individual there may be vast variations in face resolution, pose, illumination and sharpness.
- person oriented face log construction enables only high quality image(s) for each individual to be stored reducing capacity requirements and operator intervention or judgment. To accomplish this, faces need to be tracked in a scene in addition to simply being detected. By doing so, a face image history can be compiled for each individual in the scene.
- one or more high quality images are selected from his or her face history, and these images are added to the face log.
- the number of faces selected for the log depends on the confidence in the face tracking results; more faces are added to the log when the tracker results are poor.
- Constructing face image logs can be divided into four major tasks: face detection, face tracking, face quality analysis, and the selection of high quality face images to be append to the face image log.
- face detection For the purpose of tracking, techniques have been suggested that involve the use of skin color segmentation to locate candidate face regions.
- Other techniques have detected frontal faces using cascades of simple classifiers.
- Some strategies for tracking the resulting face regions include techniques based on the overlapping of bounding boxes, techniques using partial Kalman filtering for motion prediction, and techniques which use the mean shift method for tracking.
- Face image quality corresponds roughly to an image's potential to lead to a correct identification when using existing face recognition software. It is assumed that any image useful for biometric face recognition would also be useful when manually attempting to identify an individual.
- FIG. 1 shows an illustrative example of a system 100 for face tracking and face log creation.
- a scene or location is monitored by a video camera 116 operatively coupled to the video surveillance system 100 .
- the video camera 116 may be for capturing motion video or still images providing of the scene.
- An individual or subject 104 or a group of individuals 102 may enter scene covered by the video camera 116 .
- An image or frame of the scene is captured by the video (or still) camera 116 .
- the images are sent to the video surveillance system 110 for viewing and for further processing.
- the video surveillance system 110 may be associated with individual cameras or multiple cameras covering one or more scenes. Alternatively, the video surveillance system may be used for post processing of stored video surveillance footage from video cameras provide by another system or storage device.
- the system 110 comprises a central processing unit (CPU) for executing software for performing video analysis.
- Programming instructions are retrieved from memory 114 associated with the CPU 112 which intern may be retrieved from a storage device 160 .
- the storage device 160 provides storage for original video or images and processed images.
- the storage device may be any type of memory such as a solid state memory, tape, hard drive or optical based storage mediums or may contain data provided by an external medium 162 such as for example a floppy disk, digital video disc (DVD), compact disk (CD), solid state memory or external hard drive.
- the storage device may also contain the computer programmable code for operating the system.
- Input device 118 may be used by an operator of the video surveillance system for controlling cameras and for operating the surveillance system.
- An output device 120 such as a video display may be used by the operator to view the face logs or original video sequence.
- Face detection and face tracking are executed by the CPU 112 from code stored in the memory 114 .
- the face detection and tracking module 130 may be further divided into skin segmentation 132 and background segmentation 134 for analyzing the video frame and determining where the individuals face is visible and providing face images to the quality analysis module 104 and object tracking 136 for tracking movement of a candidate region between frames within the video sequence.
- the quality analysis module 140 analyzes the face image to assign a quality score to the image and to determine if it is of suitable quality for storage in the face log. Modules for assessing criteria relative to the face image are provided.
- the modules for determining criteria may include but not be limited to pose estimation 142 for determining orientation of the image, illumination quality 144 for determining dynamic range and evenness, sharpness 146 , skin content 148 for determining the amount of human skin in the image and image resolution 150 .
- Each module can determine a score for one or more criteria in addition it should be understood that the functions provided by the modules shown may be combined or divided depending on the particular system configuration.
- the face detection and tracking may be performed by separate dedicated systems or implemented in separate modules.
- the face detection and tracking module 130 may operate independently of quality analysis module 140 and may reside on distinct systems if required. Analysis of the images may be performed in a real-time or non-real-time environment depending on configuration.
- a face image 106 of an individual is provided to the quality and analysis module 140 .
- Six criteria are assessed to determine an overall quality score for the image. Pose estimation in face images
- S 1 represents the quality score awarded to the image for pose. This score has the desirable property that it decreases linearly as the estimated angle of rotation increases. Notice that values of ⁇ greater than 45° are considered to be inaccurate, and result in a score of zero.
- variations caused by changes in illumination constitute yet another significant challenge encountered by automated face recognition systems.
- certain methods of face identification are more sensitive to differences in lighting than they are to the differences between distinct individuals.
- face identification systems may use histogram equalization, or similar histogram dependent techniques, in order to normalize an image before processing. For this reason, it is very important to begin with images which make the best (maximum) use of the available dynamic range.
- This utilization ⁇ is estimated by determining the smallest range of gray intensities to which at least 95% of an image's pixels can be attributed.
- Utilization of available dynamic range is not the only desirable property of properly illuminated face images; faces should also be evenly lit. In other words, one side of the face should not appear brighter than the other.
- the evenness of the illumination can be determined by comparing the histograms of the opposing halves of the face as shown with face image 300 .
- L and R be the grey intensity histograms ( 310 and 320 respectively) attributed to the left 302 and right 304 halves of the face, normalized so that the integrals over their respective bins are both equal to one.
- the score S 3 which represents the evenness of the illumination, is then equal to the integral of the histogram resulting from the intersection L ⁇ R.
- each image may have a level of sharpness associated with the focus of the face and also related to the size of the image.
- image 410 is sharper than image 420 , which is relatively more sharp than image 430 .
- the main assumption in measuring the sharpness of face images is that these images should have comparable power spectra. Consequently, a simple global measure of sharpness defined for a given image represented by the function a(x,y), let A(u,v) be its corresponding frequency domain representation. Given the frequencies f 1 and f 2 , where f 1 ⁇ f 2 , define:
- Sh measures the global sharpness of the image. While this measure provided excelled results, it useful to further refine the procedure by omitting the frequency terms that occur at orientations within 10° of the u or v-axis. This effectively masks out the unnatural step function that occurs at the boundaries of the image as result of the cyclic nature of the Fourier transform.
- the frequencies f 1 and f 2 are chosen so that it would be highly improbable for a natural face to achieve a sharpness score greater than 1.
- an image's official sharpness score is defined as:
- Images of faces are expected to be composed mostly of flesh-toned pixels. Lack of such pixels could indicate that an image's color temperature is not properly balanced, or that the color has been washed out due to overly harsh lighting conditions. In either case, such images should score poorly in overall quality.
- the pixels representing flesh tones for example as shown in portion 530 , are generally tightly clustered in a small region of hue-saturation color space and can be obtained using a region containing all hues between ⁇ 30° and 30° and saturated between 5% and 95%.
- An image's skin score S 4 is computed as the percentage of its pixels that occur within this region. This measure is most useful when skin color segmentation is not already being used for locating candidate face regions, such as when cascades of simple classifiers are used for face detection.
- An image's resolution score is perhaps the easiest of the aforementioned criteria to measure.
- the resolution of a face image is defined to be the area of its bounding rectangle. In general, high resolution images are preferred over low resolution images.
- three face images, 610 , 620 and 630 provide different resolution images and hence varying face recognition quality. For example image 630 is of higher resolution than image 610 . Consequently, the resolution score of an image having dimensions w ⁇ h, is computed as follows assuming an optimal image resolution is 60 ⁇ 60 pixels:
- each of the six criteria discussed in the previous sections can score in the range [0, 1], however any definable range may be utilized. Although each criteria may be ranked in the same range they need not contribute equally to an image's overall quality score. For this reason, they are combined according to the following weighted sum:
- weights can be adjusted based upon the criteria that contribute the most to facial recognition. Table 1 shows an example of how each of the weights and thresholds may be used to combine the six quality criteria into a single overall score.
- a threshold T is associated to each of the individual quality criteria.
- An image is determined to be useful for identification purposes, with respect to the i th image criteria, exactly when S i >T i . For each criteria where this occurs, a value of 1.0 is added to the final score S. In this way, an image's score primarily reflects the number of criteria satisfied, while the original weighted sum is used to break ties between images satisfying an equal number of criteria.
- FIG. 7 A method of face log creation using a image quality score is shown in FIG. 7 .
- the video sequence is acquired.
- potential candidate face image regions must be defined and the region must then be tracked between frames when an individual or a group enters the scene as discussed in further detail in connection with FIG. 8 .
- face tracking algorithm With a reliable face tracking algorithm, at most one face should be detected per person per frame. In these cases, selecting the best face image for each individual is rather trivial; as each frame is processed, simply retain the highest quality face detected for the individual thus far.
- the face tracker's results are ambiguous. In such situations, care must be taken to ensure that all individuals are represented by the face image log.
- a person group (or more simply, a “group”) is introduced to define a set of people whose detected faces can not be distinguished by the face tracker.
- Single individuals can be considered person groups of size one, and do not need special consideration.
- face groups can be constructed in situations where the face tracker has indicated low confidence in its results. In these cases, it is understood that the tracker is following a group of people rather than a single individual.
- a face history is essentially a smaller face image log, containing all faces that were ever detected and associated to the group. As with the global face log, images in the face history should be labeled by the frame in which they originated. A face history is considered to be complete only after the person group has exited the scene.
- the analysis of the candidate images is performed at step 714 , as will be discussed in further detail in connection with FIG. 9 .
- a quality score is assigned to the image.
- the image is then stored to the face history log at step 716 . If multiple face images generated by the group individual scores can be assigned to each image or may be an aggregate score of the images.
- the face images may be periodically purged based upon system constraints such as a number of images (dictated capacity or resource usage of the system or storage devices) or by defining a minimum quality threshold or a combination of factors to reduce the number of face images stored at step 718 .
- step 720 it is then determined if the group has exited the scene. If the group has not exited, NO at step 720 , the method continues at step 712 . If the group has excited, or potentially if the membership of the group has changed, YES at step 720 , the face log is then purged to ensure that only face images of a minimum quality are stored in the face log as discussed in connection with FIG. 10 at step 722 . The face log may then be stored at step 724 and have metadata with the quality score associated with the images or be indexed to the original video sequence for later reference.
- FIG. 8 describes a method of obtaining the face image from a video sequence.
- the image is acquired at step 810 from a video source or from a storage device.
- the image is analyzed to determine skin segmentation visible in the image at step 812 . This can be performed by denoting the location of flesh colored pixels visible in the frame providing a binary image.
- adaptive background segmentation can be performed. This provides a binary image representing the location of foreground objects.
- a binary AND function can be performed to provide a binary image denoting the location of pixels likely representing skin at step 816 .
- the connected components are determined defining a facial image.
- a rectangular region can then be defined providing a region of interest around the face image at step 820 .
- Face detection is then applied to the region at step 822 to determine if the candidate region contains a face image.
- the rectangular region can then be tracked as movement occurs at step 824 within the video sequence and identifiers can be associated by metadata to the face image and the tracked region and the generated face images. It should be understood that alternating tracking and face detection methods may be utilized for generating face images.
- FIG. 9 illustrates a method for generating a quality score for the face image based upon image criteria assessment.
- the face image is retrieved by the quality analysis module 140 .
- a value is then determined for each of the defined criteria.
- a pose estimation value is determined by pose estimation module 142 , as discussed previously, by determining the position and orientation of the face within the frame and assigned a sharpness score.
- the illumination quality of the image is determined by the illumination quality module 144 . Illumination quality can be assessed by two criteria and therefore generates two scores, one for dynamic range and one for evenness within the image.
- the sharpness of the image is then determined at step 916 by module 146 producing a sharpness score.
- the presence of human skin is determined by human skin module 148 with an associated score.
- the image resolution is assigned a score at step 920 by module 150 .
- the score assigned to each criteria is then combined with predefined weights and thresholds, if defined, to generate an overall score at step 922 .
- the score is then associated with the face image at step 924 . It should be understood that the scoring of each criteria may be performed in any order or scored simultaneously and no particular order is required.
- the images can then be discarded or stored based upon the generated score.
- the weights and thresholds utilized may be dependent on aspects of the video system by which the face logs are acquired, such color, resolution, motion content etc as well as other processing aspects related to the tracking and face recognition software.
- FIG. 10 illustrates a method for purging a person group face history which may be called periodically when the group exits the scene.
- a sizeable number of face images may be generated and potentially stored.
- the purging of the face log may be performed by the quality analysis module 140 or by a dedicated process which determines when a group has excited the scene or membership has changed.
- the group size can be estimated as the maximum number of faces ever detected (M) and associated to the group in any single frame. This can be determined by searching a group's face history at the moment that the face history becomes complete.
- the minimum quality score for the associated face images is determined at step 1014 .
- the maximum of the determined scores is determined at step 1016 . This will be the frame with the best minimum quality face (Q). All faces in the group's face history having a quality less than Q are then be discarded at step 1016 , and any faces that remain are added to the global face log.
- the percentage of frames in which the faces of all group members are visible can be determined.
- falsely detected faces can artificially inflate the estimated group size M. Since false positives are likely to score quite poorly in quality, the situation could result in a low quality threshold Q. With a low Q, many faces will be considered of sufficient quality to be added to the global face log.
- a defined percentage value such as for example 10% of all frames in the group's face history contain M (or more) faces.
- the estimated group size can be reduced to a level that satisfies this condition. Extreme caution must be employed in order to avoid underestimating the group size.
- FIG. 11 illustrates a face log for one member of the group with varying quality images.
- the face log 1100 can be ordered in sequence of the quality score or based upon time frame index for storage.
- Each member of the group may be stored in an individual face log or they may be stored as a group or global face log.
- the purge algorithm will ensure that a group of faces is stored in which there is at least one face of each individual in the scene however there might be multiple occurrences of once face.
- the additional metadata associated with the images may be used for viewing or storage.
- Examples of suitable computing system environments or configurations that may be suitable for implementing various embodiments include: a general purpose personal computer (PC); a hand-held or laptop computer; multi-processor based systems; microprocessor based systems; programmable consumer electronics; network computers, minicomputers, mainframe computers and distributed computing environments.
- PC personal computer
- a hand-held or laptop computer multi-processor based systems
- microprocessor based systems programmable consumer electronics
- network computers minicomputers, mainframe computers and distributed computing environments.
- Components of a typical computing device include, but are not limited to, a processing unit, an input/output interface, a system memory, and a system bus.
- the system bus communicatively connects the aforementioned components and numerous other cooperatively interactive components.
- the input/output interface interacts with external components through an input/output unit (which can include keyboard, mouse-type controllers, monitors, media readers/writers and the like).
- the system memory instantiates various components and operations of the face log construction according to embodiments of the present invention.
- the description does not limit the implementation of the embodiments of the present invention to any particular computer programming language.
- the computer program product may be implemented in many computer programming languages provided that the OS (Operating System) provides the facilities that may support the requirements of the computer program product. Any limitations presented would be a result of a particular type of operating system, computer programming language, or database management system and would not be a limitation of the embodiments of the present invention described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
Description
- The present invention relates to video surveillance systems and in particular to the creation of face image logs from captured video.
- In video surveillance systems it is often useful to summarize the collected surveillance video of people in the monitored scene by images of the faces visible in the original video sequence. The sequence of faces images is referred to as face image logs. Whether reviewed by security personnel, or an automated system; processed in real-time, or upon request; these logs allow investigators to determine who was in the vicinity of the surveillance camera at any particular moment in time without having to view the video sequence itself.
- In general, face image logs need to be complete in the sense that they should contain, at the very least, one high quality image for each individual whose face appeared unobstructed in the original video. High quality images are important because they maximize the probability that an individual will be correctly identified.
- The most direct approach to constructing complete face image logs involves using existing face detection technologies to extract face images directly from video sequences, and immediately appending each of these detections to the face log. In this scenario, one face may be detected per person per frame. Surveillance footage, captured at 15 frames per second, could potentially capture 900 face images per person per minute. The high rate of detections could easily overwhelm any human operator or automated biometric face recognition system that might be trying to process the face image log in real time. Real-time or not, much of this processing is wasteful since each individual may appear in the log numerous times.
- Face image validation systems have been used for analysis of images of faces to determine if they are suitable for use in identification documents such as passports but have been limited in their applicability to real-time applications. While the technique does provide a numeric quality score for input images, it is mainly for detecting images that do not meet the criteria established by the International Civil Aviation Organization; and their decisions are inherently binary. In addition the face images are acquired in a relatively controlled manner with on a limited range of variation in subject conditions. In contrast, a more continuous appraisal system is required when selecting high quality face images from video sequences.
- Accordingly, there is a need for quality appraisal of face images for the purpose of selecting high quality faces from video sequences.
- Further features and advantages of the present invention will become apparent from the following detailed description, taken in combination with the appended drawings, in which:
-
FIG. 1 is a system for face tracking and face log creation; -
FIG. 2 is an illustrative example for determining pose estimation criteria; -
FIG. 3 is an illustrative example for determining illumination criteria; -
FIG. 4 is an illustrative example for determining sharpness criteria; -
FIG. 5 is an illustrative example for determining skin content criteria; -
FIG. 6 is an illustrative example for determining image resolution criteria; -
FIG. 7 is a method of face log creation; -
FIG. 8 is a method of face detection and tracking; -
FIG. 9 is a method of face image quality analysis; -
FIG. 10 is a method of face log purging; and -
FIG. 11 is an illustration of a face log comprising multiple face images of a single subject. - It will be noted that throughout the appended drawings, like features are identified by like reference numerals.
- Embodiments of the present invention are described below, by way of example only, with reference to
FIGS. 1-11 . - In accordance with one aspect there is provided a method of generating face logs that is concise and complete from a video sequence, the method comprising the steps of: acquiring frames from the video sequence; determining one or more face images in each frames associated with one or more subjects; determining a quality score for each of the one or more face images based upon a plurality of criteria associated with the face image; and purging face images from the face log which are below a minimum quality score.
- In accordance with another aspect there is provided a system for generating concise and complete face logs from a video sequence, the system comprising a face detection module for determining face images of one or more subjects from candidate regions within each frame of the video sequence; a quality analysis module for receiving face images from the face detection module and determining a quality score based a plurality of criteria of the face images; and a storage means for storing the analyzed face images in a face log with an associated quality score, wherein the face log is periodically purged based upon the score of the acquired images.
- In accordance with yet another aspect there is provided a computer readable medium containing instructions for generating concise and complete face logs from a video sequence, the instructions which when executed by a processor perform the steps of acquiring frames from the video sequence; determining one or more face images in each frames associated with one or more subjects; determining a quality score for each of the one or more face images based upon a plurality of criteria associated with the face image; and purging face images from the face log which are below a minimum quality score.
- The creation of face image logs from video sequences should be concise as possible to ensure adequate images are provided to identify an individual yet not unnecessarily consume processing or storage resources by providing excessive numbers of images. A concise face log need not contain every instance of every face as it only takes one good quality image to identify an individual. With each image captured of the same individual there may be vast variations in face resolution, pose, illumination and sharpness. In order to reduce the number of faces added to a face log, person oriented face log construction enables only high quality image(s) for each individual to be stored reducing capacity requirements and operator intervention or judgment. To accomplish this, faces need to be tracked in a scene in addition to simply being detected. By doing so, a face image history can be compiled for each individual in the scene. When an individual leaves the scene, one or more high quality images are selected from his or her face history, and these images are added to the face log. The number of faces selected for the log depends on the confidence in the face tracking results; more faces are added to the log when the tracker results are poor. By using person oriented face logs, it is possible to avoid overwhelming face recognition systems, when attempting to process these logs in real-time.
- There are many uncertainties when developing a person oriented face logging system. For instance, it is not obvious how the quality of face images should be assessed and compared so that high quality images can be selected. It is also unclear how the face selection process should proceed in situations where there is low confidence in the tracker results.
- Constructing face image logs can be divided into four major tasks: face detection, face tracking, face quality analysis, and the selection of high quality face images to be append to the face image log. Regarding the detection of faces for the purpose of tracking, techniques have been suggested that involve the use of skin color segmentation to locate candidate face regions. Other techniques have detected frontal faces using cascades of simple classifiers. Some strategies for tracking the resulting face regions include techniques based on the overlapping of bounding boxes, techniques using partial Kalman filtering for motion prediction, and techniques which use the mean shift method for tracking.
- In order to select high quality face images, it is necessary to develop a procedure for their appraisal. Of course, the quality of a face image is rather subjective. Face image quality corresponds roughly to an image's potential to lead to a correct identification when using existing face recognition software. It is assumed that any image useful for biometric face recognition would also be useful when manually attempting to identify an individual.
- Many criteria, weighted to varying degrees of importance, are considered when determining the quality of a face image. These criteria include image characteristics such as resolution, pose, illumination, skin content and sharpness. These criteria can be measured to determine how they contribute to the overall quality score of an image. With the exception of skin detection, each of the following scoring procedures expects a grayscale image of an upright face as input. It is important to note that, depending on the method by which faces are detected and tracked, many of the intermediate results used for assessing quality may be readily available. For example, the location of the eyes, the contour of the face, and the location of skin colored pixels may have been computed when the face tracker was locating candidate face regions. Additionally, it is also important to recognize that the scoring procedures do not directly distinguish between face and non-face images (except for in certain extreme situations) since this is precisely what the face tracker is expected to provide. In general, however, high quality face images out perform non-face images.
-
FIG. 1 shows an illustrative example of asystem 100 for face tracking and face log creation. A scene or location is monitored by avideo camera 116 operatively coupled to thevideo surveillance system 100. Thevideo camera 116 may be for capturing motion video or still images providing of the scene. An individual or subject 104 or a group ofindividuals 102 may enter scene covered by thevideo camera 116. An image or frame of the scene is captured by the video (or still)camera 116. The images are sent to thevideo surveillance system 110 for viewing and for further processing. Thevideo surveillance system 110 may be associated with individual cameras or multiple cameras covering one or more scenes. Alternatively, the video surveillance system may be used for post processing of stored video surveillance footage from video cameras provide by another system or storage device. - The
system 110 comprises a central processing unit (CPU) for executing software for performing video analysis. Programming instructions are retrieved frommemory 114 associated with theCPU 112 which intern may be retrieved from astorage device 160. Thestorage device 160 provides storage for original video or images and processed images. The storage device may be any type of memory such as a solid state memory, tape, hard drive or optical based storage mediums or may contain data provided by anexternal medium 162 such as for example a floppy disk, digital video disc (DVD), compact disk (CD), solid state memory or external hard drive. The storage device may also contain the computer programmable code for operating the system.Input device 118 may be used by an operator of the video surveillance system for controlling cameras and for operating the surveillance system. Anoutput device 120 such as a video display may be used by the operator to view the face logs or original video sequence. - Face detection and face tracking are executed by the
CPU 112 from code stored in thememory 114. The face detection andtracking module 130 may be further divided intoskin segmentation 132 andbackground segmentation 134 for analyzing the video frame and determining where the individuals face is visible and providing face images to thequality analysis module 104 and object tracking 136 for tracking movement of a candidate region between frames within the video sequence. Thequality analysis module 140 analyzes the face image to assign a quality score to the image and to determine if it is of suitable quality for storage in the face log. Modules for assessing criteria relative to the face image are provided. The modules for determining criteria may include but not be limited to poseestimation 142 for determining orientation of the image,illumination quality 144 for determining dynamic range and evenness,sharpness 146,skin content 148 for determining the amount of human skin in the image andimage resolution 150. Each module can determine a score for one or more criteria in addition it should be understood that the functions provided by the modules shown may be combined or divided depending on the particular system configuration. For example, the face detection and tracking may be performed by separate dedicated systems or implemented in separate modules. Similarly the face detection andtracking module 130 may operate independently ofquality analysis module 140 and may reside on distinct systems if required. Analysis of the images may be performed in a real-time or non-real-time environment depending on configuration. - Through the face detection and
tracking module 130, aface image 106 of an individual is provided to the quality andanalysis module 140. Six criteria are assessed to determine an overall quality score for the image. Pose estimation in face images - One of the most challenging problems encountered by face recognition systems involves properly identifying individuals despite variations in pose. Generally, faces can experience out-of-plane rotations of up to 60° before important facial features are no longer visible. As usual, an out-of-plane rotation of 0° occurs when an individual looks directly at the camera, providing a perfectly frontal view of the face. It is expected that a robust face detector should return face images whose pose falls within the slightly narrower range of 45°. Consequently, it is important that the quality score be able to distinguish between various rotations, and award higher scores to the least rotated images.
- In order to estimate pose, as shown in
FIG. 2 , using aface image 200, three columns are located on the face image: the first two columns, defined by x=l (210) and x=r (214), estimate the locations of the left and right visible edges of the face, respectively. The third column, x=c (212), approximates the face's axis of natural horizontal symmetry. Importantly, the face image is symmetric about the line x=c, only when the face is not rotated. The values of l, r and c are determined by analysis of the gradient image in order to locate the left and right sides of the face, as well as the vertical location of the eyes, y=m (222) relative to the top y=t (220), bottom y=b (224). From these values the approximate location of the eyes can be estimated, and the brightest point in this region is expected to represent the vertical center of the face. Unfortunately, this method is not effective when subjects are wearing glasses. - If the face has not experienced an out-of-plane rotation, and the values l, r, and c are accurate, then it is expected that c is equidistant from l and r. As a face experiences rotation, c deviates from it is expected position c*. The following equation can be used to estimate the angle of rotation:
-
- This is not a particularly good estimator because it models the human head as a cylinder. However, high accuracy is not needed since a it will not be used in any computations other than the following equation:
-
- where S1 represents the quality score awarded to the image for pose. This score has the desirable property that it decreases linearly as the estimated angle of rotation increases. Notice that values of θ greater than 45° are considered to be inaccurate, and result in a score of zero.
- As shown in
FIG. 3 , variations caused by changes in illumination constitute yet another significant challenge encountered by automated face recognition systems. In fact, certain methods of face identification are more sensitive to differences in lighting than they are to the differences between distinct individuals. In order to compensate for different lighting conditions, face identification systems may use histogram equalization, or similar histogram dependent techniques, in order to normalize an image before processing. For this reason, it is very important to begin with images which make the best (maximum) use of the available dynamic range. This utilization ∪ is estimated by determining the smallest range of gray intensities to which at least 95% of an image's pixels can be attributed. The score S2 is simply the percentage of the total dynamic range represented by ∪. For example, if the input is an 8-bit grayscale image, then S2=∪/256. - Utilization of available dynamic range is not the only desirable property of properly illuminated face images; faces should also be evenly lit. In other words, one side of the face should not appear brighter than the other. Using measurements obtained when estimating pose, the evenness of the illumination can be determined by comparing the histograms of the opposing halves of the face as shown with
face image 300. Let L and R be the grey intensity histograms (310 and 320 respectively) attributed to the left 302 and right 304 halves of the face, normalized so that the integrals over their respective bins are both equal to one. The score S3, which represents the evenness of the illumination, is then equal to the integral of the histogram resulting from the intersection L∩R. - It is important that a face image's score reflect its sharpness. Images marred by motion blur, for example, should score quite poorly. As shown in
FIG. 4 each image may have a level of sharpness associated with the focus of the face and also related to the size of the image. Forexample image 410 is sharper thanimage 420, which is relatively more sharp thanimage 430. The main assumption in measuring the sharpness of face images is that these images should have comparable power spectra. Consequently, a simple global measure of sharpness defined for a given image represented by the function a(x,y), let A(u,v) be its corresponding frequency domain representation. Given the frequencies f1 and f2, where f1<f2, define: -
- where Sh measures the global sharpness of the image. While this measure provided excelled results, it useful to further refine the procedure by omitting the frequency terms that occur at orientations within 10° of the u or v-axis. This effectively masks out the unnatural step function that occurs at the boundaries of the image as result of the cyclic nature of the Fourier transform. In practice, the frequencies f1 and f2 are chosen so that it would be highly improbable for a natural face to achieve a sharpness score greater than 1. Thus, an image's official sharpness score is defined as:
-
S4=min{1, Sh} - Images of faces are expected to be composed mostly of flesh-toned pixels. Lack of such pixels could indicate that an image's color temperature is not properly balanced, or that the color has been washed out due to overly harsh lighting conditions. In either case, such images should score poorly in overall quality. As shown in
FIG. 5 , withface image 500, the pixels representing flesh tones, for example as shown inportion 530, are generally tightly clustered in a small region of hue-saturation color space and can be obtained using a region containing all hues between −30° and 30° and saturated between 5% and 95%. An image's skin score S4 is computed as the percentage of its pixels that occur within this region. This measure is most useful when skin color segmentation is not already being used for locating candidate face regions, such as when cascades of simple classifiers are used for face detection. - An image's resolution score is perhaps the easiest of the aforementioned criteria to measure. The resolution of a face image is defined to be the area of its bounding rectangle. In general, high resolution images are preferred over low resolution images. As shown in
FIG. 6 , three face images, 610, 620 and 630 provide different resolution images and hence varying face recognition quality. Forexample image 630 is of higher resolution thanimage 610. Consequently, the resolution score of an image having dimensions w×h, is computed as follows assuming an optimal image resolution is 60×60 pixels: -
- Each of the six criteria discussed in the previous sections can score in the range [0, 1], however any definable range may be utilized. Although each criteria may be ranked in the same range they need not contribute equally to an image's overall quality score. For this reason, they are combined according to the following weighted sum:
-
- where the coefficients Wi determine the impact each of the quality criteria have on the final score. Weights can be adjusted based upon the criteria that contribute the most to facial recognition. Table 1 shows an example of how each of the weights and thresholds may be used to combine the six quality criteria into a single overall score.
-
i Weight (Wi) Threshold (Ti) 1 (Pose estimation) 2 0.8 2 (Dynamic range) 1 0.2 3 (Evenness) 1 0.4 4 (Sharpness) 70/3 0.13 5 (Skin content) 10/7 0.4 6 (Image resolution) 2 0.5 - When using a weighted sum alone, it becomes difficult to develop a single set of weights that works both when an image scores moderately well in all criteria, as well as when an image scores in the extremes of one or more criteria. To address this issue, a threshold T; is associated to each of the individual quality criteria.
- An image is determined to be useful for identification purposes, with respect to the ith image criteria, exactly when Si>Ti. For each criteria where this occurs, a value of 1.0 is added to the final score S. In this way, an image's score primarily reflects the number of criteria satisfied, while the original weighted sum is used to break ties between images satisfying an equal number of criteria.
- A method of face log creation using a image quality score is shown in
FIG. 7 . Atstep 710 the video sequence is acquired. At step 712 from individual frames of the video sequence potential candidate face image regions must be defined and the region must then be tracked between frames when an individual or a group enters the scene as discussed in further detail in connection withFIG. 8 . With a reliable face tracking algorithm, at most one face should be detected per person per frame. In these cases, selecting the best face image for each individual is rather trivial; as each frame is processed, simply retain the highest quality face detected for the individual thus far. Unfortunately, there are many real world situations in which the face tracker's results are ambiguous. In such situations, care must be taken to ensure that all individuals are represented by the face image log. - The concept of a person group (or more simply, a “group”) is introduced to define a set of people whose detected faces can not be distinguished by the face tracker. Single individuals can be considered person groups of size one, and do not need special consideration. More generally, face groups can be constructed in situations where the face tracker has indicated low confidence in its results. In these cases, it is understood that the tracker is following a group of people rather than a single individual.
- For the purpose of face log construction, it is assumed that a group's membership remains constant throughout its existence. This assumption can be justified if a group is assigned a new identity whenever its membership is suspected to have changed. Sudden and significant changes in the dimensions of a group's bounding box might indicate that such an event as recently occurred. So too can persistent changes in the number of faces associated to the group on a frame by frame basis. Alternatively, some trackers may be able to detect merge and split events between moving objects, and these cues could be useful for detecting changes in group membership.
- Additionally, the system is expected to maintain a face history for each person group observed by the face tracker. A face history is essentially a smaller face image log, containing all faces that were ever detected and associated to the group. As with the global face log, images in the face history should be labeled by the frame in which they originated. A face history is considered to be complete only after the person group has exited the scene.
- Once the candidate face images has been acquired at step 712 based upon object tracking within the video sequence, the analysis of the candidate images is performed at step 714, as will be discussed in further detail in connection with
FIG. 9 . Based upon the criteria associated with the image a quality score is assigned to the image. The image is then stored to the face history log atstep 716. If multiple face images generated by the group individual scores can be assigned to each image or may be an aggregate score of the images. The face images may be periodically purged based upon system constraints such as a number of images (dictated capacity or resource usage of the system or storage devices) or by defining a minimum quality threshold or a combination of factors to reduce the number of face images stored atstep 718. Atstep 720 it is then determined if the group has exited the scene. If the group has not exited, NO atstep 720, the method continues at step 712. If the group has excited, or potentially if the membership of the group has changed, YES atstep 720, the face log is then purged to ensure that only face images of a minimum quality are stored in the face log as discussed in connection withFIG. 10 at step 722. The face log may then be stored atstep 724 and have metadata with the quality score associated with the images or be indexed to the original video sequence for later reference. -
FIG. 8 describes a method of obtaining the face image from a video sequence. The image is acquired atstep 810 from a video source or from a storage device. The image is analyzed to determine skin segmentation visible in the image atstep 812. This can be performed by denoting the location of flesh colored pixels visible in the frame providing a binary image. Atstep 814 adaptive background segmentation can be performed. This provides a binary image representing the location of foreground objects. A binary AND function can be performed to provide a binary image denoting the location of pixels likely representing skin atstep 816. Atstep 818 the connected components are determined defining a facial image. A rectangular region can then be defined providing a region of interest around the face image atstep 820. Face detection is then applied to the region atstep 822 to determine if the candidate region contains a face image. The rectangular region can then be tracked as movement occurs atstep 824 within the video sequence and identifiers can be associated by metadata to the face image and the tracked region and the generated face images. It should be understood that alternating tracking and face detection methods may be utilized for generating face images. -
FIG. 9 illustrates a method for generating a quality score for the face image based upon image criteria assessment. Atstep 910 the face image is retrieved by thequality analysis module 140. A value is then determined for each of the defined criteria. At step 912 a pose estimation value is determined bypose estimation module 142, as discussed previously, by determining the position and orientation of the face within the frame and assigned a sharpness score. Atstep 914 the illumination quality of the image is determined by theillumination quality module 144. Illumination quality can be assessed by two criteria and therefore generates two scores, one for dynamic range and one for evenness within the image. The sharpness of the image is then determined atstep 916 bymodule 146 producing a sharpness score. Atstep 918 the presence of human skin is determined byhuman skin module 148 with an associated score. Finally the image resolution is assigned a score atstep 920 bymodule 150. The score assigned to each criteria is then combined with predefined weights and thresholds, if defined, to generate an overall score atstep 922. The score is then associated with the face image atstep 924. It should be understood that the scoring of each criteria may be performed in any order or scored simultaneously and no particular order is required. The images can then be discarded or stored based upon the generated score. The weights and thresholds utilized may be dependent on aspects of the video system by which the face logs are acquired, such color, resolution, motion content etc as well as other processing aspects related to the tracking and face recognition software. -
FIG. 10 illustrates a method for purging a person group face history which may be called periodically when the group exits the scene. In any scene a sizeable number of face images may be generated and potentially stored. The purging of the face log may be performed by thequality analysis module 140 or by a dedicated process which determines when a group has excited the scene or membership has changed. In order to estimate a group's size atstep 1010, it is assumed that some frames exists in which the faces of all group members are clearly visible. Consequently, the group size can be estimated as the maximum number of faces ever detected (M) and associated to the group in any single frame. This can be determined by searching a group's face history at the moment that the face history becomes complete. - Once a group's size can be estimated, it becomes possible to identify the historical frames in which all M group members were detected at
step 1012. From this subset of frames, one frame will best represent the group. For each of the frames the minimum quality score for the associated face images is determined atstep 1014. Of the minimum quality scores, the maximum of the determined scores is determined atstep 1016. This will be the frame with the best minimum quality face (Q). All faces in the group's face history having a quality less than Q are then be discarded atstep 1016, and any faces that remain are added to the global face log. - By using M, the percentage of frames in which the faces of all group members are visible can be determined. Unfortunately, falsely detected faces can artificially inflate the estimated group size M. Since false positives are likely to score quite poorly in quality, the situation could result in a low quality threshold Q. With a low Q, many faces will be considered of sufficient quality to be added to the global face log. One might suspect that the group size has been overestimated when fewer than a defined percentage value such as for example 10% of all frames in the group's face history contain M (or more) faces. In these situations the estimated group size can be reduced to a level that satisfies this condition. Extreme caution must be employed in order to avoid underestimating the group size.
- Another main concern is that a group's face history continues to grow until the group leaves the scene. If the group remains onscreen for any extended period of time, storing the face history could become prohibitively expensive. Thus, low quality face images may be removed from the face history whenever its memory requirements or number of images approaches a predetermined limit. Alternatively, low quality images may be purged on regular periodic intervals. In either case, the purge operation proceeds by using the same method described above, exactly as if the group has recently exited the scene. Care must be taken to ensure that the faces which are purged are still represented in any statistics used to estimate of group size. Additionally, images from frames containing more than M faces are never discarded. This ensures that faces from these frames will be available in case the estimate of group size is increased in future invocations of the algorithm. By definition, fewer than 10% of all frames will contain faces that are kept for this precautionary measure.
- It should be understood that although the group concept is discussed, the algorithm can easily be modified to discard face images when tracking a single individual (M=1) within the scene. Each visible face within a scene may be treated independently and not utilize the group concept.
-
FIG. 11 illustrates a face log for one member of the group with varying quality images. Theface log 1100 can be ordered in sequence of the quality score or based upon time frame index for storage. Each member of the group may be stored in an individual face log or they may be stored as a group or global face log. In a scenario where one individual is tracked over time you will have a single face log with the individual. In group scenario where the tracker can not unambiguously track individual with confidence the purge algorithm will ensure that a group of faces is stored in which there is at least one face of each individual in the scene however there might be multiple occurrences of once face. The additional metadata associated with the images may be used for viewing or storage. Implementing the quality score assessment enables the overall size of the face log to be considerably reduced compared to the potential face images that could be contained in an unedited log. Once the face log is created and is associated with an individual further purging or reduction of the face log can be performed by removing lower quality images as required therefore only keeping the best quality images. - Examples of suitable computing system environments or configurations that may be suitable for implementing various embodiments include: a general purpose personal computer (PC); a hand-held or laptop computer; multi-processor based systems; microprocessor based systems; programmable consumer electronics; network computers, minicomputers, mainframe computers and distributed computing environments.
- Components of a typical computing device include, but are not limited to, a processing unit, an input/output interface, a system memory, and a system bus. The system bus communicatively connects the aforementioned components and numerous other cooperatively interactive components. The input/output interface interacts with external components through an input/output unit (which can include keyboard, mouse-type controllers, monitors, media readers/writers and the like). The system memory instantiates various components and operations of the face log construction according to embodiments of the present invention.
- The description does not limit the implementation of the embodiments of the present invention to any particular computer programming language. The computer program product may be implemented in many computer programming languages provided that the OS (Operating System) provides the facilities that may support the requirements of the computer program product. Any limitations presented would be a result of a particular type of operating system, computer programming language, or database management system and would not be a limitation of the embodiments of the present invention described herein.
- The embodiments of the invention described above are intended to be illustrative only. The scope of the invention is therefore intended to be limited solely by the scope of the appended claims.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/741,235 US8139817B2 (en) | 2007-04-27 | 2007-04-27 | Face image log creation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/741,235 US8139817B2 (en) | 2007-04-27 | 2007-04-27 | Face image log creation |
Publications (2)
Publication Number | Publication Date |
---|---|
US20080267458A1 true US20080267458A1 (en) | 2008-10-30 |
US8139817B2 US8139817B2 (en) | 2012-03-20 |
Family
ID=39887020
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/741,235 Expired - Fee Related US8139817B2 (en) | 2007-04-27 | 2007-04-27 | Face image log creation |
Country Status (1)
Country | Link |
---|---|
US (1) | US8139817B2 (en) |
Cited By (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070098220A1 (en) * | 2005-10-31 | 2007-05-03 | Maurizio Pilu | Method of triggering a detector to detect a moving feature within a video stream |
US20110052006A1 (en) * | 2009-08-13 | 2011-03-03 | Primesense Ltd. | Extraction of skeletons from 3d maps |
US20120070070A1 (en) * | 2010-09-16 | 2012-03-22 | Primesense Ltd. | Learning-based pose estimation from depth maps |
US20130004025A1 (en) * | 2011-06-30 | 2013-01-03 | Nokia Corporation | Method and apparatus for face tracking utilizing integral gradient projections |
US8477998B1 (en) * | 2008-06-20 | 2013-07-02 | Google Inc. | Object tracking in video with visual constraints |
US20130243274A1 (en) * | 2012-03-15 | 2013-09-19 | Hiroshi Sukegawa | Person Image Processing Apparatus and Person Image Processing Method |
US20130266181A1 (en) * | 2012-04-09 | 2013-10-10 | Objectvideo, Inc. | Object tracking and best shot detection system |
US8594425B2 (en) | 2010-05-31 | 2013-11-26 | Primesense Ltd. | Analysis of three-dimensional scenes |
US20140016860A1 (en) * | 2010-06-07 | 2014-01-16 | Affectiva, Inc. | Facial analysis to detect asymmetric expressions |
US20140085499A1 (en) * | 2012-09-25 | 2014-03-27 | Samsung Electronics Co. Ltd. | Method and apparatus for photographing in portable terminal |
US8787663B2 (en) | 2010-03-01 | 2014-07-22 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US8990580B2 (en) | 2012-04-26 | 2015-03-24 | Google Inc. | Automatic user swap |
US9002099B2 (en) | 2011-09-11 | 2015-04-07 | Apple Inc. | Learning-based estimation of hand and finger pose |
US9019267B2 (en) | 2012-10-30 | 2015-04-28 | Apple Inc. | Depth mapping with enhanced resolution |
US9047507B2 (en) | 2012-05-02 | 2015-06-02 | Apple Inc. | Upper-body skeleton extraction from depth maps |
US20150161169A1 (en) * | 2008-10-03 | 2015-06-11 | Intellectual Ventures Fund 83 Llc | Interactive image selection method |
US9077891B1 (en) * | 2013-03-06 | 2015-07-07 | Amazon Technologies, Inc. | Depth determination using camera focus |
US20160086052A1 (en) * | 2014-09-19 | 2016-03-24 | Brain Corporation | Apparatus and methods for saliency detection based on color occurrence analysis |
US9311530B1 (en) * | 2012-05-15 | 2016-04-12 | Google Inc. | Summarizing a photo album in a social network system |
US20160371565A1 (en) * | 2010-03-01 | 2016-12-22 | Microsoft Technology Licensing, Llc | Ranking Based on Facial Image Analysis |
US20170061609A1 (en) * | 2015-09-02 | 2017-03-02 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
US9713982B2 (en) | 2014-05-22 | 2017-07-25 | Brain Corporation | Apparatus and methods for robotic operation using video imagery |
US20170344858A1 (en) * | 2016-05-30 | 2017-11-30 | Kabushiki Kaisha Toshiba | Control device, control method, and computer program product |
US9848112B2 (en) | 2014-07-01 | 2017-12-19 | Brain Corporation | Optical detection apparatus and methods |
US9858295B2 (en) | 2014-06-24 | 2018-01-02 | Google Llc | Ranking and selecting images for display from a set of images |
US9939253B2 (en) | 2014-05-22 | 2018-04-10 | Brain Corporation | Apparatus and methods for distance estimation using multiple image sensors |
US10043279B1 (en) | 2015-12-07 | 2018-08-07 | Apple Inc. | Robust detection and classification of body parts in a depth map |
US10057593B2 (en) | 2014-07-08 | 2018-08-21 | Brain Corporation | Apparatus and methods for distance estimation using stereo imagery |
US10194163B2 (en) | 2014-05-22 | 2019-01-29 | Brain Corporation | Apparatus and methods for real time estimation of differential motion in live video |
US10197664B2 (en) | 2015-07-20 | 2019-02-05 | Brain Corporation | Apparatus and methods for detection of objects using broadband signals |
US20190172494A1 (en) * | 2013-05-28 | 2019-06-06 | Apple Inc. | Reference and non-reference video quality evaluation |
US10366278B2 (en) | 2016-09-20 | 2019-07-30 | Apple Inc. | Curvature-based face detector |
US20190318151A1 (en) * | 2018-04-13 | 2019-10-17 | Omron Corporation | Image analysis apparatus, method, and program |
WO2020094088A1 (en) * | 2018-11-07 | 2020-05-14 | 杭州海康威视数字技术股份有限公司 | Image capturing method, monitoring camera, and monitoring system |
US11042725B2 (en) * | 2016-12-16 | 2021-06-22 | Keylemon Sa | Method for selecting frames used in face processing |
US11042775B1 (en) | 2013-02-08 | 2021-06-22 | Brain Corporation | Apparatus and methods for temporal proximity detection |
US11087798B2 (en) * | 2019-04-16 | 2021-08-10 | Honda Motor Co., Ltd. | Selective curation of user recordings |
EP3818497A4 (en) * | 2018-08-22 | 2021-08-25 | Zhejiang Dahua Technology Co., Ltd. | SYSTEMS AND PROCEDURES FOR SELECTING THE BEST FACIAL IMAGE OF A HUMAN TARGET AREA |
US20210342577A1 (en) * | 2018-10-16 | 2021-11-04 | University Of Seoul Industry Cooperation Foundation | Face recognition method and face recognition device |
US11881053B2 (en) * | 2021-05-18 | 2024-01-23 | Fortinet, Inc. | Systems and methods for hierarchical facial image clustering |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8433106B2 (en) * | 2008-07-08 | 2013-04-30 | Hewlett-Packard Development Company, L.P. | Increasing face detection speed |
WO2011041903A1 (en) | 2009-10-07 | 2011-04-14 | Telewatch Inc. | Video analytics with pre-processing at the source end |
US20110113461A1 (en) * | 2009-10-07 | 2011-05-12 | Robert Laganiere | Video analytics based control of video data storage |
CA2716705A1 (en) * | 2009-10-07 | 2011-04-07 | Telewatch Inc. | Broker mediated video analytics method and system |
WO2011041904A1 (en) | 2009-10-07 | 2011-04-14 | Telewatch Inc. | Video analytics method and system |
US9143739B2 (en) | 2010-05-07 | 2015-09-22 | Iwatchlife, Inc. | Video analytics with burst-like transmission of video data |
US8780162B2 (en) | 2010-08-04 | 2014-07-15 | Iwatchlife Inc. | Method and system for locating an individual |
US8860771B2 (en) | 2010-08-04 | 2014-10-14 | Iwatchlife, Inc. | Method and system for making video calls |
US8885007B2 (en) | 2010-08-04 | 2014-11-11 | Iwatchlife, Inc. | Method and system for initiating communication via a communication network |
US9167290B2 (en) * | 2010-12-17 | 2015-10-20 | Microsoft Technology Licensing, Llc | City scene video sharing on digital maps |
CA2822217A1 (en) | 2012-08-02 | 2014-02-02 | Iwatchlife Inc. | Method and system for anonymous video analytics processing |
US8811673B1 (en) * | 2013-04-18 | 2014-08-19 | TCL Research America Inc. | Intelligent TV system and method |
JP2018136731A (en) * | 2017-02-22 | 2018-08-30 | 東芝テック株式会社 | Information processing apparatus and program |
US10671887B2 (en) | 2017-12-05 | 2020-06-02 | Axis Ab | Best image crop selection |
US10915606B2 (en) | 2018-07-17 | 2021-02-09 | Grupiks Llc | Audiovisual media composition system and method |
KR20240086103A (en) * | 2022-12-09 | 2024-06-18 | 한국전자통신연구원 | Method and apparatus for detecting best-shot in long-range face recognition system |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050271251A1 (en) * | 2004-03-16 | 2005-12-08 | Russell Stephen G | Method for automatically reducing stored data in a surveillance system |
US7039221B1 (en) * | 1999-04-09 | 2006-05-02 | Tumey David M | Facial image verification utilizing smart-card with integrated video camera |
US20060204058A1 (en) * | 2004-12-07 | 2006-09-14 | Kim Do-Hyung | User recognition system and method thereof |
US20070097112A1 (en) * | 2005-10-31 | 2007-05-03 | Hewlett-Packard Development Company, L.P. | Method of tracking an object in a video stream |
US20070269082A1 (en) * | 2004-08-31 | 2007-11-22 | Matsushita Electric Industrial Co., Ltd. | Surveillance Recorder and Its Method |
US20080089561A1 (en) * | 2006-10-11 | 2008-04-17 | Tong Zhang | Face-based image clustering |
-
2007
- 2007-04-27 US US11/741,235 patent/US8139817B2/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7039221B1 (en) * | 1999-04-09 | 2006-05-02 | Tumey David M | Facial image verification utilizing smart-card with integrated video camera |
US20050271251A1 (en) * | 2004-03-16 | 2005-12-08 | Russell Stephen G | Method for automatically reducing stored data in a surveillance system |
US20070269082A1 (en) * | 2004-08-31 | 2007-11-22 | Matsushita Electric Industrial Co., Ltd. | Surveillance Recorder and Its Method |
US20060204058A1 (en) * | 2004-12-07 | 2006-09-14 | Kim Do-Hyung | User recognition system and method thereof |
US20070097112A1 (en) * | 2005-10-31 | 2007-05-03 | Hewlett-Packard Development Company, L.P. | Method of tracking an object in a video stream |
US20080089561A1 (en) * | 2006-10-11 | 2008-04-17 | Tong Zhang | Face-based image clustering |
Cited By (71)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8184853B2 (en) * | 2005-10-31 | 2012-05-22 | Hewlett-Packard Development Company, L.P. | Method of triggering a detector to detect a moving feature within a video stream |
US20070098220A1 (en) * | 2005-10-31 | 2007-05-03 | Maurizio Pilu | Method of triggering a detector to detect a moving feature within a video stream |
US8477998B1 (en) * | 2008-06-20 | 2013-07-02 | Google Inc. | Object tracking in video with visual constraints |
US9798741B2 (en) * | 2008-10-03 | 2017-10-24 | Monument Peak Ventures, Llc | Interactive image selection method |
US20150161169A1 (en) * | 2008-10-03 | 2015-06-11 | Intellectual Ventures Fund 83 Llc | Interactive image selection method |
US20110052006A1 (en) * | 2009-08-13 | 2011-03-03 | Primesense Ltd. | Extraction of skeletons from 3d maps |
US8565479B2 (en) | 2009-08-13 | 2013-10-22 | Primesense Ltd. | Extraction of skeletons from 3D maps |
US10296811B2 (en) * | 2010-03-01 | 2019-05-21 | Microsoft Technology Licensing, Llc | Ranking based on facial image analysis |
US20160371565A1 (en) * | 2010-03-01 | 2016-12-22 | Microsoft Technology Licensing, Llc | Ranking Based on Facial Image Analysis |
US8787663B2 (en) | 2010-03-01 | 2014-07-22 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US8824737B2 (en) | 2010-05-31 | 2014-09-02 | Primesense Ltd. | Identifying components of a humanoid form in three-dimensional scenes |
US8594425B2 (en) | 2010-05-31 | 2013-11-26 | Primesense Ltd. | Analysis of three-dimensional scenes |
US8781217B2 (en) | 2010-05-31 | 2014-07-15 | Primesense Ltd. | Analysis of three-dimensional scenes with a surface model |
US10108852B2 (en) * | 2010-06-07 | 2018-10-23 | Affectiva, Inc. | Facial analysis to detect asymmetric expressions |
US20140016860A1 (en) * | 2010-06-07 | 2014-01-16 | Affectiva, Inc. | Facial analysis to detect asymmetric expressions |
US20140037191A1 (en) * | 2010-09-16 | 2014-02-06 | Primesense Ltd. | Learning-based pose estimation from depth maps |
US8824781B2 (en) * | 2010-09-16 | 2014-09-02 | Primesense Ltd. | Learning-based pose estimation from depth maps |
US8582867B2 (en) * | 2010-09-16 | 2013-11-12 | Primesense Ltd | Learning-based pose estimation from depth maps |
US20120070070A1 (en) * | 2010-09-16 | 2012-03-22 | Primesense Ltd. | Learning-based pose estimation from depth maps |
US8873811B2 (en) * | 2011-06-30 | 2014-10-28 | Nokia Corporation | Method and apparatus for face tracking utilizing integral gradient projections |
US20130004025A1 (en) * | 2011-06-30 | 2013-01-03 | Nokia Corporation | Method and apparatus for face tracking utilizing integral gradient projections |
TWI547887B (en) * | 2011-06-30 | 2016-09-01 | 諾基亞科技公司 | Method, apparatus and computer program for face tracking utilizing integral gradient projections |
US9002099B2 (en) | 2011-09-11 | 2015-04-07 | Apple Inc. | Learning-based estimation of hand and finger pose |
US9235751B2 (en) * | 2012-03-15 | 2016-01-12 | Kabushiki Kaisha Toshiba | Method and apparatus for image detection and correction |
US20130243274A1 (en) * | 2012-03-15 | 2013-09-19 | Hiroshi Sukegawa | Person Image Processing Apparatus and Person Image Processing Method |
US11288495B2 (en) | 2012-04-09 | 2022-03-29 | Avigilon Fortress Corporation | Object tracking and best shot detection system |
US9195883B2 (en) * | 2012-04-09 | 2015-11-24 | Avigilon Fortress Corporation | Object tracking and best shot detection system |
US11704936B2 (en) | 2012-04-09 | 2023-07-18 | Motorola Solutions, Inc. | Object tracking and best shot detection system |
US9858474B2 (en) | 2012-04-09 | 2018-01-02 | Avigilon Fortress Corporation | Object tracking and best shot detection system |
US10546186B2 (en) | 2012-04-09 | 2020-01-28 | Avigilon Fortress Corporation | Object tracking and best shot detection system |
US20130266181A1 (en) * | 2012-04-09 | 2013-10-10 | Objectvideo, Inc. | Object tracking and best shot detection system |
US8990580B2 (en) | 2012-04-26 | 2015-03-24 | Google Inc. | Automatic user swap |
US9047507B2 (en) | 2012-05-02 | 2015-06-02 | Apple Inc. | Upper-body skeleton extraction from depth maps |
US9311530B1 (en) * | 2012-05-15 | 2016-04-12 | Google Inc. | Summarizing a photo album in a social network system |
US9204037B2 (en) * | 2012-09-25 | 2015-12-01 | Samsung Electronics Co., Ltd. | Method and apparatus for photographing in portable terminal |
US20140085499A1 (en) * | 2012-09-25 | 2014-03-27 | Samsung Electronics Co. Ltd. | Method and apparatus for photographing in portable terminal |
US9019267B2 (en) | 2012-10-30 | 2015-04-28 | Apple Inc. | Depth mapping with enhanced resolution |
US11042775B1 (en) | 2013-02-08 | 2021-06-22 | Brain Corporation | Apparatus and methods for temporal proximity detection |
US9661214B2 (en) | 2013-03-06 | 2017-05-23 | Amazon Technologies, Inc. | Depth determination using camera focus |
US9077891B1 (en) * | 2013-03-06 | 2015-07-07 | Amazon Technologies, Inc. | Depth determination using camera focus |
US20190172494A1 (en) * | 2013-05-28 | 2019-06-06 | Apple Inc. | Reference and non-reference video quality evaluation |
US11423942B2 (en) | 2013-05-28 | 2022-08-23 | Apple Inc. | Reference and non-reference video quality evaluation |
US10957358B2 (en) * | 2013-05-28 | 2021-03-23 | Apple Inc. | Reference and non-reference video quality evaluation |
US9939253B2 (en) | 2014-05-22 | 2018-04-10 | Brain Corporation | Apparatus and methods for distance estimation using multiple image sensors |
US9713982B2 (en) | 2014-05-22 | 2017-07-25 | Brain Corporation | Apparatus and methods for robotic operation using video imagery |
US10194163B2 (en) | 2014-05-22 | 2019-01-29 | Brain Corporation | Apparatus and methods for real time estimation of differential motion in live video |
US9858295B2 (en) | 2014-06-24 | 2018-01-02 | Google Llc | Ranking and selecting images for display from a set of images |
US10417277B2 (en) | 2014-06-24 | 2019-09-17 | Google Llc | Ranking and selecting images for display from a set of images |
US9848112B2 (en) | 2014-07-01 | 2017-12-19 | Brain Corporation | Optical detection apparatus and methods |
US10057593B2 (en) | 2014-07-08 | 2018-08-21 | Brain Corporation | Apparatus and methods for distance estimation using stereo imagery |
US10032280B2 (en) | 2014-09-19 | 2018-07-24 | Brain Corporation | Apparatus and methods for tracking salient features |
US20160086052A1 (en) * | 2014-09-19 | 2016-03-24 | Brain Corporation | Apparatus and methods for saliency detection based on color occurrence analysis |
US10055850B2 (en) | 2014-09-19 | 2018-08-21 | Brain Corporation | Salient features tracking apparatus and methods using visual initialization |
US10268919B1 (en) | 2014-09-19 | 2019-04-23 | Brain Corporation | Methods and apparatus for tracking objects using saliency |
US9870617B2 (en) * | 2014-09-19 | 2018-01-16 | Brain Corporation | Apparatus and methods for saliency detection based on color occurrence analysis |
US10197664B2 (en) | 2015-07-20 | 2019-02-05 | Brain Corporation | Apparatus and methods for detection of objects using broadband signals |
US10786196B2 (en) * | 2015-09-02 | 2020-09-29 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof for skin care analysis |
US20170061609A1 (en) * | 2015-09-02 | 2017-03-02 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
US10043279B1 (en) | 2015-12-07 | 2018-08-07 | Apple Inc. | Robust detection and classification of body parts in a depth map |
US20170344858A1 (en) * | 2016-05-30 | 2017-11-30 | Kabushiki Kaisha Toshiba | Control device, control method, and computer program product |
US10235607B2 (en) * | 2016-05-30 | 2019-03-19 | Kabushiki Kaisha Toshiba | Control device, control method, and computer program product |
US10366278B2 (en) | 2016-09-20 | 2019-07-30 | Apple Inc. | Curvature-based face detector |
US11042725B2 (en) * | 2016-12-16 | 2021-06-22 | Keylemon Sa | Method for selecting frames used in face processing |
US20190318151A1 (en) * | 2018-04-13 | 2019-10-17 | Omron Corporation | Image analysis apparatus, method, and program |
EP3818497A4 (en) * | 2018-08-22 | 2021-08-25 | Zhejiang Dahua Technology Co., Ltd. | SYSTEMS AND PROCEDURES FOR SELECTING THE BEST FACIAL IMAGE OF A HUMAN TARGET AREA |
US11557147B2 (en) | 2018-08-22 | 2023-01-17 | Zhejiang Dahua Technology Co., Ltd. | Systems and methods for selecting a best facial image of a target human face |
US20210342577A1 (en) * | 2018-10-16 | 2021-11-04 | University Of Seoul Industry Cooperation Foundation | Face recognition method and face recognition device |
US11594073B2 (en) * | 2018-10-16 | 2023-02-28 | University Of Seoul Industry Cooperation Foundation | Face recognition method and face recognition device |
WO2020094088A1 (en) * | 2018-11-07 | 2020-05-14 | 杭州海康威视数字技术股份有限公司 | Image capturing method, monitoring camera, and monitoring system |
US11087798B2 (en) * | 2019-04-16 | 2021-08-10 | Honda Motor Co., Ltd. | Selective curation of user recordings |
US11881053B2 (en) * | 2021-05-18 | 2024-01-23 | Fortinet, Inc. | Systems and methods for hierarchical facial image clustering |
Also Published As
Publication number | Publication date |
---|---|
US8139817B2 (en) | 2012-03-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8139817B2 (en) | Face image log creation | |
US8015132B2 (en) | System and method for object detection and classification with multiple threshold adaptive boosting | |
US8401250B2 (en) | Detecting objects of interest in still images | |
WO2019100888A1 (en) | Target object recognition method and device, storage medium and electronic apparatus | |
JP5261312B2 (en) | Image analysis apparatus, image analysis method, and program | |
US9477876B2 (en) | Person recognition apparatus and method thereof | |
CN101344922B (en) | Human face detection method and device | |
EP3224809A1 (en) | Fast object tracking framework for sports video recognition | |
RU2713876C1 (en) | Method and system for detecting alarm events when interacting with self-service device | |
US20090060346A1 (en) | Method And System For Automatically Determining The Orientation Of A Digital Image | |
JP2010231254A (en) | Image analyzing device, method of analyzing image, and program | |
CN112651996B (en) | Target detection tracking method, device, electronic equipment and storage medium | |
Fourney et al. | Constructing face image logs that are both complete and concise | |
CN108647629A (en) | Image path recognition methods and device in a kind of image recognition localization method, video | |
RU2175148C1 (en) | Method for recognizing person identity | |
CN114979567B (en) | Object and region interaction method and system applied to video intelligent monitoring | |
CN108334811B (en) | Face image processing method and device | |
US11763595B2 (en) | Method and system for identifying, tracking, and collecting data on a person of interest | |
Hbali et al. | Object detection based on HOG features: Faces and dual-eyes augmented reality | |
Nautiyal et al. | An automated technique for criminal face identification using biometric approach | |
CN117636402B (en) | Pedestrian re-identification-based passenger analysis method and device and computer storage medium | |
Adouani et al. | Comparative study of face detection methods in spontaneous videos | |
JP7540500B2 (en) | GROUP IDENTIFICATION DEVICE, GROUP IDENTIFICATION METHOD, AND PROGRAM | |
Nguyen et al. | Automated conformance testing for ISO/IEC 19794-5 Standard on facial photo specifications | |
US20230316562A1 (en) | Causal interaction detection apparatus, control method, and computer-readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: UNIVERSITY OF OTTAWA, CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LAGANIERE, ROBERT;FOURNEY, ADAM;REEL/FRAME:019228/0571;SIGNING DATES FROM 20070426 TO 20070427 Owner name: UNIVERSITY OF OTTAWA, CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LAGANIERE, ROBERT;FOURNEY, ADAM;SIGNING DATES FROM 20070426 TO 20070427;REEL/FRAME:019228/0571 |
|
AS | Assignment |
Owner name: VISUAL CORTEK INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UNIVERSITY OF OTTAWA;REEL/FRAME:025013/0131 Effective date: 20100729 Owner name: TELEWATCH INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VISUAL CORTEK INC.;REEL/FRAME:025013/0421 Effective date: 20100917 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
REMI | Maintenance fee reminder mailed | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
SULP | Surcharge for late payment | ||
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20200320 |