+

WO2016177259A1 - Procédé et dispositif de reconnaissance d'images similaires - Google Patents

Procédé et dispositif de reconnaissance d'images similaires Download PDF

Info

Publication number
WO2016177259A1
WO2016177259A1 PCT/CN2016/079158 CN2016079158W WO2016177259A1 WO 2016177259 A1 WO2016177259 A1 WO 2016177259A1 CN 2016079158 W CN2016079158 W CN 2016079158W WO 2016177259 A1 WO2016177259 A1 WO 2016177259A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
identified
feature
normalized
distance
Prior art date
Application number
PCT/CN2016/079158
Other languages
English (en)
Chinese (zh)
Inventor
陈岳峰
Original Assignee
阿里巴巴集团控股有限公司
陈岳峰
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司, 陈岳峰 filed Critical 阿里巴巴集团控股有限公司
Publication of WO2016177259A1 publication Critical patent/WO2016177259A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/761Proximity, similarity or dissimilarity measures

Definitions

  • the present application relates to the field of communications technologies, and in particular, to a similar image recognition method.
  • the application also relates to a similar image recognition device.
  • the traditional face authentication method is mainly based on SIFT (Scale-invariant feature transform), LBP (Local Binary Patterns) and other features to describe the photos and the faces in the existing photos. Then, through the classifier, it is judged whether the two faces are the same person, wherein SIFT is a local feature descriptor for the field of image processing, and the description has scale invariance and can detect key points in the image, and the SIFT feature is Regardless of the size and rotation of the image based on some local appearance of interest points on the object, the tolerance for light, noise, and slight viewing angle changes is also quite high; LBP is an effective texture description operator that can measure and extract images. Local texture information, which is invariant to illumination.
  • SIFT Scale-invariant feature transform
  • LBP Local Binary Patterns
  • the prior art has the following disadvantages: the traditional feature-based face authentication algorithm often extracts high-dimensional features from the face region and uses a classifier. Face authentication. Such algorithms are often only targeted at face features. Don't notice the pictures or photos to be effective. In the case that the background is complicated and the face changes greatly, the recognition technology in the prior art often cannot accurately pass the images in the two photos to be the same person. Therefore, how to ensure the recognition accuracy is The fast and efficient recognition of the image to be detected and the existing image has become a technical problem to be solved by those skilled in the art.
  • the present application provides a similar image recognition method for quickly and efficiently identifying an image to be detected and an existing image under the premise of ensuring accuracy, and the method includes:
  • a metric distance between the normalized image and a normalized image of the second image to be recognized the metric distance being characterized according to the normalized image and the normalized image of the second image to be recognized Distance generation in space, wherein the distance of the similar normalized image in the feature space is less than the distance of the non-similar normalized image in the feature space;
  • the metric distance is less than or equal to the threshold, confirming that the first to-be-identified image is similar to a specified feature of the second to-be-identified image.
  • the area to be compared corresponding to the specified feature in the first to-be-identified image is obtained, specifically:
  • the key point coordinates corresponding to the plurality of key point features of the specified feature in the to-be-contrast area are obtained by a preset key point regression model.
  • the image in the area to be identified is aligned with a preset standard image, specifically:
  • the parameter M is generated according to coordinates of each key point of the standard image and key point coordinates of an image corresponding to the specified feature in the labeled image.
  • the method further includes:
  • the resolution of the normalized image is adjusted to a preset resolution.
  • determining a metric distance between the normalized image and the normalized image of the second to-be-identified image is specifically:
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are trained according to an annotated image comprising a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • the present application also proposes a similar image recognition device, including:
  • An acquiring module configured to acquire an area to be compared corresponding to the specified feature in the first to-be-identified image
  • An alignment module configured to align an image in the area to be identified with a preset standard image, And using the aligned image as a normalized image of the first to-be-identified image, the standard image corresponding to the designated feature;
  • a determining module configured to determine a metric distance between the normalized image of the first to-be-identified image and a normalized image of the second to-be-identified image, the metric distance according to the normalized image and the Calculating a distance of the normalized image of the second image to be recognized in the feature space, wherein a distance of the similar normalized image in the feature space is smaller than a distance of the non-similar normalized image in the feature space;
  • An identification module configured to confirm, when the metric distance is greater than a preset threshold, that the first to-be-identified image is not similar to a specified feature of the second to-be-identified image, and that the metric distance is less than or equal to the threshold It is confirmed that the first to-be-identified image is similar to the specified feature of the second to-be-identified image.
  • the determining module is specifically configured to:
  • the alignment module is specifically configured to:
  • the method further comprises:
  • an adjustment module configured to adjust a resolution of the normalized image to a preset resolution.
  • the obtaining module is specifically configured to:
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are trained according to an annotated image comprising a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • the present application also proposes a similar image recognition method, which is applied to a client, and includes the following steps:
  • the identity authentication request carries the first to-be-identified image uploaded by the user and the authentication information of the user;
  • the client presents the authentication result to the user according to the identity authentication response.
  • the user identity authentication request is received, specifically:
  • the identity authentication response is an identity authentication success response or an identity authentication failure response, and further includes:
  • the identity authentication success response is generated by the server after confirming that the first to-be-identified image is similar to the specified feature of the second to-be-identified image;
  • the identity authentication failure response is generated by the server after confirming that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image.
  • the authentication result is displayed to the user according to the identity authentication response, specifically:
  • the user When receiving the identity authentication failure response, the user is presented with a preset interface corresponding to the identity authentication failure response, and prompting the user whether the manual verification is required.
  • the method further includes:
  • the identity authentication request is sent to a preset server.
  • the application also proposes a client, including:
  • a receiving module configured to receive an identity authentication request of the user, where the identity authentication request carries the first to-be-identified image uploaded by the user and the authentication information of the user;
  • a sending module configured to send the identity authentication request to the server, so that the server acquires a second to-be-identified image corresponding to the user according to the authentication information;
  • a receiving module configured to receive an identity authentication response sent by the server
  • a display module configured to display the authentication result to the user according to the identity authentication response.
  • the receiving module is specifically configured to:
  • the identity authentication response is an identity authentication success response or an identity authentication failure response. Also includes:
  • the identity authentication success response is generated by the server after confirming that the first to-be-identified image is similar to the specified feature of the second to-be-identified image;
  • the identity authentication failure response is generated by the server after confirming that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image.
  • the displaying module is configured to: when the receiving module receives the identity authentication success response, display the preset interface corresponding to the identity authentication success response to the user;
  • the displaying module is specifically configured to: when the receiving module receives the identity authentication failure response, display, to the user, a preset interface corresponding to the identity authentication failure response, and to the user Show tips for manual verification.
  • the receiving module when the display module displays the preset interface corresponding to the identity authentication failure response to the user and prompts the user whether the manual verification is required, the receiving module further receives the The manual verification request of the user, the receiving module instructing the sending module to send the identity authentication request to a preset server.
  • the present application also proposes a similar image recognition method, which is applied to a server, and includes the following steps:
  • the metric distance is generated according to a distance of the normalized image and the normalized image of the second image to be recognized in the feature space, wherein a distance of the similar normalized image in the feature space is less than a non-similar The normalized image of the distance in the feature space;
  • metric distance is greater than a preset threshold, confirm that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image, and return an identity verification failure response to the client;
  • the metric distance is less than or equal to the threshold, confirm that the first to-be-identified image is similar to the specified feature of the second to-be-identified image, and return an identity verification success response to the client.
  • the area to be compared corresponding to the specified feature in the first to-be-identified image is obtained, specifically:
  • the key point coordinates corresponding to the plurality of key point features of the specified feature in the to-be-contrast area are obtained by a preset key point regression model.
  • the image in the area to be identified is aligned with a preset standard image, specifically:
  • the parameter M is generated according to coordinates of each key point of the standard image and key point coordinates of an image corresponding to the specified feature in the labeled image.
  • the method further includes:
  • the resolution of the normalized image is adjusted to a preset resolution.
  • determining a metric distance between the normalized image and the normalized image of the second to-be-identified image is specifically:
  • the specified feature is mapped to Solving the feature value after the space, and using the feature value as the feature value of the normalized image;
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are trained according to an annotated image comprising a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • a server including:
  • a querying module configured to query, according to the authentication information, a second to-be-identified image corresponding to the user
  • An acquiring module configured to acquire an area to be compared corresponding to the specified feature in the first to-be-identified image
  • An alignment module configured to align an image in the to-be-identified area with a preset standard image, and use the aligned image as a normalized image of the first to-be-identified image, the standard image and the Specify the feature correspondence;
  • a determining module configured to determine a metric distance between the normalized image of the first to-be-identified image and a normalized image of the second to-be-identified image, the metric distance according to the normalized image and the Calculating a distance of the normalized image of the second image to be recognized in the feature space, wherein a distance of the similar normalized image in the feature space is smaller than a distance of the non-similar normalized image in the feature space;
  • An identification module configured to confirm, when the metric distance is greater than a preset threshold, that the first to-be-identified image is not similar to a specified feature of the second to-be-identified image, and that the metric distance is less than or When the threshold is equal to the confirmation, the first to-be-identified image is similar to the specified feature of the second to-be-identified image;
  • a sending module configured to return an identity verification failure response to the client when the identification module confirms that the first to-be-identified image is not similar to a specified feature of the second to-be-identified image, and confirm in the identification module And returning the identity verification success response to the client when the first to-be-identified image is similar to the specified feature of the second to-be-identified image.
  • the determining module is specifically configured to:
  • the alignment module is specifically configured to map coordinates of each key point of the to-be-identified area to key point coordinates of the aligned image according to the parameter M, wherein the parameter M is a key according to the standard image. Point coordinates and keypoint coordinates of the image corresponding to the specified feature in the annotated image.
  • the method further comprises:
  • an adjustment module configured to adjust a resolution of the normalized image to a preset resolution.
  • the obtaining module is specifically configured to:
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are obtained according to the labeled image training, and the labeled
  • the annotation image includes a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • the area to be identified is to be recognized.
  • the resolution of the aligned image is adjusted to a preset resolution, and the adjusted image is used as a normalized image, and finally the normalized image of the first image to be recognized and the normalized image of the second image to be recognized are acquired.
  • the metric distance between the images determines whether the specified features of the first to-be-identified image and the second to-be-identified image are similar according to a size between the metric distance and the preset threshold. Therefore, under the premise of ensuring accuracy, the similarity between the image to be detected and another image to be detected is quickly and efficiently identified, which provides a reference for improving the security of the existing system.
  • FIG. 1 is a schematic flow chart of a similar image recognition method proposed in the present application.
  • FIG. 2 is a structural diagram of a convolutional neural network for training face feature point positioning in a specific embodiment of the present application
  • FIG. 3 is a schematic flowchart of performing depth metric learning in a specific embodiment of the present application.
  • FIG. 4 is a structural diagram of a convolutional neural network for training face authentication in a specific embodiment of the present application
  • FIG. 5 is a schematic flowchart of a similar image recognition performed by a client in a specific embodiment of the present application.
  • FIG. 6 is a schematic flowchart of performing similar image recognition by a server in a specific embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of a similar image recognition device according to the present application.
  • FIG. 8 is a schematic structural diagram of a client according to the present application.
  • FIG. 9 is a schematic structural diagram of a server according to the present application.
  • the present application proposes a recognition method for similar images, which can be implemented by means of a computer device in a network environment.
  • the main purpose of judging the similarity is the server in the background of the system, and the client-oriented client is either a mobile device compatible with key input and touch screen input, or a PC device, and the client and the server are wired or A wireless way to achieve network connectivity.
  • a schematic flowchart of a similar image recognition method proposed in the present application includes the following steps:
  • the area to be compared can be obtained by determining the key point coordinates (related to the face). Specifically, in determining a region to be compared corresponding to the specified feature in the first to-be-identified image, the to-be-contrast region in the first to-be-identified image may be first determined according to a detection algorithm corresponding to the specified feature.
  • the key point coordinates corresponding to the plurality of key point features of the specified feature in the to-be-compared area are obtained by using a preset key point regression model, so as to accurately determine the area to be compared.
  • the designated feature may be a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • a deep convolutional neural network is used to implement regression of face key points.
  • the structure of the neural network in this embodiment is as shown in FIG. 2, and includes four convolution layers and two fully connected layers.
  • the first three convolutional layers contain the maximum pooling operation, and the last convolutional layer contains only the convolution operation.
  • the first fully connected layer contains 100 nodes, and the second is fully connected to have 10 nodes, representing the coordinates of the five key points of the face.
  • Regression uses the Euclidean distance as the loss function, and the expression is as follows:
  • x represents the coordinates of the key points of the annotation, representing the coordinates of the key points predicted by the convolutional neural network.
  • the specific embodiment uses a stochastic gradient descent algorithm to optimize the parameters in the model, thereby training to obtain a model for predicting key points of the face.
  • this step maps the coordinates of each key point of the to-be-identified area to the key point coordinates of the aligned image according to the parameter M, wherein the parameter M is the coordinate of each key point according to the standard image and the labeled image
  • the key point coordinates of the image corresponding to the specified feature are generated.
  • the specific embodiment defines five key point positions in the standard human face, which are the left eye, the right eye, the nose, the left mouth corner, and the right mouth corner position, and are detected by The face is rotated, panned, and scaled to align to a standard face. Assuming that the position of the feature point in the standard face is (x, y) and the position of the predicted feature point is (x', y'), then the relationship between the two is:
  • 4 equations are needed.
  • the specific embodiment maps five points, establishes a linear equation group, and calculates a system of linear equations by the least squares method of the linear equations. details as follows:
  • the above process is a detailed generation process of the parameter M.
  • Those skilled in the art can perform alignment processing on the face image according to the parameter M.
  • other improved implementation manners that can obtain the parameter M are all within the protection scope of the present application.
  • the comparison result is more accurate.
  • the application needs to align the image in the area to be identified with the preset standard image.
  • this step aligns the face to a standard face, and those skilled in the art can set the face of the standard based on existing comparison criteria, which are all within the scope of the present application. .
  • the present application adjusts the resolution of the normalized image to a preset resolution.
  • the step will be the face key.
  • the point coordinate information is normalized to the 39x39 scale space.
  • a specified feature in the normalized image is first extracted by a convolutional neural network, and then the specified feature is mapped to the convolutional neural network and the distance metric loss function.
  • the feature value after the feature space, and the feature value is used as the feature value of the normalized image, and finally determining the feature value of the normalized image and the feature of the normalized image of the second image to be recognized
  • the Euclidean distance between the values, the Euclidean distance is taken as the metric distance.
  • the technical solution of the present application combines a deep convolutional neural network and metric learning to train a face authentication model.
  • Deep convolutional neural networks are widely used in the field of image understanding, including image classification, image retrieval, target detection, and face recognition.
  • the convolutional neural network has a feature self-learning, and the model generalization ability is good.
  • Metric learning is a linear or non-linear mapping of feature spaces such that the same face feature distance is less than a different face feature distance.
  • the convolutional neural network parameters are obtained according to the labeled image training, and the labeled image includes a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • sample pairing is performed in a pair-wise manner, and each sample includes two portrait images, if two images are in the image Not the same person, indicating a negative sample, if the same person is expressed as Positive sample.
  • a positive sample is generated by combining two pictures belonging to the same person, and a negative sample is generated by a picture that is not the same person.
  • the depth metric learning mainly consists of two parts. One of them is the parameter W, which represents the parameters of the resulting convolutional neural network that need to be trained, and the other is the distance metric loss function. Different from traditional face recognition, the input of this application is 2 faces, and the final loss is also the distance between 2 faces in the feature space.
  • the structure of the convolutional neural network used in this embodiment is as shown in FIG.
  • metric learning is to find a transformation space in which the distance of similar samples is reduced, and the distance between different types of samples is increased. Therefore, this step first searches for a nonlinear transformation through metric learning, transforming the face from the original pixel to a feature space, so that the similar face distance is small and the dissimilar face distance is large in this space. Then the facial features are extracted by the deep convolutional neural network. Finally, the features learned by the convolutional neural network are mapped to a feature space in combination with metric learning. Since the convolutional neural network continues the non-linear mapping of the face image, the feature expression thus obtained is more robust than the artificially designed feature, and the face authentication accuracy is higher.
  • the resulting 100-dimensional feature will be subjected to metric learning, and the loss function employed in the process is as follows:
  • the parameter W of the model can be obtained by minimizing the loss function.
  • the technician can use the chain derivation rule to obtain the gradient of the corresponding parameter and use the stochastic gradient descent method (SGD) to optimize the parameters of the calculation model.
  • SGD stochastic gradient descent method
  • Other calculation models capable of achieving the optimization effect are also within the protection scope of the present application.
  • the faces in the map are respectively detected, and feature point extraction and face alignment are performed according to the face region.
  • the feature extraction is performed and the feature is mapped into the space of 100 dimensions.
  • the Euclidean distance of the two facial features is calculated. When the distance is greater than or equal to ⁇ , it means that it is not the same person, otherwise it is the same person.
  • the process can be completed by the client and the server.
  • the user can use a mobile terminal such as a smart phone or a tablet to upload images and information, and also upload images and related information through the PC terminal.
  • the server may be a data server or a web server pre-configured by the system operator.
  • the client As a link between the user and the server, the client is mainly used to forward the user's input content to the server, and the server verifies the identity of the user according to the content input by the user, and finally the client displays the verification result according to the server.
  • the following first introduces the similar image recognition method on the client side, as shown in FIG. 5, including the following steps:
  • S501 Receive an identity authentication request of the user, where the identity authentication request carries the first to-be-identified image uploaded by the user and the authentication information of the user.
  • the form of the client is not limited, and the client may be a PC device or a mobile device. However, both can provide the user with the function of image uploading and information input. Specifically, the client first acquires the image uploaded by the user and the information input by the user, and then uses the image as the first image to be recognized, and Using the information as the authentication information, and finally generating the identity authentication request according to the first to-be-identified image and the authentication information.
  • the server may obtain the second to-be-identified image corresponding to the user according to the data.
  • the server queries the database for the image in the user ID according to the identity information provided by the user, and uses the image as the second image to be identified, thereby determining whether the image uploaded by the user is the same as the ID card. Image matching.
  • the identity authentication response is an identity authentication success response or an identity authentication failure response, wherein the identity authentication success response is that the server confirms that the first to-be-identified image is similar to the specified feature of the second to-be-identified image. And then generated, and the identity authentication failure response is generated by the server after confirming that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image.
  • the client When the client receives the identity authentication success response, the client displays the preset interface corresponding to the identity authentication success response to the user;
  • the client when the client receives the identity authentication failure response, the client displays the preset interface corresponding to the identity authentication failure response to the user, and shows whether the user needs to perform Manually verified prompt information.
  • the present application automatically judges a group of images to be recognized by the device, in order to further avoid the influence of the error, when the failure response is returned to the user, the prompt information indicating whether manual verification is required is simultaneously displayed to the user. . If the user thinks that the manual review needs to be resubmitted, then the client re-advertises the manual verification request, and after receiving the manual verification request from the user, the client sends the identity authentication request to the preset server.
  • the above is the process of the client, and is mainly used to implement the interaction between the user and the server.
  • the following embodiment is a similar image recognition method on the server side, as shown in FIG. 6, including the following steps:
  • S605. Determine a metric distance between the normalized image and a normalized image of the second to-be-identified image, the metric distance according to the normalized image and the normalized image of the second to-be-identified image a distance generation in the feature space, wherein a distance of the similar normalized image in the feature space is less than a distance of the non-similar normalized image in the feature space;
  • the present application also proposes a similar image recognition device, as shown in FIG. 7, comprising:
  • the obtaining module 710 is configured to obtain an area to be compared corresponding to the specified feature in the first to-be-identified image
  • the aligning module 720 is configured to align an image in the to-be-identified area with a preset standard image, and use the aligned image as a normalized image of the first to-be-identified image, the standard image and the Corresponding to the specified feature;
  • a determining module 730 configured to determine a metric distance between the normalized image of the first to-be-identified image and a normalized image of the second to-be-identified image, the metric distance according to the normalized image and A normalized image of the second image to be identified is generated in a distance in the feature space, wherein a distance of the similar normalized image in the feature space is smaller than a non-similar normalized image in the feature Distance of space;
  • the identification module 740 is configured to confirm that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image when the metric distance is greater than a preset threshold, and that the metric distance is less than or equal to the The threshold is confirmed to be similar to the specified feature of the second image to be recognized.
  • the determining module is specifically configured to:
  • the alignment module is specifically configured to:
  • an adjustment module configured to adjust a resolution of the normalized image to a preset resolution.
  • the acquiring module is specifically configured to:
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are obtained according to the labeled image, and the labeled image includes normalized images with specified features being similar to each other and designated features are not mutually exclusive. A similar normalized image.
  • the application also proposes a client, as shown in FIG. 8, comprising:
  • the receiving module 810 is configured to receive an identity authentication request of the user, where the identity authentication request carries the first to-be-identified image uploaded by the user and the authentication information of the user;
  • the sending module 820 is configured to send the identity authentication request to the server, so that the server acquires a second to-be-identified image corresponding to the user according to the authentication information;
  • the receiving module 810 is further configured to receive an identity authentication response sent by the server.
  • the displaying module 830 is configured to display the authentication result to the user according to the identity authentication response.
  • the receiving module is specifically configured to:
  • the identity authentication response is an identity authentication success response or an identity authentication failure response, and further includes:
  • the identity authentication success response is generated by the server after confirming that the first to-be-identified image is similar to the specified feature of the second to-be-identified image;
  • the identity authentication failure response is generated by the server after confirming that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image.
  • the displaying module is specifically configured to: when the receiving module receives the identity authentication success response, display the preset interface corresponding to the identity authentication success response to the user; Or the displaying module is specifically configured to: when the receiving module receives the identity authentication failure response, display, to the user, a preset interface corresponding to the identity authentication failure response, and to the user Show tips for manual verification.
  • the receiving module displays to the user in the display module Receiving a manual verification request corresponding to the identity authentication failure response and a prompt to the user to indicate whether manual verification is required, the receiving module instructing the sending module to The identity authentication request is sent to the preset server.
  • the embodiment of the present application further provides a server, as shown in FIG. 9, including:
  • the receiving module 910 is configured to receive an identity authentication request sent by the client, where the identity authentication request carries the first to-be-identified image uploaded by the user and the authentication information of the user;
  • the querying module 920 is configured to query, according to the authentication information, a second to-be-identified image corresponding to the user;
  • the obtaining module 930 is configured to acquire an area to be compared corresponding to the specified feature in the first to-be-identified image
  • An alignment module 940 configured to align an image in the to-be-identified area with a preset standard image, and use the aligned image as a normalized image of the first to-be-identified image, the standard image and the Corresponding to the specified feature;
  • a determining module 950 configured to determine a metric distance between the normalized image of the first to-be-identified image and the normalized image of the second to-be-identified image, the metric distance according to the normalized image and a normalized image of the second image to be identified is generated in a distance in the feature space, wherein a distance of the similar normalized image in the feature space is smaller than a distance of the non-similar normalized image in the feature space ;
  • the identification module 960 is configured to confirm that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image when the metric distance is greater than a preset threshold, and that the metric distance is less than or equal to the Confirming, at the threshold, that the first to-be-identified image is similar to a specified feature of the second to-be-identified image;
  • the sending module 970 is configured to: when the identifying module confirms that the first to-be-identified image is not similar to the specified feature of the second to-be-identified image, return an identity verification failure response to the client, and And when the identifying module confirms that the first to-be-identified image is similar to the specified feature of the second to-be-identified image, returning an identity verification success response to the client.
  • the determining module is specifically configured to:
  • the alignment module is specifically configured to:
  • an adjustment module configured to adjust a resolution of the normalized image to a preset resolution.
  • the acquiring module is specifically configured to:
  • the designated feature is specifically a face region
  • the key feature includes at least a left eye region, a right eye region, a nose region, a left corner region, and a right corner region.
  • the convolutional neural network parameters are trained according to the labeled image, and the labeled image includes a normalized image in which the specified features are similar to each other and a normalized image in which the specified features are not similar to each other.
  • the present application can be implemented in hardware or by means of software plus the necessary general hardware platform.
  • the technical solution of the present application may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a USB flash drive, a mobile hard disk, etc.), including several The instructions are for causing a computer device (which may be a personal computer, server, or network device, etc.) to perform the methods described in various implementation scenarios of the present application.
  • modules in the apparatus in the implementation scenario may be distributed in the apparatus for implementing the scenario according to the implementation scenario description, or may be correspondingly changed in one or more devices different from the implementation scenario.
  • the modules of the above implementation scenarios may be combined into one module, or may be further split into multiple sub-modules.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Collating Specific Patterns (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un procédé de reconnaissance d'images similaires. Après détermination d'une zone à comparer, correspondant à une caractéristique spécifiée, dans une première image à reconnaître, et alignement d'une image dans la zone à reconnaître sur une image standard prédéfinie, la résolution de l'image alignée dans la zone à reconnaître est ajustée à une résolution prédéfinie et l'image ajustée est utilisée comme image normalisée, et enfin, une distance métrique entre l'image normalisée de la première image à reconnaître et une image normalisée d'une seconde image à reconnaître est acquise, et il est déterminé que les caractéristiques spécifiées de la première image à reconnaître et de la seconde image à reconnaître sont similaires en fonction de la différence entre la distance métrique et un seuil prédéfini. Ainsi, sur la prémisse que la précision est assurée, la similarité entre une image à détecter et une autre image à détecter est reconnue de manière efficace et rapide, ce qui permet d'obtenir une base de référence en vue d'améliorer la sécurité des systèmes existants.
PCT/CN2016/079158 2015-05-07 2016-04-13 Procédé et dispositif de reconnaissance d'images similaires WO2016177259A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510229654.6 2015-05-07
CN201510229654.6A CN106203242B (zh) 2015-05-07 2015-05-07 一种相似图像识别方法及设备

Publications (1)

Publication Number Publication Date
WO2016177259A1 true WO2016177259A1 (fr) 2016-11-10

Family

ID=57217488

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/079158 WO2016177259A1 (fr) 2015-05-07 2016-04-13 Procédé et dispositif de reconnaissance d'images similaires

Country Status (2)

Country Link
CN (1) CN106203242B (fr)
WO (1) WO2016177259A1 (fr)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108804996A (zh) * 2018-03-27 2018-11-13 腾讯科技(深圳)有限公司 人脸验证方法、装置、计算机设备及存储介质
CN109345770A (zh) * 2018-11-14 2019-02-15 深圳市尼欧科技有限公司 一种孩童遗留车内报警系统及孩童遗留车内报警方法
CN110084161A (zh) * 2019-04-17 2019-08-02 中山大学 一种人体骨骼关键点的快速检测方法及系统
CN111079644A (zh) * 2019-12-13 2020-04-28 四川新网银行股份有限公司 基于距离和关节点识别外力辅助拍照的方法及存储介质
CN112464689A (zh) * 2019-09-06 2021-03-09 佳能株式会社 生成神经网络的方法、装置和系统及存储指令的存储介质
CN113568571A (zh) * 2021-06-28 2021-10-29 西安电子科技大学 基于残差神经网络的图像去重方法
CN113688737A (zh) * 2017-12-15 2021-11-23 北京市商汤科技开发有限公司 人脸图像处理方法、装置、电子设备、存储介质及程序
CN113744769A (zh) * 2021-09-06 2021-12-03 盐城市聚云网络科技有限公司 一种计算机信息安全产品用存储装置
WO2022242713A1 (fr) * 2021-05-21 2022-11-24 北京字跳网络技术有限公司 Procédé et dispositif d'alignement d'image
US12314342B2 (en) 2019-03-26 2025-05-27 Huawei Technologies Co., Ltd. Object recognition method and apparatus

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106897390B (zh) * 2017-01-24 2019-10-15 北京大学 基于深度度量学习的目标精确检索方法
CN108428242B (zh) 2017-02-15 2022-02-08 宏达国际电子股份有限公司 图像处理装置及其方法
CN108573201A (zh) * 2017-03-13 2018-09-25 金德奎 一种基于人脸识别技术的用户身份识别匹配方法
EP3510524A4 (fr) * 2017-06-30 2019-08-21 Beijing Didi Infinity Technology and Development Co., Ltd. Systèmes et procédés permettant de vérifier l'authenticité d'une photo d'identification
CN107451965B (zh) * 2017-07-24 2019-10-18 深圳市智美达科技股份有限公司 畸变人脸图像校正方法、装置、计算机设备和存储介质
CN108012080B (zh) * 2017-12-04 2020-02-04 Oppo广东移动通信有限公司 图像处理方法、装置、电子设备及计算机可读存储介质
CN108932727B (zh) * 2017-12-29 2021-08-27 浙江宇视科技有限公司 人脸跟踪方法和装置
CN110110189A (zh) * 2018-02-01 2019-08-09 北京京东尚科信息技术有限公司 用于生成信息的方法和装置
CN108921209A (zh) * 2018-06-21 2018-11-30 杭州骑轻尘信息技术有限公司 图片识别方法、装置及电子设备
CN109459873A (zh) * 2018-11-12 2019-03-12 广州小鹏汽车科技有限公司 一种测试方法、装置、自动测试系统及存储介质
CN110781917B (zh) * 2019-09-18 2021-03-02 北京三快在线科技有限公司 重复图像的检测方法、装置、电子设备及可读存储介质
US11610391B2 (en) 2019-12-30 2023-03-21 Industrial Technology Research Institute Cross-domain image comparison method and system using semantic segmentation
CN112508109B (zh) * 2020-12-10 2023-05-19 锐捷网络股份有限公司 一种图像识别模型的训练方法及装置
CN112560971B (zh) * 2020-12-21 2024-07-16 上海明略人工智能(集团)有限公司 一种主动学习自迭代的图像分类方法和系统
CN114359933B (zh) * 2021-11-18 2022-09-20 珠海读书郎软件科技有限公司 一种封面图像的识别方法
CN114596594A (zh) * 2022-01-20 2022-06-07 北京极豪科技有限公司 一种指纹图像匹配方法、设备、介质及程序产品

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030146913A1 (en) * 2002-02-07 2003-08-07 Hong Shen Object-correspondence identification without full volume registration
CN101669824A (zh) * 2009-09-22 2010-03-17 浙江工业大学 基于生物特征识别的人与身份证同一性检验装置
CN102629320A (zh) * 2012-03-27 2012-08-08 中国科学院自动化研究所 基于特征层定序测量统计描述的人脸识别方法
CN103207898A (zh) * 2013-03-19 2013-07-17 天格科技(杭州)有限公司 一种基于局部敏感哈希的相似人脸快速检索方法
CN103678984A (zh) * 2013-12-20 2014-03-26 湖北微模式科技发展有限公司 一种利用摄像头实现用户身份验证的方法

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101510257B (zh) * 2009-03-31 2011-08-10 华为技术有限公司 一种人脸相似度匹配方法及装置
CN102375970B (zh) * 2010-08-13 2016-03-30 北京中星微电子有限公司 一种基于人脸的身份认证方法和认证装置
CN103824051B (zh) * 2014-02-17 2017-05-03 北京旷视科技有限公司 一种基于局部区域匹配的人脸搜索方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030146913A1 (en) * 2002-02-07 2003-08-07 Hong Shen Object-correspondence identification without full volume registration
CN101669824A (zh) * 2009-09-22 2010-03-17 浙江工业大学 基于生物特征识别的人与身份证同一性检验装置
CN102629320A (zh) * 2012-03-27 2012-08-08 中国科学院自动化研究所 基于特征层定序测量统计描述的人脸识别方法
CN103207898A (zh) * 2013-03-19 2013-07-17 天格科技(杭州)有限公司 一种基于局部敏感哈希的相似人脸快速检索方法
CN103678984A (zh) * 2013-12-20 2014-03-26 湖北微模式科技发展有限公司 一种利用摄像头实现用户身份验证的方法

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113688737A (zh) * 2017-12-15 2021-11-23 北京市商汤科技开发有限公司 人脸图像处理方法、装置、电子设备、存储介质及程序
CN108804996A (zh) * 2018-03-27 2018-11-13 腾讯科技(深圳)有限公司 人脸验证方法、装置、计算机设备及存储介质
CN108804996B (zh) * 2018-03-27 2022-03-04 腾讯科技(深圳)有限公司 人脸验证方法、装置、计算机设备及存储介质
CN109345770A (zh) * 2018-11-14 2019-02-15 深圳市尼欧科技有限公司 一种孩童遗留车内报警系统及孩童遗留车内报警方法
US12314342B2 (en) 2019-03-26 2025-05-27 Huawei Technologies Co., Ltd. Object recognition method and apparatus
CN110084161A (zh) * 2019-04-17 2019-08-02 中山大学 一种人体骨骼关键点的快速检测方法及系统
CN112464689A (zh) * 2019-09-06 2021-03-09 佳能株式会社 生成神经网络的方法、装置和系统及存储指令的存储介质
CN111079644A (zh) * 2019-12-13 2020-04-28 四川新网银行股份有限公司 基于距离和关节点识别外力辅助拍照的方法及存储介质
WO2022242713A1 (fr) * 2021-05-21 2022-11-24 北京字跳网络技术有限公司 Procédé et dispositif d'alignement d'image
CN113568571A (zh) * 2021-06-28 2021-10-29 西安电子科技大学 基于残差神经网络的图像去重方法
CN113568571B (zh) * 2021-06-28 2024-06-04 西安电子科技大学 基于残差神经网络的图像去重方法
CN113744769A (zh) * 2021-09-06 2021-12-03 盐城市聚云网络科技有限公司 一种计算机信息安全产品用存储装置

Also Published As

Publication number Publication date
CN106203242A (zh) 2016-12-07
CN106203242B (zh) 2019-12-24

Similar Documents

Publication Publication Date Title
WO2016177259A1 (fr) Procédé et dispositif de reconnaissance d'images similaires
US10755084B2 (en) Face authentication to mitigate spoofing
WO2020207189A1 (fr) Procédé et dispositif d'authentification d'identité, support de mémoire et dispositif informatique
WO2021012526A1 (fr) Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage
US20210064900A1 (en) Id verification with a mobile device
JP6754619B2 (ja) 顔認識方法及び装置
US11727053B2 (en) Entity recognition from an image
WO2018086607A1 (fr) Procédé de suivi de cible, dispositif électronique et support d'informations
WO2022188697A1 (fr) Procédé et appareil d'extraction de caractéristique biologique, dispositif, support et produit programme
CN110472460B (zh) 人脸图像处理方法及装置
WO2019071664A1 (fr) Procédé et appareil de reconnaissance de visage humain combinés à des informations de profondeur, et support de stockage
WO2019033571A1 (fr) Procédé de détection de point de caractéristique faciale, appareil et support de stockage
US10956738B1 (en) Identity authentication using an inlier neural network
CN111091075A (zh) 人脸识别方法、装置、电子设备及存储介质
CA3040971A1 (fr) Authentification faciale permettant d'attenuer la mystification
US12192207B2 (en) Media data based user profiles
WO2019033567A1 (fr) Procédé de capture de mouvement de globe oculaire, dispositif et support d'informations
WO2022078168A1 (fr) Procédé et appareil de vérification d'identité fondés sur l'intelligence artificielle, dispositif informatique et support de mémoire
TW201822150A (zh) 基於圖片的判別方法及裝置和計算設備
CN113591603A (zh) 证件的验证方法、装置、电子设备及存储介质
CN118552826A (zh) 基于双流注意力的可见光和红外图像目标检测方法及装置
CN114743277A (zh) 活体检测方法、装置、电子设备、存储介质及程序产品
KR20220000851A (ko) 딥러닝 기술을 활용한 피부과 시술 추천 시스템 및 방법
CN118887689A (zh) 手写电子签名的真实性验证方法、装置
US9786030B1 (en) Providing focal length adjustments

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16789253

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16789253

Country of ref document: EP

Kind code of ref document: A1

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载