WO2018121690A1 - Procédé et dispositif de détection d'attribut d'objet, procédé et dispositif d'apprentissage de réseau neuronal et procédé et dispositif de détection zonale - Google Patents
Procédé et dispositif de détection d'attribut d'objet, procédé et dispositif d'apprentissage de réseau neuronal et procédé et dispositif de détection zonale Download PDFInfo
- Publication number
- WO2018121690A1 WO2018121690A1 PCT/CN2017/119535 CN2017119535W WO2018121690A1 WO 2018121690 A1 WO2018121690 A1 WO 2018121690A1 CN 2017119535 W CN2017119535 W CN 2017119535W WO 2018121690 A1 WO2018121690 A1 WO 2018121690A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- neural network
- target area
- training
- sample
- Prior art date
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 224
- 238000012549 training Methods 0.000 title claims abstract description 222
- 238000000034 method Methods 0.000 title claims abstract description 125
- 238000001514 detection method Methods 0.000 title claims abstract description 54
- 230000006870 function Effects 0.000 claims description 24
- 238000013527 convolutional neural network Methods 0.000 claims description 15
- 238000005070 sampling Methods 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 9
- 238000012935 Averaging Methods 0.000 claims 2
- 238000004891 communication Methods 0.000 description 27
- 238000010586 diagram Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 7
- 230000006399 behavior Effects 0.000 description 6
- 239000011521 glass Substances 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 230000002787 reinforcement Effects 0.000 description 5
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Definitions
- the present application relates to artificial intelligence technology, and in particular, to an object attribute detection method and apparatus, a neural network training method and apparatus, and an area detection method and apparatus, and an electronic apparatus.
- Convolutional neural networks are an important research field for computer vision and pattern recognition. They use computer-like biological brain thinking to inspire information processing similar to humans to specific objects. Through the convolutional neural network, object detection and recognition can be performed efficiently. With the development of Internet technology and the rapid increase of information volume, convolutional neural networks are more and more widely used in the field of object detection and recognition to find out the actual information needed from a large amount of information.
- the embodiment of the present application provides an object attribute detection scheme, a neural network training scheme, and an area detection scheme.
- an object attribute detecting method includes: inputting an image to be detected into an attention neural network for area detection, and obtaining an image of the object to be detected in the image to be detected. At least one target area; inputting the to-be-detected image and the at least one target area into an attribute classification neural network for attribute detection, and obtaining object attribute information of the to-be-detected image.
- a neural network training method including: inputting a training sample image into an attention neural network for area training, obtaining probability information of a candidate target area; and according to the candidate target area
- the probability information is used to sample the candidate target region of the training sample image to obtain the sampled image sample; and input the attribute information of the target region and the image sample into the auxiliary classification network for attribute training to obtain the image sample.
- the accuracy information of the candidate target area; the attribute information of the target area is attribute information of the target area marked for the training sample image; and the parameter of the attention neural network is adjusted according to the accuracy information.
- a region detecting method includes: acquiring a target image to be detected, wherein the target image includes a still image or a video image; and detecting the target image by using an attention neural network Obtaining a target area of the target image; wherein the attention neural network is trained using a neural network training method as described in any of the embodiments of the present application.
- an object attribute detecting apparatus comprising: a first acquiring module, configured to input an image to be detected into an attention neural network for area detection, and obtain the image to be detected. At least one target area associated with the object attribute of the target; a second obtaining module, configured to input the image to be detected and the at least one target area into an attribute classification neural network for attribute detection, and obtain the image to be detected Object property information.
- a neural network training apparatus including: a sixth acquiring module, configured to input a training sample image into an attention neural network for area training, and obtain probability information of a candidate target area.
- a seventh acquiring module configured to perform candidate target region sampling on the training sample image according to probability information of the candidate target region, to obtain a sampled image sample
- an eighth obtaining module configured to: perform attribute information of the target region and And inputting the image sample into the auxiliary classification network for attribute training, obtaining accuracy information of the candidate target area in the image sample; and the attribute information of the target area is attribute information of the target area marked for the training sample image
- the second parameter adjustment module is configured to adjust parameters of the attention neural network according to the accuracy information.
- an area detecting apparatus including: a ninth obtaining module, configured to acquire a target image to be detected, where the target image includes a still image or a video image; a module, configured to detect the target image by using an attention neural network to obtain a target area of the target image; wherein the attention neural network adopts a neural network training method or a neural network according to any embodiment of the present application
- the training device is trained.
- an electronic device including:
- the memory is configured to store at least one executable instruction, the executable instruction causing the processor to perform an operation corresponding to the object attribute detecting method according to any one of the embodiments of the present application; or the memory is configured to store at least one An executable instruction, the executable instruction causing the processor to perform an operation corresponding to the neural network training method according to any one of the embodiments of the present application; or the memory is configured to store at least one executable instruction, the executable The instructions cause the processor to perform operations corresponding to the area detecting method described in any of the embodiments of the present application.
- another electronic device including:
- the processor and the object attribute detecting apparatus according to any one of the embodiments of the present application; when the processor runs the object attribute detecting apparatus, the unit in the object attribute detecting apparatus according to any one of the embodiments of the present application is operated; or
- the processor and the neural network training device according to any one of the embodiments of the present application; when the processor runs the neural network training device, the unit in the neural network training device according to any one of the embodiments of the present application is operated; or
- a computer program comprising computer readable code, when the computer readable code is run on a device, the processor in the device is operative to implement any of the present application An instruction of each step in the object attribute detecting method described in the embodiment; or
- the processor in the device executes instructions for implementing the steps in the neural network training method described in any of the embodiments of the present application.
- the processor in the device executes instructions for implementing the steps in the region detecting method of any of the embodiments of the present application.
- a computer readable storage medium for storing computer readable instructions, and when the instructions are executed, implementing an object attribute detecting method according to any one of the embodiments of the present application.
- an Attention neural network is used to detect an area of a target in an image to be inspected, and then an image region detected by the attention neural network is input into an attribute classification neural network to perform attribute detection of the target. Obtain the corresponding object attribute information.
- the trained attention neural network can accurately detect the target area in the image, and perform targeted attribute detection on the area to obtain more accurate object attribute information.
- FIG. 1 is a flowchart of an object attribute detecting method according to an embodiment of the present application.
- FIG. 2 is a flowchart of an object attribute detecting method according to an embodiment of the present application.
- FIG. 3 is a flowchart of a neural network training method according to an embodiment of the present application.
- FIG. 5 is a structural block diagram of an object attribute detecting apparatus according to an embodiment of the present application.
- FIG. 6 is a structural block diagram of a neural network training apparatus according to an embodiment of the present application.
- FIG. 7 is a structural block diagram of an area detecting apparatus according to an embodiment of the present application.
- FIG. 8 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
- Embodiments of the present application can be applied to electronic devices such as terminal devices, computer systems, servers, etc., which can operate with numerous other general purpose or special purpose computing system environments or configurations.
- Examples of well-known terminal devices, computing systems, environments, and/or configurations suitable for use with electronic devices such as terminal devices, computer systems, servers, and the like include, but are not limited to, personal computer systems, server computer systems, thin clients, thick clients Machines, handheld or laptop devices, microprocessor-based systems, set-top boxes, programmable consumer electronics, networked personal computers, small computer systems, mainframe computer systems, and distributed cloud computing technology environments including any of the above, and the like.
- Electronic devices such as terminal devices, computer systems, servers, etc., can be described in the general context of computer system executable instructions (such as program modules) being executed by a computer system.
- program modules may include routines, programs, target programs, components, logic, data structures, and the like that perform particular tasks or implement particular abstract data types.
- the computer system/server can be implemented in a distributed cloud computing environment where tasks are performed by remote processing devices that are linked through a communication network.
- program modules may be located on a local or remote computing system storage medium including storage devices.
- the object attribute detecting method of this embodiment includes the following steps:
- Step S102 input the image to be detected into the attention neural network for area detection, and obtain at least one local area associated with the object attribute of the target in the image to be detected as the target area.
- the image to be examined in each embodiment of the present application may include a still image or a video image.
- the object attribute of the target in the image to be detected is a preset attribute to be detected.
- the detection of the face attribute in the image to be inspected includes, but is not limited to, one or more of the following: whether or not the glasses are worn, whether or not there is Wearing a hat with or without a mask; for example, detection of vehicle attributes in the image to be examined, including but not limited to: vehicle color, style, license plate number, and the like.
- the attention neural network is applied to image recognition of deep learning, and when the image is imitated by a person, the focus of the eye moves on different objects.
- the neural network recognizes the image, it is more accurate to focus on each feature at a time.
- the attention neural network can calculate the weight of each feature, and then weight the sum of the features. The larger the weight, the greater the contribution of the feature to the current recognition.
- the target area is a partial area of the image to be detected.
- the trained attentional neural network has automatic target area detection capability, and the image to be detected is input into the attention neural network, and the corresponding target area can be obtained, and the target area can be one or multiple, such as multiple faces. Area, so that attribute detection can be performed on multiple faces at the same time.
- the attention neural network may be a neural network that has been trained to be directly used by a third party, or may be an attention neural network obtained through sample training, such as the training obtained by the method described in the following embodiments of the present application. Force neural network.
- the step S102 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by the first acquisition module 502 being executed by the processor.
- Step S104 Input the image to be detected and the at least one target area into the attribute classification neural network for attribute detection, and obtain object attribute information of the image to be detected.
- the attribute classification neural network may adopt any appropriate network form, such as VGG-16 neural network, GoogleNet neural network, etc., and the training may also adopt a conventional training method, so that the trained network has attribute classification and recognition functions. For example, the gender, age, wearing, etc. of the pedestrian can be identified.
- the input of the attribute classification neural network is the entire image to be detected and the target area determined by the attention neural network, such as the head area of the human body, and the output is the value of the attribute of the target area, such as the value of the attribute of the head.
- the step S104 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a second acquisition module 504 being executed by the processor.
- the method further includes: displaying the object attribute information in the image to be detected.
- the operation of displaying the object attribute information in the image to be examined may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a display module 506 executed by the processor.
- the attention neural network is used to detect the target region in the image to be inspected, and then the image region detected by the attention neural network is input into the attribute classification neural network to perform target attribute detection, and corresponding object attribute information is obtained.
- the trained attention neural network can accurately detect the target region (ie, the target region) in the image, and perform targeted attribute detection on the target region to obtain more accurate target object attribute information.
- the attention neural network for detecting the region corresponding to the target may be trained, and then the object property detection is performed using the trained attention neural network.
- the object attribute detecting method of this embodiment includes the following steps:
- Step S202 Using the training sample image and the auxiliary classification network, the attention neural network is trained as a neural network for detecting a target area in the image.
- the step S202 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by the first training module 508 being executed by the processor.
- this step S202 may include:
- Step S2022 Input the training sample image into the attention neural network for area training, and obtain probability information of the candidate target area.
- the training sample image may be appropriately selected by a person skilled in the art according to actual needs, and may include, for example, but not limited to, a person sample image and a vehicle sample image.
- the attention neural network in the embodiments of the present application can be considered as a convolution network that introduces an attention mechanism. After the attention mechanism is introduced, the convolutional network can determine the degree of influence of each candidate target region in the image on the final target region during image training. This degree of influence is usually expressed in the form of probability, that is, candidates. Probability information for the target area.
- a plurality of candidate target regions are usually included.
- the probability values of each candidate target region in the image may be the final target region.
- all images in the training sample set are processed by the attention neural network to obtain probability values for which each candidate target region in the respective image may be the final target region.
- a plurality of candidate regions are probability values of the head region.
- the attention sample neural network is taken as an example to realize the attention neural network to the corresponding target area of the character, such as the head area, the upper body area, the lower body area, the foot area, the hand area, and the like.
- auto recognition Those skilled in the art can refer to the training of the sample image of the person to realize the training of other sample images, such as the image of the vehicle sample, such as the automatic attention of the attention neural network to the corresponding target area of the vehicle, such as the vehicle brand area, the vehicle sign area, the vehicle body area, and the like. Identification.
- Step S2024 Perform sampling of the candidate target region on the training sample image according to the probability information of the candidate target region, and obtain the sampled sampled image.
- the candidate target region with a larger probability value is more likely to be sampled.
- it is sampled, and some or a plurality of candidate target regions of the sample image may be acquired.
- the number of samples may be appropriately set by a person skilled in the art according to actual needs, and the embodiment of the present application does not limit this.
- the polynomial distribution corresponding to the probability value of the candidate target region may be determined first; then, according to the polynomial distribution, the candidate target region is sampled for each training sample image, and the sampled image sample is obtained.
- information of the target area in the collected training sample image can be obtained, and the feature map of the corresponding target area can be obtained by using the information.
- Step S2026 input attribute information and image samples of the target area into the auxiliary classification network for attribute training, obtain accuracy information of the candidate target area in the image sample, and adjust network parameters of the attention neural network according to the accuracy information
- Network parameters may include, for example, but are not limited to, weights, biases, and the like.
- the attribute information of the target area is attribute information of the target area marked for the training sample image.
- the attribute information of the target area is used to represent the attribute of the object of the target area.
- the attribute information may include, but is not limited to, one or more of the following: gender, age, hairstyle, whether Wear glasses, wear masks, etc.
- the sampled image sample contains information of the sampled area, including which area is collected, and the corresponding feature map of the area.
- the attribute information of the target area needs to be acquired first.
- the attribute information of the target area may be input together with the training sample image at the initial time, wherein the training sample image is input to the attention.
- the neural network, and the attribute information of the target area is input to the auxiliary classification network.
- the information of the target area may also be input into the attention neural network together with the training sample image, and then transmitted to the auxiliary classification network by the attention neural network; or may be temporarily passed when the sampled sample is input. Get it in the right way.
- the auxiliary classification network is used to implement the reinforcement learning of the attention neural network.
- the auxiliary classification network may adopt any suitable network capable of implementing reinforcement learning.
- Reinforcement learning as a problem of Sequential Decision Making, continuously selects some behaviors and obtains the greatest return from these behaviors as the best result. It does not have a label to tell the algorithm what to do, by first trying to make some behavior, and then getting a result, by judging whether the result is right or wrong to feedback the previous behavior. This feedback is used to adjust the previous behavior, and through continuous adjustment algorithms, we can learn under what circumstances what kind of behavior can be selected to get the best results.
- the auxiliary classification network determines whether the probability estimation of the corresponding candidate target region by the attention neural network is accurate by calculating the reward value of each candidate target region in each sampled image sample, and then determines How to adjust the network parameters of the attention neural network to make the prediction of the attention neural network more accurate.
- the attribute information and the image sample of the target area are input into the auxiliary classification network for attribute training, and the loss value of the attribute information of the candidate target area in the image sample is obtained by the loss function of the auxiliary classification network.
- the loss function is determined according to the attribute information of the target area; then, according to the obtained loss value, the reward value of the candidate target area in the image sample is determined, and the reward value is the accuracy information.
- the loss values of at least one candidate target region of the at least one image sample may be first averaged to obtain an average value; and the candidate in the sampled image sample is determined according to the relationship between the average value and the obtained loss value.
- the return value of the target area may be first averaged to obtain an average value; and the candidate in the sampled image sample is determined according to the relationship between the average value and the obtained loss value. The return value of the target area.
- the return value of the candidate target area corresponding to the loss value is set as the first return value; otherwise, the candidate corresponding to the loss value is The return value of the target area is set to the second return value.
- the setting criterion may be that the loss value is less than X times of the average value (for example, in an actual application, the X value may be 0.5), and the return value of the candidate target area corresponding to the loss value is set to 1; otherwise Set the return value of the candidate target area corresponding to the loss value to 0.
- the setting criterion may be appropriately set by a person skilled in the art according to actual conditions, for example, the loss value may be less than 0.5 times of the average value, and the first N or the like among the loss values from large to small, N is greater than An integer of 0.
- the adjusted network parameters of the attention neural network can make the target area obtained by the auxiliary classification network have a return value of 1 and the non-target area has a return value of 0, then the attention neural network training can be considered completed. Otherwise, the parameters of the attention neural network are continuously adjusted according to the reward value until the target area obtained by the auxiliary classification network has a return value of 1 and the non-target area has a return value of 0.
- Step S204 Input the image to be detected into the attention neural network for area detection, and obtain at least one local area associated with the object attribute of the target in the image to be detected as the target area.
- the trained attention neural network is capable of performing target region detection to detect at least one target region associated with the target object attribute
- the step S204 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by the first acquisition module 502 being executed by the processor.
- Step S206 Input the image to be detected and the at least one target area into the attribute classification neural network for attribute detection, and obtain object attribute information of the image to be inspected.
- the step S206 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a second acquisition module 504 being executed by the processor.
- the attention neural network is used to detect the target region in the image to be inspected, and then the image region detected by the attention neural network is input into the attribute classification neural network to perform target attribute detection, and corresponding object attribute information is obtained.
- the trained attention neural network can accurately detect the target area in the image, and perform targeted attribute detection on the area to obtain more accurate object attribute information.
- FIG. 3 a flow chart of a neural network training method in accordance with an embodiment of the present application is shown.
- the neural network training method of this embodiment includes the following steps:
- Step S302 Input the training sample image into the attention neural network for area training, and obtain probability information of the candidate target area.
- the attention sample neural network is still taken as an example to realize the automatic recognition of the corresponding target area of the character by the attention neural network.
- the probability information of the candidate target area may include a probability value of the candidate target area.
- the step S302 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a third acquisition module 5082 being executed by the processor.
- Step S304 Perform sampling of the candidate target region on the training sample image according to the probability information of the candidate target region, and obtain the sampled sampled image.
- the candidate target region with a larger probability value is more likely to be sampled.
- it is sampled, and some or a plurality of candidate target regions of the sample image may be acquired.
- the number of samples may be appropriately set by a person skilled in the art according to actual needs, and the embodiment of the present application does not limit this.
- the polynomial distribution corresponding to the probability value of the candidate target region may be determined first; then, according to the polynomial distribution, the candidate target region is sampled by the training sample image, and the sampled image sample is obtained.
- the step S304 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a fourth acquisition module 5084 being executed by the processor.
- Step S306 Input attribute information and image samples of the target area into the auxiliary classification network for attribute training, and obtain accuracy information of the candidate target area in the image sample.
- the attribute information of the target area is attribute information of the target area marked for the training sample image.
- the attribute information of the target area is used to represent the attributes of the object of the target area.
- the attribute information may include, but is not limited to, gender, age, hairstyle, whether to wear glasses, whether to wear a mask, or the like.
- the sampled image sample contains information of the sampled area, including which area is collected, and the corresponding feature map of the area.
- the auxiliary classification network determines whether the attention neural network determines the probability of the corresponding candidate target region accurately by calculating the return value of each candidate target region in each sampled sample, and then determines how to adjust the attention.
- the network parameters of the neural network to make the prediction of the attention neural network more accurate.
- the attribute information and the image of the target region may be obtained.
- the sample is input into the auxiliary classification network for attribute training, and the loss value of the attribute information of the candidate target area in the image sample is obtained by the loss function of the auxiliary classification network, wherein the loss function is determined according to the attribute information of the target area;
- the loss value determines a reward value of the candidate target area in the image sample, and the reward value is the accuracy information.
- the loss values of at least one candidate target region of the at least one image sample may be first averaged (eg, the loss values of the respective candidate target regions of all image samples are averaged) to obtain an average value; and then obtained according to the average value
- the relationship of the loss values determines the return value of the candidate target region in the sampled image sample.
- the obtained loss value is less than 0.5 times of the average value, and the loss value satisfies the set standard, the return value of the candidate target area corresponding to the loss value is set to 1; otherwise, the loss value is corresponding.
- the return value of the candidate target area is set to zero.
- the setting criteria may be appropriately set by a person skilled in the art according to actual conditions, such as setting the first N or the like among the loss values from large to small, and N is an integer greater than 0.
- step S306 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a fifth acquisition module 5086 being executed by the processor.
- Step S308 Adjust parameters of the attention neural network according to the accuracy information.
- the parameters of the adjusted attention neural network may include, for example, but are not limited to, network parameters such as weight parameters and offset parameters.
- the adjusted network parameters of the attention neural network can make the target area obtained by the auxiliary classification network have a return value of 1 and the non-target area has a return value of 0, then the attention neural network training can be considered completed. Otherwise, the parameters of the attention neural network are continuously adjusted according to the reward value until the target area obtained by the auxiliary classification network has a return value of 1 and the non-target area has a return value of 0.
- the training convergence condition of the above attention neural network is only one of the implementation solutions. It can be understood that, in practical applications, the attention neural network of the embodiment of the present application may also set other training convergence conditions, and the above training convergence conditions are An example should not be construed as the only implementation.
- step S308 may be performed by the processor invoking a corresponding instruction stored in the memory, or may be performed by the first parameter adjustment module 5088 being executed by the processor.
- an optional method is to separately train the attention neural network for different target areas, for example, in one training, only the attention neural network is trained on the head region of the character. In another training, only the attentional neural network is trained to predict the upper body area of the character.
- the following alternatives can be performed: using the trained attentional neural network to detect the training sample image, obtaining the target region of the training sample image; using the training sample image , a target area of each training sample image, and an attribute information training attribute classification neural network of each target area.
- the attribute classification neural network may adopt any appropriate network form, such as a convolutional neural network, and the training may also adopt a conventional training method.
- the recognition of the target area in the training sample image can be effectively learned and trained, and the attribute information of each target area can be effectively used in the target area in the recognized person image.
- the properties of the object are learned and trained.
- the attention neural network in the embodiment of the present application may be a full convolutional neural network.
- the convolutional layer parameters required by the full convolutional neural network are less. Training is faster.
- the attributes of the subject object in the image are often only related to certain areas of the subject, and do not require the characteristics of the entire image.
- the pedestrian attribute is often only related to certain body areas of the pedestrian, and does not require a whole pedestrian image.
- Features such as wearing glasses, wearing a hat, or wearing a mask are only required for the characteristics of the pedestrian's head.
- the attention mechanism based on the Reinforcement Learning method is adopted to enable the algorithm to automatically select the associated region of each attribute in the image, and then the feature of the associated region may be extracted, thereby utilizing the feature and
- the global characteristics of the image are used to predict the corresponding attributes, which not only saves the cost of manual labeling, but also finds areas that are better for training.
- the neural network training method of the present embodiment will be exemplarily illustrated with an optional example.
- the training attention neural network is used to identify the head region of a person as an example.
- the training process is as follows:
- the pedestrian attributes to be identified are manually classified according to their associated body parts, and the same attributes of the associated areas are classified into one category, such as wearing glasses, wearing a hat, and wearing a mask. These attributes only relate to the head of the pedestrian; Types, backpacks These attributes only relate to the upper body of the pedestrian.
- the attention neural network selects a batch of images as input.
- the attention neural network inputs a part of the data of the entire data set for training, that is, a batch of images, and the next batch of images is input in the next iteration.
- the attention neural network outputs a feature map for each image, at least one position in the feature map satisfies the polynomial distribution, and the value of the at least one position corresponds to Probability; then, each image randomly samples M regions from this polynomial distribution, and the probability of sampling each region in the M regions is the probability value corresponding to the region in the feature map, where M is an integer greater than 0,
- M is an integer greater than 0.
- the technical personnel in the field are appropriately set according to actual needs; each area sampled will pass through the auxiliary classification network, and the loss of the classification of an attribute is obtained by the loss function of the attribute classification in the auxiliary classification network; the loss of L is N ⁇ M areas
- the average value, N represents the number of image samples, and the loss of the M regions selected for each image is from small to large.
- Sorting if each area is located before the sorted queue top_k months, and less than 0.5L (ie: half the average loss), the return value of the region 1, and 0 otherwise.
- the top_k may be appropriately set by a person skilled in the art according to actual needs, and the embodiment of the present application does not limit this.
- each attribute is a multi-valued attribute
- each attribute can use the loss function (softmax function) to calculate the loss
- the final loss is the sum of the losses of all attributes.
- a loss function for attribute classification in an auxiliary classification network is as follows:
- the true label of the kth attribute of the nth image sample (determined according to the attribute value of the input header area),
- the label for this attribute output for the network is The probability, n and N are the number of image samples, and k is the number of attributes of the image sample.
- the input of the attentional neural network after training is a whole pedestrian image
- the output is the probability that each possible region in the image is the head
- the attention neural network is a full convolutional neural network.
- it can be 2 convolution layers plus one Softmax layer, and each convolution layer is followed by a ReLU layer.
- the output of the last convolutional layer before the Softmax layer is a feature map of a single channel, and then after passing through the Softmax layer, the value of each position of the feature map is that the location corresponding to the location in the original image can be selected as the header.
- the probability, the area with the highest probability, can be selected as the head area.
- the attention neural network in this embodiment is optimized by using reinforcement learning.
- the attention neural network does not calculate the loss immediately, but evaluates the return value of each region. The ultimate goal is to maximize the return value.
- the area is re-entered into an auxiliary classification network, and the loss function of the auxiliary classification network is the classification loss of the attributes related to the head area.
- the return value of each possible selection as the head region is determined by the classification effect of the region on the attributes of the header region after passing through the auxiliary classification network.
- the attentional neural network is trained by training the sample image, and each training sample image may include a plurality of different candidate target regions, and each candidate target region may be obtained through the attention neural network.
- a probability value of the target area further, after obtaining the probability value, sampling the corresponding training sample image according to the probability value, wherein a region with a larger probability value is more likely to be collected; after sampling, the target is
- the attribute information of the region is input into the auxiliary classification network together with the sampled image, and the reward value of the collected region is calculated by the auxiliary classification network, and then the network parameters of the attention neural network are adjusted according to the reward value until the attention neural network is satisfied. Convergence conditions, complete the training of attention neural networks.
- the attention neural network can automatically find the image.
- the region with the largest response to the attribute corresponding to the attribute information it is not necessary to manually mark the training sample, which not only saves the cost of manual labeling, but also finds the best area corresponding to the attribute information, and reduces the cost of the convolutional neural network training process. Reduced training time.
- the neural network training method of this embodiment may be performed by any suitable device having data processing capabilities, including but not limited to: a PC or the like.
- the area detecting method of this embodiment includes the following steps:
- Step S402 Acquire a target image to be detected.
- the target image may include a still image or a video image.
- the video image may include a pedestrian image or a vehicle image in video surveillance.
- the corresponding target area such as the head area of a certain human body or the area where a certain vehicle is located, may be first located, and corresponding attribute recognition is performed for the target area.
- the step S402 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a ninth acquisition module 702 being executed by the processor.
- Step S404 detecting the target image by using an attention neural network to obtain a target area of the target image.
- an attention neural network trained by the method shown in any of the above embodiments is employed.
- the target area of the target image can be quickly and accurately located, and the target area can be processed according to actual needs, such as attribute recognition, image information acquisition, and area positioning.
- the target area when the target image is a person image, the target area may include, but is not limited to, any one or more of the following: a head, an upper body, a lower body, a foot, a hand; when the target image is a vehicle image
- the target area may include, for example but not limited to, any one or more of the following: a vehicle brand area, a vehicle sign area, and a body area.
- the step S404 may be performed by a processor invoking a corresponding instruction stored in the memory, or may be performed by a tenth acquisition module 704 that is executed by the processor.
- the area detection method in this embodiment can accurately and effectively detect and locate the target area in the image, reduce the target area positioning cost, and improve the target area positioning efficiency.
- the area detecting method of this embodiment may be performed by any suitable device having data processing capabilities, including but not limited to: a PC or the like.
- any of the methods provided by the embodiments of the present application may be performed by any suitable device having data processing capabilities, including but not limited to: a terminal device, a server, and the like.
- any of the methods provided by the embodiments of the present application may be executed by a processor, such as the processor, by executing a corresponding instruction stored in the memory to perform any of the methods mentioned in the embodiments of the present application. This will not be repeated below.
- the foregoing program may be stored in a computer readable storage medium, and the program is executed when executed.
- the foregoing steps include the steps of the foregoing method embodiments; and the foregoing storage medium includes: a medium that can store program codes, such as a ROM, a RAM, a magnetic disk, or an optical disk.
- the object attribute detecting apparatus of the embodiment includes: a first acquiring module 502, configured to input the image to be detected into the attention neural network for area detection, and obtain at least one local area in the image to be detected that is associated with the object attribute of the target.
- the second obtaining module 504 is configured to input the to-be-detected image and the at least one target area into the attribute classification neural network for attribute detection, and obtain the to-be-checked object attribute information.
- the object attribute detecting apparatus of the embodiment further includes: a display module 506, configured to display the object attribute information in the image to be detected.
- the target area may include any one or more of the following: a head, an upper body, a lower body, a foot, a hand; and/or, when the target image is a vehicle image,
- the target area may include, for example, any one or more of the following: a vehicle brand area, a vehicle sign area, and a body area.
- the image to be examined may include a still image or a video image.
- the video image may include a pedestrian image and/or a vehicle image in video surveillance.
- the object attribute detecting apparatus of the embodiment further includes: a first training module 508, configured to use the training sample image and before the first acquiring module 502 inputs the image to be detected into the attention neural network for area detection.
- the auxiliary classification network trains the attention neural network as a neural network for detecting a target area in the image.
- the first training module 508 includes: a third obtaining module 5082, configured to input the training sample image into the attention neural network for area training, to obtain probability information of the candidate target area; and a fourth obtaining module 5084, configured to: The candidate target region is sampled according to the probability information of the candidate target region, and the sampled image sample is obtained.
- the fifth obtaining module 5086 is configured to input the attribute information and the image sample of the target region into the auxiliary classification network for attribute training. And obtaining the accuracy information of the candidate target area in the image sample; the attribute information of the target area is attribute information of the target area marked for the training sample image; the first parameter adjustment module 5088 is configured to adjust the attention neural network according to the accuracy information Network parameters.
- the fifth obtaining module 5086 includes: a first loss obtaining module 50862, configured to input attribute information and image samples of the target area into the auxiliary classification network for attribute training, and obtain image samples by using a loss function of the auxiliary classification network. And a loss value of the attribute information of the candidate target area, wherein the loss function is determined according to the attribute information of the target area; the first report obtaining module 50864 is configured to determine, according to the obtained loss value, a return value of the candidate target area in the image sample The return value is the accuracy information.
- the first report obtaining module 50864 is configured to average the loss values of the at least one candidate target region of the at least one image sample to obtain an average value; and determine the candidate in the image sample according to the relationship between the average value and the obtained loss value. The return value of the target area.
- the first report obtaining module 50864 is configured to average the loss values of the at least one candidate target region of the at least one image sample to obtain an average value; if the obtained loss value satisfies the setting criterion, the candidate corresponding to the loss value is obtained.
- the return value of the target area is set to the first return value; otherwise, the return value of the candidate target area corresponding to the loss value is set as the second return value.
- the fourth obtaining module 5084 is configured to determine a polynomial distribution corresponding to the probability value of the candidate target region; according to the polynomial distribution, the candidate target region is sampled by the training sample image, and the sampled image sample is obtained.
- the attention neural network comprises a full convolutional neural network.
- the object attribute detecting apparatus of the embodiment further includes: a second training module 510, configured to detect the training sample image by using the trained attention neural network to obtain a target area of the training sample image; and use the training sample image, at least A target area of the training sample image, and an attribute information training attribute classification neural network of at least one target area.
- a second training module 510 configured to detect the training sample image by using the trained attention neural network to obtain a target area of the training sample image; and use the training sample image, at least A target area of the training sample image, and an attribute information training attribute classification neural network of at least one target area.
- the object attribute detecting apparatus of the present embodiment can be used to implement the corresponding object attribute detecting method in the foregoing multiple method embodiments, and has the beneficial effects of the corresponding method embodiments, and details are not described herein again.
- the neural network training device of the present embodiment includes: a sixth obtaining module 602, configured to input a training sample image into the attention neural network for area training, and obtain probability information of the candidate target region; and a seventh obtaining module 604, configured to The probability information of the candidate target region is used to sample the candidate target region of the training sample image to obtain the sampled image sample.
- the eighth obtaining module 606 is configured to input the attribute information and the image sample of the target region into the auxiliary classification network for attribute training. Obtaining accuracy information of the candidate target region in the image sample; the attribute information of the target region is attribute information of the target region marked for the training sample image; and the second parameter adjustment module 608 is configured to adjust the attention neural network according to the accuracy information parameter.
- the eighth obtaining module 606 includes: a second loss obtaining module 6062, configured to input attribute information and image samples of the target area into the auxiliary classification network for attribute training, and obtain an image sample by using a loss function of the auxiliary classification network.
- the second report obtaining module 6064 is configured to determine the return value of the candidate target area in the image sample according to the obtained loss value
- the return value is the accuracy information.
- the second report obtaining module 6064 is configured to average the loss values of the at least one candidate target region of the at least one image sample to obtain an average value; and determine the candidate in the image sample according to the relationship between the average value and the obtained loss value. The return value of the target area.
- the second report obtaining module 6064 is configured to average the loss values of the at least one candidate target region of the at least one image sample to obtain an average value; if the obtained loss value satisfies the setting criterion, the candidate corresponding to the loss value is obtained.
- the return value of the target area is set to the first return value; otherwise, the return value of the candidate target area corresponding to the loss value is set as the second return value.
- the seventh obtaining module 604 is configured to determine a polynomial distribution corresponding to the probability value of the candidate target region; according to the polynomial distribution, the candidate target region is sampled by the training sample image, and the sampled image sample is obtained.
- the attention neural network is a full convolutional neural network.
- the neural network training device of the embodiment further includes: a third training module 610, configured to detect the training sample image by using the trained attention neural network to obtain a target region of the training sample image; and use the training sample image, at least A target area of the training sample image, and an attribute information training attribute classification neural network of at least one target area.
- a third training module 610 configured to detect the training sample image by using the trained attention neural network to obtain a target region of the training sample image; and use the training sample image, at least A target area of the training sample image, and an attribute information training attribute classification neural network of at least one target area.
- the neural network training device of the present embodiment is used to implement the corresponding neural network training method in the foregoing multiple method embodiments, and has the beneficial effects of the corresponding method embodiments, and details are not described herein again.
- the area detecting device of this embodiment includes: a ninth obtaining module 702, configured to acquire a target image to be detected, wherein the target image includes a still image or a video image; and a tenth acquiring module 704 is configured to detect the target by using an attention neural network.
- the image is obtained from the target area of the target image; wherein the attention neural network is trained by using the neural network training method or the neural network training device according to any of the above embodiments of the present application.
- the target area may include any one or more of the following: a head, an upper body, a lower body, a foot, a hand; when the target image is a vehicle image, the target area may include the following: Any one or more of: vehicle grade area, vehicle sign area, body area.
- the video image includes a pedestrian image or a vehicle image in video surveillance.
- the area detecting device of the present embodiment can be used to implement the corresponding area detecting method in the foregoing multiple method embodiments, and has the beneficial effects of the corresponding method embodiments, and details are not described herein again.
- an embodiment of the present application further provides an electronic device, including: a processor and a memory;
- the memory is configured to store at least one executable instruction, the executable instruction causing the processor to perform an operation corresponding to the object attribute detecting method according to any one of the foregoing embodiments of the present application; or
- the memory is configured to store at least one executable instruction, the executable instruction causing the processor to perform an operation corresponding to the neural network training method described in any one of the foregoing embodiments of the present application; or
- the memory is configured to store at least one executable instruction, the executable instruction causing the processor to perform an operation corresponding to the area detecting method according to any one of the foregoing embodiments of the present application.
- the embodiment of the present application further provides another electronic device, including:
- the processor and the object attribute detecting apparatus according to any one of the above embodiments of the present application; when the processor runs the object attribute detecting apparatus, the unit in the object attribute detecting apparatus according to any one of the above embodiments of the present application is operated; or
- the processor and the neural network training device according to any one of the above embodiments of the present application; when the processor runs the neural network training device, the unit in the neural network training device according to any of the above embodiments of the present application is operated; or
- the processor and the area detecting device according to any of the above embodiments of the present application; when the processor runs the area detecting device, the unit in the area detecting device according to any of the above embodiments of the present application is operated.
- the embodiment of the present application further provides an electronic device, such as a mobile terminal, a personal computer (PC), a tablet computer, a server, and the like.
- electronic device 800 includes one or more first processors, first communication elements, etc., such as one or more central processing units (CPUs) 801, and / or one or more image processor (GPU) 813 or the like, the first processor may be loaded into the random access memory (RAM) 803 according to executable instructions stored in read only memory (ROM) 802 or from storage portion 808.
- the executable instructions execute various appropriate actions and processes.
- the first read only memory 802 and the random access memory 803 are collectively referred to as a first memory.
- the first communication component includes a communication component 812 and/or a communication interface 809.
- the communication component 812 can include, but is not limited to, a network card.
- the network card can include, but is not limited to, an IB (Infiniband) network card.
- the communication interface 809 includes a communication interface of a network interface card such as a LAN card, a modem, etc., and the communication interface 809 is via an Internet interface.
- the network performs communication processing.
- the first processor can communicate with read only memory 802 and/or random access memory 803 to execute executable instructions, connect to communication component 812 via first communication bus 804, and communicate with other target devices via communication component 812 to complete
- the operation corresponding to any object attribute detecting method provided by the embodiment of the present application, for example, inputting the image to be detected into the attention neural network for area detection, and obtaining at least one target area in the image to be detected that is associated with the object attribute of the target. Entering the image to be inspected and the at least one target area into the attribute classification neural network for attribute detection, and obtaining object attribute information of the image to be inspected.
- the first processor can communicate with read only memory 802 and/or random access memory 803 to execute executable instructions, connect to communication component 812 via first communication bus 804, and communicate with other target devices via communication component 812,
- the operation corresponding to any of the neural network training methods provided by the embodiment of the present application is completed, for example, the training sample image is input into the attention neural network for area training, and probability information of the candidate target area is obtained; according to the candidate target area Probabilistic information is used to sample candidate image regions of the training sample image to obtain sampled image samples; and input attribute information of the target region and the image samples into an auxiliary classification network for attribute training to obtain candidates in the image samples.
- the accuracy information of the target area; the attribute information of the target area is attribute information of the target area marked for the training sample image; and the parameter of the attention neural network is adjusted according to the accuracy information.
- the first processor can communicate with read only memory 802 and/or random access memory 803 to execute executable instructions, connect to communication component 812 via first communication bus 804, and communicate with other target devices via communication component 812,
- the operation corresponding to any area detection method provided by the embodiment of the present application is completed, for example, acquiring a target image to be detected, where the target image includes a still image or a video image; and the target image is detected by using an attention neural network.
- RAM 803 various programs and data required for the operation of the device can be stored.
- the CPU 801 or GPU 813, the ROM 802, and the RAM 803 are connected to each other through the first communication bus 804.
- ROM 802 is an optional module.
- the RAM 803 stores executable instructions or writes executable instructions to the ROM 802 at runtime, the executable instructions causing the first processor to perform operations corresponding to the above-described communication methods.
- An input/output (I/O) interface 805 is also coupled to the first communication bus 804.
- the communication component 812 can be integrated or can be configured to have multiple sub-modules (e.g., multiple IB network cards) and be on a communication bus link.
- the following components are connected to the I/O interface 805: an input portion 806 including a keyboard, a mouse, etc.; an output portion 807 including, for example, a cathode ray tube (CRT), a liquid crystal display (LCD), and the like, and a storage portion 808 including a hard disk or the like. And a communication interface 809 including a network interface card such as a LAN card, modem, or the like.
- Driver 810 is also coupled to I/O interface 805 as needed.
- a removable medium 811 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like, is mounted on the drive 810 as needed so that a computer program read therefrom is installed into the storage portion 808 as needed.
- FIG. 8 is only an optional implementation manner.
- the number and type of components in FIG. 8 may be selected, deleted, added, or replaced according to actual needs;
- Different function components can also be implemented in separate settings or integrated settings, such as GPU and CPU detachable settings or GPU can be integrated on the CPU, communication components can be separated, or integrated on the CPU or GPU. ,and many more.
- These alternative embodiments are all within the scope of the present application.
- embodiments of the present application include a computer program product comprising a computer program tangibly embodied on a machine readable medium, the computer program comprising program code for executing the method illustrated in the flowchart, the program code comprising the corresponding execution
- the instructions corresponding to the method steps provided by any embodiment of the present application may include an instruction corresponding to the following steps provided in the embodiment of the present application: inputting the image to be detected into the attention neural network for area detection, and obtaining at least one of the to-be-detected image associated with the object attribute of the target.
- Target area inputting the image to be inspected and the at least one target area into the attribute classification neural network for attribute detection, and obtaining object attribute information of the image to be inspected.
- the program code may include an instruction corresponding to the following steps provided in the embodiment of the present application: inputting the training sample image into the attention neural network for area training, and obtaining probability information of the candidate target area; according to the candidate target area The probability information is used to sample the candidate target region of the training sample image to obtain the sampled image sample; and input the attribute information of the target region and the image sample into the auxiliary classification network for attribute training to obtain the image sample.
- the program code may include an instruction corresponding to the following steps provided in the embodiment of the present application: acquiring a target image to be detected, wherein the target image includes a still image or a video image; and detecting the target by using an attention neural network And obtaining a target area of the target image; wherein the attention neural network is trained using a neural network training method as described in any of the embodiments of the present application.
- the computer program can be downloaded and installed from the network via a communication component, and/or installed from the removable media 811. The above-described functions defined in the method of any of the embodiments of the present application are executed when the computer program is executed by the first processor.
- the embodiment of the present application further provides a computer program, including computer readable code, when the computer readable code is run on a device, the processor in the device executes to implement any of the embodiments of the present application.
- the processor in the device executes instructions for implementing the steps in the neural network training method described in any of the embodiments of the present application.
- the processor in the device executes instructions for implementing the steps in the region detecting method as described in any of the embodiments of the present application.
- the embodiment of the present application further provides a computer readable storage medium, which is configured to store computer readable instructions, and when the instructions are executed, implement the steps in the object attribute detecting method according to any embodiment of the present application.
- a computer readable storage medium configured to store computer readable instructions, and when the instructions are executed, implement the steps in the object attribute detecting method according to any embodiment of the present application.
- At least one embodiment of the present specification is described in a progressive manner, and at least one embodiment focuses on differences from other embodiments, and the same or similar parts between at least one embodiment may be referred to each other.
- the description is relatively simple, and the relevant parts can be referred to the description of the method embodiment.
- the methods, apparatus, and apparatus of the present application may be implemented in a number of ways.
- the method, apparatus, and apparatus of the embodiments of the present application can be implemented by software, hardware, firmware, or any combination of software, hardware, and firmware.
- the above-described sequence of steps for the method is for illustrative purposes only, and the steps of the method of the embodiments of the present application are not limited to the order of the above optional description unless otherwise specified.
- the present application may also be embodied as a program recorded in a recording medium, the programs including machine readable instructions for implementing a method in accordance with embodiments of the present application.
- the present application also covers a recording medium storing a program for executing the method according to an embodiment of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Mathematical Physics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Software Systems (AREA)
- Evolutionary Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Multimedia (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201611246395.9 | 2016-12-29 | ||
CN201611246395.9A CN108229267B (zh) | 2016-12-29 | 2016-12-29 | 对象属性检测、神经网络训练、区域检测方法和装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018121690A1 true WO2018121690A1 (fr) | 2018-07-05 |
Family
ID=62657290
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2017/119535 WO2018121690A1 (fr) | 2016-12-29 | 2017-12-28 | Procédé et dispositif de détection d'attribut d'objet, procédé et dispositif d'apprentissage de réseau neuronal et procédé et dispositif de détection zonale |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN108229267B (fr) |
WO (1) | WO2018121690A1 (fr) |
Cited By (82)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110046632A (zh) * | 2018-11-09 | 2019-07-23 | 阿里巴巴集团控股有限公司 | 模型训练方法和装置 |
CN110059721A (zh) * | 2019-03-16 | 2019-07-26 | 平安城市建设科技(深圳)有限公司 | 户型图区域识别方法、装置、设备及计算机可读存储介质 |
CN110443222A (zh) * | 2019-08-14 | 2019-11-12 | 北京百度网讯科技有限公司 | 用于训练脸部关键点检测模型的方法和装置 |
CN110458829A (zh) * | 2019-08-13 | 2019-11-15 | 腾讯医疗健康(深圳)有限公司 | 基于人工智能的图像质控方法、装置、设备及存储介质 |
CN110766152A (zh) * | 2018-07-27 | 2020-02-07 | 富士通株式会社 | 用于训练深度神经网络的方法和装置 |
CN110766129A (zh) * | 2018-07-27 | 2020-02-07 | 杭州海康威视数字技术股份有限公司 | 一种神经网络训练系统及显示数据的方法 |
CN110969657A (zh) * | 2018-09-29 | 2020-04-07 | 杭州海康威视数字技术股份有限公司 | 一种枪球坐标关联方法、装置、电子设备及存储介质 |
CN110969173A (zh) * | 2018-09-28 | 2020-04-07 | 杭州海康威视数字技术股份有限公司 | 目标分类方法及装置 |
CN111160429A (zh) * | 2019-12-17 | 2020-05-15 | 平安银行股份有限公司 | 图像检测模型的训练方法、图像检测方法、装置及设备 |
CN111191526A (zh) * | 2019-12-16 | 2020-05-22 | 汇纳科技股份有限公司 | 行人属性识别网络训练方法、系统、介质及终端 |
CN111241869A (zh) * | 2018-11-28 | 2020-06-05 | 杭州海康威视数字技术股份有限公司 | 物料盘点的方法、装置及计算机可读存储介质 |
CN111242951A (zh) * | 2020-01-08 | 2020-06-05 | 上海眼控科技股份有限公司 | 车辆检测方法、装置、计算机设备和存储介质 |
CN111259701A (zh) * | 2018-12-03 | 2020-06-09 | 杭州海康威视数字技术股份有限公司 | 行人再识别方法、装置及电子设备 |
CN111259763A (zh) * | 2020-01-13 | 2020-06-09 | 华雁智能科技(集团)股份有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN111292331A (zh) * | 2020-02-23 | 2020-06-16 | 华为技术有限公司 | 图像处理的方法与装置 |
CN111291597A (zh) * | 2018-12-07 | 2020-06-16 | 杭州海康威视数字技术股份有限公司 | 一种基于图像的人群态势分析方法、装置、设备及系统 |
CN111310775A (zh) * | 2018-12-11 | 2020-06-19 | Tcl集团股份有限公司 | 数据训练方法、装置、终端设备及计算机可读存储介质 |
CN111340090A (zh) * | 2020-02-21 | 2020-06-26 | 浙江每日互动网络科技股份有限公司 | 图像特征比对方法及装置、设备、计算机可读存储介质 |
CN111357014A (zh) * | 2018-09-19 | 2020-06-30 | 华为技术有限公司 | 一种ai模型的开发方法及装置 |
CN111368923A (zh) * | 2020-03-05 | 2020-07-03 | 上海商汤智能科技有限公司 | 神经网络训练方法及装置、电子设备和存储介质 |
CN111428671A (zh) * | 2020-03-31 | 2020-07-17 | 杭州博雅鸿图视频技术有限公司 | 人脸结构化信息识别方法、系统、装置及存储介质 |
CN111428536A (zh) * | 2019-01-09 | 2020-07-17 | 北京京东尚科信息技术有限公司 | 用于检测物品类别和位置的检测网络的训练方法和装置 |
CN111435364A (zh) * | 2019-01-14 | 2020-07-21 | 阿里巴巴集团控股有限公司 | 电子病历质检方法和装置 |
CN111435432A (zh) * | 2019-01-15 | 2020-07-21 | 北京市商汤科技开发有限公司 | 网络优化方法及装置、图像处理方法及装置、存储介质 |
CN111444749A (zh) * | 2019-01-17 | 2020-07-24 | 杭州海康威视数字技术股份有限公司 | 路面导向标志的识别方法、装置及存储介质 |
CN111459675A (zh) * | 2020-03-31 | 2020-07-28 | 拉扎斯网络科技(上海)有限公司 | 一种数据处理方法、装置、可读存储介质和电子设备 |
CN111507958A (zh) * | 2020-04-15 | 2020-08-07 | 全球能源互联网研究院有限公司 | 目标检测方法、检测模型的训练方法及电子设备 |
CN111523600A (zh) * | 2020-04-26 | 2020-08-11 | 上海商汤临港智能科技有限公司 | 神经网络训练、目标检测、及智能设备控制的方法及装置 |
CN111539481A (zh) * | 2020-04-28 | 2020-08-14 | 北京市商汤科技开发有限公司 | 图像标注方法、装置、电子设备及存储介质 |
CN111539452A (zh) * | 2020-03-26 | 2020-08-14 | 深圳云天励飞技术有限公司 | 多任务属性的图像识别方法、装置、电子设备及存储介质 |
CN111539947A (zh) * | 2020-04-30 | 2020-08-14 | 上海商汤智能科技有限公司 | 图像检测方法及相关模型的训练方法和相关装置、设备 |
CN111582107A (zh) * | 2020-04-28 | 2020-08-25 | 浙江大华技术股份有限公司 | 目标重识别模型的训练方法、识别方法、电子设备及装置 |
CN111598902A (zh) * | 2020-05-20 | 2020-08-28 | 北京字节跳动网络技术有限公司 | 图像分割方法、装置、电子设备及计算机可读介质 |
CN111612732A (zh) * | 2020-04-02 | 2020-09-01 | 深圳大学 | 图像质量评估方法、装置、计算机设备及存储介质 |
CN111832368A (zh) * | 2019-04-23 | 2020-10-27 | 长沙智能驾驶研究院有限公司 | 可行驶区域检测模型的训练方法、训练装置及应用 |
CN111860573A (zh) * | 2020-06-04 | 2020-10-30 | 北京迈格威科技有限公司 | 模型训练方法、图像类别检测方法、装置和电子设备 |
CN111967597A (zh) * | 2020-08-18 | 2020-11-20 | 上海商汤临港智能科技有限公司 | 神经网络训练及图像分类方法、装置、存储介质、设备 |
CN112016630A (zh) * | 2020-09-03 | 2020-12-01 | 平安科技(深圳)有限公司 | 基于图像分类模型的训练方法、装置、设备及存储介质 |
CN112101282A (zh) * | 2020-09-25 | 2020-12-18 | 北京瞰天科技有限公司 | 水上目标识别方法、装置及电子设备和存储介质 |
CN112101169A (zh) * | 2020-09-08 | 2020-12-18 | 平安科技(深圳)有限公司 | 基于注意力机制的道路图像目标检测方法及相关设备 |
CN112163545A (zh) * | 2020-10-12 | 2021-01-01 | 北京易华录信息技术股份有限公司 | 一种头部特征提取方法、装置、电子设备及存储介质 |
CN112184635A (zh) * | 2020-09-10 | 2021-01-05 | 上海商汤智能科技有限公司 | 目标检测方法、装置、存储介质及设备 |
CN112257604A (zh) * | 2020-10-23 | 2021-01-22 | 北京百度网讯科技有限公司 | 图像检测方法、装置、电子设备和存储介质 |
CN112307850A (zh) * | 2019-08-01 | 2021-02-02 | 浙江商汤科技开发有限公司 | 神经网络训练方法、车道线检测方法、装置和电子设备 |
CN112418261A (zh) * | 2020-09-17 | 2021-02-26 | 电子科技大学 | 一种基于先验原型注意力机制的人体图像多属性分类方法 |
CN112464785A (zh) * | 2020-11-25 | 2021-03-09 | 浙江大华技术股份有限公司 | 一种目标检测方法、装置、计算机设备及存储介质 |
CN112487226A (zh) * | 2020-11-06 | 2021-03-12 | 北京三快在线科技有限公司 | 图片分类模型获取方法、装置、电子设备及可读存储介质 |
CN112528995A (zh) * | 2020-12-22 | 2021-03-19 | 北京百度网讯科技有限公司 | 用于训练目标检测模型的方法、目标检测方法及装置 |
CN112529839A (zh) * | 2020-11-05 | 2021-03-19 | 西安交通大学 | 一种核磁共振图像中颈动脉血管中心线的提取方法及系统 |
CN112614117A (zh) * | 2020-12-28 | 2021-04-06 | 广州绿怡信息科技有限公司 | 设备区域提取模型训练方法、设备区域提取方法及装置 |
CN112700002A (zh) * | 2020-12-23 | 2021-04-23 | 北京三快在线科技有限公司 | 图神经网络的训练方法、装置、电子设备及可读存储介质 |
CN112712088A (zh) * | 2020-12-31 | 2021-04-27 | 洛阳语音云创新研究院 | 动物膘情检测方法、装置及计算机可读存储介质 |
CN112733578A (zh) * | 2019-10-28 | 2021-04-30 | 普天信息技术有限公司 | 车辆重识别方法及系统 |
CN112749609A (zh) * | 2020-07-23 | 2021-05-04 | 腾讯科技(深圳)有限公司 | 人体图像分割方法、装置、计算机设备及存储介质 |
CN112861858A (zh) * | 2021-02-19 | 2021-05-28 | 首都师范大学 | 显著性真值图的生成方法及显著性检测模型的训练方法 |
CN112906651A (zh) * | 2021-03-25 | 2021-06-04 | 中国联合网络通信集团有限公司 | 目标检测方法及装置 |
CN112906685A (zh) * | 2021-03-04 | 2021-06-04 | 重庆赛迪奇智人工智能科技有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
CN112925938A (zh) * | 2021-01-28 | 2021-06-08 | 上海商汤智能科技有限公司 | 一种图像标注方法、装置、电子设备及存储介质 |
CN112949767A (zh) * | 2021-04-07 | 2021-06-11 | 北京百度网讯科技有限公司 | 样本图像增量、图像检测模型训练及图像检测方法 |
CN113012176A (zh) * | 2021-03-17 | 2021-06-22 | 北京百度网讯科技有限公司 | 样本图像的处理方法、装置、电子设备及存储介质 |
CN113052175A (zh) * | 2021-03-26 | 2021-06-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN113454649A (zh) * | 2021-06-17 | 2021-09-28 | 商汤国际私人有限公司 | 目标检测方法、装置、电子设备和计算机可读存储介质 |
CN113469121A (zh) * | 2021-07-21 | 2021-10-01 | 浙江大华技术股份有限公司 | 一种车辆状态的识别方法及装置 |
CN113516144A (zh) * | 2020-12-01 | 2021-10-19 | 阿里巴巴集团控股有限公司 | 目标检测方法及装置、计算设备 |
CN113516013A (zh) * | 2021-04-09 | 2021-10-19 | 阿波罗智联(北京)科技有限公司 | 目标检测方法、装置、电子设备、路侧设备和云控平台 |
CN113573044A (zh) * | 2021-01-19 | 2021-10-29 | 腾讯科技(深圳)有限公司 | 视频数据处理方法、装置、计算机设备及可读存储介质 |
CN113642431A (zh) * | 2021-07-29 | 2021-11-12 | 北京百度网讯科技有限公司 | 目标检测模型的训练方法及装置、电子设备和存储介质 |
CN113688933A (zh) * | 2019-01-18 | 2021-11-23 | 北京市商汤科技开发有限公司 | 分类网络的训练方法及分类方法和装置、电子设备 |
CN113723159A (zh) * | 2021-02-26 | 2021-11-30 | 腾讯科技(深圳)有限公司 | 场景识别模型训练方法、场景识别方法及模型训练装置 |
CN113743535A (zh) * | 2019-05-21 | 2021-12-03 | 北京市商汤科技开发有限公司 | 神经网络训练方法及装置以及图像处理方法及装置 |
CN113742562A (zh) * | 2020-05-27 | 2021-12-03 | 北京达佳互联信息技术有限公司 | 视频推荐方法、装置、电子设备及存储介质 |
CN113822111A (zh) * | 2021-01-19 | 2021-12-21 | 北京京东振世信息技术有限公司 | 人群检测模型训练方法、装置以及人群计数方法、装置 |
CN113902999A (zh) * | 2021-09-24 | 2022-01-07 | 青岛海信网络科技股份有限公司 | 一种跟踪方法、装置、设备及介质 |
CN113963249A (zh) * | 2021-10-29 | 2022-01-21 | 山东大学 | 一种星系图像的检测方法和系统 |
CN113989772A (zh) * | 2021-10-27 | 2022-01-28 | 广州小鹏自动驾驶科技有限公司 | 一种交通灯检测方法、装置、车辆和可读存储介质 |
CN114387649A (zh) * | 2022-01-11 | 2022-04-22 | 北京百度网讯科技有限公司 | 图像处理方法、装置、电子设备以及存储介质 |
CN114571472A (zh) * | 2020-12-01 | 2022-06-03 | 北京小米移动软件有限公司 | 用于足式机器人的地面属性检测方法及驱动方法及其装置 |
CN114764919A (zh) * | 2021-01-14 | 2022-07-19 | 郑州信大先进技术研究院 | 一种基于深度学习的行人属性识别方法 |
CN109800654B (zh) * | 2018-12-24 | 2023-04-07 | 百度在线网络技术(北京)有限公司 | 车载摄像头检测处理方法、装置及车辆 |
CN111414930B (zh) * | 2019-01-07 | 2023-10-27 | 中国移动通信有限公司研究院 | 深度学习模型训练方法及装置、电子设备及存储介质 |
CN117037218A (zh) * | 2023-10-08 | 2023-11-10 | 腾讯科技(深圳)有限公司 | 对象属性识别方法、相关装置、设备和介质 |
CN118587746A (zh) * | 2024-08-06 | 2024-09-03 | 浙江大华技术股份有限公司 | 人体目标检测方法、装置、计算机设备和存储介质 |
Families Citing this family (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689030A (zh) * | 2018-07-04 | 2020-01-14 | 佳能株式会社 | 属性识别装置和方法及存储介质 |
CN108875076B (zh) * | 2018-07-10 | 2021-07-20 | 重庆大学 | 一种基于Attention机制和卷积神经网络的快速商标图像检索方法 |
CN109376757B (zh) * | 2018-09-06 | 2020-09-08 | 苏州飞搜科技有限公司 | 一种多标签分类方法及系统 |
CN111103629A (zh) * | 2018-10-25 | 2020-05-05 | 杭州海康威视数字技术股份有限公司 | 一种目标检测方法、装置、nvr设备及安检系统 |
CN111325052A (zh) * | 2018-12-13 | 2020-06-23 | 北京嘀嘀无限科技发展有限公司 | 一种目标检测方法及装置 |
CN109886072B (zh) * | 2018-12-25 | 2021-02-26 | 中国科学院自动化研究所 | 基于双向Ladder结构的人脸属性分类系统 |
CN111382734B (zh) * | 2018-12-29 | 2022-08-23 | 阿里巴巴集团控股有限公司 | 电话号码的检测及识别方法、装置及存储介质 |
CN111435452B (zh) * | 2019-01-11 | 2023-11-03 | 百度在线网络技术(北京)有限公司 | 模型训练方法、装置、设备和介质 |
CN110069997B (zh) * | 2019-03-22 | 2021-07-20 | 北京字节跳动网络技术有限公司 | 场景分类方法、装置及电子设备 |
CN111753857B (zh) * | 2019-03-26 | 2024-08-02 | 北京地平线机器人技术研发有限公司 | 一种应用于目标物的自动分类的模型训练方法、装置及电子设备 |
CN110059577B (zh) * | 2019-03-26 | 2022-02-18 | 北京迈格威科技有限公司 | 行人属性信息提取方法及装置 |
CN112001211B (zh) * | 2019-05-27 | 2024-04-19 | 商汤集团有限公司 | 对象检测方法、装置、设备及计算机可读存储介质 |
CN110210561B (zh) * | 2019-05-31 | 2022-04-01 | 北京市商汤科技开发有限公司 | 神经网络的训练方法、目标检测方法及装置、存储介质 |
CN110338835B (zh) * | 2019-07-02 | 2023-04-18 | 深圳安科高技术股份有限公司 | 一种智能扫描立体监测方法及系统 |
CN110378895A (zh) * | 2019-07-25 | 2019-10-25 | 山东浪潮人工智能研究院有限公司 | 一种基于深度注意力学习的乳腺癌图像识别方法 |
CN110458077B (zh) * | 2019-08-05 | 2022-05-03 | 高新兴科技集团股份有限公司 | 一种车辆颜色识别方法及系统 |
EP3779799A1 (fr) * | 2019-08-14 | 2021-02-17 | Robert Bosch GmbH | Procédé et système de classification fiable à l'aide d'un réseau neuronal |
CN110738211B (zh) * | 2019-10-17 | 2024-09-03 | 腾讯科技(深圳)有限公司 | 一种对象检测的方法、相关装置以及设备 |
CN112836549B (zh) * | 2019-11-22 | 2024-07-26 | 虹软科技股份有限公司 | 用户信息的检测方法及系统、电子设备 |
CN111144313A (zh) * | 2019-12-27 | 2020-05-12 | 创新奇智(青岛)科技有限公司 | 一种基于多感受野动态结合的人脸检测方法及系统 |
CN111274945B (zh) * | 2020-01-19 | 2023-08-08 | 北京百度网讯科技有限公司 | 一种行人属性的识别方法、装置、电子设备和存储介质 |
CN111401359A (zh) * | 2020-02-25 | 2020-07-10 | 北京三快在线科技有限公司 | 目标识别方法、装置、电子设备和存储介质 |
CN111753702A (zh) * | 2020-06-18 | 2020-10-09 | 上海高德威智能交通系统有限公司 | 目标检测方法、装置及设备 |
CN111510752B (zh) * | 2020-06-18 | 2021-04-23 | 平安国际智慧城市科技股份有限公司 | 数据传输方法、装置、服务器及存储介质 |
CN112152821B (zh) * | 2020-09-23 | 2023-03-28 | 青岛海尔科技有限公司 | 定向通信方法和装置、存储介质及电子设备 |
CN113065592B (zh) * | 2021-03-31 | 2025-02-25 | 上海商汤智能科技有限公司 | 图像分类方法、装置、电子设备及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101419671A (zh) * | 2008-11-10 | 2009-04-29 | 北方工业大学 | 基于模糊支持向量机的人脸性别识别方法 |
CN102880859A (zh) * | 2012-08-30 | 2013-01-16 | 华南理工大学 | 一种车牌识别方法 |
CN104134079A (zh) * | 2014-07-31 | 2014-11-05 | 中国科学院自动化研究所 | 一种基于极值区域和极限学习机的车牌识别方法 |
CN105512676A (zh) * | 2015-11-30 | 2016-04-20 | 华南理工大学 | 一种智能终端上的食物识别方法 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2535613B (en) * | 2014-12-19 | 2019-01-02 | Apical Ltd | Sensor noise profile |
CN105447529B (zh) * | 2015-12-30 | 2020-11-03 | 商汤集团有限公司 | 一种服饰检测及其属性值识别的方法和系统 |
-
2016
- 2016-12-29 CN CN201611246395.9A patent/CN108229267B/zh active Active
-
2017
- 2017-12-28 WO PCT/CN2017/119535 patent/WO2018121690A1/fr active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101419671A (zh) * | 2008-11-10 | 2009-04-29 | 北方工业大学 | 基于模糊支持向量机的人脸性别识别方法 |
CN102880859A (zh) * | 2012-08-30 | 2013-01-16 | 华南理工大学 | 一种车牌识别方法 |
CN104134079A (zh) * | 2014-07-31 | 2014-11-05 | 中国科学院自动化研究所 | 一种基于极值区域和极限学习机的车牌识别方法 |
CN105512676A (zh) * | 2015-11-30 | 2016-04-20 | 华南理工大学 | 一种智能终端上的食物识别方法 |
Cited By (137)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110766152B (zh) * | 2018-07-27 | 2023-08-04 | 富士通株式会社 | 用于训练深度神经网络的方法和装置 |
CN110766152A (zh) * | 2018-07-27 | 2020-02-07 | 富士通株式会社 | 用于训练深度神经网络的方法和装置 |
CN110766129A (zh) * | 2018-07-27 | 2020-02-07 | 杭州海康威视数字技术股份有限公司 | 一种神经网络训练系统及显示数据的方法 |
CN111357014A (zh) * | 2018-09-19 | 2020-06-30 | 华为技术有限公司 | 一种ai模型的开发方法及装置 |
CN110969173A (zh) * | 2018-09-28 | 2020-04-07 | 杭州海康威视数字技术股份有限公司 | 目标分类方法及装置 |
CN110969173B (zh) * | 2018-09-28 | 2023-10-24 | 杭州海康威视数字技术股份有限公司 | 目标分类方法及装置 |
CN110969657A (zh) * | 2018-09-29 | 2020-04-07 | 杭州海康威视数字技术股份有限公司 | 一种枪球坐标关联方法、装置、电子设备及存储介质 |
CN110969657B (zh) * | 2018-09-29 | 2023-11-03 | 杭州海康威视数字技术股份有限公司 | 一种枪球坐标关联方法、装置、电子设备及存储介质 |
CN110046632B (zh) * | 2018-11-09 | 2023-06-02 | 创新先进技术有限公司 | 模型训练方法和装置 |
CN110046632A (zh) * | 2018-11-09 | 2019-07-23 | 阿里巴巴集团控股有限公司 | 模型训练方法和装置 |
CN111241869A (zh) * | 2018-11-28 | 2020-06-05 | 杭州海康威视数字技术股份有限公司 | 物料盘点的方法、装置及计算机可读存储介质 |
CN111241869B (zh) * | 2018-11-28 | 2024-04-02 | 杭州海康威视数字技术股份有限公司 | 物料盘点的方法、装置及计算机可读存储介质 |
CN111259701A (zh) * | 2018-12-03 | 2020-06-09 | 杭州海康威视数字技术股份有限公司 | 行人再识别方法、装置及电子设备 |
CN111259701B (zh) * | 2018-12-03 | 2023-04-25 | 杭州海康威视数字技术股份有限公司 | 行人再识别方法、装置及电子设备 |
CN111291597A (zh) * | 2018-12-07 | 2020-06-16 | 杭州海康威视数字技术股份有限公司 | 一种基于图像的人群态势分析方法、装置、设备及系统 |
CN111291597B (zh) * | 2018-12-07 | 2023-10-13 | 杭州海康威视数字技术股份有限公司 | 一种基于图像的人群态势分析方法、装置、设备及系统 |
CN111310775A (zh) * | 2018-12-11 | 2020-06-19 | Tcl集团股份有限公司 | 数据训练方法、装置、终端设备及计算机可读存储介质 |
CN111310775B (zh) * | 2018-12-11 | 2023-08-25 | Tcl科技集团股份有限公司 | 数据训练方法、装置、终端设备及计算机可读存储介质 |
CN109800654B (zh) * | 2018-12-24 | 2023-04-07 | 百度在线网络技术(北京)有限公司 | 车载摄像头检测处理方法、装置及车辆 |
CN111414930B (zh) * | 2019-01-07 | 2023-10-27 | 中国移动通信有限公司研究院 | 深度学习模型训练方法及装置、电子设备及存储介质 |
CN111428536B (zh) * | 2019-01-09 | 2024-04-19 | 北京京东乾石科技有限公司 | 用于检测物品类别和位置的检测网络的训练方法和装置 |
CN111428536A (zh) * | 2019-01-09 | 2020-07-17 | 北京京东尚科信息技术有限公司 | 用于检测物品类别和位置的检测网络的训练方法和装置 |
CN111435364B (zh) * | 2019-01-14 | 2023-04-18 | 阿里巴巴集团控股有限公司 | 电子病历质检方法和装置 |
CN111435364A (zh) * | 2019-01-14 | 2020-07-21 | 阿里巴巴集团控股有限公司 | 电子病历质检方法和装置 |
CN111435432A (zh) * | 2019-01-15 | 2020-07-21 | 北京市商汤科技开发有限公司 | 网络优化方法及装置、图像处理方法及装置、存储介质 |
CN111435432B (zh) * | 2019-01-15 | 2023-05-26 | 北京市商汤科技开发有限公司 | 网络优化方法及装置、图像处理方法及装置、存储介质 |
CN111444749A (zh) * | 2019-01-17 | 2020-07-24 | 杭州海康威视数字技术股份有限公司 | 路面导向标志的识别方法、装置及存储介质 |
CN111444749B (zh) * | 2019-01-17 | 2023-09-01 | 杭州海康威视数字技术股份有限公司 | 路面导向标志的识别方法、装置及存储介质 |
CN113688933B (zh) * | 2019-01-18 | 2024-05-24 | 北京市商汤科技开发有限公司 | 分类网络的训练方法及分类方法和装置、电子设备 |
CN113688933A (zh) * | 2019-01-18 | 2021-11-23 | 北京市商汤科技开发有限公司 | 分类网络的训练方法及分类方法和装置、电子设备 |
CN110059721A (zh) * | 2019-03-16 | 2019-07-26 | 平安城市建设科技(深圳)有限公司 | 户型图区域识别方法、装置、设备及计算机可读存储介质 |
CN111832368A (zh) * | 2019-04-23 | 2020-10-27 | 长沙智能驾驶研究院有限公司 | 可行驶区域检测模型的训练方法、训练装置及应用 |
CN113743535B (zh) * | 2019-05-21 | 2024-05-24 | 北京市商汤科技开发有限公司 | 神经网络训练方法及装置以及图像处理方法及装置 |
CN113743535A (zh) * | 2019-05-21 | 2021-12-03 | 北京市商汤科技开发有限公司 | 神经网络训练方法及装置以及图像处理方法及装置 |
CN112307850A (zh) * | 2019-08-01 | 2021-02-02 | 浙江商汤科技开发有限公司 | 神经网络训练方法、车道线检测方法、装置和电子设备 |
CN110458829A (zh) * | 2019-08-13 | 2019-11-15 | 腾讯医疗健康(深圳)有限公司 | 基于人工智能的图像质控方法、装置、设备及存储介质 |
CN110458829B (zh) * | 2019-08-13 | 2024-01-30 | 腾讯医疗健康(深圳)有限公司 | 基于人工智能的图像质控方法、装置、设备及存储介质 |
CN110443222A (zh) * | 2019-08-14 | 2019-11-12 | 北京百度网讯科技有限公司 | 用于训练脸部关键点检测模型的方法和装置 |
CN110443222B (zh) * | 2019-08-14 | 2022-09-09 | 北京百度网讯科技有限公司 | 用于训练脸部关键点检测模型的方法和装置 |
CN112733578B (zh) * | 2019-10-28 | 2024-05-24 | 普天信息技术有限公司 | 车辆重识别方法及系统 |
CN112733578A (zh) * | 2019-10-28 | 2021-04-30 | 普天信息技术有限公司 | 车辆重识别方法及系统 |
CN111191526B (zh) * | 2019-12-16 | 2023-10-10 | 汇纳科技股份有限公司 | 行人属性识别网络训练方法、系统、介质及终端 |
CN111191526A (zh) * | 2019-12-16 | 2020-05-22 | 汇纳科技股份有限公司 | 行人属性识别网络训练方法、系统、介质及终端 |
CN111160429A (zh) * | 2019-12-17 | 2020-05-15 | 平安银行股份有限公司 | 图像检测模型的训练方法、图像检测方法、装置及设备 |
CN111160429B (zh) * | 2019-12-17 | 2023-09-05 | 平安银行股份有限公司 | 图像检测模型的训练方法、图像检测方法、装置及设备 |
CN111242951A (zh) * | 2020-01-08 | 2020-06-05 | 上海眼控科技股份有限公司 | 车辆检测方法、装置、计算机设备和存储介质 |
CN111259763A (zh) * | 2020-01-13 | 2020-06-09 | 华雁智能科技(集团)股份有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN111259763B (zh) * | 2020-01-13 | 2024-02-02 | 华雁智能科技(集团)股份有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN111340090B (zh) * | 2020-02-21 | 2023-08-01 | 每日互动股份有限公司 | 图像特征比对方法及装置、设备、计算机可读存储介质 |
CN111340090A (zh) * | 2020-02-21 | 2020-06-26 | 浙江每日互动网络科技股份有限公司 | 图像特征比对方法及装置、设备、计算机可读存储介质 |
CN111292331A (zh) * | 2020-02-23 | 2020-06-16 | 华为技术有限公司 | 图像处理的方法与装置 |
CN111292331B (zh) * | 2020-02-23 | 2023-09-12 | 华为云计算技术有限公司 | 图像处理的方法与装置 |
CN111368923A (zh) * | 2020-03-05 | 2020-07-03 | 上海商汤智能科技有限公司 | 神经网络训练方法及装置、电子设备和存储介质 |
CN111368923B (zh) * | 2020-03-05 | 2023-12-19 | 上海商汤智能科技有限公司 | 神经网络训练方法及装置、电子设备和存储介质 |
CN111539452B (zh) * | 2020-03-26 | 2024-03-26 | 深圳云天励飞技术有限公司 | 多任务属性的图像识别方法、装置、电子设备及存储介质 |
CN111539452A (zh) * | 2020-03-26 | 2020-08-14 | 深圳云天励飞技术有限公司 | 多任务属性的图像识别方法、装置、电子设备及存储介质 |
CN111428671A (zh) * | 2020-03-31 | 2020-07-17 | 杭州博雅鸿图视频技术有限公司 | 人脸结构化信息识别方法、系统、装置及存储介质 |
CN111459675A (zh) * | 2020-03-31 | 2020-07-28 | 拉扎斯网络科技(上海)有限公司 | 一种数据处理方法、装置、可读存储介质和电子设备 |
CN111459675B (zh) * | 2020-03-31 | 2023-09-15 | 拉扎斯网络科技(上海)有限公司 | 一种数据处理方法、装置、可读存储介质和电子设备 |
CN111612732A (zh) * | 2020-04-02 | 2020-09-01 | 深圳大学 | 图像质量评估方法、装置、计算机设备及存储介质 |
CN111612732B (zh) * | 2020-04-02 | 2023-07-18 | 深圳大学 | 图像质量评估方法、装置、计算机设备及存储介质 |
CN111507958A (zh) * | 2020-04-15 | 2020-08-07 | 全球能源互联网研究院有限公司 | 目标检测方法、检测模型的训练方法及电子设备 |
CN111507958B (zh) * | 2020-04-15 | 2023-05-26 | 全球能源互联网研究院有限公司 | 目标检测方法、检测模型的训练方法及电子设备 |
CN111523600A (zh) * | 2020-04-26 | 2020-08-11 | 上海商汤临港智能科技有限公司 | 神经网络训练、目标检测、及智能设备控制的方法及装置 |
CN111523600B (zh) * | 2020-04-26 | 2023-12-19 | 上海商汤临港智能科技有限公司 | 神经网络训练、目标检测、及智能设备控制的方法及装置 |
CN111539481A (zh) * | 2020-04-28 | 2020-08-14 | 北京市商汤科技开发有限公司 | 图像标注方法、装置、电子设备及存储介质 |
CN111539481B (zh) * | 2020-04-28 | 2024-03-08 | 北京市商汤科技开发有限公司 | 图像标注方法、装置、电子设备及存储介质 |
CN111582107B (zh) * | 2020-04-28 | 2023-09-29 | 浙江大华技术股份有限公司 | 目标重识别模型的训练方法、识别方法、电子设备及装置 |
CN111582107A (zh) * | 2020-04-28 | 2020-08-25 | 浙江大华技术股份有限公司 | 目标重识别模型的训练方法、识别方法、电子设备及装置 |
CN111539947A (zh) * | 2020-04-30 | 2020-08-14 | 上海商汤智能科技有限公司 | 图像检测方法及相关模型的训练方法和相关装置、设备 |
CN111539947B (zh) * | 2020-04-30 | 2024-03-29 | 上海商汤智能科技有限公司 | 图像检测方法及相关模型的训练方法和相关装置、设备 |
CN111598902B (zh) * | 2020-05-20 | 2023-05-30 | 抖音视界有限公司 | 图像分割方法、装置、电子设备及计算机可读介质 |
CN111598902A (zh) * | 2020-05-20 | 2020-08-28 | 北京字节跳动网络技术有限公司 | 图像分割方法、装置、电子设备及计算机可读介质 |
CN113742562B (zh) * | 2020-05-27 | 2023-10-10 | 北京达佳互联信息技术有限公司 | 视频推荐方法、装置、电子设备及存储介质 |
CN113742562A (zh) * | 2020-05-27 | 2021-12-03 | 北京达佳互联信息技术有限公司 | 视频推荐方法、装置、电子设备及存储介质 |
CN111860573B (zh) * | 2020-06-04 | 2024-05-10 | 北京迈格威科技有限公司 | 模型训练方法、图像类别检测方法、装置和电子设备 |
CN111860573A (zh) * | 2020-06-04 | 2020-10-30 | 北京迈格威科技有限公司 | 模型训练方法、图像类别检测方法、装置和电子设备 |
CN112749609B (zh) * | 2020-07-23 | 2024-03-19 | 腾讯科技(深圳)有限公司 | 人体图像分割方法、装置、计算机设备及存储介质 |
CN112749609A (zh) * | 2020-07-23 | 2021-05-04 | 腾讯科技(深圳)有限公司 | 人体图像分割方法、装置、计算机设备及存储介质 |
CN111967597A (zh) * | 2020-08-18 | 2020-11-20 | 上海商汤临港智能科技有限公司 | 神经网络训练及图像分类方法、装置、存储介质、设备 |
CN112016630B (zh) * | 2020-09-03 | 2024-03-19 | 平安科技(深圳)有限公司 | 基于图像分类模型的训练方法、装置、设备及存储介质 |
CN112016630A (zh) * | 2020-09-03 | 2020-12-01 | 平安科技(深圳)有限公司 | 基于图像分类模型的训练方法、装置、设备及存储介质 |
CN112101169A (zh) * | 2020-09-08 | 2020-12-18 | 平安科技(深圳)有限公司 | 基于注意力机制的道路图像目标检测方法及相关设备 |
CN112101169B (zh) * | 2020-09-08 | 2024-04-05 | 平安科技(深圳)有限公司 | 基于注意力机制的道路图像目标检测方法及相关设备 |
CN112184635A (zh) * | 2020-09-10 | 2021-01-05 | 上海商汤智能科技有限公司 | 目标检测方法、装置、存储介质及设备 |
CN112418261A (zh) * | 2020-09-17 | 2021-02-26 | 电子科技大学 | 一种基于先验原型注意力机制的人体图像多属性分类方法 |
CN112101282A (zh) * | 2020-09-25 | 2020-12-18 | 北京瞰天科技有限公司 | 水上目标识别方法、装置及电子设备和存储介质 |
CN112101282B (zh) * | 2020-09-25 | 2024-04-26 | 北京瞰天科技有限公司 | 水上目标识别方法、装置及电子设备和存储介质 |
CN112163545A (zh) * | 2020-10-12 | 2021-01-01 | 北京易华录信息技术股份有限公司 | 一种头部特征提取方法、装置、电子设备及存储介质 |
CN112257604A (zh) * | 2020-10-23 | 2021-01-22 | 北京百度网讯科技有限公司 | 图像检测方法、装置、电子设备和存储介质 |
CN112529839B (zh) * | 2020-11-05 | 2023-05-02 | 西安交通大学 | 一种核磁共振图像中颈动脉血管中心线的提取方法及系统 |
CN112529839A (zh) * | 2020-11-05 | 2021-03-19 | 西安交通大学 | 一种核磁共振图像中颈动脉血管中心线的提取方法及系统 |
CN112487226A (zh) * | 2020-11-06 | 2021-03-12 | 北京三快在线科技有限公司 | 图片分类模型获取方法、装置、电子设备及可读存储介质 |
CN112464785A (zh) * | 2020-11-25 | 2021-03-09 | 浙江大华技术股份有限公司 | 一种目标检测方法、装置、计算机设备及存储介质 |
CN114571472A (zh) * | 2020-12-01 | 2022-06-03 | 北京小米移动软件有限公司 | 用于足式机器人的地面属性检测方法及驱动方法及其装置 |
CN113516144A (zh) * | 2020-12-01 | 2021-10-19 | 阿里巴巴集团控股有限公司 | 目标检测方法及装置、计算设备 |
CN114571472B (zh) * | 2020-12-01 | 2024-01-23 | 北京小米机器人技术有限公司 | 用于足式机器人的地面属性检测方法及驱动方法及其装置 |
CN112528995B (zh) * | 2020-12-22 | 2023-08-04 | 北京百度网讯科技有限公司 | 用于训练目标检测模型的方法、目标检测方法及装置 |
CN112528995A (zh) * | 2020-12-22 | 2021-03-19 | 北京百度网讯科技有限公司 | 用于训练目标检测模型的方法、目标检测方法及装置 |
CN112700002A (zh) * | 2020-12-23 | 2021-04-23 | 北京三快在线科技有限公司 | 图神经网络的训练方法、装置、电子设备及可读存储介质 |
CN112614117A (zh) * | 2020-12-28 | 2021-04-06 | 广州绿怡信息科技有限公司 | 设备区域提取模型训练方法、设备区域提取方法及装置 |
CN112712088A (zh) * | 2020-12-31 | 2021-04-27 | 洛阳语音云创新研究院 | 动物膘情检测方法、装置及计算机可读存储介质 |
CN112712088B (zh) * | 2020-12-31 | 2023-02-14 | 洛阳语音云创新研究院 | 动物膘情检测方法、装置及计算机可读存储介质 |
CN114764919A (zh) * | 2021-01-14 | 2022-07-19 | 郑州信大先进技术研究院 | 一种基于深度学习的行人属性识别方法 |
CN113822111B (zh) * | 2021-01-19 | 2024-05-24 | 北京京东振世信息技术有限公司 | 人群检测模型训练方法、装置以及人群计数方法、装置 |
CN113573044B (zh) * | 2021-01-19 | 2022-12-09 | 腾讯科技(深圳)有限公司 | 视频数据处理方法、装置、计算机设备及可读存储介质 |
CN113822111A (zh) * | 2021-01-19 | 2021-12-21 | 北京京东振世信息技术有限公司 | 人群检测模型训练方法、装置以及人群计数方法、装置 |
CN113573044A (zh) * | 2021-01-19 | 2021-10-29 | 腾讯科技(深圳)有限公司 | 视频数据处理方法、装置、计算机设备及可读存储介质 |
CN112925938A (zh) * | 2021-01-28 | 2021-06-08 | 上海商汤智能科技有限公司 | 一种图像标注方法、装置、电子设备及存储介质 |
CN112861858B (zh) * | 2021-02-19 | 2024-06-07 | 北京龙翼风科技有限公司 | 显著性真值图的生成方法及显著性检测模型的训练方法 |
CN112861858A (zh) * | 2021-02-19 | 2021-05-28 | 首都师范大学 | 显著性真值图的生成方法及显著性检测模型的训练方法 |
CN113723159A (zh) * | 2021-02-26 | 2021-11-30 | 腾讯科技(深圳)有限公司 | 场景识别模型训练方法、场景识别方法及模型训练装置 |
CN112906685B (zh) * | 2021-03-04 | 2024-03-26 | 重庆赛迪奇智人工智能科技有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
CN112906685A (zh) * | 2021-03-04 | 2021-06-04 | 重庆赛迪奇智人工智能科技有限公司 | 一种目标检测方法、装置、电子设备及存储介质 |
CN113012176A (zh) * | 2021-03-17 | 2021-06-22 | 北京百度网讯科技有限公司 | 样本图像的处理方法、装置、电子设备及存储介质 |
CN113012176B (zh) * | 2021-03-17 | 2023-12-15 | 阿波罗智联(北京)科技有限公司 | 样本图像的处理方法、装置、电子设备及存储介质 |
CN112906651A (zh) * | 2021-03-25 | 2021-06-04 | 中国联合网络通信集团有限公司 | 目标检测方法及装置 |
CN112906651B (zh) * | 2021-03-25 | 2023-07-11 | 中国联合网络通信集团有限公司 | 目标检测方法及装置 |
CN113052175B (zh) * | 2021-03-26 | 2024-03-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN113052175A (zh) * | 2021-03-26 | 2021-06-29 | 北京百度网讯科技有限公司 | 目标检测方法、装置、电子设备及可读存储介质 |
CN112949767A (zh) * | 2021-04-07 | 2021-06-11 | 北京百度网讯科技有限公司 | 样本图像增量、图像检测模型训练及图像检测方法 |
CN112949767B (zh) * | 2021-04-07 | 2023-08-11 | 北京百度网讯科技有限公司 | 样本图像增量、图像检测模型训练及图像检测方法 |
CN113516013A (zh) * | 2021-04-09 | 2021-10-19 | 阿波罗智联(北京)科技有限公司 | 目标检测方法、装置、电子设备、路侧设备和云控平台 |
CN113516013B (zh) * | 2021-04-09 | 2024-05-14 | 阿波罗智联(北京)科技有限公司 | 目标检测方法、装置、电子设备、路侧设备和云控平台 |
CN113454649A (zh) * | 2021-06-17 | 2021-09-28 | 商汤国际私人有限公司 | 目标检测方法、装置、电子设备和计算机可读存储介质 |
CN113454649B (zh) * | 2021-06-17 | 2024-05-24 | 商汤国际私人有限公司 | 目标检测方法、装置、电子设备和计算机可读存储介质 |
CN113469121A (zh) * | 2021-07-21 | 2021-10-01 | 浙江大华技术股份有限公司 | 一种车辆状态的识别方法及装置 |
CN113642431A (zh) * | 2021-07-29 | 2021-11-12 | 北京百度网讯科技有限公司 | 目标检测模型的训练方法及装置、电子设备和存储介质 |
CN113642431B (zh) * | 2021-07-29 | 2024-02-06 | 北京百度网讯科技有限公司 | 目标检测模型的训练方法及装置、电子设备和存储介质 |
CN113902999A (zh) * | 2021-09-24 | 2022-01-07 | 青岛海信网络科技股份有限公司 | 一种跟踪方法、装置、设备及介质 |
CN113989772A (zh) * | 2021-10-27 | 2022-01-28 | 广州小鹏自动驾驶科技有限公司 | 一种交通灯检测方法、装置、车辆和可读存储介质 |
CN113963249B (zh) * | 2021-10-29 | 2024-04-09 | 山东大学 | 一种星系图像的检测方法和系统 |
CN113963249A (zh) * | 2021-10-29 | 2022-01-21 | 山东大学 | 一种星系图像的检测方法和系统 |
CN114387649A (zh) * | 2022-01-11 | 2022-04-22 | 北京百度网讯科技有限公司 | 图像处理方法、装置、电子设备以及存储介质 |
CN117037218A (zh) * | 2023-10-08 | 2023-11-10 | 腾讯科技(深圳)有限公司 | 对象属性识别方法、相关装置、设备和介质 |
CN117037218B (zh) * | 2023-10-08 | 2024-03-15 | 腾讯科技(深圳)有限公司 | 对象属性识别方法、相关装置、设备和介质 |
CN118587746A (zh) * | 2024-08-06 | 2024-09-03 | 浙江大华技术股份有限公司 | 人体目标检测方法、装置、计算机设备和存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN108229267B (zh) | 2020-10-16 |
CN108229267A (zh) | 2018-06-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2018121690A1 (fr) | Procédé et dispositif de détection d'attribut d'objet, procédé et dispositif d'apprentissage de réseau neuronal et procédé et dispositif de détection zonale | |
EP3767536B1 (fr) | Code latent pour adaptation de domaines non supervisée | |
CN112990432B (zh) | 目标识别模型训练方法、装置及电子设备 | |
Xu et al. | Investigating bias and fairness in facial expression recognition | |
US11657602B2 (en) | Font identification from imagery | |
US11270124B1 (en) | Temporal bottleneck attention architecture for video action recognition | |
US20210117760A1 (en) | Methods and apparatus to obtain well-calibrated uncertainty in deep neural networks | |
Bendale et al. | Towards open set deep networks | |
CN108304775B (zh) | 遥感图像识别方法、装置、存储介质以及电子设备 | |
US11468266B2 (en) | Target identification in large image data | |
CN113537630B (zh) | 业务预测模型的训练方法及装置 | |
CN108229673B (zh) | 卷积神经网络的处理方法、装置和电子设备 | |
KR20190029083A (ko) | 신경망 학습 방법 및 이를 적용한 장치 | |
US20240127153A1 (en) | Systems and methods for automated risk assessment in machine learning | |
CN113780145A (zh) | 精子形态检测方法、装置、计算机设备和存储介质 | |
CN112215831B (zh) | 一种用于人脸图像质量的评价方法和系统 | |
CN115471714A (zh) | 数据处理方法、装置、计算设备和计算机可读存储介质 | |
CN116958615A (zh) | 图片识别方法、装置、设备和介质 | |
Ojo et al. | Real-time Face-based Gender Identification System Using Pelican Support Vector Machine | |
Jiang et al. | Fair selection through kernel density estimation | |
Thevarasa et al. | Weighted Ensemble Algorithm for Aerial Imaging Based Mosquito Breeding Sites Classification | |
CN116872961A (zh) | 用于智能驾驶车辆的控制系统 | |
Jamshed et al. | Grey wolf optimization (GWO) with the convolution neural network (CNN)-based pattern recognition system | |
CN116561663A (zh) | 地面无人装备人工干预程度分级方法、系统、计算设备 | |
Jadon | Ethical AI development: Mitigating bias in generative models |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17888499 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17888499 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 17.12.2019) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17888499 Country of ref document: EP Kind code of ref document: A1 |