+

CN119180988A - A visual recognition method and device based on computer processing - Google Patents

A visual recognition method and device based on computer processing Download PDF

Info

Publication number
CN119180988A
CN119180988A CN202411219159.2A CN202411219159A CN119180988A CN 119180988 A CN119180988 A CN 119180988A CN 202411219159 A CN202411219159 A CN 202411219159A CN 119180988 A CN119180988 A CN 119180988A
Authority
CN
China
Prior art keywords
wafer
defect
image
model
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202411219159.2A
Other languages
Chinese (zh)
Inventor
孙小霞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jingdezhen University
Original Assignee
Jingdezhen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jingdezhen University filed Critical Jingdezhen University
Priority to CN202411219159.2A priority Critical patent/CN119180988A/en
Publication of CN119180988A publication Critical patent/CN119180988A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0475Generative networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/094Adversarial learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/52Scale-space analysis, e.g. wavelet analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/766Arrangements for image or video recognition or understanding using pattern recognition or machine learning using regression, e.g. by projecting features on hyperplanes
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/70Labelling scene content, e.g. deriving syntactic or semantic representations
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/30Noise filtering
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/36Applying a local operator, i.e. means to operate on image points situated in the vicinity of a given point; Non-linear local filtering operations, e.g. median filtering
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/06Recognition of objects for industrial automation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Testing Or Measuring Of Semiconductors Or The Like (AREA)
  • Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)

Abstract

The invention provides a visual identification method and a visual identification device based on computer processing, wherein the method comprises the steps of S1, collecting picture data based on a wafer generation line, preprocessing and label judging the collected picture data to obtain a wafer defect detection data set, S2, constructing a feature extraction and feature enhancement network aiming at the wafer defect detection data set obtained in the S1 to obtain a multi-level feature map, S3, designing a wafer defect detection model based on the multi-level feature map obtained in the S2 to obtain a semiconductor wafer defect detection model, S4, constructing a training strategy combining with antagonism collaborative learning based on the semiconductor wafer defect detection model obtained in the S3, training to obtain a final wafer defect detection model, S5, deploying the wafer defect detection model obtained in the S4 on a production process equipment visual module, and detecting and feeding back defect problems in real time. The invention has the following beneficial effects that the accuracy is maintained and the detection accuracy and the robustness are obviously improved.

Description

Visual recognition method and device for computer processing
Technical Field
The invention belongs to the field of computer vision, and particularly relates to a multitasking and collaborative visual detection method for defects of a semiconductor wafer.
Background
With the rapid development of semiconductor technology, the types of defects occurring in the wafer fabrication process are becoming more and more complex and diverse. These defects not only affect the yield of the wafer, but also have profound effects on the performance and reliability of the final product. Accurate defect detection is a critical element in order to ensure high quality wafer fabrication. However, with the continuous progress of the process technology, the conventional inspection method is faced with small and complex defects, which gradually exposes the problem that it is difficult to meet the modern manufacturing requirements. Therefore, there is an urgent need to develop more advanced inspection technologies to cope with the increasing inspection demands and to ensure the production efficiency and the product quality.
Existing PCB semiconductor wafer defect detection mainly covers the following three methods, namely a template matching-based method, an image segmentation-based method and a machine learning-based method:
(1) Template matching-based method:
the method detects defects by pixel-level comparison of the wafer image to be inspected with a standard defect-free wafer image. Although simple to implement, there are the following disadvantages:
a) The requirement on image alignment is extremely high, and small deviation can lead to false detection;
b) The ability to detect non-repetitive defects is poor;
c) The calculation complexity is high, and the real-time requirement of online detection is difficult to meet.
(2) Image segmentation-based method:
The method firstly segments the wafer image and then analyzes the segmented region characteristics to identify defects. Its major drawbacks include:
a) The robustness of the segmentation algorithm is insufficient, and the segmentation algorithm is easily influenced by image noise and illumination variation;
b) The detection effect on micro defects and low-contrast defects is poor;
c) It is difficult to effectively distinguish defects from normal structural changes.
(3) A machine learning based method:
Such methods utilize traditional machine learning algorithms (e.g., SVM, random forest, etc.) to classify the extracted image features. Although improved over the former two methods, the following problems remain:
a) The feature engineering relies on manual design, so that complex defect features are difficult to comprehensively capture;
b) The generalization capability is limited, and the adaptability to the type of the defects which are not found is poor;
c) It is difficult to achieve both high accuracy and high efficiency detection.
Aiming at the limitations of the existing semiconductor wafer defect detection methods, the problems of insufficient precision, low efficiency, weak generalization capability and the like can be seen when the defects of the complex and diversified wafers are processed. Particularly, in the prior process, the traditional method is difficult to cope with increasingly miniaturized and diversified defect types, and cannot meet the requirements of modern semiconductor manufacturing on high-precision and high-efficiency defect detection. Therefore, there is a need to develop a new semiconductor wafer defect detection method to overcome these drawbacks and realize high-precision, high-efficiency and powerful semiconductor wafer defect detection.
Disclosure of Invention
In order to solve the problems, the invention provides a visual recognition method for computer processing, which effectively realizes high-precision and high-efficiency wafer defect detection, and particularly shows strong adaptability and robustness in the presence of complex and diversified defect types.
A visual recognition method for computer processing, the method comprising:
S1, acquiring original semiconductor wafer image data based on a real semiconductor wafer generation line, and cutting, denoising and data enhancement data preprocessing methods for an acquired original image data design center to ensure smoothness of image data, and performing label cutting to obtain a complete wafer defect detection data set;
S2, constructing a dynamic multi-scale feature extraction and feature enhancement network of the wafer defect image according to the wafer defect detection data set obtained in the S1, and obtaining a multi-level feature map containing local details and global semantic information of the wafer defect image;
S3, designing a multitasking wafer defect detection model based on a multi-stage feature map of the wafer defects to obtain classification, regression and circular boundary prediction results of the wafer defects, thereby realizing visual detection of the semiconductor wafer defects;
s4, constructing a training strategy combining the antagonism collaborative learning based on the semiconductor wafer defect detection model, and performing multistage progressive training to obtain a final wafer defect detection model;
And S5, deploying a model, namely deploying the wafer defect detection model obtained in the step S4 on a visual module of process equipment for producing the semiconductor wafer, and detecting and feeding back whether the produced semiconductor wafer has defect or not in real time.
The visual recognition method as described above, the collecting and manufacturing process of the data set in S1 includes:
(1) Raw data acquisition
The source and acquisition method of the semiconductor wafer defect image data is that the original data acquisition is carried out by installing high-precision industrial cameras on a semiconductor wafer production line. An ace series industrial camera (model: acA4112-20 um) from Basler corporation, germany was used, which had a4112 x 3008 pixel resolution, up to 20 frames/sec acquisition speed. The camera is mounted above the wafer transport track and cooperates with the light source system to ensure that a clear image of the wafer surface is captured.
To ensure image quality, an LED ring light source (model: CCS LDR2-70SW2-LA 1) is used to provide uniform and stable illumination conditions. In the image acquisition process, triggering of the camera and the light source is precisely controlled by a PLC (programmable logic controller) so as to ensure that each wafer can be completely captured.
The raw image data is saved in a 16-bit TIFF format, with 65536 gray levels per pixel, to ensure high dynamic range and detail retention of the image. The file size of each image is about 24MB, and the naming convention is "wafer_ YYYYMMDD _ HHMMSS _sequence number. Tiff", wherein YYYYMMDD represents the date, HHMMSS represents the time, and the sequence number is the sequential number collected on the current day.
Data diversity and representativeness analysis to ensure diversity and representativeness of data sets, wafer samples were collected from different production lots, from different process stages. The method specifically comprises the following steps:
wafers of different sizes 6 inches, 8 inches and 12 inches;
wafers of different process nodes are 28nm, 14nm and 7nm;
Wafers at different process stages, namely after photoetching, etching and polishing;
different defect types, scratch, crack, flaking and foreign object attachment;
different defect levels, ranging from mild to severe.
A total of 100,000 raw wafer images were acquired, containing various defect types and defect-free samples. Through data analysis, the distribution of various defects in a data set is basically balanced, and the defect-free sample accounts for about 60 percent so as to simulate the occurrence frequency of the defects in actual production. The acquired raw image data is denoted as dataset.
(2) Data preprocessing
ROI region clipping and its effect on model accuracy considering the circular nature of the wafer, a circular ROI (region of interest) clipping algorithm is designed:
Firstly detecting the edge of a wafer by using Hough circle transformation to obtain the center coordinate and the radius of the wafer, secondly taking the detected center as the center of a circle, taking a circular region with the radius slightly smaller than the detection radius (98 percent) as the ROI, and finally setting the pixel value of the region outside the ROI as 0, and reserving the original image information in the ROI. The cutting method can effectively remove the interference information of the wafer edge and improve the attention of the model to the defects of the central area.
Image noise reduction processing in wafer images, gaussian white noise and speckle noise are generally present due to thermal noise of sensor readout circuits in ace-series industrial cameras and the influence of high-frequency components of the images. In order to reduce the influence of these noises on the image quality and the subsequent processing, image denoising processing is required.
First, gaussian white noise is removed by gaussian filtering. Gaussian filtering by convolution with the gaussian kernel can effectively smooth the image and remove noise. Specifically, the calculation formula of the gaussian filter is:
Where σ is the standard deviation of the gaussian kernel, x and y represent the pixel positions in the image, respectively, to represent the offset of each pixel point to the center of the filter. This offset is used to calculate the value of the gaussian distribution function to determine the weight of the pixel. Gaussian filtering can effectively reduce gaussian white noise in an image by convolving the image.
Data enhancement technology in order to increase the diversity of data and generalization capability of a model, the following data enhancement technology is adopted:
Random rotation, namely randomly rotating the image within the range of [ -10 degrees, 10 degrees ];
Random scaling, namely randomly scaling the image in the range of [0.9,1.1 ];
random translation, namely randomly translating within the range of +/-5% of the image width and height;
random brightness and contrast adjustment, wherein the brightness adjustment range is [0.8,1.2] and the contrast adjustment range is [0.8,1.2];
adding Gaussian noise with the mean value of 0 and standard deviation within the range of [0,0.05] into the random Gaussian noise;
random horizontal and vertical overturn, namely overturning at 50% probability;
These data enhancement techniques are implemented through Albumentations libraries and applied in real-time during the training process. Each original image would generate 5 enhanced images, expanding the dataset to 500,000. This enhancement strategy significantly improves the model's ability to adapt to various lighting conditions and defect morphologies.
The image data after the data preprocessing is recorded as data.
(3) Label arbitration
The manual labeling of the collected raw image data is to provide accurate training data for subsequent defect detection tasks. Industrial-level labeling software LabelImg is employed that has an intuitive user interface and rich functionality. The user can draw a bounding box on the image to annotate the target object through simple operation, and corresponding label information is added for each annotation box. The labeling procedure follows the following rules:
strict pixel level labeling, namely accurately drawing a boundary box of a defect area on an image by labeling personnel so as to ensure the accuracy and precision of labeling;
The defect classification, regression and circular boundaries are used as labels, wherein the classification labels are labeled label cls and correspond to five defect types (no defect, scratch, crack, flaking and foreign object adhesion), the value of each dimension of the classification labels represents the confidence score of the defect of the type, the regression labels are labeled label reg and contain the boundary frame coordinate information of a prediction target and are used for accurately positioning the position of the defect, and the circular boundary labels are labeled label cir and give circular parameters closely related to the defect boundary, so that the accuracy of boundary positioning can be further improved.
Each image may contain 0 or more defect labels, and labeling personnel need to label all existing defect areas in the image according to actual conditions;
and (3) storing the labeling result, namely storing the labeling result according to a PASCAL VOC format, wherein the format can be conveniently integrated and processed with the wafer defect detection model.
The labeling process requires labeling personnel to have certain expertise and experience so as to ensure the accuracy and usability of labeling results.
(4) Data set generation
In order to divide the marked data into a training set, a verification set and a test set, and prepare for training, evaluation and testing of the model, a data set needs to be constructed, namely the marked data set is marked as (data), and the training set, the verification set and the test set are divided according to the proportion of 8:1:1.
In order to divide the annotated data into a training set, a validation set and a test set, so as to prepare for training, evaluation and testing of the model, the data set (data, label) needs to be divided according to the proportion of 8:1:1, and the specific steps are as follows:
The labeled dataset (data, label) is divided into 80% of the dataset used as a training Set (TRAINING SET), 10% of the dataset used as a Validation Set (Validation Set), and 10% of the dataset used as a Test Set (Test Set).
Such partitioning can ensure that there is sufficient data for parameter adjustment and optimization during model training, and evaluation of the validation set and final performance validation of the test set are performed after model training to ensure generalization and validity of the model.
The visual recognition method as described above, the image dynamic multi-scale feature extraction and feature enhancement network construction process in S2 includes:
The scheme creatively provides a wafer image dynamic multi-scale feature enhancement network, and breaks through the limitation of traditional single-scale feature extraction. Through multi-scale feature extraction and enhancement and combination of a dynamic weight distribution mechanism, the network can capture local details and global semantics at the same time, and the richness and the robustness of feature representation are remarkably improved. Particularly, the module for dynamically adjusting the feature weight effectively highlights key information, suppresses irrelevant noise, and enhances the adaptability of the model in complex production line environments. In addition, the network design pays attention to the calculation efficiency, and the method is suitable for semiconductor wafer production line equipment by taking the performance and the instantaneity into consideration through the characteristic reuse and the light-weight structure. The method not only improves the accuracy of defect detection, but also enhances the robustness of the system under various visual interference conditions.
Processing an input wafer image from a bottom layer to a top layer, firstly extracting feature graphs { C2, C3, C4, C5} with different scales, and then constructing a multi-scale feature enhancement network according to the following steps:
Starting from the deepest P5 layer, an upper wafer feature map is generated by upsampling, and this high-level semantic information is transferred to the top layer of the multi-scale feature enhancement network:
M5=C5
wherein M5 is the top-most feature map in the multi-scale feature enhancement network, from the deepest level C5 feature map, and then performing 1×1 convolutional channel number adjustment on M5:
P5=Conv1×1(M5)
then constructing a multi-scale feature enhancement network from top to bottom:
M4=C4+Upsample(P5)
P4=Conv3×3(M4)
Conv3×3 is a3×3 convolution, and is used for fusing features of different scales, upsample is an up-sampling operation implemented by deconvolution, the steps are repeated to gradually generate { P3, P4, P5}, and finally P3, P4, P5 is a multi-scale feature enhancement network fused with multi-scale information;
The dynamic characteristic weight distribution module is integrated in the network to enhance the adaptability to the multi-scale characteristics and learn the correlation among the characteristics at different positions in the wafer characteristic diagram. By highlighting important features and suppressing irrelevant features, the mechanism first generates a spatial attention sub-module, i.e. an attention weight for each location, and then combines the spatial attention sub-module with the original feature map to obtain an enhanced feature representation. The method comprises the following specific steps:
The input of the feature weight dynamic allocation module is a wafer feature map X, the dimension is (C X H X W), two 1X 1 convolution checks X are used for channel compression, compressed wafer feature maps A and B are generated, and the dimension of the compressed wafer feature map is (C' X H X W):
A=X·Wa(C′×H×W)
B=X·Wb(C′×H×W)
Wherein, W a and W b are the weights of two 1×1 convolution kernels, respectively, and the similarity between the positions is calculated by using a and B to generate a spatial attention submodule map M, and the dimension is (h×w×h×w):
M=Softmax(AT·B)
Then, element level multiplication is carried out on the original input X and M, and a wafer characteristic diagram X' after weighting of the space attention submodule is obtained:
X′=X·M
and finally, adding X' and X, and outputting through a convolution layer:
O=γ·X′+X·Wc
Wherein, gamma is a leachable scaling factor for scaling the enhancement effect of the spatial attention sub-module, W c is a1×1 convolution kernel weight, and O is the final output wafer feature map of the feature weight dynamic allocation module.
The visual identification method as described above, the process of constructing the wafer defect detection model by the multitasking cooperation in S3 includes:
In the defect detection task, classification and regression are two interrelated but often separately processed sub-tasks. The traditional method respectively models and optimizes two tasks, and lacks knowledge fusion and mutual promotion between the two tasks. To solve this problem, the present invention devised a multitasking collaborative detection algorithm (Multi-task Collaborative Detection Algorithm, MCDA).
The key point of the MCDA is that a multi-task cooperative coding mechanism is introduced to mutually fuse the classification information and the regression information. The model cascades the feature graphs of the two subtasks through the convolution layer and mutually codes the information of the other party so as to obtain the classification and regression features fused with the two-way relation. In addition, the MCDA introduces a circular boundary prediction branch, and circular parameters related to the defect boundary can be predicted more finely through cascading classification and regression prediction results, so that the accuracy of boundary positioning is improved. Compared with the traditional separate modeling mode, the MCDA can better mine the inherent relation among classification, regression and shape, so that the consistency and accuracy of overall detection are improved. The method specifically comprises the following steps:
And (3) sending the dynamic multiscale feature extraction and the feature O output by the feature enhancement network into the MCDA, and simultaneously completing classification, regression and circular boundary prediction:
clspred,regpred,cirpred=MCDA(O)
Cls pred,regpred,cirpred is a classification prediction result, a regression prediction result, and a circular boundary prediction result of wafer defect detection, respectively.
The specific calculation process of the MCDA is as follows:
1) Performing basic convolution on O to obtain an initial predicted value { cls init,reginit }:
clsinit,reginit=Conv(O)
2) Performing relation coding on reg init and cls init to obtain regression characteristics reg fused fusing the relation between the reg init and cls init:
regfused=reginit+Conv(Concat(reginit,clsinit))
3) Meanwhile, cls init and reg init are subjected to relational coding to obtain classification characteristics cls fused fusing the relation between the cls init and reg init:
clsfused=clsinit+Conv(Concat(clsinit,reginit))
4) Convolution was performed on reg fused and cls fused, respectively, to yield the final classification prediction cls pred and regression prediction reg pred:
5) Cascading cls pred and reg pred, through an additional convolution branch, predicts the defect boundary circle parameter cir pred closely related to the defect boundary:
cirpred=Conv(Concat(clspred,regpred))
Through the multi-task collaborative coding, classification and regression knowledge can be fully fused, so that two tasks are mutually promoted. While introducing circular boundary prediction branches to more finely describe the shape and location of wafer defects.
In the existing target detection method, common loss functions such as FocalLoss, CIoULoss and the like only focus on optimizing indexes of a single task, and lack of consideration on interaction among multiple tasks and knowledge guidance. To solve this problem, the present invention proposes a multitasking synergy loss function.
The innovation of the loss function is to integrate three parts of classification, regression and circular boundary so as to realize joint optimization of three closely related subtasks of classification, regression and shape. Compared with the traditional single loss function, the multi-task cooperative loss function fully utilizes the correlation among the subtasks, realizes the mutual promotion among the tasks, and simultaneously integrates the guidance of priori knowledge, so that the model benefits from additional knowledge transfer in the optimization process, and the overall performance and the robustness of detection are improved.
Therefore, the invention adopts a multi-task cooperative loss function, and simultaneously optimizes three tasks of classification, regression and circular boundary prediction, namely:
L=λ1Lcls2Lreg3Lcir
Wherein L cls is a classification loss, focalLoss is used, L reg is a regression loss, CIoULoss is used, L cir is a circular boundary loss, polygonLoss is used, and lambda 123 is a super parameter for balancing each loss term.
Under the optimization of the multi-task cooperative loss function, the target detection network can output three prediction results, namely a classification prediction result cls pred, a regression prediction result reg pred and a circular boundary prediction result cir pred.
The classification predictor cls pred is a vector, and corresponds to four defect types (scratch, crack, flaking, foreign object adhesion), and the value of each dimension represents the confidence score of the defect type. The defect type in the current prediction frame can be determined by acquiring the dimension index with the highest score;
The regression prediction result reg pred contains the boundary frame coordinate information of the prediction target and is used for accurately positioning the position of the defect;
The circular boundary prediction result cir pred gives out a circular parameter closely related to the defect boundary, and the accuracy of boundary positioning is further improved.
After the three prediction results are obtained, the defect type can be judged, and accurate position and shape information is combined, so that accurate detection of various defects of the wafer is realized, and an important basis is provided for subsequent quality control and defect repair.
The visual recognition method as described above, wherein the step S4 includes:
The labeled training data set is input into the model, and the model is trained by the scheme by providing an innovative antagonistic collaborative learning strategy. The strategy is innovative multi-stage progressive training, the local detail and global semantic information in wafer image data and the dynamic relation of the local detail and the global semantic information in the fusion process are fully considered, and the performance of the model is improved through multi-scale sequential consistency and depth collaborative optimization.
(1) Multiscale timing consistency loss:
To enhance the model's learning of different time batches of data information, the present proposal proposes a multi-scale timing consistency penalty. The loss function measures the predicted consistency across different time scales:
Ltc=∑sλs*DKL(P(Y|Xt),P(Y|X{t-s}))
Where s represents different time scales, λ s is the corresponding weight, D KL is the KL divergence, and P (y|x t) represents the detection distribution at time step t, including classification detection, regression detection, and circular boundary detection, as summarized for one total detection result. This loss encourages the model to maintain predictive consistency across different time scales, thereby improving the modeling ability of different time batch data.
(2) Resistance training strategy:
to increase the robustness of the training, the present proposal introduces an antagonistic training strategy. In particular, a discriminator D is designed in an attempt to distinguish whether a feature is from local detail or global semantic information. The converged network F is then trained to spoof the arbiter:
Ladv=E[log(D(F(Xv,Xa)))]+E[log(1-D(Xv))+log(1-D(Xa))]
Wherein X v and X a are local detail and global semantic information inputs, respectively. Such countermeasure training forces the network to generate more indistinguishable, more tightly fused representations of features.
(3) Course learning and difficulty adaptation
According to the (2) resistance strategy, a course learning strategy based on sample difficulty is provided. The sample difficulty D (x) is defined as:
D(x)=1-exp(-γ(wvLv+waLa))
where L v and L a are loss of local detail and global semantic information, respectively, and γ is an adjustable parameter. In the training process, the proportion of difficult samples is gradually increased:
Wherein T is the current number of training steps, T is the total number of training steps, p 0 is the initial difficult sample proportion, and mu controls the rate of increase of difficulty.
(4) Dynamic batch normalization
The present proposal proposes a dynamic batch normalization (Dynamic Batch Normalization, DBN) technique to account for variability in illumination and noise conditions in semiconductor wafer production lines. The DBN dynamically adjusts normalization parameters according to the input statistical characteristics:
y=γ(x)(x-μ(x))/(σ(x)+ε)+β(x)
where γ (x) and β (x) are input dependent scaling and offset parameters. The method can be better suitable for the characteristic distribution change under different input conditions.
(5) Model training
Based on the method, the specific flow of model training is as follows:
1) Initializing a model parameter theta, wherein theta represents a parameter to be updated and comprises all weight matrixes and bias vectors in the wafer defect detection model;
2) For each training step t:
a. sampling a batch from the dataset, containing difficult samples in a proportion of p hard (t);
b. forward propagation is carried out to obtain multi-scale characteristics and prediction results;
c. Calculating a multi-scale time sequence consistency loss L tc;
d. performing an antagonism training strategy, and updating the discriminator D and the fusion network F;
e. applying dynamic batch normalization;
f. calculating the total loss L total=λtc*Ltcadv*Ladv;
g. back propagation, updating model parameters: Wherein the method comprises the steps of Is the learning rate.
3) And (2) repeating the step (2), and terminating the training process when the loss cannot be reduced by N continuous epochs. And storing final model parameters at the end of training to serve as a deployment model for wafer defect detection.
The visual inspection method as described above, the wafer defect inspection model deployment process in S5 includes:
(1) Hardware platform selection and configuration:
a. Selecting proper edge computing equipment, such as NVIDIAJetson series or Intel NUC high-performance embedded systems, according to the actual requirements of a semiconductor wafer production line;
b. Configuring necessary deep learning frames and dependency libraries, such as CUDA, CUDNN and the like, on a selected hardware platform to support efficient operation of the model;
c. Optimizing hardware resource allocation, and reasonably setting GPU memory use limit and CPU thread number to balance detection performance and system stability.
(2) Model integration and interface development:
a. Designing and realizing a model reasoning interface, which comprises an image preprocessing function module, a model reasoning function module and a result post-processing function module;
b. developing a communication interface with a production line control system to realize real-time transmission and feedback of detection results;
c. and a caching mechanism is constructed, so that data stream processing is optimized, and the influence of I/O operation on the detection speed is reduced.
(3) Real-time image acquisition and pretreatment:
a. the wafer image is acquired in real time through a high-speed industrial camera, so that the image quality and the acquisition frequency are ensured to meet the detection requirement;
b. Realizing an image preprocessing pipeline, including center clipping, denoising and data enhancement operations, so as to improve the quality of an input image;
c. and an asynchronous processing mechanism is adopted, and preprocessing is performed while the image is acquired, so that the computing resource is utilized to the maximum extent.
(4) Defect detection and result output:
a. Inputting the preprocessed image into a deployed wafer defect detection model, and executing reasoning operation;
b. outputting an analysis model, and extracting defect types, boundary frame coordinate information and circular parameter information;
c. and screening and grading the detection result according to a preset threshold value to reduce false alarm and missing report.
(5) And (3) visualizing and storing detection results:
a. developing a real-time visual interface, and intuitively displaying detection results, wherein the detection results comprise defect types, boundary frame coordinate information and circular parameter information;
b. the local storage function of the detection result is realized, and the local storage function comprises an original image, the detection result and related data;
(6) Linkage and feedback of the production line:
a. Transmitting the detection result to a production line control system in real time for automatic decision making;
b. Developing an alarm mechanism, and timely notifying related personnel and triggering emergency response of a production line when a serious defect is detected;
c. And the correlation analysis of the detection result and the production parameter is realized, and data support is provided for process optimization.
(7) Performance evaluation and continuous optimization:
a. Establishing a periodic performance evaluation mechanism, including statistical analysis of key indexes of detection accuracy, recall and processing speed;
b. developing an automatic test flow, and evaluating the performance of the model in an actual production environment by using a standard test set;
c. And continuously optimizing model parameters and deployment strategies according to the performance evaluation result and the production feedback, and continuously improving the overall performance of the detection system.
The invention also provides a visual recognition device for computer processing, which uses the visual recognition method.
Compared with the prior art, the invention has the following beneficial effects:
(1) Multitasking collaborative optimization:
The invention introduces a multi-task collaborative optimization strategy in wafer defect detection, and integrates classification, regression and circular boundary prediction tasks into an integral frame. Conventional detection methods often focus on a single task, such as performing only defect classification or region segmentation, resulting in limitations and deviations in the detection results. Through the multi-task collaborative optimization, the method and the system can utilize the relevance and complementarity among different tasks to improve the overall performance of the model. Specifically, the cooperative mechanism enables the model to improve the detection capability of tiny and edge defects while maintaining the accuracy when processing complex and diversified defects, and remarkably enhances the comprehensiveness and robustness of detection.
(2) Dynamic multi-scale feature extraction and enhancement:
The present invention innovatively proposes a dynamic multi-scale feature extraction and enhancement network architecture to address the different scale defects on semiconductor wafers. Conventional inspection methods typically employ fixed-scale feature extraction, which tends to lose critical information in the face of complex wafer defects, particularly when dealing with defects of varying sizes and morphologies. According to the invention, by introducing a dynamic multi-scale feature extraction network, the extraction scale can be dynamically adjusted according to the geometric characteristics and the position of the defect, so that the global structure and the local detail of the defect are fully captured in the multi-scale feature fusion process. Meanwhile, the feature enhancement mechanism further optimizes the expression capability of key features, so that the model can accurately identify fine defects under a complex background, and the accuracy and the robustness of detection are improved.
(3) Antagonistic collaborative learning training strategy:
In order to further improve the generalization capability of the model, the invention adopts a unique antagonistic collaborative learning training strategy. In the strategy, the model performs collaborative optimization on multiple tasks by simulating various complex conditions in an actual production environment in the training process. By introducing a contrast factor in the multi-stage progressive training, the model is able to effectively cope with different types of wafer defects, especially in the face of unknown or very challenging defects, yet still be able to maintain a high level of inspection accuracy. The strategy remarkably enhances the resistance of the model to various interference factors, so that the model shows stronger robustness and reliability in practical application.
Drawings
FIG. 1 is a diagram of a wafer image dynamic multi-scale feature enhancement network;
Fig. 2 is an overall block diagram of semiconductor wafer defect detection.
Detailed Description
FIG. 1 is a diagram of a wafer image dynamic multi-scale feature enhancement network incorporating feature weight dynamic allocation modules in the network to enhance adaptability to multi-scale features and learn correlations between features at different locations in the wafer feature map. By highlighting important features and suppressing irrelevant features, the mechanism first generates a spatial attention sub-module, i.e. an attention weight for each location, and then combines the spatial attention sub-module with the original feature map to obtain an enhanced feature representation.
Fig. 2 is a block diagram of a semiconductor wafer defect detection system, which includes the steps of firstly, collecting original semiconductor wafer image data based on a real semiconductor wafer generation line, preprocessing the data, including center cutting, denoising and data enhancement, ensuring the smoothness of the image data, finally obtaining a complete wafer defect detection data set, secondly, constructing a dynamic multi-scale feature extraction and feature enhancement network of a wafer defect image aiming at the data set, thereby generating a multi-stage feature map containing local details and global semantic information, then, designing a multi-stage feature map based on the multi-stage feature map, outputting a multi-stage collaborative wafer defect detection model, outputting classification, regression and circular boundary prediction results of wafer defects, realizing visual detection of the semiconductor wafer defects, and finally, performing progressive training by combining training strategies of antagonism collaborative learning, finally obtaining a high-efficiency wafer defect detection model, and disposing the high-efficiency wafer defect detection model into a visual module of semiconductor wafer production equipment, so as to realize real-time defect detection and feedback.
The above description is only of the preferred embodiments of the present application and is not intended to limit the present application, but various modifications and variations can be made to the present application by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the protection scope of the present application.
While the foregoing describes the embodiments of the present invention, it should be understood that the present invention is not limited to the embodiments, and that various modifications and changes can be made by those skilled in the art without any inventive effort.

Claims (7)

1.一种基于计算机处理用视觉识别方法,其特征在于,该方法包括:1. A computer-based visual recognition method, characterized in that the method comprises: S1、基于真实的半导体晶圆生成线采集原始的半导体晶圆图片数据,对于所采集的原始图像数据设计中心裁剪、去噪和数据增强的数据预处理方法,并进行标签裁定,得到完整的晶圆缺陷检测数据集;S1. Collect original semiconductor wafer image data based on a real semiconductor wafer production line, design a data preprocessing method of center cropping, denoising and data enhancement for the collected original image data, and perform label determination to obtain a complete wafer defect detection data set; S2、针对S1所得晶圆缺陷检测数据集,构建晶圆缺陷图像动态多尺度特征提取与特征增强网络,得到包含晶圆缺陷图像的局部细节和全局语义信息的多级特征图;S2. For the wafer defect detection dataset obtained in S1, a dynamic multi-scale feature extraction and feature enhancement network for wafer defect images is constructed to obtain a multi-level feature map containing local details and global semantic information of the wafer defect image; S3、基于S2所得晶圆缺陷的多级特征图,设计多任务协同的晶圆缺陷检测模型,得到晶圆缺陷的分类、回归和圆形边界预测结果,从而得到半导体晶圆缺陷检测模型;S3. Based on the multi-level feature map of wafer defects obtained in S2, a multi-task collaborative wafer defect detection model is designed to obtain the classification, regression and circular boundary prediction results of wafer defects, thereby obtaining a semiconductor wafer defect detection model; S4、基于S3所得半导体晶圆缺陷检测模型,构造结合对抗性协同学习的训练策略,进行多阶段渐进式训练得到最终的晶圆缺陷检测模型;S4, based on the semiconductor wafer defect detection model obtained in S3, construct a training strategy combined with adversarial collaborative learning, and perform multi-stage progressive training to obtain the final wafer defect detection model; S5、模型部署,将S4获得的晶圆缺陷检测模型部署到半导体晶圆生产的过程设备视觉模块上,实时检测并反馈生产的半导体晶圆是否存在缺陷问题。S5, model deployment, deploying the wafer defect detection model obtained in S4 to the vision module of the process equipment for semiconductor wafer production, to detect and provide feedback in real time whether there are defects in the produced semiconductor wafers. 2.根据权利要求1的视觉识别方法,其特征在于,所述S1中数据集的采集与制作过程包括:2. The visual recognition method according to claim 1, characterized in that the process of collecting and making the data set in S1 comprises: (1)原始数据采集:(1) Raw data collection: 将高精度工业相机安装在晶圆传输轨道上方,与光源系统协同工作,确保捕获清晰的晶圆表面图像;原始图像数据以16位TIFF格式保存,每个像素包含65536个灰度级,以保证图像的高动态范围和细节保留;采集不同生产批次、不同工艺阶段的晶圆样本;具体包括:不同尺寸、工艺节点、制程阶段、缺陷类型、缺陷程度;总所采集的原始图像数据记为dataset;A high-precision industrial camera is installed above the wafer transport track and works with the light source system to ensure the capture of clear wafer surface images; the original image data is saved in 16-bit TIFF format, and each pixel contains 65536 gray levels to ensure the high dynamic range and detail retention of the image; wafer samples from different production batches and different process stages are collected; specifically, they include: different sizes, process nodes, process stages, defect types, and defect degrees; the total collected original image data is recorded as dataset; (2)数据预处理:(2) Data preprocessing: ROI区域裁剪及其对模型精度的影响:考虑到晶圆的圆形特性,设计一个圆形ROI(感兴趣区域)裁剪算法:ROI area cropping and its impact on model accuracy: Considering the circular characteristics of the wafer, a circular ROI (region of interest) cropping algorithm is designed: 首先使用Hough圆变换检测晶圆边缘,得到晶圆中心坐标和半径;其次以检测到的中心为圆心,半径略小于检测半径(取98%)的圆形区域作为ROI;最后将ROI以外的区域像素值设为0,保留ROI内的原始图像信息;这种裁剪方法可以有效去除晶圆边缘的干扰信息,提高模型对中心区域缺陷的关注度;First, the wafer edge is detected using Hough circle transform to obtain the wafer center coordinates and radius; secondly, a circular area with the detected center as the center and a radius slightly smaller than the detection radius (98%) is used as the ROI; finally, the pixel values of the area outside the ROI are set to 0, and the original image information within the ROI is retained; this cropping method can effectively remove the interference information of the wafer edge and increase the model's attention to defects in the central area; 图像降噪处理,首先,采用高斯滤波来去除高斯白噪声;高斯滤波通过与高斯核进行卷积,可以有效地平滑图像并去除噪声;具体地,高斯滤波器的计算公式为:Image denoising, first, Gaussian filtering is used to remove Gaussian white noise; Gaussian filtering can effectively smooth the image and remove noise by convolving with the Gaussian kernel; specifically, the calculation formula of the Gaussian filter is: 其中,σ是高斯核的标准差,x和y分别表示图像中的像素位置,用来表示每个像素点到滤波器中心的偏移;这个偏移量用于计算高斯分布函数的值,以确定像素的权重;通过与图像进行卷积操作,高斯滤波可以有效地降低图像中的高斯白噪声;Where σ is the standard deviation of the Gaussian kernel, x and y represent the pixel positions in the image, which are used to represent the offset of each pixel to the center of the filter; this offset is used to calculate the value of the Gaussian distribution function to determine the weight of the pixel; by performing convolution operations with the image, Gaussian filtering can effectively reduce the Gaussian white noise in the image; 通过Albumentations库实现数据增强技术,并在训练过程中实时应用:Data augmentation techniques are implemented through the Albumentations library and applied in real time during training: 随机旋转:在[-10°,10°]范围内随机旋转图像;随机缩放:在[0.9,1.1]范围内随机缩放图像;随机平移:在图像宽高的±5%范围内随机平移;随机亮度和对比度调整:亮度调整范围为[0.8,1.2],对比度调整范围为[0.8,1.2];随机高斯噪声:添加均值为0,标准差在[0,0.05]范围内的高斯噪声;随机水平和垂直翻转:各50%的概率进行翻转;Random rotation: randomly rotate the image in the range of [-10°, 10°]; Random scaling: randomly scale the image in the range of [0.9, 1.1]; Random translation: randomly translate the image within the range of ±5% of the image width and height; Random brightness and contrast adjustment: the brightness adjustment range is [0.8, 1.2], and the contrast adjustment range is [0.8, 1.2]; Random Gaussian noise: add Gaussian noise with a mean of 0 and a standard deviation in the range of [0, 0.05]; Random horizontal and vertical flip: flip with a probability of 50% each; 每张原始图像会生成5张增强后的图像,从而将数据集扩充至500,000张;经过数据预处理后的图像数据记为data;Each original image will generate 5 enhanced images, thus expanding the data set to 500,000 images; the image data after data preprocessing is recorded as data; (3)标签裁定:(3) Labeling ruling: 采用工业级别的标注软件LabelImg,标注过程遵循以下规则:The industrial-grade labeling software LabelImg is used, and the labeling process follows the following rules: 严格的像素级别标注:标注人员需要在图像上精确地绘制出缺陷区域的边界框,以确保标注的准确性和精度;Strict pixel-level annotation: Annotators need to accurately draw the bounding box of the defect area on the image to ensure the accuracy and precision of the annotation; 缺陷分类、回归和圆形边界作为标签:分类标签记为labelcls,对应着五种缺陷类型(无缺陷、划痕、裂纹、剥落和异物附着),其每一维的值表示该类型缺陷的置信度得分;回归标签记为labelreg包含了预测目标的边界框坐标信息,用于精确定位缺陷的位置;圆形边界标签记为labelcir给出了与缺陷边界紧密相关的圆形参数,可以进一步提高了边界定位的精度;Defect classification, regression and circular boundaries as labels: The classification label is recorded as label cls , which corresponds to five defect types (no defect, scratch, crack, peeling and foreign matter attachment). The value of each dimension represents the confidence score of the defect type; the regression label is recorded as label reg, which contains the bounding box coordinate information of the predicted target, which is used to accurately locate the position of the defect; the circular boundary label is recorded as label cir , which gives the circular parameters closely related to the defect boundary, which can further improve the accuracy of boundary positioning; 多个缺陷标注:每个图像可能包含0个或多个缺陷标注,标注人员需要根据实际情况在图像中标注出所有存在的缺陷区域;Multiple defect annotations: Each image may contain 0 or more defect annotations. The annotator needs to mark all defect areas in the image according to the actual situation. 标注结果保存:标注结果按照PASCAL VOC格式进行保存;Annotation result saving: Annotation results are saved in PASCAL VOC format; 标注过程需要标注人员具备一定的专业知识和经验,以确保标注结果的准确性和可用性;The annotation process requires the annotator to have certain professional knowledge and experience to ensure the accuracy and usability of the annotation results; (4)数据集制作:(4) Dataset preparation: 将标注后的数据集记为(data,label),按8:1:1的比例划分训练集、验证集和测试集,具体为:The labeled data set is recorded as (data, label), and the training set, validation set, and test set are divided into 8:1:1 ratios, specifically: 80%的数据用作训练集(Training Set)、10%的数据用作验证集(Validation Set)、10%的数据用作测试集(Test Set)。80% of the data is used as a training set, 10% of the data is used as a validation set, and 10% of the data is used as a test set. 3.根据权利要求1的视觉识别方法,其特征在于,所述S2中的图像动态多尺度特征提取与特征增强网络构建过程过程包括:3. The visual recognition method according to claim 1, characterized in that the image dynamic multi-scale feature extraction and feature enhancement network construction process in S2 comprises: 从底层到顶层对输入晶圆图像进行处理,首先提取不同尺度的特征图{C2,C3,C4,C5},然后按照以下步骤构建多尺度特征增强网络:The input wafer image is processed from the bottom layer to the top layer. First, feature maps {C2, C3, C4, C5} of different scales are extracted, and then a multi-scale feature enhancement network is constructed according to the following steps: 从最深的P5层开始,通过上采样生成上层的晶圆特征图,将这种高级语义信息传递到了多尺度特征增强网络的顶层:Starting from the deepest P5 layer, the upper-level wafer feature map is generated by upsampling, and this high-level semantic information is passed to the top layer of the multi-scale feature enhancement network: M5=C5M5=C5 其中M5是多尺度特征增强网络中最顶层的特征图,来自最深层次C5特征图,然后对M5进行1×1卷积通道数调整:Among them, M5 is the top-level feature map in the multi-scale feature enhancement network, which comes from the deepest C5 feature map, and then the number of 1×1 convolution channels of M5 is adjusted: P5=Conv1×1(M5)P5=Conv1×1(M5) 然后自顶向下构建多尺度特征增强网络:Then build a multi-scale feature enhancement network from top to bottom: M4=C4+Upsample(P5)M4=C4+Upsample(P5) P4=Conv3×3(M4)P4=Conv3×3(M4) 其中,Conv3×3为3×3卷积,用于融合不同尺度的特征,Upsample是反卷积实现的上采样操作,重复以上的步骤逐步生成{P3,P4,P5},最终P3,P4,P5为融合了多尺度信息的多尺度特征增强网络;Among them, Conv3×3 is a 3×3 convolution, which is used to fuse features of different scales. Upsample is an upsampling operation implemented by deconvolution. Repeat the above steps to gradually generate {P3, P4, P5}. Finally, P3, P4, P5 are multi-scale feature enhancement networks that fuse multi-scale information. 在该网络中融入了特征权重动态分配模块,并学习晶圆特征图中不同位置特征之间的相关性;通过突出重要特征并抑制不相关特征,该机制首先生成空间注意力子模块,即每个位置对应的注意力权重,然后将空间注意力子模块与原始特征图相结合,从而获得增强的特征表示;具体步骤如下:The network incorporates a feature weight dynamic allocation module and learns the correlation between features at different positions in the wafer feature map. By highlighting important features and suppressing irrelevant features, the mechanism first generates a spatial attention submodule, that is, the attention weight corresponding to each position, and then combines the spatial attention submodule with the original feature map to obtain an enhanced feature representation. The specific steps are as follows: 特征权重动态分配模块的输入是一个晶圆特征图X,维度为(C×H×W),使用两个1×1卷积核对X进行通道压缩,生成压缩晶圆特征图A和B,压缩后的晶圆特征图维度为(C′×H×W):The input of the feature weight dynamic allocation module is a wafer feature map X with a dimension of (C×H×W). Two 1×1 convolution kernels are used to perform channel compression on X to generate compressed wafer feature maps A and B. The dimensions of the compressed wafer feature map are (C′×H×W): A=X·Wa(C′×H×W)A=X·W a (C′×H×W) B=X·Wb(C′×H×W)B=X·W b (C′×H×W) 其中,Wa和Wb分别为两个1×1卷积核的权重,利用A和B计算位置间的相似度,生成空间注意力子模块映射M,维度为(H×W×H×W):Among them, Wa and Wb are the weights of two 1×1 convolution kernels, respectively. A and B are used to calculate the similarity between positions to generate the spatial attention submodule map M with a dimension of (H×W×H×W): M=Softmax(AT·B)M = Softmax( AT ·B) 然后对原始输入X和M做元素级相乘,得到空间注意力子模块加权后的晶圆特征图X′:Then, the original input X and M are element-wise multiplied to obtain the wafer feature map X′ weighted by the spatial attention submodule: X′=X·MX′=X·M 最后将X′和X相加,再经过一个卷积层输出:Finally, X′ and X are added and then output through a convolutional layer: O=γ·X′+X·Wc O=γ·X′+X·W c 其中,γ是可学习的放缩因子,用于放缩空间注意力子模块增强的效果,Wc是一个1×1卷积核权重,O是特征权重动态分配模块的最终输出晶圆特征图。Among them, γ is a learnable scaling factor used to scale the enhanced effect of the spatial attention submodule, Wc is a 1×1 convolution kernel weight, and O is the final output wafer feature map of the feature weight dynamic allocation module. 4.根据权利要求1的视觉识别方法,其特征在于,所述S3中多任务协同的晶圆缺陷检测模型构建的过程包括:4. The visual recognition method according to claim 1, characterized in that the process of constructing the multi-task collaborative wafer defect detection model in S3 comprises: 模型通过卷积层级联两个子任务的特征图,并相互编码对方的信息,以获得融合了双向关系的分类和回归特征;使用多任务协同检测算法(Multi-task CollaborativeDetection Algorithm,MCDA)引入圆形边界预测分支,通过级联分类和回归预测结果,具体包括以下过程:The model cascades the feature maps of the two subtasks through the convolution layer and encodes each other's information to obtain classification and regression features that integrate bidirectional relationships; the multi-task collaborative detection algorithm (MCDA) is used to introduce a circular boundary prediction branch, and the classification and regression prediction results are cascaded. The specific process includes the following: 将动态多尺度特征提取与特征增强网络输出的特征O送入MCDA,同时完成分类、回归和圆形边界预测:The feature O output by the dynamic multi-scale feature extraction and feature enhancement network is fed into MCDA to complete classification, regression and circular boundary prediction at the same time: clspred,regpred,cirpred=MCDA(O)cls pred ,reg pred ,cir pred =MCDA(O) 其中,clspred,regpred,cirpred分别为晶圆缺陷检测的分类预测结果、回归预测结果以及圆形边界预测结果;Among them, cls pred , reg pred , cir pred are the classification prediction results, regression prediction results, and circular boundary prediction results of wafer defect detection respectively; MCDA的具体计算过程如下:The specific calculation process of MCDA is as follows: 1)对0进行基础卷积,获得分类和回归的初始预测值{clsinit,reginit}:1) Perform basic convolution on 0 to obtain the initial prediction values for classification and regression {cls init , reg init }: clsinit,reginit=Conv(O)cls init ,reg init =Conv(O) 2)将reginit与clsinit进行关系编码,得到融合两者关系的回归特征regfused2) Encode the relationship between reg init and cls init to obtain the regression feature reg fused that integrates the relationship between the two: regfused=reginit+Conv(Concat(reginit,clsinit))reg fused =reg init +Conv(Concat(reg init , cls init )) 3)同时将clsinit与reginit进行关系编码,得到融合两者关系的分类特征clsfused3) At the same time, the relationship between cls init and reg init is encoded to obtain the classification feature cls fused that integrates the relationship between the two: clsfused=clsinit+Conv(Concat(clsinit,reginit))cls fused =cls init +Conv(Concat(cls init ,reg init )) 4)分别对regfused和clsfused进行卷积,得到最终的分类预测clspred和回归预测regpred4) Convolve reg fused and cls fused respectively to obtain the final classification prediction cls pred and regression prediction reg pred : 5)将clspred和regpred进行级联,经过一个额外的卷积分支,预测与缺陷边界紧密相关的缺陷边界圆形参数cirpred5) Cascade cls pred and reg pred , and pass through an additional convolution branch to predict the defect boundary circular parameter cir pred which is closely related to the defect boundary: cirpred=Conv(Concat(clspred,regpred))cir pred =Conv(Concat(cls pred ,reg pred )) 采用多任务协同损失函数,同时优化分类、回归和圆形边界预测三个任务,即:A multi-task collaborative loss function is used to simultaneously optimize the three tasks of classification, regression, and circular boundary prediction, namely: L=λ1Lcls2Lreg3Lcir L=λ 1 L cls2 L reg3 L cir 其中,Lcls为分类损失,使用FocalLoss;Lreg为回归损失,使用CIoULoss;Lcir为环状边界损失,使用PolygonLoss;λ1,λ2,λ3分别为平衡各损失项的超参数;Among them, L cls is the classification loss, using FocalLoss; L reg is the regression loss, using CIoULoss; L cir is the ring boundary loss, using PolygonLoss; λ 1 , λ 2 , λ 3 are hyperparameters for balancing each loss term; 在多任务协同损失函数的优化下,该目标检测网络可同时输出三种预测结果:分类预测结果clspred、回归预测结果regpred和圆形边界预测结果cirpred;Under the optimization of multi-task collaborative loss function, the object detection network can simultaneously output three prediction results: classification prediction result clspred, regression prediction result regpred and circular boundary prediction result cirpred; 其中,分类预测结果clspred为一个向量,对应着四种缺陷类型(划痕、裂纹、剥落、异物附着),其每一维的值表示该类型缺陷的置信度得分;Among them, the classification prediction result clspred is a vector, corresponding to four defect types (scratches, cracks, peeling, and foreign matter attachment), and the value of each dimension represents the confidence score of the defect type; 通过获取分数最高的维度索引,即可确定当前预测框内的缺陷类型;回归预测结果regpred包含了预测目标的边界框坐标信息;圆形边界预测结果cirpred给出了与缺陷边界紧密相关的圆形参数;By obtaining the dimension index with the highest score, the defect type in the current prediction box can be determined; the regression prediction result regpred contains the bounding box coordinate information of the predicted target; the circular boundary prediction result cirpred gives the circular parameters that are closely related to the defect boundary; 在获得上述三种预测结果后,可以对缺陷类型进行判定,并结合精确的位置和形状信息,实现了晶圆各类缺陷的准确检测,为后续的质量控制和缺陷修复提供了重要依据。After obtaining the above three prediction results, the defect type can be determined, and combined with precise position and shape information, accurate detection of various wafer defects can be achieved, providing an important basis for subsequent quality control and defect repair. 5.根据权利要求1的视觉识别方法,其特征在于,所述S4中对抗性协同学习训练策略过程包括:5. The visual recognition method according to claim 1, characterized in that the adversarial collaborative learning training strategy process in S4 comprises: (1)计算多尺度时序一致性损失:(1) Calculate multi-scale temporal consistency loss: 本发明提出了多尺度时序一致性损失,该损失函数在不同时间尺度上衡量预测的一致性:This paper proposes a multi-scale temporal consistency loss, which measures the consistency of predictions at different time scales: Ltc=∑sλs*DKL(P(Y|Xt),P(Y|X{t-s}))L tc =∑ s λ s *D KL (P(Y|X t ),P(Y|X{ts})) 其中,s表示不同的时间尺度,λs是对应的权重,DKL是KL散度,P(Y|Xt)表示在时间步t的检测分布,包含分类检测、回归检测和圆形边界检测,为一个总检测结果概括;这个损失鼓励模型在不同时间尺度上保持预测的一致性,从而提高对不同时间批次数据的建模能力;Among them, s represents different time scales, λ s is the corresponding weight, D KL is the KL divergence, and P(Y|X t ) represents the detection distribution at time step t, including classification detection, regression detection and circular boundary detection, which is a summary of the total detection results; this loss encourages the model to maintain consistency in predictions at different time scales, thereby improving the modeling ability of different time batch data; (2)对抗性训练策略:(2) Adversarial training strategy: 引入了对抗性训练策略,具体为:设计一个判别器D,试图区分特征是来自局部细节还是全局语义信息;融合网络F则被训练来欺骗判别器:An adversarial training strategy is introduced, specifically: a discriminator D is designed to try to distinguish whether the features come from local details or global semantic information; the fusion network F is trained to deceive the discriminator: Ladv=E[log(D(F(Xv,Xa)))]+E[log(1-D(Xv))+log(1-D(Xa))]L adv =E[log(D(F(X v ,X a )))]+E[log(1-D(X v ))+log(1-D(X a ))] 其中,Xv和Xa分别是局部细节和全局语义信息输入;Among them, Xv and Xa are local details and global semantic information input respectively; (3)课程学习与难度自适应:(3) Course learning and difficulty adaptation: 根据(2)对抗性策略,提出一种基于样本难度的课程学习策略;样本难度D(x)定义为:According to (2) adversarial strategy, a curriculum learning strategy based on sample difficulty is proposed; the sample difficulty D(x) is defined as: D(x)=1-exp(-γ(wvLv+waLa))D(x)=1-exp(-γ(w v L v +w a L a )) 其中,Lv和La分别是局部细节和全局语义信息的损失,γ是一个可调节的参数;训练过程中,逐步增加难样本的比例:Among them, Lv and La are the losses of local details and global semantic information respectively, and γ is an adjustable parameter. During the training process, the proportion of difficult samples is gradually increased: 其中,t是当前训练步数,T是总训练步数,p0是初始难样本比例,μ控制难度增加速率;Where t is the current training step, T is the total training step, p0 is the initial difficult sample ratio, and μ controls the rate at which difficulty increases; (4)动态批归一化:(4) Dynamic Batch Normalization: 使用动态批归一化(Dynamic Batch Normalization,DBN)技术,DBN根据输入的统计特性动态调整归一化参数:Using Dynamic Batch Normalization (DBN) technology, DBN dynamically adjusts the normalization parameters according to the statistical characteristics of the input: y=γ(x)(x-μ(x))/(σ(x)+ε)+β(x)y=γ(x)(x-μ(x))/(σ(x)+ε)+β(x) 其中,γ(x)和β(x)是输入依赖的缩放和偏移参数;Where γ(x) and β(x) are input-dependent scaling and offset parameters; (5)模型训练:(5) Model training: 基于上述方法,模型训练的具体流程为:Based on the above method, the specific process of model training is as follows: 1)初始化模型参数θ,其中,θ表示待更新的参数,包含晶圆缺陷检测模型中所有的权重矩阵和偏置向量;1) Initializing the model parameters θ, where θ represents the parameters to be updated, including all weight matrices and bias vectors in the wafer defect detection model; 2)对于每个训练步骤t:2) For each training step t: a.从数据集中采样一个批次,按phard(t)的比例包含难样本;a. Sample a batch from the dataset, containing hard samples in proportion to phard(t); b.进行前向传播,得到多尺度特征和预测结果;b. Perform forward propagation to obtain multi-scale features and prediction results; c.计算多尺度时序一致性损失Ltcc. Calculate the multi-scale temporal consistency loss L tc ; d.进行对抗性训练策略,更新判别器D和融合网络F;d. Perform adversarial training strategy to update the discriminator D and fusion network F; e.应用动态批归一化;e. Apply dynamic batch normalization; f.计算总损失:Ltotal=λtc*Ltcadv*Ladvf. Calculate the total loss: L total = λ tc *L tc + λ adv *L adv ; g.反向传播,更新模型参数:其中为学习率;g. Back propagation, update model parameters: in is the learning rate; 3)重复步骤2),当连续N个epoch无法降低损失时,终止训练过程;在训练结束时保存最终的模型参数,作为晶圆缺陷检测的部署模型。3) Repeat step 2) and terminate the training process when the loss cannot be reduced after N consecutive epochs; save the final model parameters at the end of the training as the deployment model for wafer defect detection. 6.根据权利要求1的视觉识别方法,其特征在于,所述S5中晶圆缺陷检测模型部署过程包括:6. The visual recognition method according to claim 1, characterized in that the wafer defect detection model deployment process in S5 comprises: (1)硬件平台选择与配置:(1) Hardware platform selection and configuration: a.根据半导体晶圆生产线的实际需求,选择适合的边缘计算设备,如NVIDIA Jetson系列或Intel NUC高性能嵌入式系统;a. Select suitable edge computing devices, such as NVIDIA Jetson series or Intel NUC high-performance embedded systems, based on the actual needs of the semiconductor wafer production line; b.在选定的硬件平台上配置必要的深度学习框架和依赖库:CUDA、cuDNN等;b. Configure the necessary deep learning frameworks and dependent libraries on the selected hardware platform: CUDA, cuDNN, etc.; c.优化硬件资源分配,合理设置GPU内存使用限制和CPU线程数;c. Optimize hardware resource allocation and reasonably set GPU memory usage limits and CPU thread numbers; (2)模型集成与接口开发:(2) Model integration and interface development: a.设计并实现模型推理接口,包括图像预处理、模型推理和结果后处理功能模块;a. Design and implement the model reasoning interface, including image preprocessing, model reasoning and result post-processing functional modules; b.开发与生产线控制系统的通信接口,实现检测结果的实时传输和反馈;b. Develop a communication interface with the production line control system to achieve real-time transmission and feedback of test results; c.构建缓存机制,优化数据流处理,减少I/O操作对检测速度的影响;c. Build a cache mechanism, optimize data stream processing, and reduce the impact of I/O operations on detection speed; (3)实时图像采集与预处理:(3) Real-time image acquisition and preprocessing: a.通过高速工业相机实时采集晶圆图像;a. Real-time acquisition of wafer images through high-speed industrial cameras; b.实现图像预处理流水线,包括中心裁剪、去噪和数据增强操作;b. Implement the image preprocessing pipeline, including center cropping, denoising, and data augmentation operations; c.采用异步处理机制,在图像采集的同时进行预处理;c. Use asynchronous processing mechanism to perform preprocessing while image acquisition; (4)缺陷检测与结果输出:(4) Defect detection and result output: a.将预处理后的图像输入部署的晶圆缺陷检测模型,执行推理操作;a. Input the preprocessed image into the deployed wafer defect detection model to perform inference operations; b.解析模型输出,提取缺陷类别、边界框坐标信息和圆形参数信息;b. Parse the model output to extract defect category, bounding box coordinate information, and circle parameter information; c.根据预设阈值,对检测结果进行筛选和分级;c. Screen and classify the test results according to the preset threshold; (5)检测结果可视化与存储:(5) Visualization and storage of test results: a.开发实时可视化界面,直观展示检测结果,包括缺陷类别、边界框坐标信息和圆形参数信息;a. Develop a real-time visualization interface to intuitively display the inspection results, including defect categories, bounding box coordinate information, and circle parameter information; b.实现检测结果的本地存储功能,包括原始图像、检测结果及相关数据;b. Realize the local storage function of the test results, including original images, test results and related data; (6)生产线联动与反馈:(6) Production line linkage and feedback: a.将检测结果实时传输至生产线控制系统,用于自动化决策;a. Transmit the test results to the production line control system in real time for automated decision-making; b.开发报警机制,当检测到严重缺陷时,及时通知相关人员并触发生产线应急响应;b. Develop an alarm mechanism to promptly notify relevant personnel and trigger an emergency response on the production line when a serious defect is detected; c.实现检测结果与生产参数的关联分析,为工艺优化提供数据支持;c. Realize the correlation analysis between test results and production parameters to provide data support for process optimization; (7)性能评估与持续优化:(7) Performance evaluation and continuous optimization: a.建立定期性能评估机制,包括检测准确率、召回率和处理速度关键指标的统计分析;a. Establish a regular performance evaluation mechanism, including statistical analysis of key indicators of detection accuracy, recall rate and processing speed; b.开发自动化测试流程,使用标准测试集评估模型在实际生产环境中的表现;b. Develop an automated testing process and use standard test sets to evaluate the performance of the model in the actual production environment; c.根据性能评估结果和生产反馈,持续优化模型参数和部署策略,不断提升检测系统的整体性能。c. Based on performance evaluation results and production feedback, continuously optimize model parameters and deployment strategies to continuously improve the overall performance of the detection system. 7.一种基于计算机处理用视觉识别装置,其特征在于,所述视觉识别装置使用权利要求1~6任一项视觉检测方法。7. A visual recognition device based on computer processing, characterized in that the visual recognition device uses any one of the visual detection methods of claims 1 to 6.
CN202411219159.2A 2024-09-02 2024-09-02 A visual recognition method and device based on computer processing Pending CN119180988A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202411219159.2A CN119180988A (en) 2024-09-02 2024-09-02 A visual recognition method and device based on computer processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202411219159.2A CN119180988A (en) 2024-09-02 2024-09-02 A visual recognition method and device based on computer processing

Publications (1)

Publication Number Publication Date
CN119180988A true CN119180988A (en) 2024-12-24

Family

ID=93900659

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202411219159.2A Pending CN119180988A (en) 2024-09-02 2024-09-02 A visual recognition method and device based on computer processing

Country Status (1)

Country Link
CN (1) CN119180988A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119380166A (en) * 2024-12-30 2025-01-28 杭州宇泛智能科技股份有限公司 Safety monitoring method based on hybrid large model and neural network algorithm
CN119619175A (en) * 2025-02-12 2025-03-14 北京嘉海鼎盛科技有限公司 A semiconductor testing equipment intelligent calibration method and system
CN119762733A (en) * 2025-03-10 2025-04-04 合肥综合性国家科学中心能源研究院(安徽省能源实验室) Wire rope defect identification and positioning system and method based on real-time target detection
CN119963555A (en) * 2025-04-10 2025-05-09 广东索鲁达科技有限公司 Method and system for detecting wafer defects based on attention pyramid changes
CN120451176A (en) * 2025-07-14 2025-08-08 杭州映图智能科技有限公司 AI-based bottle cap 360-degree full-view defect high-speed detection method and system

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119380166A (en) * 2024-12-30 2025-01-28 杭州宇泛智能科技股份有限公司 Safety monitoring method based on hybrid large model and neural network algorithm
CN119619175A (en) * 2025-02-12 2025-03-14 北京嘉海鼎盛科技有限公司 A semiconductor testing equipment intelligent calibration method and system
CN119762733A (en) * 2025-03-10 2025-04-04 合肥综合性国家科学中心能源研究院(安徽省能源实验室) Wire rope defect identification and positioning system and method based on real-time target detection
CN119963555A (en) * 2025-04-10 2025-05-09 广东索鲁达科技有限公司 Method and system for detecting wafer defects based on attention pyramid changes
CN120451176A (en) * 2025-07-14 2025-08-08 杭州映图智能科技有限公司 AI-based bottle cap 360-degree full-view defect high-speed detection method and system
CN120451176B (en) * 2025-07-14 2025-09-05 杭州映图智能科技有限公司 AI-based bottle cap 360-degree full-view defect high-speed detection method and system

Similar Documents

Publication Publication Date Title
CN119180988A (en) A visual recognition method and device based on computer processing
US11816149B2 (en) Electronic device and control method thereof
CN114663346A (en) Strip steel surface defect detection method based on improved YOLOv5 network
CN113822880A (en) A deep learning-based crack identification method
CN107133569A (en) The many granularity mask methods of monitor video based on extensive Multi-label learning
CN115439458A (en) Industrial image defect target detection algorithm based on depth map attention
CN113808079B (en) Industrial product surface defect self-adaptive detection method based on deep learning model AGLNet
CN113487610B (en) Herpes image recognition method and device, computer equipment and storage medium
Cui et al. Real-time detection of wood defects based on SPP-improved YOLO algorithm
CN114627366A (en) Industrial intelligent visual algorithm cloud service system and method
Lin et al. Integrated circuit board object detection and image augmentation fusion model based on YOLO
CN120014500A (en) Road crack analysis method based on improved YOLO algorithm
CN119559435A (en) Multi-modal small sample image classification method and system based on multi-scale dynamic feature fusion
CN119625440B (en) Flotation froth anomaly classification method and system based on multi-scale feature fusion
CN119863700A (en) Visual-based marine industry equipment identification method
CN110097603B (en) Fashionable image dominant hue analysis method
CN119323567B (en) Module lens dirt detection and glue line detection method
Liu et al. An improved YOLOv8 model and mask convolutional autoencoder for multi-scale defect detection of ceramic tiles
CN110889418A (en) Gas contour identification method
CN119274122A (en) A sanitation work evaluation method based on MRS-YOLO model
Shi et al. Positive Anchor Area Merge Algorithm: A Knowledge Distillation Algorithm for Fruit Detection Tasks Based on Yolov8
He et al. A review of fabric defect detection in textile manufacturing
CN119068179B (en) Remote sensing image aerial airplane detection method based on ambiguity guiding region screening
Samuthiram et al. Surface defect prediction on printed circuit boards using a novel deep learning model with spatial and channel attention-based DenseNet
Wei et al. YOLO11n-Screw: A Recognition Method for Aircraft Screw Tail Marks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载