CN114639124A - A Pedestrian Re-identification Method Based on Weighted Feature Fusion - Google Patents
A Pedestrian Re-identification Method Based on Weighted Feature Fusion Download PDFInfo
- Publication number
- CN114639124A CN114639124A CN202210321956.6A CN202210321956A CN114639124A CN 114639124 A CN114639124 A CN 114639124A CN 202210321956 A CN202210321956 A CN 202210321956A CN 114639124 A CN114639124 A CN 114639124A
- Authority
- CN
- China
- Prior art keywords
- features
- pedestrian
- loss function
- global
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于加权特征融合的行人重识别方法,步骤为:输入行人图像,通过主干网络ResNeSt‑50提取到行人的特征;将从主干网络获取的不同特征输入到加权特征融合金字塔模型中,得到多尺度的包含高层语义信息和低层高分辨率信息的特征;在全局分支网络中将高层语义特征作为全局特征;在局部分支网络中将低层高分辨率特征作为局部特征;联合Softmax损失函数、三元组损失函数和中心损失函数分别对全局特征和局部特征进行优化训练。本发明在加权特征金字塔网络中进行特征的融合,将高层语义特征语作为全局特征;联合Softmax损失函数、三元组损失函数和中心损失函数分别对全局特征和局部特征进行优化训练,提高了行人识别的精度和准确率。
The invention discloses a pedestrian re-identification method based on weighted feature fusion. The steps are: inputting pedestrian images, and extracting pedestrian features through a backbone network ResNeSt-50; inputting different features obtained from the backbone network into a weighted feature fusion pyramid model In the process, multi-scale features containing high-level semantic information and low-level high-resolution information are obtained; high-level semantic features are used as global features in the global branch network; low-level high-resolution features are used as local features in the local branch network; joint Softmax loss function, triple loss function and center loss function to optimize training on global features and local features, respectively. The present invention performs feature fusion in a weighted feature pyramid network, and uses high-level semantic feature words as global features; jointly optimizes and trains global features and local features by combining Softmax loss function, triple loss function and central loss function, and improves pedestrian performance. Recognition accuracy and accuracy.
Description
技术领域technical field
本发明属于行人重识别的技术领域,尤其涉及一种基于加权特征融合的行人重识别方法。The invention belongs to the technical field of pedestrian re-identification, and in particular relates to a pedestrian re-identification method based on weighted feature fusion.
背景技术Background technique
行人重识别是指在跨摄像头跨场景的情况下,寻找特定的行人视频序列或图像。随着智能视频系统和智能安防应用的领域越来越多,行人重识别也吸引着越来越多的学者进行研究。由于摄像头参数不同和场景的不同,会有光照、遮挡或多姿态变化等问题,导致同一个人拍摄出来的行人图像也会有很大的差异,这就会为网络提取特征增加了难度。Pedestrian re-identification refers to finding a specific pedestrian video sequence or image in the case of cross-camera and cross-scene. With more and more fields of intelligent video systems and intelligent security applications, pedestrian re-identification is also attracting more and more scholars to conduct research. Due to different camera parameters and different scenes, there will be problems such as illumination, occlusion or multi-pose changes, resulting in great differences in pedestrian images captured by the same person, which will increase the difficulty for the network to extract features.
目前的行人重识别方法大多采用全局特征和局部特征融合的方法,一般会引进特征金字塔加强高层语义特征和低层高分辨率特征的融合,但是在融合的过程中没有考虑在特征融合时,不同特征融合的贡献度问题。Most of the current pedestrian re-identification methods use the fusion of global features and local features. Generally, feature pyramids are introduced to strengthen the fusion of high-level semantic features and low-level high-resolution features. However, in the process of fusion, different features are not considered during feature fusion. Contribution of Fusion.
发明内容SUMMARY OF THE INVENTION
基于以上现有技术的不足,本发明所解决的技术问题在于提供一种基于加权特征融合的行人重识别方法,通过ResNeSt-50网络提取特征,在加权特征金字塔网络中进行特征的融合,将高层语义特征语作为全局特征,将低层高分辨率特征作为局部特征;联合Softmax损失函数、三元组损失函数和中心损失函数分别对全局特征和局部特征进行优化训练,提高了行人识别的精度和准确率。Based on the above deficiencies of the prior art, the technical problem solved by the present invention is to provide a pedestrian re-identification method based on weighted feature fusion. Semantic feature words are used as global features, and low-level high-resolution features are used as local features. Combined with Softmax loss function, triple loss function and center loss function, the global and local features are optimized and trained respectively, which improves the accuracy and accuracy of pedestrian recognition. Rate.
为了解决上述技术问题,本发明通过以下技术方案来实现:In order to solve the above-mentioned technical problems, the present invention realizes through the following technical solutions:
本发明的基于加权特征融合的行人重识别方法,包括以下步骤:The pedestrian re-identification method based on weighted feature fusion of the present invention comprises the following steps:
步骤S1:输入行人图像,通过主干网络ResNeSt-50提取到行人的特征;Step S1: Input the pedestrian image, and extract the characteristics of the pedestrian through the backbone network ResNeSt-50;
步骤S2:将从主干网络获取的不同特征输入到加权特征融合金字塔模型中,得到多尺度的包含高层语义信息和低层高分辨率信息的特征;Step S2: Input different features obtained from the backbone network into the weighted feature fusion pyramid model to obtain multi-scale features including high-level semantic information and low-level high-resolution information;
步骤S3:在全局分支网络中将高层语义特征作为全局特征;在局部分支网络中将低层高分辨率特征作为局部特征;Step S3: use high-level semantic features as global features in the global branch network; use low-level high-resolution features as local features in the local branch network;
步骤S4:联合Softmax损失函数、三元组损失函数和中心损失函数分别对全局特征和局部特征进行优化训练。Step S4: Combine the Softmax loss function, the triplet loss function and the center loss function to optimize and train the global features and local features respectively.
优选的,在步骤S1中的行人图像会随机的进行随机擦除和水平翻转操作,增加了行人数据量;当使用ResNeSt-50网络提取行人特征时,会获得下采样3、4和5倍的特征图。Preferably, the pedestrian image in step S1 will be randomly erased and flipped horizontally, which increases the amount of pedestrian data; when using the ResNeSt-50 network to extract pedestrian features, downsampling 3, 4, and 5 times will be obtained. feature map.
可选的,在步骤2中,将步骤1中获得的下采样3、4和5倍的特征图作为输入到加权特征金字塔网络中,获得下采样3、4、5和6倍的经过加权融合的特征图。Optionally, in step 2, the feature maps downsampled by 3, 4, and 5 times obtained in
可选的,在步骤3中,将步骤2中下采样5倍和6倍的特征图在通道维度上进行连接,再经过全局平均池化获得全局特征;下采样3倍和4倍的特征图在通道维度上进行连接,将特征图水平均匀的划分为4份,然后对这4个特征图分别进行广义均值池化,并将结果连接起来获得局部特征。Optionally, in step 3, the feature maps downsampled by 5 times and 6 times in step 2 are connected in the channel dimension, and then global features are obtained through global average pooling; the feature maps are downsampled by 3 times and 4 times. Connect in the channel dimension, divide the feature map horizontally into 4 parts, and then perform generalized mean pooling on these 4 feature maps respectively, and connect the results to obtain local features.
由上,本发明的基于加权特征融合的行人重识别方法至少具有如下有益效果:From the above, the pedestrian re-identification method based on weighted feature fusion of the present invention has at least the following beneficial effects:
1、使用ResNeSt-50网络提取特征,可以使网络更加关注到行人图像的重点区域,减少对非重点区域的关注;1. Using the ResNeSt-50 network to extract features can make the network pay more attention to the key areas of pedestrian images and reduce the attention to non-key areas;
2、使用带加权操作的特征金字塔网络,可以更好地获得更加丰富的层级特征信息,从而提高网络的泛化能力和鲁棒性;2. Using the feature pyramid network with weighted operation can better obtain richer hierarchical feature information, thereby improving the generalization ability and robustness of the network;
3、使用联合Softmax损失函数、三元组损失函数和中心损失函数的损失函数,可以更好的对模型进行优化。3. Using the loss function of the joint Softmax loss function, triple loss function and center loss function, the model can be better optimized.
上述说明仅是本发明技术方案的概述,为了能够更清楚了解本发明的技术手段,而可依照说明书的内容予以实施,并且为了让本发明的上述和其他目的、特征和优点能够更明显易懂,以下结合优选实施例,并配合附图,详细说明如下。The above description is only an overview of the technical solution of the present invention, in order to be able to understand the technical means of the present invention more clearly, it can be implemented according to the content of the description, and in order to make the above and other objects, features and advantages of the present invention more obvious and easy to understand , the following detailed description is given in conjunction with the preferred embodiments and in conjunction with the accompanying drawings.
附图说明Description of drawings
为了更清楚地说明本发明实施例的技术方案,下面将对实施例的附图作简单地介绍。In order to illustrate the technical solutions of the embodiments of the present invention more clearly, the accompanying drawings of the embodiments will be briefly introduced below.
图1是ResNeSt-50分裂结构图;Fig. 1 is ResNeSt-50 splitting structure diagram;
图2是加权特征金字塔网络结构图;Fig. 2 is a weighted feature pyramid network structure diagram;
图3是分支网络结构图;Fig. 3 is a branch network structure diagram;
图4是本发明的基于加权特征融合的行人重识别方法的流程图。FIG. 4 is a flow chart of the pedestrian re-identification method based on weighted feature fusion of the present invention.
具体实施方式Detailed ways
下面结合附图详细说明本发明的具体实施方式,其作为本说明书的一部分,通过实施例来说明本发明的原理,本发明的其他方面、特征及其优点通过该详细说明将会变得一目了然。在所参照的附图中,不同的图中相同或相似的部件使用相同的附图标号来表示。The specific embodiments of the present invention will be described in detail below with reference to the accompanying drawings. As a part of this specification, the principles of the present invention will be illustrated by examples. Other aspects, features and advantages of the present invention will become apparent from the detailed description. In the figures to which reference is made, the same reference numerals are used for the same or similar parts in different figures.
下面参照图1-4对本发明的基于加权特征融合的行人重识别方法进行详细说明。本发明的基于加权特征融合的行人重识别方法包括以下步骤:The pedestrian re-identification method based on weighted feature fusion of the present invention will be described in detail below with reference to FIGS. 1-4 . The pedestrian re-identification method based on weighted feature fusion of the present invention comprises the following steps:
步骤1:输入行人图像,通过主干网络ResNeSt-50提取到行人的特征;Step 1: Input the pedestrian image, and extract the characteristics of the pedestrian through the backbone network ResNeSt-50;
步骤2:将从主干网络获取的不同特征输入到加权特征融合金字塔模型中,得到多尺度的包含高层语义信息和低层高分辨率信息的特征;Step 2: Input the different features obtained from the backbone network into the weighted feature fusion pyramid model to obtain multi-scale features containing high-level semantic information and low-level high-resolution information;
步骤3:在全局分支网络中将高层语义特征语作为全局特征;在局部分支网络中将低层高分辨率特征作为局部特征,分支网络图如图3所示;Step 3: In the global branch network, high-level semantic features are used as global features; in the local branch network, low-level high-resolution features are used as local features, and the branch network diagram is shown in Figure 3;
步骤4:联合Softmax损失函数、三元组损失函数和中心损失函数分别对全局特征和局部特征进行优化训练。Step 4: Combine the Softmax loss function, the triple loss function and the center loss function to optimize the training of global features and local features respectively.
具体的,所述步骤1的步骤如下:Specifically, the steps of
步骤1.1:对行人图片随机的进行随机擦除和水平翻转,然后将图片的大小调整为384×128;Step 1.1: Randomly erase and flip the pedestrian image randomly, and then resize the image to 384×128;
步骤1.2:输入到ResNeSt-50网络中进行行人特征的提取,最终会提取到下采样3倍、下采样4倍和下采样5倍的特征,分别记为C3、C4和C5,ResNeSt-50网络中分裂结构如图1所示,ResNeSt-50网络可以使网络更加关注到行人图像的重点区域,减少对非重点区域的关注。Step 1.2: Input into the ResNeSt-50 network to extract pedestrian features, and finally extract features with downsampling 3 times, downsampling 4 times, and downsampling 5 times, respectively denoted as C 3 , C 4 and C 5 , ResNeSt The split structure in the -50 network is shown in Figure 1. The ResNeSt-50 network can make the network pay more attention to the key areas of pedestrian images and reduce the attention to non-key areas.
具体的,步骤2的步骤如下:Specifically, the steps of step 2 are as follows:
步骤2.1:将步骤1.2中获得C3、C4和C5输入到加权特征金字塔网络中,加权特征金字塔网络如图2所示;Step 2.1: Input C 3 , C 4 and C 5 obtained in step 1.2 into the weighted feature pyramid network, the weighted feature pyramid network is shown in Figure 2;
步骤2.2:C3、C4和C5经过卷积操作后获得统一通道数为512的特征图P3_in、P4_in和P5_in,然后对P5_in进行最大池化操作,获取下采样6倍的特征图P6_in。然后对P6_in进行卷积操作获得特征图P6_td,对P6_td上采样后与P5_in相加得到P5_td,对P5_td上采样后与P4_in相加获得P4_td,每一层级以此类推,最终获得特征图P6_out、P5_out、P4_out和P3_out。以P5_td和P5_out为例,运算公式如下所示:Step 2.2: C 3 , C 4 and C 5 obtain feature maps P 3_in , P 4_in and P 5_in with a uniform channel number of 512 after the convolution operation, and then perform a maximum pooling operation on P 5_in to obtain 6 times of downsampling Feature map P 6_in . Then perform a convolution operation on P 6_in to obtain a feature map P 6_td , upsample P 6_td and add it to P 5_in to obtain P 5_td , upsample P 5_td and add it to P 4_in to obtain P 4_td , and so on for each level , and finally obtain feature maps P 6_out , P 5_out , P 4_out and P 3_out . Taking P 5_td and P 5_out as an example, the operation formula is as follows:
其中ωi是可学习的参数,0≤ωi≤1,ε是一个非常小的值,主要作用是为了避免出现分母为0的情况。Where ω i is a learnable parameter, 0≤ω i ≤1, ε is a very small value, the main function is to avoid the situation where the denominator is 0.
具体的,所述步骤3的步骤如下:Specifically, the steps of step 3 are as follows:
步骤3.1:将P6_out上采样成与P5_out相同大小的特征图,然后在通道维度上与P5_out进行连接,然后输入到全局平均池化(Global Average Pooling,GAP)中获得行人图像的全局特征。Step 3.1: Upsample P6_out into a feature map of the same size as P5_out , then connect it with P5_out in the channel dimension, and then input it into Global Average Pooling (GAP) to obtain global features of pedestrian images .
步骤3.2:将P3_out进行下采样得到与P4_out相同大小的特征图,然后在通道维度中与P4_out连接,获得局部特征图,然后将局部特征图沿水平方向均匀的划分为4份,然后对这4个特征图分别进行广义均值池化(Generalized Mean Pooling,GeM),并将结果连接起来,最后经过一个1×1的卷积对特征图进行降维获得局部特征图。局部特征图中包含行人图像的高分辨率特征。在局部分支中,广义均值池化的公式为:Step 3.2: Downsample P 3_out to obtain a feature map of the same size as P 4_out , and then connect it with P 4_out in the channel dimension to obtain a local feature map, and then divide the local feature map into 4 evenly in the horizontal direction, and then Generalized Mean Pooling (GeM) is performed on the four feature maps respectively, and the results are connected. Finally, a 1×1 convolution is performed to reduce the dimension of the feature map to obtain a local feature map. The local feature maps contain high-resolution features of pedestrian images. In the local branch, the formula for generalized mean pooling is:
其中pk为超参数,当pk=1时,f(g)为平均池化;当pk→∞时,f(g)为最大池化。在训练的过程中,广义均值池化会自动学习超参数pk,可以让模型更好捕获特征差异性。where p k is a hyperparameter, f (g) is average pooling when p k = 1, and f (g) is max pooling when p k → ∞. During the training process, generalized mean pooling will automatically learn the hyperparameter p k , which allows the model to better capture feature differences.
以上所述是本发明的优选实施方式而已,当然不能以此来限定本发明之权利范围,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和变动,这些改进和变动也视为本发明的保护范围。The above are only the preferred embodiments of the present invention, of course, the scope of the rights of the present invention cannot be limited by this. Several improvements and changes are made, and these improvements and changes are also regarded as the protection scope of the present invention.
Claims (4)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202210321956.6A CN114639124A (en) | 2022-03-29 | 2022-03-29 | A Pedestrian Re-identification Method Based on Weighted Feature Fusion |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202210321956.6A CN114639124A (en) | 2022-03-29 | 2022-03-29 | A Pedestrian Re-identification Method Based on Weighted Feature Fusion |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN114639124A true CN114639124A (en) | 2022-06-17 |
Family
ID=81951755
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202210321956.6A Pending CN114639124A (en) | 2022-03-29 | 2022-03-29 | A Pedestrian Re-identification Method Based on Weighted Feature Fusion |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN114639124A (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN115223193A (en) * | 2022-06-19 | 2022-10-21 | 浙江爱达科技有限公司 | Capsule endoscope image focus identification method based on focus feature importance |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113408492A (en) * | 2021-07-23 | 2021-09-17 | 四川大学 | Pedestrian re-identification method based on global-local feature dynamic alignment |
| WO2021203801A1 (en) * | 2020-04-08 | 2021-10-14 | 苏州浪潮智能科技有限公司 | Person re-identification method and apparatus, electronic device, and storage medium |
| CN113516012A (en) * | 2021-04-09 | 2021-10-19 | 湖北工业大学 | Pedestrian re-identification method and system based on multi-level feature fusion |
| CN113591697A (en) * | 2021-07-30 | 2021-11-02 | 上海电科智能系统股份有限公司 | Video pedestrian re-identification method based on triple pyramid model and migration fusion |
-
2022
- 2022-03-29 CN CN202210321956.6A patent/CN114639124A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2021203801A1 (en) * | 2020-04-08 | 2021-10-14 | 苏州浪潮智能科技有限公司 | Person re-identification method and apparatus, electronic device, and storage medium |
| CN113516012A (en) * | 2021-04-09 | 2021-10-19 | 湖北工业大学 | Pedestrian re-identification method and system based on multi-level feature fusion |
| CN113408492A (en) * | 2021-07-23 | 2021-09-17 | 四川大学 | Pedestrian re-identification method based on global-local feature dynamic alignment |
| CN113591697A (en) * | 2021-07-30 | 2021-11-02 | 上海电科智能系统股份有限公司 | Video pedestrian re-identification method based on triple pyramid model and migration fusion |
Non-Patent Citations (2)
| Title |
|---|
| 孙劲光 等: "基于多尺度加权特征融合的行人重识别方法研究", 《信 号处理》, 10 June 2022 (2022-06-10), pages 1 - 12 * |
| 满船清梦压星河HK: "【论文笔记】EfficientDet(BiFPN)(2020)", pages 1 - 6, Retrieved from the Internet <URL:https://blog.csdn.net/qq_38253797/article/details/118439965> * |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN115223193A (en) * | 2022-06-19 | 2022-10-21 | 浙江爱达科技有限公司 | Capsule endoscope image focus identification method based on focus feature importance |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN110188817B (en) | A real-time high-performance semantic segmentation method for street view images based on deep learning | |
| CN111539887B (en) | A Neural Network Image Dehazing Method Based on Hybrid Convolutional Attention Mechanism and Hierarchical Learning | |
| CN109389556B (en) | Multi-scale cavity convolutional neural network super-resolution reconstruction method and device | |
| WO2023056889A1 (en) | Model training and scene recognition method and apparatus, device, and medium | |
| CN110059538B (en) | Water body identification method based on deep dense neural network | |
| WO2023212997A1 (en) | Knowledge distillation based neural network training method, device, and storage medium | |
| CN114936605A (en) | A neural network training method, equipment and storage medium based on knowledge distillation | |
| CN116386104A (en) | Self-supervision facial expression recognition method combining contrast learning and mask image modeling | |
| CN116664397B (en) | TransSR-Net structured image super-resolution reconstruction method | |
| CN116524307A (en) | Self-supervision pre-training method based on diffusion model | |
| CN113298084B (en) | Feature map extraction method and system for semantic segmentation | |
| CN115631513B (en) | Transformer-based multi-scale pedestrian re-identification method | |
| CN110706239A (en) | Scene segmentation method fusing full convolution neural network and improved ASPP module | |
| CN111414988A (en) | Remote sensing image super-resolution method based on multi-scale feature self-adaptive fusion network | |
| CN110555461A (en) | scene classification method and system based on multi-structure convolutional neural network feature fusion | |
| CN117315752A (en) | Training method, device, equipment and medium for face emotion recognition network model | |
| CN114639124A (en) | A Pedestrian Re-identification Method Based on Weighted Feature Fusion | |
| CN114596488B (en) | Lightweight remote sensing target detection method based on dense feature fusion network | |
| CN116363149A (en) | An Improved Medical Image Segmentation Method Based on U-Net | |
| CN115797684A (en) | Infrared small target detection method and system based on context information | |
| CN113393385B (en) | Multi-scale fusion-based unsupervised rain removing method, system, device and medium | |
| CN113222016B (en) | Change detection method and device based on cross enhancement of high-level and low-level features | |
| CN108961270A (en) | A kind of Bridge Crack Image Segmentation Model based on semantic segmentation | |
| CN118262256A (en) | Multi-scale feature fusion small target detection algorithm for unmanned aerial vehicle aerial image | |
| CN118279587A (en) | Multi-scale feature fusion remote sensing image semantic segmentation method based on RemoteCLIP image encoder |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination |