+

CN114821042A - An R-FCN knife gate detection method combining local and global features - Google Patents

An R-FCN knife gate detection method combining local and global features Download PDF

Info

Publication number
CN114821042A
CN114821042A CN202210453322.6A CN202210453322A CN114821042A CN 114821042 A CN114821042 A CN 114821042A CN 202210453322 A CN202210453322 A CN 202210453322A CN 114821042 A CN114821042 A CN 114821042A
Authority
CN
China
Prior art keywords
prediction
global
fcn
local
knife gate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210453322.6A
Other languages
Chinese (zh)
Other versions
CN114821042B (en
Inventor
肖振远
宗起振
陶征勇
李佑文
褚红健
曾清旋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Sac Rail Traffic Engineering Co ltd
Original Assignee
Nanjing Sac Rail Traffic Engineering Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Sac Rail Traffic Engineering Co ltd filed Critical Nanjing Sac Rail Traffic Engineering Co ltd
Priority to CN202210453322.6A priority Critical patent/CN114821042B/en
Publication of CN114821042A publication Critical patent/CN114821042A/en
Application granted granted Critical
Publication of CN114821042B publication Critical patent/CN114821042B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01RMEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
    • G01R31/00Arrangements for testing electric properties; Arrangements for locating electric faults; Arrangements for electrical testing characterised by what is being tested not provided for elsewhere
    • G01R31/327Testing of circuit interrupters, switches or circuit-breakers
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an R-FCN disconnecting link detection method combining local features and global features. The method comprises the steps that a network camera centralized control software is embedded into a transformer substation auxiliary monitoring system to achieve knife switch image acquisition of cameras in multiple positions and multiple angles in different outdoor weather environments and different backgrounds, and diverse knife switch data are constructed; by parallelly connecting the R-FCN output prediction network with the global feature prediction module, the defect of insufficient receptive field generated by the original network only through predicting the disconnecting link by local features is supplemented, the accuracy of detecting the partially shielded disconnecting link is improved, and the omission factor and the false detection factor of the disconnecting link in a complex background are reduced; the prediction results are accumulated after the local feature prediction results and the global feature prediction results are regularized, so that the local feature prediction results are supplemented by the global feature prediction results, and the requirements of remote detection and unattended operation of the disconnecting link of the transformer substation are met.

Description

一种联合局部特征和全局特征的R-FCN刀闸检测方法An R-FCN knife gate detection method combining local and global features

技术领域technical field

本发明涉及变电站辅助监控系统中刀闸检测的技术领域,且涉及一种联合局部特征和全局特征的R-FCN刀闸检测方法。The invention relates to the technical field of knife gate detection in a substation auxiliary monitoring system, and relates to an R-FCN knife gate detection method combining local features and global features.

背景技术Background technique

变电站是高压输电中重要的中转场所,站内运行着各种至关重要的通断控制器件。以往变电站出现中断工作时,相关作业人员需要逐个器件进行排查寻找故障,故障排查时间的长短决定着变电站何时正常工作,严重影响居民的日常生活。刀闸作为变电站的重要控电开关,其开合状态直接决定电气化电路整体是否运行。传统对变电站出现停电状态排查,作业人员首先需要进入变电站高压区域查看刀闸是否断开,然后逐一排查其他连接器件,容易发生触电危险。随着变电站无人化值守的发展需求,研究对抓拍图像中刀闸状态的自动化检测将有利于快速识别刀闸的当前状态,缩短变电站的故障检测效率,保证作业人员的生命安全,有效提高电网的安全运行。The substation is an important transfer place in high-voltage transmission, and there are various vital on-off control devices running in the station. In the past, when the substation was interrupted, the relevant operators needed to check each device to find the fault. The length of the troubleshooting time determines when the substation works normally, which seriously affects the daily life of residents. As an important electrical control switch in a substation, the opening and closing state of the knife switch directly determines whether the entire electrified circuit operates. Traditionally, to check the power failure status of the substation, the operator first needs to enter the high-voltage area of the substation to check whether the knife switch is disconnected, and then check other connecting devices one by one, which is prone to electric shock. With the development demand of unmanned duty in substations, researching the automatic detection of the state of the knife switch in the snapshot image will help to quickly identify the current state of the knife switch, shorten the fault detection efficiency of the substation, ensure the life safety of the operators, and effectively improve the power grid. safe operation.

对于刀闸的自动化检测,目前经常遇到的问题:刀闸工作环境位于室外,存在多种和刀闸同为金属制品的设备,颜色相近,不容易区分;受天气环境影响,处于同一位置的刀闸在不同的环境下识别难度不同;由于室外存在多种电压控制设备和导线以及刀闸长条矩形形状影响,在对变电站部署辅助监控系统时,经常出现采集到的刀闸受其他对象设备遮挡或不能完整采集到整个特征。For the automatic detection of knife gates, there are often problems encountered at present: the working environment of knife gates is located outdoors, and there are many kinds of equipment that are made of the same metal products as the knife gates, with similar colors and are not easy to distinguish; Knife switches are difficult to identify in different environments; due to the presence of various voltage control devices and wires outdoors, as well as the influence of the long rectangular shape of the knife switches, when the auxiliary monitoring system is deployed in the substation, it often occurs that the collected knife switches are affected by other object equipment. Occlusion or incomplete capture of the entire feature.

针对刀闸检测方法,分为两种:1)基于图像处理的方法;2)基于深度学习的方法。受刀闸数据采集影响,目前多数方法采用的是基于图像处理的方法,首先对固定角度位置采集的标准图像建立刀闸模板,然后使用特征点匹配算法或者模板匹配算法等图像处理算法进行定位以及状态识别,这种方法在短时间内良好的天气环境情况下可以获得较高的监测准确率,但随着天气状况的频繁改变,相机位置在预先建立模板图像角度基础上会发生偏移,进而无法识别出刀闸的状态,因此需要经常人工查看调整相机位置或者从新建立模板。采用深度学习的方法,需要依赖大量不同角度、不同天气的数据集,数据集的缺少会限制了相关工作的进度,除此之外,目前多数深度学习网络模型通过全局感受野进行预测,如:Faster-Rcnn系列网络模型、Yolo系列网络模型等,当对刀闸进行检测时,如果刀闸存在遮挡或采集对象不完整,会导致刀闸整体特征信息缺少,产生漏检或者误检;部分深度学习网络模型通过局部特征进行预测,如:R-FCN网络模型、VIT 网络模型等,当对刀闸进行检测时,虽然能有效检测到不完整的刀闸,但当刀闸充满整幅图像时,局部特征预测也会出现漏检情况。There are two types of knife gate detection methods: 1) methods based on image processing; 2) methods based on deep learning. Affected by the data collection of the knife gate, most of the current methods are based on image processing. First, the knife gate template is established for the standard image collected at a fixed angle position, and then image processing algorithms such as feature point matching algorithm or template matching algorithm are used to locate and locate. Status recognition, this method can achieve high monitoring accuracy in a short period of good weather conditions, but with frequent changes in weather conditions, the camera position will shift based on the pre-established template image angle, and then The state of the knife gate cannot be recognized, so it is necessary to manually check and adjust the camera position or create a new template. The deep learning method needs to rely on a large number of data sets from different angles and different weathers. The lack of data sets will limit the progress of related work. In addition, most deep learning network models currently make predictions through the global receptive field, such as: Faster-Rcnn series network models, Yolo series network models, etc., when the knife gate is detected, if the knife gate is blocked or the collection object is incomplete, the overall feature information of the knife gate will be missing, resulting in missed detection or false detection; partial depth The learning network model predicts through local features, such as: R-FCN network model, VIT network model, etc. When detecting the knife gate, although the incomplete knife gate can be effectively detected, when the knife gate is full of the whole image , the local feature prediction will also be missed.

发明内容SUMMARY OF THE INVENTION

针对上述技术问题,本发明的目的是:基于R-FCN网络模型,提出一种联合局部特征和全局特征的基于区域的全卷积目标检测网络R-FCN(R-FCN:Object Detection viaRegion-based Fully Convolutional Networks)刀闸检测方法,对R-FCN 网络模型嵌入全局特征预测,联合局部特征和全局特征预测的优点提升刀闸检测的准确率,提升现有刀闸检测方法的准确率。In view of the above technical problems, the purpose of the present invention is: based on the R-FCN network model, to propose a region-based fully convolutional target detection network R-FCN (R-FCN: Object Detection viaRegion-based) that combines local features and global features Fully Convolutional Networks) knife gate detection method, which embeds global feature prediction into the R-FCN network model, and combines the advantages of local features and global feature prediction to improve the accuracy of knife gate detection and the accuracy of existing knife gate detection methods.

为了实现上述目的,本发明所采用的技术方案为:一种联合局部特征和全局特征的R-FCN刀闸检测,包括以下步骤:In order to achieve the above object, the technical solution adopted in the present invention is: a kind of R-FCN knife gate detection combining local features and global features, comprising the following steps:

步骤1:搭建变电站辅助监控系统采集刀闸图像;Step 1: Build a substation auxiliary monitoring system to collect knife gate images;

步骤2:划分、清洗并标注图像数据集;Step 2: Divide, clean and label the image dataset;

步骤3:选择R-FCN主干特征提取网络;Step 3: Select the R-FCN backbone feature extraction network;

步骤4:调整R-FCN的主干特征提取网络;Step 4: Adjust the backbone feature extraction network of R-FCN;

步骤5:构建局部特征预测分支;Step 5: Build a local feature prediction branch;

步骤6:构建全局特征预测分支;Step 6: Build a global feature prediction branch;

步骤7:融合局部预测结果和全局预测结果;Step 7: Fusion of local prediction results and global prediction results;

步骤8:训练保存模型。Step 8: Train and save the model.

进一步的,步骤1中的变电站辅助监控系统为变电站中的所有网络监控摄像机和相关控制设备的集成,通过调整各相机的角度,实现采集不同时刻、天气、背景的包含刀闸和未包含刀闸的图像数据集,且包含刀闸的图像数据集数量应均匀包含刀闸的开关两种状态。Further, the substation auxiliary monitoring system in step 1 is the integration of all network monitoring cameras and related control equipment in the substation. By adjusting the angle of each camera, it is possible to collect different time, weather, and backgrounds including knife gates and without knife gates. , and the number of image datasets containing the knife switch should evenly include the two states of the switch.

进一步的,步骤2中的划分数据集指的是将采集的图像分成包含刀闸的图像和未包含刀闸的图像,并将包含刀闸图像的数据集划分成训练集和测试集比例为 8:2;清洗数据集指的是将采集到的数据集中存在模糊的数据集进行剔除;标注数据只标注包含刀闸对象的图像。Further, dividing the data set in step 2 refers to dividing the collected images into images containing knife gates and images not containing knife gates, and dividing the data set containing knife gate images into training sets and test sets with a ratio of 8. : 2; Cleaning the data set refers to removing the blurred data set in the collected data set; the labeling data only labels the images containing the knife gate object.

进一步的,步骤3中选择R-FCN主干特征提取网络,实现方式为:采用分类网络ResNet101作为主干特征提取网络,在ResNet101第四组卷积层后使用区域建议网络(RPN:Region Proposal Network)操作产生感兴趣区域(用于后期刀闸的检测),抛弃使用ResNet101第五组卷积层后面的池化层和全连接层,最终输出的通道数为2048个。所述感兴趣区域为存在目标的区域。Further, in step 3, the R-FCN backbone feature extraction network is selected, and the implementation method is as follows: the classification network ResNet101 is used as the backbone feature extraction network, and the region proposal network (RPN: Region Proposal Network) is used after the fourth convolutional layer of ResNet101. Generate a region of interest (for later detection of knife gates), discard the pooling layer and fully connected layer after the fifth set of convolutional layers of ResNet101, and the final output channel number is 2048. The region of interest is an area where a target exists.

进一步的,步骤4中调整R-FCN的主干特征提取网络,实现方式为:在步骤 3选择修整的主干特征提取网络ResNet101后添加卷积核大小为1×1的卷积层将通道数降低为1024个,以降低特征数据维度但不改变特征图的大小,提升计算速度。Further, in step 4, the backbone feature extraction network of R-FCN is adjusted, and the implementation method is as follows: after selecting the trimmed backbone feature extraction network ResNet101 in step 3, a convolutional layer with a convolution kernel size of 1×1 is added to reduce the number of channels to 1024, in order to reduce the dimension of feature data without changing the size of the feature map, and improve the calculation speed.

进一步的,步骤5中构建局部特征预测分支,实现方式为:采用原始具有局部特征预测的网络模型R-FCN输出局部预测结果,且局部特征预测采用RPN建议区域划分为7×7个局部区域进行预测。Further, the local feature prediction branch is constructed in step 5, and the implementation method is as follows: the original network model R-FCN with local feature prediction is used to output the local prediction result, and the local feature prediction is divided into 7 × 7 local regions by using the RPN recommendation area. predict.

进一步的,步骤6中构建全局特征预测分支,实现方式为:对提取的语义特征进行池化操作,统一提取的语义特征大小;然后,串联使用卷积核大小为7×7 和1×1的卷积层输出全局预测结果。Further, a global feature prediction branch is constructed in step 6, and the implementation method is as follows: perform a pooling operation on the extracted semantic features, and unify the extracted semantic feature sizes; The convolutional layer outputs the global prediction result.

进一步的,步骤7中融合局部预测结果和全局预测结果,是对步骤5和步骤 6中的输出预测结果要使用L2正则化,数学表达式为:

Figure RE-GDA0003694899550000031
式中x、y为向量,x=(x0,x1,x2...xn),表示预测输出结果,y=(y0,y1,y2...yn),表示正则化后的预测输出结果,最终将两种预测结果数值统一缩放至0到1区间,并将两种预测输出结果进行向量相加,然后进行Softmax操作,输出最终预测结果,Softmax的数学表达式为:
Figure RE-GDA0003694899550000032
Ri为输出结果向量中第i个输出值。Further, the fusion of local prediction results and global prediction results in step 7 is to use L2 regularization for the output prediction results in steps 5 and 6, and the mathematical expression is:
Figure RE-GDA0003694899550000031
In the formula, x and y are vectors, x=(x 0 , x 1 , x 2 ... x n ), representing the predicted output result, y=(y 0 , y 1 , y 2 ... y n ), representing The normalized prediction output results, and finally the two prediction results are uniformly scaled to the range of 0 to 1, and the two prediction output results are vector added, and then the Softmax operation is performed to output the final prediction result, the mathematical expression of Softmax for:
Figure RE-GDA0003694899550000032
R i is the ith output value in the output result vector.

进一步的,步骤8中训练模型,训练使用划分的两种数据集进行两次训练:第一次训练只使用包含刀闸的清晰数据集,第二次训练使用包含刀闸的和未包含刀闸但类似刀闸的数据数量比例为1:1的模糊数据集。Further, in step 8, the model is trained, and the training is performed twice using the divided two data sets: the first training uses only the clear data set containing the knife gate, and the second training uses the knife gate and does not include the knife gate. But a fuzzy dataset with a 1:1 ratio of data quantity like knife gate.

进一步的,步骤8中训练模型,训练使用的损失函数数学表达式为:

Figure RE-GDA0003694899550000033
其中L(s,t)表示分类损失和回归损失的总损失,s为类别预测概率,
Figure RE-GDA0003694899550000041
表示类别c*的预测概率,t表示模型预测的回归框, Lcls(s)为分类损失,数学表达式为:
Figure RE-GDA0003694899550000042
λ多任务平衡因子,c*为类别真实标签,当c*=0表示背景类别,c*≠0表示对应的对别,[c*>0]组成指导因子,其值取为1,表示回归框只对非背景类别对象进行调整,Lreg(t,t*)为回归损失,t表示模型预测的目标框,t*表示人工标注的真实框。Further, in the training model in step 8, the mathematical expression of the loss function used in the training is:
Figure RE-GDA0003694899550000033
where L(s,t) represents the total loss of classification loss and regression loss, s is the class prediction probability,
Figure RE-GDA0003694899550000041
represents the predicted probability of class c * , t represents the regression box predicted by the model, L cls (s) is the classification loss, and the mathematical expression is:
Figure RE-GDA0003694899550000042
λ multi-task balance factor, c * is the true label of the category, when c * =0 represents the background category, c * ≠0 represents the corresponding pair, [c * >0] constitutes the guidance factor, and its value is 1, indicating regression The box is only adjusted for non-background category objects, L reg (t, t * ) is the regression loss, t represents the target box predicted by the model, and t * represents the ground-truth box that is manually annotated.

与现有技术相比,本发明取得的有益效果为:对于存在部分特征遮挡的以及较大特征的刀闸检测,基于联合局部特征和全局特征的R-FCN网络相比于基于原始局部特征预测,均能预测判断。基于联合局部特征和全局特征的R-FCN刀闸检测方法的提出,有效降低了变电站复杂场景中刀闸检测的误检率和漏检率。Compared with the prior art, the beneficial effects obtained by the present invention are: for the detection of knife gates with partial feature occlusion and larger features, the R-FCN network based on joint local features and global features is compared with prediction based on original local features. , can be predicted and judged. The R-FCN knife switch detection method based on joint local features and global features is proposed, which effectively reduces the false detection rate and missed detection rate of knife switch detection in complex scenes of substations.

附图说明Description of drawings

为更清楚的说明本发明的技术方案,下面结合对实施例中所需的附图做进一步说明。In order to illustrate the technical solutions of the present invention more clearly, further description will be given below with reference to the accompanying drawings required in the embodiments.

图1为本发明的联合局部特征和全局特征的R-FCN刀闸检测方法操作流程示意图。FIG. 1 is a schematic diagram of the operation flow of the R-FCN knife gate detection method combining local features and global features of the present invention.

图2为本发明的联合局部特征和全局特征的R-FCN模型的模块示意图。FIG. 2 is a schematic diagram of a module of an R-FCN model combining local features and global features of the present invention.

图3为本发明的局部特征预测的模块示意图。FIG. 3 is a schematic diagram of a module for local feature prediction according to the present invention.

图4为本发明的全局特征预测的模块示意图。FIG. 4 is a schematic diagram of a module for global feature prediction according to the present invention.

具体实施方式Detailed ways

下面结合附图1-4对本发明进行实施方案的详细说明。The following describes the embodiments of the present invention in detail with reference to the accompanying drawings 1-4.

如图1所示,一种联合局部特征和全局特征的R-FCN刀闸检测方法,其步骤如下:As shown in Figure 1, an R-FCN knife gate detection method combining local features and global features, the steps are as follows:

步骤1:变电站辅助监控系统采集刀闸图像时,从不同角度、不同视野、不同背景等自然因素中采集,保证采集到图片的多样性。Step 1: When the substation auxiliary monitoring system collects the knife gate images, it is collected from natural factors such as different angles, different fields of view, and different backgrounds, so as to ensure the diversity of the collected images.

步骤2:将步骤1采集到的数据集进行清洗,去除损坏的图像数据,并分成两种:只包含刀闸的图像数据集和未包含刀闸但与刀闸类似的数据集。本发明采集到的数据集包含刀闸的图像总数为800张,未包含刀闸但与刀闸类似的800 张,分别对两种进行划分,其中训练集和测试集之比为8:2,并将只包含刀闸图像的数据集进行标注刀闸位置和状态,第一次训练,采用样本的数量如表1 所示,第二次训练,采用样本的数量如表2所示。Step 2: Clean the data set collected in Step 1, remove damaged image data, and divide it into two types: the image data set that only contains the knife gate and the data set that does not contain the knife gate but is similar to the knife gate. The data set collected by the present invention includes a total of 800 images of the knife gate, and 800 images that do not include the knife gate but are similar to the knife gate. The data set containing only knife gate images is marked with the knife gate position and state. The number of samples used in the first training is shown in Table 1, and the number of samples used in the second training is shown in Table 2.

步骤3:调整R-FCN主干特征提取网络:采用分类网络ResNet101作为主干特征提取网络,在ResNet101第四组卷积层后使用RPN操作产生感兴趣区域(用于后期刀闸的检测),抛弃使用ResNet101第五组卷积层后面的池化层和全连接层,最终输出的通道数为2048个,如图2所示。Step 3: Adjust the R-FCN backbone feature extraction network: use the classification network ResNet101 as the backbone feature extraction network, and use the RPN operation after the fourth convolutional layer of ResNet101 to generate the region of interest (for later detection of knife gates), discard the use of The pooling layer and fully connected layer after the fifth set of convolutional layers of ResNet101, the final output channel number is 2048, as shown in Figure 2.

步骤4:在步骤3选择修整的主干特征提取网络ResNet101后添加卷积核大小为1×1的卷积层将通道数降低为1024个,以降低特征数据维度但不改变特征图的大小,如图2所示。Step 4: After selecting the trimmed backbone feature extraction network ResNet101 in Step 3, add a convolutional layer with a convolution kernel size of 1×1 to reduce the number of channels to 1024 to reduce the dimension of feature data without changing the size of the feature map, such as shown in Figure 2.

步骤5:构建局部特征预测分支:采用原始具有局部特征预测的网络模型 R-FCN输出局部预测结果,且局部特征预测采用RPN建议区域划分为7×7个局部区域进行预测,如图3所示。Step 5: Build a local feature prediction branch: use the original network model R-FCN with local feature prediction to output the local prediction result, and the local feature prediction is divided into 7 × 7 local regions for prediction using the RPN recommendation area, as shown in Figure 3 .

步骤6:构建全局特征预测分支:对提取的语义特征进行池化操作,统一提取的语义特征大小;然后,串联使用卷积核大小为7×7和1×1的卷积层输出全局预测结果,如图4所示。Step 6: Build a global feature prediction branch: perform a pooling operation on the extracted semantic features to unify the size of the extracted semantic features; then, use convolutional layers with convolution kernel sizes of 7×7 and 1×1 in series to output the global prediction result ,As shown in Figure 4.

步骤7:融合局部预测结果和全局预测结果:对局部预测结果和全局预测结果进行正则化,统一缩放至同一数值区间并进行相加,完成信息融合预测。Step 7: Fusion of local prediction results and global prediction results: The local prediction results and the global prediction results are regularized, uniformly scaled to the same numerical range and added to complete the information fusion prediction.

进一步的,数学表达式为:

Figure RE-GDA0003694899550000051
式中x、y为向量, x=(x0,x1,x2...xn),表示预测输出结果,y=(y0,y1,y2...yn),表示正则化后的预测输出结果,最终将两种预测结果数值统一缩放至0到1区间,并将两种预测输出结果进行向量相加,然后进行Softmax操作,输出最终预测结果,Softmax的数学表达式为:
Figure RE-GDA0003694899550000052
Ri为输出结果向量中第i个输出值。Further, the mathematical expression is:
Figure RE-GDA0003694899550000051
In the formula, x and y are vectors, x=(x 0 , x 1 , x 2 ... x n ), representing the predicted output result, y=(y 0 , y 1 , y 2 ... y n ), representing The normalized prediction output results, and finally the two prediction results are uniformly scaled to the range of 0 to 1, and the two prediction output results are vector added, and then the Softmax operation is performed to output the final prediction result, the mathematical expression of Softmax for:
Figure RE-GDA0003694899550000052
R i is the ith output value in the output result vector.

步骤8:训练模型:训练时模型主干特征提取网络ResNet101的初始化参数使用在ImageNet数据集上训练好的模型权重,模型训练使用的损失函数选择与 R-FCN网络相同的损失函数,进行两次训练:第一次使用只包含有刀闸图像的数据训练,至损失函数收敛,保存网络模型;第二次基于第一次训练的模型参数基础之上,使用包含刀闸和未包含刀闸类似刀闸的数据进行训练,进一步特征网络模型的刀闸辨别能力,最终得到模型的测试结果如表3所示。Step 8: Train the model: The initialization parameters of the model backbone feature extraction network ResNet101 during training use the model weights trained on the ImageNet dataset, and the loss function used for model training selects the same loss function as the R-FCN network, and conducts two trainings : For the first time, use the data that only contains the knife gate image to train, until the loss function converges, and save the network model; the second time, based on the model parameters of the first training, use the similar knife that contains the knife gate and does not include the knife gate The data of the gate is used for training, and the identification ability of the knife gate of the network model is further characterized, and the test results of the model are finally obtained as shown in Table 3.

进一步的,步骤8中训练模型,训练使用的损失函数数学表达式为:

Figure RE-GDA0003694899550000061
其中L(s,t)表示分类损失和回归损失的总损失,s为类别预测概率,
Figure RE-GDA0003694899550000062
表示类别c*的预测概率,t表示模型预测的回归框, Lcls(s)为分类损失,数学表达式为:
Figure RE-GDA0003694899550000063
λ多任务平衡因子,c*为类别真实标签,当c*=0表示背景类别,c*≠0表示对应的对别,[c*>0]组成指导因子,其值取为1,表示回归框只对非背景类别对象进行调整,Lreg(t,t*)为回归损失,t表示模型预测的目标框,t*表示人工标注的真实框。Further, in the training model in step 8, the mathematical expression of the loss function used in the training is:
Figure RE-GDA0003694899550000061
where L(s,t) represents the total loss of classification loss and regression loss, s is the class prediction probability,
Figure RE-GDA0003694899550000062
represents the predicted probability of class c * , t represents the regression box predicted by the model, L cls (s) is the classification loss, and the mathematical expression is:
Figure RE-GDA0003694899550000063
λ multi-task balance factor, c * is the true label of the category, when c * =0 represents the background category, c * ≠0 represents the corresponding pair, [c * >0] constitutes the guidance factor, and its value is 1, indicating regression The box is only adjusted for non-background category objects, L reg (t, t * ) is the regression loss, t represents the target box predicted by the model, and t * represents the ground-truth box that is manually annotated.

表1是第一次训练集样本数量的统计结果。Table 1 is the statistical results of the number of samples in the first training set.

Figure RE-GDA0003694899550000064
Figure RE-GDA0003694899550000064

表2是第二次训练集样本数量的统计结果。Table 2 is the statistical results of the number of samples in the second training set.

Figure RE-GDA0003694899550000065
Figure RE-GDA0003694899550000065

表3是联合局部特征和全局特征的R-FCN刀闸检测准确率对比结果。Table 3 shows the comparison results of the R-FCN knife gate detection accuracy by combining local features and global features.

Figure RE-GDA0003694899550000066
Figure RE-GDA0003694899550000066

尽管已经展示和描述了本发明的实施例,对于本领域的普通技术人员而言,可以理解在不脱离本发明的原理和精神的情况下可以对这些实施例进行多种变化、修改、替换和变型,本发明的保护范围由所附权利要求及其等同物限定。Although embodiments of the present invention have been shown and described, it will be understood by those skilled in the art that various changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principle and spirit of the invention Variations, the scope of protection of the present invention is defined by the appended claims and their equivalents.

Claims (9)

1.一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于,包括以下步骤:1. an R-FCN knife gate detection method of joint local feature and global feature, is characterized in that, comprises the following steps: 步骤1:搭建变电站辅助监控系统,从不同的相机角度和不同时刻的天气情况对目标进行图像采集;Step 1: Build a substation auxiliary monitoring system to collect images from different camera angles and weather conditions at different times; 步骤2:对步骤1采集的图像划分为训练集图像、测试集图像,并对图像数据集进行清洗以及标注;Step 2: Divide the images collected in step 1 into training set images and test set images, and clean and label the image data set; 步骤3:构建具有联合局部特征和全局特征的R-FCN刀闸检测模型:调整R-FCN主干特征提取网络:采用分类网络ResNet101作为主干特征提取网络,在ResNet101第四组卷积层后使用区域建议网络RPN操作产生感兴趣区域,抛弃使用ResNet101第五组卷积层后面的池化层和全连接层;Step 3: Build an R-FCN Knife Gate Detection Model with Joint Local Features and Global Features: Adjust the R-FCN Backbone Feature Extraction Network: Use the classification network ResNet101 as the backbone feature extraction network, and use the region after the fourth set of convolutional layers of ResNet101 It is recommended that the network RPN operation generates a region of interest, and discards the pooling layer and fully connected layer after the fifth set of convolutional layers of ResNet101; 步骤4:基于步骤3调整的主干特征提取网络,输出的通道数为2048个,在其后面附加卷积核大小为1×1的卷积层将通道数降低为1024个;Step 4: Based on the backbone feature extraction network adjusted in Step 3, the number of output channels is 2048, and a convolutional layer with a convolution kernel size of 1×1 is attached to reduce the number of channels to 1024; 步骤5:构建局部特征预测分支:基于步骤4操作,将提取的语义特征信息分两个方向进行并联预测:基于原始网络的局部特征预测和全局特征的预测,局部特征预测直接采用原始预测网络模型输出局部预测结果;Step 5: Build a local feature prediction branch: Based on the operation in step 4, the extracted semantic feature information is divided into two directions for parallel prediction: local feature prediction and global feature prediction based on the original network, and the local feature prediction directly adopts the original prediction network model Output local prediction results; 步骤6:构建全局特征预测分支:全局特征预测基于步骤4操作提取的语义特征,首先对提取的语义特征进行池化操作,统一提取的语义特征大小;然后,串联使用卷积核大小为7×7和1×1的卷积层输出全局预测结果;Step 6: Build a global feature prediction branch: The global feature prediction is based on the semantic features extracted in step 4. First, the extracted semantic features are pooled to unify the size of the extracted semantic features; then, the convolution kernel size is 7× 7 and 1×1 convolutional layers output global prediction results; 步骤7:融合局部预测结果和全局预测结果:对局部预测结果和全局预测结果进行正则化,统一缩放至同一数值区间并进行相加,完成信息融合预测;Step 7: Fusion of local prediction results and global prediction results: Regularize the local prediction results and global prediction results, uniformly scale them to the same numerical range and add them to complete the information fusion prediction; 步骤8:训练模型:模型训练使用的损失函数选择与R-FCN网络相同的损失函数,用于指导模型参数的优化;网络训练参数更新至损失函数收敛,保存网络模型。Step 8: Train the model: The loss function used for model training selects the same loss function as the R-FCN network, which is used to guide the optimization of model parameters; the network training parameters are updated until the loss function converges, and the network model is saved. 2.根据权利要求1所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述步骤1中的辅助监控系统是由多个网络摄像机组合对变电站各角度实行全面覆盖、实时监控,所述辅助监控系统嵌入相机集控程序,通过多个网络摄像机实现不同角度、不同背景、不同天气因素的数据采集。2. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 1, is characterized in that: the auxiliary monitoring system in the described step 1 is to combine each angle of the substation by a plurality of network cameras Implement comprehensive coverage and real-time monitoring. The auxiliary monitoring system is embedded in the camera centralized control program, and data collection from different angles, different backgrounds, and different weather factors is realized through multiple network cameras. 3.根据权利要求1所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述步骤2中的清洗数据集具体为:将相机震动造成的模糊图像进行剔除,将包含刀闸的图像归为一个数据集,将未包含刀闸但存在类似刀闸的图像归为一个数据集。3. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 1, is characterized in that: the cleaning data set in the described step 2 is specifically: the blurred image caused by camera vibration is carried out. For culling, images containing knife gates are grouped into a dataset, and images that do not contain knife gates but have similar knife gates are grouped into a dataset. 4.根据权利要求1所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述步骤5中的局部特征预测是将RPN建议区域划分为7×7个局部区域进行预测。4. The R-FCN knife gate detection method combining local features and global features according to claim 1, wherein the local feature prediction in the step 5 is to divide the RPN suggested area into 7×7 local area for prediction. 5.根据权利要求1所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述步骤7中的融合局部预测结果和全局预测结果具体为:使用正则化方式分别对局部预测结果和全局预测结果进行正则化,且使用的是L2正则化,数学表达式为:
Figure RE-FDA0003694899540000021
式中x、y为向量,x=(x0,x1,x2...xn),表示预测输出结果,y=(y0,y1,y2...yn),表示正则化预测输出结果,最终将预测结果数值统一缩放至0到1区间。
5. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 1, is characterized in that: the fusion local prediction result and the global prediction result in the described step 7 are specifically: use regularization The method normalizes the local prediction results and the global prediction results respectively, and uses L2 regularization. The mathematical expression is:
Figure RE-FDA0003694899540000021
In the formula, x and y are vectors, x=(x 0 , x 1 , x 2 ... x n ), representing the predicted output result, y=(y 0 , y 1 , y 2 ... y n ), representing Regularize the prediction output results, and finally scale the prediction results uniformly to the range of 0 to 1.
6.根据权利要求6所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于,使用的融合方式为:对相同维度的向量相加。6 . The R-FCN knife gate detection method combining local features and global features according to claim 6 , wherein the fusion method used is: adding vectors of the same dimension. 7 . 7.根据权利要求1所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述步骤8中模型训练使用的数据分为两种:包含刀闸的清晰数据集和混乱困难容易样本比例为1:1的模糊数据集。7. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 1, is characterized in that: the data that model training uses in described step 8 is divided into two kinds: the clear that comprises knife gate Datasets and clutter-hard easy-to-sample ratios are 1:1 fuzzy datasets. 8.根据权利要求7所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述训练模型具体为:训练时使用迁移学习思想,对主干特征提取网络采用再ImageNet数据集上训练好的模型权重;所述清晰数据集指的是所有训练图像至少包含一个刀闸图像,所述模糊数据集指的是包含刀闸图像和未包含刀闸且类似刀闸的图像各占50%。8. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 7, it is characterized in that: described training model is specifically: use migration learning thought during training, adopt main feature extraction network to adopt The weight of the model trained on the ImageNet dataset; the clear dataset means that all training images contain at least one knife gate image, and the fuzzy data set refers to the image that contains the knife gate and does not contain the knife gate and is similar to the knife gate 50% of the images each. 9.根据权利要求7所述的一种联合局部特征和全局特征的R-FCN刀闸检测方法,其特征在于:所述训练模型:训练使用的损失函数数学表达式为:
Figure RE-FDA0003694899540000023
其中L(s,t)表示分类损失和回归损失的总损失,s为类别预测概率,
Figure RE-FDA0003694899540000022
表示类别c*的预测概率,t表示模型预测的回归框,Lcls(s)为分类损失,λ多任务平衡因子,c*为类别真实标签,Lreg(t,t*)为回归损失,t*表示人工标注的真实框。
9. the R-FCN knife gate detection method of a kind of joint local feature and global feature according to claim 7, is characterized in that: described training model: the loss function mathematical expression that training uses is:
Figure RE-FDA0003694899540000023
where L(s,t) represents the total loss of classification loss and regression loss, s is the class prediction probability,
Figure RE-FDA0003694899540000022
Represents the predicted probability of class c * , t represents the regression box predicted by the model, L cls (s) is the classification loss, λ multi-task balance factor, c * is the class true label, L reg (t, t * ) is the regression loss, t * denotes the ground-truth box annotated by humans.
CN202210453322.6A 2022-04-27 2022-04-27 R-FCN knife switch detection method combining local features and global features Active CN114821042B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210453322.6A CN114821042B (en) 2022-04-27 2022-04-27 R-FCN knife switch detection method combining local features and global features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210453322.6A CN114821042B (en) 2022-04-27 2022-04-27 R-FCN knife switch detection method combining local features and global features

Publications (2)

Publication Number Publication Date
CN114821042A true CN114821042A (en) 2022-07-29
CN114821042B CN114821042B (en) 2025-07-22

Family

ID=82509689

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210453322.6A Active CN114821042B (en) 2022-04-27 2022-04-27 R-FCN knife switch detection method combining local features and global features

Country Status (1)

Country Link
CN (1) CN114821042B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115359441A (en) * 2022-08-29 2022-11-18 安徽大学 Anomaly detection method for spilled objects based on Vit network heuristic self-supervised training
CN118015555A (en) * 2024-04-10 2024-05-10 南京国电南自轨道交通工程有限公司 A switch state recognition method based on visual detection and mask image direction vector

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150104102A1 (en) * 2013-10-11 2015-04-16 Universidade De Coimbra Semantic segmentation method with second-order pooling
CN106447658A (en) * 2016-09-26 2017-02-22 西北工业大学 Significant target detection method based on FCN (fully convolutional network) and CNN (convolutional neural network)
CN113221969A (en) * 2021-04-25 2021-08-06 浙江师范大学 Semantic segmentation system and method based on Internet of things perception and based on dual-feature fusion

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150104102A1 (en) * 2013-10-11 2015-04-16 Universidade De Coimbra Semantic segmentation method with second-order pooling
CN106447658A (en) * 2016-09-26 2017-02-22 西北工业大学 Significant target detection method based on FCN (fully convolutional network) and CNN (convolutional neural network)
CN113221969A (en) * 2021-04-25 2021-08-06 浙江师范大学 Semantic segmentation system and method based on Internet of things perception and based on dual-feature fusion

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
莫蓓蓓: ""基于Mask R-CNN的电力设备红外图像分割技术的研究与实现"", 《中国优秀硕士学位论文全文数据库工程科技Ⅱ辑》, 15 June 2021 (2021-06-15) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115359441A (en) * 2022-08-29 2022-11-18 安徽大学 Anomaly detection method for spilled objects based on Vit network heuristic self-supervised training
CN118015555A (en) * 2024-04-10 2024-05-10 南京国电南自轨道交通工程有限公司 A switch state recognition method based on visual detection and mask image direction vector

Also Published As

Publication number Publication date
CN114821042B (en) 2025-07-22

Similar Documents

Publication Publication Date Title
CN109101906A (en) A kind of converting station electric power equipment infrared image exception real-time detection method and device
CN106595551B (en) Ice covering thickness detection method in powerline ice-covering image based on deep learning
CN113486877B (en) Power equipment infrared image real-time detection and diagnosis method based on light artificial intelligent model
CN114821042A (en) An R-FCN knife gate detection method combining local and global features
CN110378221A (en) A kind of power grid wire clamp detects and defect identification method and device automatically
CN111814742A (en) Knife switch state recognition method based on deep learning
CN112697798A (en) Infrared image-oriented diagnosis method and device for current-induced thermal defects of power transformation equipment
CN118411724A (en) Ice morphology segmentation method for overhead transmission lines
CN116191680B (en) Monitoring management system applied to photovoltaic power generation
CN119813201B (en) Adaptive reclosing optimization method and system for power distribution network
CN112149877A (en) A multi-source data-driven fault prediction method and system for multi-element and complex urban power grids
Kumar et al. Deep Learning based Fault Detection in Power Transmission Lines
CN120032193B (en) Power equipment defect detection method and system
CN113780224A (en) Transformer substation unmanned inspection method and system
Shan et al. Research on efficient detection method of foreign objects on transmission lines based on improved YOLOv4 network
CN117132531A (en) Lightweight-based YOLOv5 insulator defect detection method
CN110555460B (en) A bird detection method for mobile transmission lines based on image slices
CN118100797B (en) Photovoltaic module fault detection method
CN114332771A (en) Method and system for detecting state of disconnecting link based on real-time monitoring and intelligent terminal
CN116958841B (en) Unmanned aerial vehicle inspection system for power distribution line based on image recognition
CN118097202A (en) A mine fire image recognition and fire extinguishing method
CN117949887A (en) Fault detection and prevention method for electric energy metering system of energy controller
CN117725993A (en) Topological structure optimization method for establishing graph deep learning model based on knowledge graph
CN116205905A (en) Power distribution network construction safety and quality image detection method and system based on mobile terminal
Tian et al. Overview of UAV-Based Overhead Power Line Defect Detection Based on Deep Learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载