+

CN116321008A - Indoor position fingerprint positioning method based on triplet self-encoder - Google Patents

Indoor position fingerprint positioning method based on triplet self-encoder Download PDF

Info

Publication number
CN116321008A
CN116321008A CN202310392222.1A CN202310392222A CN116321008A CN 116321008 A CN116321008 A CN 116321008A CN 202310392222 A CN202310392222 A CN 202310392222A CN 116321008 A CN116321008 A CN 116321008A
Authority
CN
China
Prior art keywords
fingerprint
network
triplet
training
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310392222.1A
Other languages
Chinese (zh)
Inventor
袁正午
徐发鹏
徐水英
陈望
邵文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202310392222.1A priority Critical patent/CN116321008A/en
Publication of CN116321008A publication Critical patent/CN116321008A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • H04W4/029Location-based management or tracking services
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • H04W4/025Services making use of location information using location based information parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/30Services specially adapted for particular environments, situations or purposes
    • H04W4/33Services specially adapted for particular environments, situations or purposes for indoor environments, e.g. buildings
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Collating Specific Patterns (AREA)

Abstract

The invention relates to an indoor position fingerprint positioning method based on a triplet self-encoder, and belongs to the technical field of positioning. The method comprises the following steps: acquiring a group of WIFI signal intensity data at each position by using a WIFI receiver to acquire the WIFI signal intensity of each position point; preprocessing the acquired data by using a Gaussian filtering mode; constructing a self-encoder-based triplet neural network and a loss function and a BP neural network for final positioning; training the constructed triplet self-encoder by the pretreatment indoor position fingerprint data set, and reserving the structure and parameters of a network model; extracting the preprocessed data set by using a feature extraction network to obtain a dimensionality-reduced data set, training a BP neural network, and storing the feature extraction network and the trained BP neural network as a Tri-Sae model; the method solves the problems of high feature extraction difficulty, low positioning accuracy, more bad data and the like.

Description

一种基于三元组自编码器的室内位置指纹定位方法An Indoor Location Fingerprint Location Method Based on Triplet Autoencoder

技术领域technical field

本发明属于定位技术领域,涉及一种基于三元组自编码器的室内位置指纹定位方法。The invention belongs to the technical field of positioning, and relates to an indoor position fingerprint positioning method based on a triplet autoencoder.

背景技术Background technique

室内位置指纹定位是基于模式匹配、机器学习和深度学习等技术,借助于计算机处理技术,对数据库中位置指纹进行实时匹配的技术。目前,室内位置指纹定位技术主要是使用神经网络和机器学习的方式来实现的,使用BP神经网络的方式来捕捉位置指纹深层的特征,再使用机器学习的方式,对比特征之间的欧氏距离,实现最终的位置定位。但使用BP神经网络作为室内定位系统的特征提取层难以捕捉样本之间的关系,使得最后系统定位精度很不理想。Indoor location fingerprint positioning is based on pattern matching, machine learning, deep learning and other technologies, and with the help of computer processing technology, it performs real-time matching of location fingerprints in the database. At present, indoor location fingerprint positioning technology is mainly realized by using neural network and machine learning, using BP neural network to capture the deep features of location fingerprints, and then using machine learning to compare the Euclidean distance between features , to achieve the final position positioning. However, using BP neural network as the feature extraction layer of the indoor positioning system is difficult to capture the relationship between samples, which makes the positioning accuracy of the final system very unsatisfactory.

发明内容Contents of the invention

有鉴于此,本发明的目的在于提供一种基于三元组自编码器的室内位置指纹定位方法。In view of this, the object of the present invention is to provide a method for indoor position fingerprint positioning based on a triplet autoencoder.

为达到上述目的,本发明提供如下技术方案:To achieve the above object, the present invention provides the following technical solutions:

一种基于三元组自编码器的室内位置指纹定位方法,该方法包括以下步骤:A method for indoor position fingerprint positioning based on a triple autoencoder, the method comprising the following steps:

S1:利用WIFI接收器在每个位置获得一组WIFI信号强度数据,得到每个位置点的WIFI信号强弱;S1: Use the WIFI receiver to obtain a set of WIFI signal strength data at each location, and obtain the WIFI signal strength of each location point;

S2:使用高斯滤波的方式对已经采集到的数据进行预处理;S2: Use Gaussian filtering to preprocess the collected data;

S3:构造基于自编码器的三元组神经网络及损失函数和用于最后定位的BP神经网络;S3: Construct the triplet neural network and loss function based on the autoencoder and the BP neural network for final positioning;

S4:经过预处理的室内位置指纹数据集训练构造好的三元组自编码器,对训练后的模型剪枝得到特征提取网络,并保留网络模型的结构和参数;S4: The preprocessed indoor location fingerprint data set trains the constructed triplet autoencoder, prunes the trained model to obtain the feature extraction network, and retains the structure and parameters of the network model;

S5:使用特征提取网络提取预处理后的数据集,得到降维后的数据集,并训练BP神经网络,将特征提取网络和训练后的BP神经网络保存为Tri-Sae模型。S5: Use the feature extraction network to extract the preprocessed data set, obtain the data set after dimension reduction, and train the BP neural network, and save the feature extraction network and the trained BP neural network as a Tri-Sae model.

S6:使用Tri-Sae模型和距离度量的方式联合投票定位,最后评价室内定位系统的精度。S6: Use the Tri-Sae model and the distance measure to jointly vote for positioning, and finally evaluate the accuracy of the indoor positioning system.

可选的,所述S2中,对采集到的数据采取高斯滤波的处理,去除样本中的坏数据,从RSSI值的高概率区域中选择有效值,并计算输出值的平均值,以提高测距精度,它的过程是:Optionally, in said S2, Gaussian filtering is performed on the collected data, bad data in the sample is removed, effective values are selected from the high probability area of the RSSI value, and the average value of the output value is calculated to improve the measurement efficiency. From precision, its procedure is:

设n个RSSI值的总数服从高斯分布,RSSI值的平均值为μ,RSSI值的方差为σ2,RSSI值的概率密度函数为f(x),因此

Figure BDA0004176195540000011
RSSI值的均值μ的计算公式为/>
Figure BDA0004176195540000012
σ2的计算公式为/>
Figure BDA0004176195540000013
如果RSSI值在[μ-σ,μ+σ]的范围内,则为高概率置信区间。通过高斯滤波器计算出的平均RSSI值减少环境干扰的影响。Suppose the total number of n RSSI values obeys Gaussian distribution, the mean value of RSSI value is μ, the variance of RSSI value is σ 2 , and the probability density function of RSSI value is f(x), so
Figure BDA0004176195540000011
The formula for calculating the mean value μ of the RSSI value is />
Figure BDA0004176195540000012
The calculation formula of σ 2 is />
Figure BDA0004176195540000013
If the RSSI value is in the range [μ-σ, μ+σ], then it is a high-probability confidence interval. The average RSSI value calculated by a Gaussian filter reduces the influence of environmental interference.

可选的,所述S3中,三元组网络构造分为设计子网络,添加损失函数loss构造完整网络的过程:Optionally, in the S3, the triplet network construction is divided into design sub-networks, and the process of adding a loss function loss to construct a complete network:

S31:三元组网络的子网络是堆叠自编码器,设计使用[输入层+[256,128,64,128,256]+输出层]的网络结构,自编码器分为上中下层结构,其中上下两层结构共享权重,每层的激活函数设置为

Figure BDA0004176195540000021
S31: The sub-network of the triplet network is a stacked autoencoder, and the network structure of [input layer + [256, 128, 64, 128, 256] + output layer] is designed. The autoencoder is divided into upper, middle and lower layers. The upper and lower layers share weights, and the activation function of each layer is set to
Figure BDA0004176195540000021

S32:定义新的神经网络模型的损失函数loss,如图1展示了损失函数loss的计算思路和三元组神经网络架构,它们用来完成了神经网络模型的构建,神经网络模型的损失函数loss的过程如下:S32: Define the loss function loss of the new neural network model. Figure 1 shows the calculation idea of the loss function loss and the triplet neural network architecture, which are used to complete the construction of the neural network model. The loss function loss of the neural network model The process is as follows:

将位置指纹数据输入堆叠自编码器,提取到每个样本数据的特征向量,而且每一条样本数据的特征维度都是相同的。每次训练选取三条样本作为训练数据,分别是每个类别的代表特征anchor(通过简单的计算平均值得到),anchor的正样本positive(与anchor来自同一位置),anchor的负样本negative(与anchor来自于不同的位置),将他们输入到堆叠自编码器(三元组网络的子网络),得到所对应的特征向量。Input the location fingerprint data into the stacked autoencoder to extract the feature vector of each sample data, and the feature dimension of each sample data is the same. Three samples are selected for each training as the training data, which are the representative feature anchor of each category (obtained by simply calculating the average value), the positive sample positive of the anchor (from the same position as the anchor), and the negative sample negative of the anchor (with the anchor from different positions), and input them to the stacked autoencoder (a sub-network of the triplet network) to obtain the corresponding feature vector.

本发明想要的是让anchor和positive得到的向量的欧氏距离越小越好;让anchor和negative得到的向量的欧氏距离越大越好;可使得公式What the present invention wants is to allow the Euclidean distance of the vector obtained by anchor and positive to be as small as possible; to allow the Euclidean distance of the vector obtained by anchor and negative to be as large as possible; the formula can be made

Figure BDA0004176195540000022
成立,;同时positive和negative的样本来自原始数据集本身,经过了堆叠自编码器后,还原到了与输入数据相同的维度,想让输入结果和输出结果尽量保持一致,就需要满足/>
Figure BDA0004176195540000023
β为任意大于零的实数,模型就具有了较好的效果,所以loss用公式表达为
Figure BDA0004176195540000022
Established; at the same time, the positive and negative samples come from the original data set itself. After passing through the stacked autoencoder, they are restored to the same dimension as the input data. If you want to keep the input results and output results as consistent as possible, you need to satisfy />
Figure BDA0004176195540000023
β is any real number greater than zero, and the model has a better effect, so the loss is expressed as

Figure BDA0004176195540000024
Figure BDA0004176195540000024

S33:设计了一种BP神经网络模型,用于最后的定位功能,此神经网络是简单的多层结构,其网络结构是[输入层+[128,128]+输出层],输出层使用sigmoid激活函数,其余层使用ReLU激活函数,并且添加Dropout层,丢弃率为0.1,以防止过拟合;S33: A BP neural network model is designed for the final positioning function. This neural network is a simple multi-layer structure, and its network structure is [input layer + [128, 128] + output layer], and the output layer uses sigmoid Activation function, the remaining layers use the ReLU activation function, and add a Dropout layer with a discard rate of 0.1 to prevent overfitting;

可选的,所述S4包括训练过程和剪枝过程:Optionally, the S4 includes a training process and a pruning process:

S41,在训练模型阶段,设共有N个位置可供定位,每个位置点有T个样本,经过数据处理,为位置Ri生成了训练数据集

Figure BDA0004176195540000025
和验证数据集/>
Figure BDA0004176195540000026
其中训练数据集/>
Figure BDA0004176195540000027
中含有n个张量样本,用式表达为/>
Figure BDA0004176195540000028
则训练数据集
Figure BDA0004176195540000031
其中验证数据集/>
Figure BDA0004176195540000032
中含有T-n个张量样本,用式表达为/>
Figure BDA0004176195540000033
则训练数据集/>
Figure BDA0004176195540000034
然后要构建训练数据集,本方案中首先利用每个位置的训练数据集构造该位置的anchor指纹张量,构造anchor指纹张量的过程为:将Ri的训练数据集/>
Figure BDA0004176195540000035
中的RSSI指纹张量求均值,得到Fi,并作为模型输入之一的anchor,其计算公式为/>
Figure BDA0004176195540000036
上式中的Fi即为Ri的anchor张量,所有N个位置的anchor即构成了指纹集合F,用式表达为F={F1,F2,…,Fi,…,FN},然后为了保证训练的充分性,选择将训练集中的指纹张量与指纹集合F中的指纹张量两两配对得到训练数据集K1和验证数据集K2,K1表达式为K1={T1,1,1,Ti,j,k,TN,n,N},其中元素Ti,j,k表示由训练集中i个位置Ri的第j个指纹张量/>
Figure BDA0004176195540000037
与第k个位置Rk的anchor指纹张量Fk构造的样本对,可以用式表达为/>
Figure BDA0004176195540000038
K2表达式为K2={V1,1,1,Vi,j,k,VN,T-n,N},其中元素Vi,j,k表示由验证集中Ri的第j个指纹张量/>
Figure BDA0004176195540000039
与Rk的anchor指纹张量Fk构造的样本对,可以用式表达为/>
Figure BDA00041761955400000310
进而,使用新数据集训练三元组网络模型;S41, in the training model stage, assume that there are N positions available for positioning, and each position point has T samples. After data processing, a training data set is generated for the position R i
Figure BDA0004176195540000025
and the validation dataset />
Figure BDA0004176195540000026
where the training dataset />
Figure BDA0004176195540000027
Contains n tensor samples, expressed as />
Figure BDA0004176195540000028
Then the training data set
Figure BDA0004176195540000031
where the validation dataset />
Figure BDA0004176195540000032
Contains Tn tensor samples, expressed as />
Figure BDA0004176195540000033
Then the training data set />
Figure BDA0004176195540000034
Then to construct the training data set, in this scheme, first use the training data set of each position to construct the anchor fingerprint tensor of the position, the process of constructing the anchor fingerprint tensor is: the training data set of R i
Figure BDA0004176195540000035
Calculate the mean value of the RSSI fingerprint tensor in F i , and use it as one of the model input anchors, the calculation formula is />
Figure BDA0004176195540000036
F i in the above formula is the anchor tensor of R i , and all the anchors of N positions constitute the fingerprint set F, expressed as F={F 1 , F 2 ,..., F i ,...,F N }, and then in order to ensure the adequacy of training, choose to pair the fingerprint tensor in the training set with the fingerprint tensor in the fingerprint set F to obtain the training data set K 1 and the verification data set K 2 , the expression of K 1 is K 1 = {T 1, 1, 1 , T i, j, k , T N, n, N }, where the element T i, j, k represents the jth fingerprint tensor from the i position Ri in the training set />
Figure BDA0004176195540000037
The sample pair constructed with the anchor fingerprint tensor F k of the k-th position R k can be expressed as />
Figure BDA0004176195540000038
The expression of K 2 is K 2 ={V 1, 1, 1 , V i, j, k , V N, Tn, N }, where the element V i, j, k represents the jth fingerprint sheet of Ri in the verification set Quantity />
Figure BDA0004176195540000039
The sample pair constructed with the anchor fingerprint tensor F k of R k can be expressed as />
Figure BDA00041761955400000310
Furthermore, the triplet network model is trained using the new dataset;

S42:训练好三元组网络后,剪去三元组的主体网络,只保留堆叠自编码器的前半部分,这部分保存为室内定位系统的特征提取网络。S42: After the triplet network is trained, cut off the main network of the triplet, and only keep the first half of the stacked autoencoder, which is saved as the feature extraction network of the indoor positioning system.

可选的,所述S5中,构建完整的神经网络模型就是将保存下来的特征提取网络连接到已经训练好的BP神经网络,最后保存模型及其参数。Optionally, in S5, constructing a complete neural network model is to connect the saved feature extraction network to the trained BP neural network, and finally save the model and its parameters.

可选的,所述S6中,测试阶段,在待测位置Li获取y1组指纹张量后,使用S2中相同的高斯滤波方法,预处理新数据得到y2个样本。首先将得到的数据依次输入到Tri-Sae模型中,最后将y2个定位结果;然后将计算出的anchor指纹集合F={F1,F2,…,Fi,…,FN}与y2个样本构成测试样本对集D={D1,…,Dj,…,Dy2},其中Dj={k1,k2,…,ki,…,kN},其中ki=(t,Fi),t为y2个样本中的某一条数据,然后将此测试对输入到两个已经训练好的堆叠式编码其中,对比两个测试对的欧式距离,最小的即为此测试对的定位结果,然后采取投票的方式得到样本t的最终结果,迭代y2次,将会得到y2个结构;综合上述,每个待测位置Li将会得到2×y2个结果,采用投票的方式,得到待测位置Li的最终预测标签。Optionally, in S6, in the testing phase, after obtaining y 1 sets of fingerprint tensors at the location L i to be tested, use the same Gaussian filtering method in S2 to preprocess new data to obtain y 2 samples. First, input the obtained data into the Tri-Sae model in turn, and finally put y 2 positioning results; then combine the calculated anchor fingerprint set F={F 1 , F 2 ,..., F i ,..., F N } with y 2 samples constitute the test sample pair set D={D 1 ,...,D j ,...,D y2 }, where D j ={k 1 ,k 2 ,..., ki ,...,k N }, where k i = (t, F i ), t is a piece of data in the 2 samples of y, and then input this test pair into two well-trained stacked codes, compare the Euclidean distance of the two test pairs, the smallest That is, the positioning result of this test pair, and then vote to get the final result of sample t, iterate y 2 times, and y 2 structures will be obtained; based on the above, each position L i to be tested will get 2×y 2 results, by voting, to get the final predicted label of the position L i to be tested.

本发明的有益效果在于:The beneficial effects of the present invention are:

1、提出一种基于三元组网络的自动编码器训练方式,为提取室内位置指纹特征向量提供了一种新的技术,不再通过简单的训练自编码器来提取样本特征,解决了训练集少、特征提取难度大、定位精度低等问题。1. An autoencoder training method based on triplet network is proposed, which provides a new technology for extracting indoor location fingerprint feature vectors. It no longer extracts sample features through simple training of autoencoders, and solves the problem of training set Few, difficult feature extraction, low positioning accuracy and other issues.

2、位置指纹数据库不再利用位置指纹的原始数据来存储位置信息,转换成位置所对应的特征向量来存储,不光保护了室内空间隐私,减小了存储空间,同时在定位过程中。2. The location fingerprint database no longer uses the original data of the location fingerprint to store the location information, but converts it into the feature vector corresponding to the location for storage, which not only protects the privacy of the indoor space, reduces the storage space, but also in the positioning process.

3、随着后续数据的持续更新,可以继续训练完善特征提取模型,只需要将训练数据通过特征提取网络,就可以保存到用于训练定位模型的数据库里面。3. With the continuous update of subsequent data, the feature extraction model can be continuously trained and improved. It only needs to pass the training data through the feature extraction network, and then it can be saved in the database used to train the positioning model.

4、在定位的离线阶段,使用了投票的方式,提高了室内定位系统的整体精度,避免了个别没被去除的坏数据对最后的定位结果造成影响。4. In the offline stage of positioning, voting is used to improve the overall accuracy of the indoor positioning system and avoid the impact of individual bad data that has not been removed on the final positioning result.

本发明的其他优点、目标和特征在某种程度上将在随后的说明书中进行阐述,并且在某种程度上,基于对下文的考察研究对本领域技术人员而言将是显而易见的,或者可以从本发明的实践中得到教导。本发明的目标和其他优点可以通过下面的说明书来实现和获得。Other advantages, objects and features of the present invention will be set forth in the following description to some extent, and to some extent, will be obvious to those skilled in the art based on the investigation and research below, or can be obtained from It is taught in the practice of the present invention. The objects and other advantages of the invention may be realized and attained by the following specification.

附图说明Description of drawings

为了使本发明的目的、技术方案和优点更加清楚,下面将结合附图对本发明作优选的详细描述,其中:In order to make the purpose of the present invention, technical solutions and advantages clearer, the present invention will be described in detail below in conjunction with the accompanying drawings, wherein:

图1是本发明损失函数loss的求值流程示意图和网络架构示意图;Fig. 1 is a schematic diagram of the evaluation process and a schematic diagram of the network architecture of the loss function loss of the present invention;

图2是本发明的定位流程示意图。Fig. 2 is a schematic diagram of the positioning process of the present invention.

具体实施方式Detailed ways

以下通过特定的具体实例说明本发明的实施方式,本领域技术人员可由本说明书所揭露的内容轻易地了解本发明的其他优点与功效。本发明还可以通过另外不同的具体实施方式加以实施或应用,本说明书中的各项细节也可以基于不同观点与应用,在没有背离本发明的精神下进行各种修饰或改变。需要说明的是,以下实施例中所提供的图示仅以示意方式说明本发明的基本构想,在不冲突的情况下,以下实施例及实施例中的特征可以相互组合。Embodiments of the present invention are described below through specific examples, and those skilled in the art can easily understand other advantages and effects of the present invention from the content disclosed in this specification. The present invention can also be implemented or applied through other different specific implementation modes, and various modifications or changes can be made to the details in this specification based on different viewpoints and applications without departing from the spirit of the present invention. It should be noted that the diagrams provided in the following embodiments are only schematically illustrating the basic concept of the present invention, and the following embodiments and the features in the embodiments can be combined with each other in the case of no conflict.

其中,附图仅用于示例性说明,表示的仅是示意图,而非实物图,不能理解为对本发明的限制;为了更好地说明本发明的实施例,附图某些部件会有省略、放大或缩小,并不代表实际产品的尺寸;对本领域技术人员来说,附图中某些公知结构及其说明可能省略是可以理解的。Wherein, the accompanying drawings are for illustrative purposes only, and represent only schematic diagrams, rather than physical drawings, and should not be construed as limiting the present invention; in order to better illustrate the embodiments of the present invention, some parts of the accompanying drawings may be omitted, Enlargement or reduction does not represent the size of the actual product; for those skilled in the art, it is understandable that certain known structures and their descriptions in the drawings may be omitted.

本发明实施例的附图中相同或相似的标号对应相同或相似的部件;在本发明的描述中,需要理解的是,若有术语“上”、“下”、“左”、“右”、“前”、“后”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本发明和简化描述,而不是指示或暗示所指的装置或元件必须具有特定的方位、以特定的方位构造和操作,因此附图中描述位置关系的用语仅用于示例性说明,不能理解为对本发明的限制,对于本领域的普通技术人员而言,可以根据具体情况理解上述术语的具体含义。In the drawings of the embodiments of the present invention, the same or similar symbols correspond to the same or similar components; , "front", "rear" and other indicated orientations or positional relationships are based on the orientations or positional relationships shown in the drawings, which are only for the convenience of describing the present invention and simplifying the description, rather than indicating or implying that the referred devices or elements must It has a specific orientation, is constructed and operated in a specific orientation, so the terms describing the positional relationship in the drawings are for illustrative purposes only, and should not be construed as limiting the present invention. For those of ordinary skill in the art, the understanding of the specific meaning of the above terms.

本发明所揭示的是一种基于三元组自编码器的室内位置指纹定位技术,如图2具体展示了详情,详细步骤如下:What the present invention discloses is an indoor position fingerprint positioning technology based on a triplet autoencoder, as shown in Figure 2 in detail, and the detailed steps are as follows:

S1:在室内环境里均匀安装无线信号发射装置,利用手机自带的信号读取设备获取每个位置的RSSI指纹张量,设置信号接收极弱或者接收不到的节点的RSSI值为-110,最后完成构造原始的指纹数据库;S1: Install wireless signal transmitters evenly in the indoor environment, use the signal reading device that comes with the mobile phone to obtain the RSSI fingerprint tensor of each location, and set the RSSI value of the nodes with extremely weak or unreceived signals to -110, Finally, construct the original fingerprint database;

S2:原始数据库中所有样本的标签分别对应不同的位置,设有N个位置,位置集合R的表达式为R={R1,R2,…,Ri,…,RN},每个位置Ri都有n个样本,假设n个样本服从于高斯分布,然后计算出均值μ和方差σ2,选取[μ-σ,μ+σ]区间的数组RSSI指纹张量作为Ri的训练样本,依次迭代N次,直至所有位置都使用高斯滤波处理,过滤出来的RSSI指纹张量构成新的训练数据集。S2: The labels of all the samples in the original database correspond to different locations. There are N locations. The expression of the location set R is R={R 1 , R 2 ,...,R i ,...,R N }, each There are n samples at position R i , assuming that n samples obey the Gaussian distribution, then calculate the mean value μ and variance σ 2 , and select the array RSSI fingerprint tensor in the interval [μ-σ, μ+σ] as the training of R i The sample is iterated N times in turn until all positions are processed by Gaussian filtering, and the filtered RSSI fingerprint tensor forms a new training data set.

S3:首先构造堆叠式自编码器作为三元组网络的子网络,设计使用[输入层+[256,128,64,128,256]+输出层]的网络结构,每层的激活函数设置为

Figure BDA0004176195540000051
此编码器分为上中下层结构,其中上下两层结构共享权重,节省了参数量,提升了性能;然后就是设计使用三元组网络,此网络结构有三个子网络,且都是上述的堆叠式自编码器,三个子网络间共享权值,网络的输入是特别构造过的样本对,样本对会被拆分成三份,分别喂入到三个子网络里面,最后计算loss函数,使用Adam算法,持续更新参数;最后设计BP神经网络,构造[输入层+[128,128]+输出尾]的网络结构,最后一层使用sigmoid激活函数,其余层都使用ReLU激活函数。S3: First construct a stacked autoencoder as a sub-network of the triplet network, design a network structure using [input layer + [256, 128, 64, 128, 256] + output layer], and set the activation function of each layer to
Figure BDA0004176195540000051
This encoder is divided into an upper, middle and lower layer structure, in which the upper and lower layers share weights, which saves the amount of parameters and improves performance; then it is designed to use a triplet network. This network structure has three sub-networks, and they are all stacked above. Autoencoder, the weights are shared between the three sub-networks. The input of the network is a specially constructed sample pair. The sample pair will be split into three parts and fed into the three sub-networks respectively. Finally, the loss function is calculated and the Adam algorithm is used. , continuously update the parameters; finally design the BP neural network, construct the network structure of [input layer + [128, 128] + output tail], the last layer uses the sigmoid activation function, and the rest of the layers use the ReLU activation function.

S4:使用平均公式

Figure BDA0004176195540000052
计算每个位置Ri的anchor指纹张量,获得指纹集F={F1,F2,…,Fi,…,FN},再使用逐一匹配的方法构造训练数据集K={T1,1,1,Ti,j,k,TN,n,N},将构造好的数据集K喂到构造好的三元组网络中,逐轮训练,使用梯度下降算法Adam持续更新模型的参数,模型训练完成后,只保留三元组模型的子网络的前半部分作为特征提取网络。S4: Use the average formula
Figure BDA0004176195540000052
Calculate the anchor fingerprint tensor of each position R i , obtain the fingerprint set F={F 1 , F 2 ,..., F i ,...,F N }, and then use the one-by-one matching method to construct the training data set K={T 1 , 1 , 1 , T i, j, k , T N, n, N }, feed the constructed data set K to the constructed triplet network, train round by round, and use the gradient descent algorithm Adam to continuously update the model After the model training is completed, only the first half of the sub-network of the triplet model is reserved as the feature extraction network.

S5:将构造好的BP神经网络连接到特征提取网络后面,使用预处理后的数据集训练此网络,只使用梯度下降算法Adam更新BP神经网络的参数,迭代完成后,模型最后保存为Tri-Sae网络模型;S5: Connect the constructed BP neural network to the back of the feature extraction network, use the preprocessed data set to train the network, and only use the gradient descent algorithm Adam to update the parameters of the BP neural network. After the iteration is completed, the model is finally saved as Tri- Sae network model;

S6:首先将预处理后的多条实时RSSI数据喂到Tri-Sae模型中,将得到n个实验结果;然后将预处理后的多条实时RSSI数据与anchor指纹集合构造成测试样本对ki=(t,Fi)并输入到特征提取网络中,对比两样本间的欧式距离

Figure BDA0004176195540000053
取dist最小的为最终结果,经过多次迭代,最终也将获得n个实验结果,然后通过投票的方式,裁定最终定位结构,根据待测位置标签来评估室内定位系统的精度。S6: First feed the preprocessed pieces of real-time RSSI data into the Tri-Sae model, and get n experimental results; then construct the preprocessed pieces of real-time RSSI data and the anchor fingerprint set into a test sample pair k i =(t, F i ) and input to the feature extraction network, compare the Euclidean distance between the two samples
Figure BDA0004176195540000053
The smallest dist is taken as the final result. After multiple iterations, n experimental results will be finally obtained, and then the final positioning structure will be determined by voting, and the accuracy of the indoor positioning system will be evaluated according to the location label to be tested.

最后说明的是,以上实施例仅用以说明本发明的技术方案而非限制,尽管参照较佳实施例对本发明进行了详细说明,本领域的普通技术人员应当理解,可以对本发明的技术方案进行修改或者等同替换,而不脱离本技术方案的宗旨和范围,其均应涵盖在本发明的权利要求范围当中。Finally, it is noted that the above embodiments are only used to illustrate the technical solutions of the present invention without limitation. Although the present invention has been described in detail with reference to the preferred embodiments, those of ordinary skill in the art should understand that the technical solutions of the present invention can be carried out Modifications or equivalent replacements, without departing from the spirit and scope of the technical solution, should be included in the scope of the claims of the present invention.

Claims (6)

1.一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:该方法包括以下步骤:1. A method for indoor position fingerprint positioning based on triplet autoencoder, characterized in that: the method may further comprise the steps: S1:利用WIFI接收器在每个位置获得一组WIFI信号强度数据,得到每个位置点的WIFI信号强弱;S1: Use the WIFI receiver to obtain a set of WIFI signal strength data at each location, and obtain the WIFI signal strength of each location point; S2:使用高斯滤波的方式对已经采集到的数据进行预处理;S2: Use Gaussian filtering to preprocess the collected data; S3:构造基于自编码器的三元组神经网络及损失函数和用于最后定位的BP神经网络;S3: Construct the triplet neural network and loss function based on the autoencoder and the BP neural network for final positioning; S4:经过预处理的室内位置指纹数据集训练构造好的三元组自编码器,对训练后的模型剪枝得到特征提取网络,并保留网络模型的结构和参数;S4: The preprocessed indoor location fingerprint data set trains the constructed triplet autoencoder, prunes the trained model to obtain the feature extraction network, and retains the structure and parameters of the network model; S5:使用特征提取网络提取预处理后的数据集,得到降维后的数据集,并训练BP神经网络,将特征提取网络和训练后的BP神经网络保存为Tri-Sae模型;S5: Use the feature extraction network to extract the preprocessed data set, obtain the data set after dimension reduction, and train the BP neural network, save the feature extraction network and the trained BP neural network as a Tri-Sae model; S6:使用Tri-Sae模型和距离度量的方式联合投票定位,最后评价室内定位系统的精度。S6: Use the Tri-Sae model and the distance measure to jointly vote for positioning, and finally evaluate the accuracy of the indoor positioning system. 2.根据权利要求1所述的一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:所述S2中,对采集到的数据采取高斯滤波的处理,去除样本中的坏数据,从RSSI值的高概率区域中选择有效值,并计算输出值的平均值,以提高测距精度,它的过程是:2. A kind of indoor position fingerprint location method based on triplet autoencoder according to claim 1, it is characterized in that: in described S2, the processing of Gaussian filter is taken to the data collected, removes the bad part in the sample. Data, select effective values from the high probability area of RSSI value, and calculate the average value of the output value to improve the ranging accuracy, its process is: 设n个RSSI值的总数服从高斯分布,RSSI值的平均值为μ,RSSI值的方差为σ2,RSSI值的概率密度函数为f(x),因此
Figure FDA0004176195530000011
RSSI值的均值μ的计算公式为
Figure FDA0004176195530000012
σ2的计算公式为/>
Figure FDA0004176195530000013
如果RSSI值在[μ-σ,μ+σ]的范围内,则为高概率置信区间;通过高斯滤波器计算出的平均RSSI值减少环境干扰的影响。
Suppose the total number of n RSSI values obeys Gaussian distribution, the mean value of RSSI value is μ, the variance of RSSI value is σ 2 , and the probability density function of RSSI value is f(x), so
Figure FDA0004176195530000011
The formula for calculating the mean μ of the RSSI value is
Figure FDA0004176195530000012
The calculation formula of σ 2 is />
Figure FDA0004176195530000013
If the RSSI value is in the range of [μ-σ, μ+σ], it is a high-probability confidence interval; the average RSSI value calculated by the Gaussian filter reduces the influence of environmental interference.
3.根据权利要求1所述的一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:所述S3中,三元组网络构造分为设计子网络,添加损失函数loss构造完整网络的过程:3. a kind of indoor position fingerprint location method based on triplet autoencoder according to claim 1, it is characterized in that: in described S3, triplet network structure is divided into design sub-network, adds loss function loss structure The process of the complete network: S31:三元组网络的子网络是堆叠自编码器,设计使用[输入层+[256,128,64,128,256]+输出层]的网络结构,自编码器分为上中下层结构,其中上下两层结构共享权重,每层的激活函数设置为
Figure FDA0004176195530000014
S31: The sub-network of the triplet network is a stacked autoencoder, and the network structure of [input layer + [256, 128, 64, 128, 256] + output layer] is designed. The autoencoder is divided into upper, middle and lower layers. The upper and lower layers share weights, and the activation function of each layer is set to
Figure FDA0004176195530000014
S32:定义新的神经网络模型的损失函数loss,神经网络模型的损失函数loss的过程如下:S32: Define the loss function loss of the new neural network model, the process of the loss function loss of the neural network model is as follows: 将位置指纹数据输入堆叠自编码器,提取到每个样本数据的特征向量,而且每一条样本数据的特征维度都是相同的;每次训练选取三条样本作为训练数据,分别是每个类别的代表特征anchor,anchor的正样本positive,anchor的负样本negative,将他们输入到堆叠自编码器,得到所对应的特征向量;Input the location fingerprint data into the stacked autoencoder to extract the feature vector of each sample data, and the feature dimension of each sample data is the same; each training selects three samples as training data, which are representatives of each category The feature anchor, the positive sample positive of the anchor, and the negative sample negative of the anchor, input them to the stacked autoencoder to obtain the corresponding feature vector; 使得公式
Figure FDA0004176195530000021
成立;同时positive和negative的样本来自原始数据集本身,经过堆叠自编码器后,还原到与输入数据相同的维度,想让输入结果和输出结果尽量保持一致,满足/>
Figure FDA0004176195530000022
β为任意大于零的实数,loss用公式表达为:
makes the formula
Figure FDA0004176195530000021
Established; at the same time, positive and negative samples come from the original data set itself. After stacking the autoencoder, it is restored to the same dimension as the input data. I want to keep the input and output results as consistent as possible, satisfying />
Figure FDA0004176195530000022
β is any real number greater than zero, and the loss is expressed as:
Figure FDA0004176195530000023
Figure FDA0004176195530000023
S33:设计BP神经网络模型,用于最后的定位功能,此神经网络是简单的多层结构,其网络结构是[输入层+[128,128]+输出层],输出层使用sigmoid激活函数,其余层使用ReLU激活函数,并且添加Dropout层,丢弃率为0.1,以防止过拟合。S33: Design a BP neural network model for the final positioning function. This neural network is a simple multi-layer structure, and its network structure is [input layer + [128, 128] + output layer], and the output layer uses a sigmoid activation function. The remaining layers use the ReLU activation function, and add a Dropout layer with a dropout rate of 0.1 to prevent overfitting.
4.根据权利要求1所述的一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:所述S4包括训练过程和剪枝过程:4. a kind of indoor position fingerprint positioning method based on triplet autoencoder according to claim 1, is characterized in that: described S4 comprises training process and pruning process: S41,在训练模型阶段,设共有N个位置可供定位,每个位置点有T个样本,经过数据处理,为位置Ri生成了训练数据集
Figure FDA0004176195530000024
和验证数据集/>
Figure FDA0004176195530000025
其中训练数据集/>
Figure FDA0004176195530000026
中含有n个张量样本,用式表达为/>
Figure FDA0004176195530000027
则训练数据集
Figure FDA0004176195530000028
其中验证数据集/>
Figure FDA0004176195530000029
中含有T-n个张量样本,用式表达为/>
Figure FDA00041761955300000210
则训练数据集/>
Figure FDA00041761955300000211
然后要构建训练数据集,首先利用每个位置的训练数据集构造该位置的anchor指纹张量,构造anchor指纹张量的过程为:将Ri的训练数据集/>
Figure FDA00041761955300000212
中的RSSI指纹张量求均值,得到Fi,并作为模型输入之一的anchor,其计算公式为/>
Figure FDA00041761955300000213
上式中的Fi即为Ri的anchor张量,所有N个位置的anchor即构成了指纹集合F,用式表达为F={F1,F2,…,Fi,…,FN},然后为了保证训练的充分性,选择将训练集中的指纹张量与指纹集合F中的指纹张量两两配对得到训练数据集K1和验证数据集K2,K1表达式为K1={T1,1,1,Ti,j,k,TN,n,N},其中元素Ti,j,k表示由训练集中i个位置Ri的第j个指纹张量/>
Figure FDA00041761955300000214
与第k个位置Rk的anchor指纹张量Fk构造的样本对,用式表达为/>
Figure FDA00041761955300000215
K2表达式为K2={V1,1,1,Vi,j,k,VN,T-n,N},其中元素Vi,j,k表示由验证集中Ri的第j个指纹张量/>
Figure FDA00041761955300000216
与Rk的anchor指纹张量Fk构造的样本对,用式表达为
Figure FDA00041761955300000217
使用新数据集训练三元组网络模型;
S41, in the training model stage, assume that there are N positions available for positioning, and each position point has T samples. After data processing, a training data set is generated for the position R i
Figure FDA0004176195530000024
and the validation dataset />
Figure FDA0004176195530000025
where the training dataset />
Figure FDA0004176195530000026
Contains n tensor samples, expressed as />
Figure FDA0004176195530000027
Then the training data set
Figure FDA0004176195530000028
where the validation dataset />
Figure FDA0004176195530000029
Contains Tn tensor samples, expressed as />
Figure FDA00041761955300000210
Then the training data set />
Figure FDA00041761955300000211
Then to construct the training data set, first use the training data set of each position to construct the anchor fingerprint tensor of the position, the process of constructing the anchor fingerprint tensor is: the training data set of R i />
Figure FDA00041761955300000212
Calculate the mean value of the RSSI fingerprint tensor in F i , and use it as one of the model input anchors, the calculation formula is />
Figure FDA00041761955300000213
F i in the above formula is the anchor tensor of R i , and all the anchors of N positions constitute the fingerprint set F, expressed as F={F 1 , F 2 ,..., F i ,...,F N }, and then in order to ensure the adequacy of training, choose to pair the fingerprint tensor in the training set with the fingerprint tensor in the fingerprint set F to obtain the training data set K 1 and the verification data set K 2 , the expression of K 1 is K 1 = {T 1, 1, 1 , T i, j, k , T N, n, N }, where the element T i, j, k represents the jth fingerprint tensor from the i position Ri in the training set />
Figure FDA00041761955300000214
The sample pair constructed with the anchor fingerprint tensor F k of the k-th position R k is expressed as />
Figure FDA00041761955300000215
The expression of K 2 is K2={V 1, 1, 1 , V i, j, k , V N, Tn, N }, where the element V i, j, k represents the jth fingerprint tensor of Ri in the validation set />
Figure FDA00041761955300000216
The sample pair constructed with the anchor fingerprint tensor F k of R k is expressed as
Figure FDA00041761955300000217
Train a triplet network model using the new dataset;
S42:训练好三元组网络后,剪去三元组的主体网络,只保留堆叠自编码器的前半部分,这部分保存为室内定位系统的特征提取网络。S42: After the triplet network is trained, cut off the main network of the triplet, and only keep the first half of the stacked autoencoder, which is saved as the feature extraction network of the indoor positioning system.
5.根据权利要求1所述的一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:所述S5中,构建完整的神经网络模型就是将保存下来的特征提取网络连接到已经训练好的BP神经网络,最后保存模型及其参数。5. a kind of indoor position fingerprint location method based on triplet autoencoder according to claim 1, it is characterized in that: in described S5, constructing complete neural network model is exactly that the feature extraction network that preserves is connected to The trained BP neural network, and finally save the model and its parameters. 6.根据权利要求1所述的一种基于三元组自编码器的室内位置指纹定位方法,其特征在于:所述S6中,测试阶段,在待测位置Li获取y1组指纹张量后,使用S2中相同的高斯滤波方法,预处理新数据得到y2个样本;首先将得到的数据依次输入到Tri-Sae模型中,最后将y2个定位结果;然后将计算出的anchor指纹集合F={F1,F2,…,Fi,…,FN}与y2个样本构成测试样本对集D={D1,…,Dj,…,Dy2},其中Dj={k1,k2,…,ki,…,kN},其中ki=(t,Fi),t为y2个样本中的某一条数据,然后将此测试对输入到两个已经训练好的堆叠式编码其中,对比两个测试对的欧式距离,最小的即为此测试对的定位结果,然后采取投票的方式得到样本t的最终结果,迭代y2次,将会得到y2个结构;每个待测位置Li将会得到2×y2个结果,采用投票的方式,得到待测位置Li的最终预测标签。6. a kind of indoor position fingerprint positioning method based on triplet autoencoder according to claim 1, is characterized in that: in described S6, test stage, obtains y 1 group fingerprint tensors at position L to be tested Finally, use the same Gaussian filtering method in S2 to preprocess the new data to obtain y 2 samples; firstly, input the obtained data into the Tri-Sae model in turn, and finally y 2 positioning results; then the calculated anchor fingerprint The set F={F 1 , F 2 ,...,F i ,...,F N } and y 2 samples constitute the test sample pair set D={D 1 ,...,D j ,...,D y2 }, where D j ={k 1 , k 2 ,..., ki ,...,k N }, where ki =(t, F i ), t is a piece of data in y 2 samples, and then this test pair is input into two In a stacked code that has been trained, compare the Euclidean distances of two test pairs, the smallest one is the positioning result of this test pair, and then vote to get the final result of sample t, and iterate y 2 times, you will get y 2 structures; each location L i to be tested will get 2×y 2 results, and the final predicted label of the location L i to be tested will be obtained by voting.
CN202310392222.1A 2023-04-13 2023-04-13 Indoor position fingerprint positioning method based on triplet self-encoder Pending CN116321008A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310392222.1A CN116321008A (en) 2023-04-13 2023-04-13 Indoor position fingerprint positioning method based on triplet self-encoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310392222.1A CN116321008A (en) 2023-04-13 2023-04-13 Indoor position fingerprint positioning method based on triplet self-encoder

Publications (1)

Publication Number Publication Date
CN116321008A true CN116321008A (en) 2023-06-23

Family

ID=86822386

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310392222.1A Pending CN116321008A (en) 2023-04-13 2023-04-13 Indoor position fingerprint positioning method based on triplet self-encoder

Country Status (1)

Country Link
CN (1) CN116321008A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108810799A (en) * 2018-05-28 2018-11-13 湖南大学 A kind of more floor indoor orientation methods and system based on linear discriminant analysis
CN109444813A (en) * 2018-10-26 2019-03-08 南京邮电大学 A kind of RFID indoor orientation method based on BP and DNN amphineura network
US10655971B1 (en) * 2019-05-01 2020-05-19 Mapsted Corp. Maintaining a trained neural network for mobile device RSS fingerprint based indoor navigation
CN111461251A (en) * 2020-04-10 2020-07-28 桂林电子科技大学 Indoor positioning method of WiFi fingerprint based on random forest and self-encoder
WO2021007812A1 (en) * 2019-07-17 2021-01-21 深圳大学 Deep neural network hyperparameter optimization method, electronic device and storage medium
CN114818800A (en) * 2022-04-19 2022-07-29 浙江工业大学 Indoor positioning floor recognition method based on stacking self-encoder deep learning framework

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108810799A (en) * 2018-05-28 2018-11-13 湖南大学 A kind of more floor indoor orientation methods and system based on linear discriminant analysis
CN109444813A (en) * 2018-10-26 2019-03-08 南京邮电大学 A kind of RFID indoor orientation method based on BP and DNN amphineura network
US10655971B1 (en) * 2019-05-01 2020-05-19 Mapsted Corp. Maintaining a trained neural network for mobile device RSS fingerprint based indoor navigation
WO2021007812A1 (en) * 2019-07-17 2021-01-21 深圳大学 Deep neural network hyperparameter optimization method, electronic device and storage medium
CN111461251A (en) * 2020-04-10 2020-07-28 桂林电子科技大学 Indoor positioning method of WiFi fingerprint based on random forest and self-encoder
CN114818800A (en) * 2022-04-19 2022-07-29 浙江工业大学 Indoor positioning floor recognition method based on stacking self-encoder deep learning framework

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
刘震宇;李嘉俊;王昆;: "一种基于深度自编码器的指纹匹配定位方法", 广东工业大学学报, no. 05, 31 December 2017 (2017-12-31) *
张永宏;王逸飞;赵晓平;吴家新;王丽华;: "基于深度度量学习的电机故障诊断", 测控技术, 第07期, 31 December 2020 (2020-12-31), pages 1 *

Similar Documents

Publication Publication Date Title
CN113378632B (en) An unsupervised domain adaptation person re-identification method based on pseudo-label optimization
Jang et al. Indoor localization with WiFi fingerprinting using convolutional neural network
Shanthamallu et al. A brief survey of machine learning methods and their sensor and IoT applications
CN110536245B (en) An indoor wireless positioning method and system based on deep learning
CN109101938B (en) Multi-label age estimation method based on convolutional neural network
CN113095370A (en) Image recognition method and device, electronic equipment and storage medium
CN109444813A (en) A kind of RFID indoor orientation method based on BP and DNN amphineura network
CN108629380A (en) A kind of across scene wireless signal cognitive method based on transfer learning
CN113642596B (en) A brain network classification method based on community detection and dual-path autoencoding
CN101980250A (en) Object Recognition Method Based on Dimensionality Reduction Local Feature Descriptor and Hidden Conditional Random Field
CN115311502A (en) A small sample scene classification method for remote sensing images based on multi-scale dual-stream architecture
CN114708637B (en) A meta-learning-based facial action unit detection method
CN111586605B (en) KNN indoor target positioning method based on adjacent weighted self-adaptive k value
CN111726765A (en) A WIFI indoor positioning method and system for large-scale complex scenes
CN111461251A (en) Indoor positioning method of WiFi fingerprint based on random forest and self-encoder
CN105760900A (en) Hyperspectral image classification method based on affinity propagation clustering and sparse multiple kernel learning
CN111783688B (en) A classification method of remote sensing image scene based on convolutional neural network
CN112766229A (en) Human face point cloud image intelligent identification system and method based on attention mechanism
CN114818800A (en) Indoor positioning floor recognition method based on stacking self-encoder deep learning framework
CN113128518A (en) Sift mismatch detection method based on twin convolution network and feature mixing
KR20240121445A (en) Apparatus and method for measuring user location based on a cnn model using markov transition field
CN114724245A (en) CSI-based incremental learning human body action identification method
CN107679484A (en) A kind of Remote Sensing Target automatic detection and recognition methods based on cloud computing storage
CN116959035A (en) Gait-based identity recognition method, device, chip and terminal
CN107967454B (en) A two-way convolutional neural network remote sensing classification method considering the spatial neighborhood relationship

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载