WO2021007812A1 - Deep neural network hyperparameter optimization method, electronic device and storage medium - Google Patents
Deep neural network hyperparameter optimization method, electronic device and storage medium Download PDFInfo
- Publication number
- WO2021007812A1 WO2021007812A1 PCT/CN2019/096379 CN2019096379W WO2021007812A1 WO 2021007812 A1 WO2021007812 A1 WO 2021007812A1 CN 2019096379 W CN2019096379 W CN 2019096379W WO 2021007812 A1 WO2021007812 A1 WO 2021007812A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- task
- neural network
- network model
- conditional
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- the invention relates to a hyperparameter optimization method, in particular to a deep neural network hyperparameter optimization method, electronic equipment and storage medium.
- neural network hyperparameters have always been a problem that limits the performance of network models.
- neural network models are becoming more and more complex, and there are more and more types of hyperparameters, which often results in failure to select appropriate hyperparameter combinations according to the relationship between hyperparameters.
- the most commonly used hyperparameter optimization method is the Bayesian optimization method. This method can find a good decision combination from the decision space under the actual evaluation times as much as possible. The main idea is to construct a proxy model of the entire problem process based on historical data.
- the main proxy model in Bayesian optimization is Gaussian Process (GPs), but generally only a single problem is optimized at a time, or runs in parallel at the expense of hardware to ensure full use of data and information within a certain period of time.
- GPs Gaussian Process
- Single-task learning learns from scratch independently, ignoring the relevant information of other similar tasks to in-depth study of data characteristics, and often encounters problems such as large noise, high data dimensionality, or small amount of data that have a greater impact on the results.
- one of the objectives of the present invention is to provide a deep neural network hyperparameter optimization method, which can solve the problems of high calculation cost and long running time for hyperparameter optimization in the prior art.
- the second object of the present invention is to provide an electronic device that can solve the problems of high calculation cost and long running time when optimizing hyperparameters in the prior art.
- the third object of the present invention is to provide a computer storage medium, which can solve the problems of high calculation cost and long running time when optimizing hyperparameters in the prior art.
- a deep neural network hyperparameter optimization method including the following steps:
- Parameter setting steps set the number of training tasks, the objective function of each task, the number of initial values of each task, and the maximum number of iterations;
- Model training steps Obtain the observation set of each task according to the number of initial values of each task and the objective function, and train the network model according to the observation set of each task to obtain a multi-task conditional neural network model;
- Prediction step predict any point in the unknown area according to the multi-task conditional neural network model to obtain the objective function value of each point corresponding to each task;
- Screening step Screen the objective function value of each task according to the particle swarm algorithm and all points in the unknown area to obtain candidate points for each task;
- Iterative steps Substitute the candidate points of each task into the multi-task conditional neural network model to calculate the true value, and add the candidate points of each task and the calculated true value to the observation set of each task to form each The new observation set of the task, and then according to the new observation set, the model training step, the prediction step, the screening step and the iteration step are executed in sequence; until the maximum number of iterations is reached, the observation set formed by the last iteration is selected from the observation set corresponding to the largest response value
- the multi-task neural conditional network model learns from each other during the training of the conditional neural network model through multi-task learning, and then forms the multi-task conditional neural network model.
- the multi-task conditional neural network model combines the output layers of the conditional neural network model of multiple tasks with a similarity network layer to form a multi-task conditional neural network model; the similarity network layer is composed of a fully connected network.
- model training step specifically includes:
- Step S11 Set the number of model training iterations, the maximum number of algorithm iterations, and the minimized loss function; the minimized loss function is the sum of the minimized loss functions of all tasks; the minimized loss function of each task is: minimize the negative conditional logarithm Probability
- Step S12 randomly select the corresponding initial value from the observation set of each task according to the preset ratio, generate the decoder training set of the corresponding task, and use the observation set of each task as the encoder training set of the corresponding task;
- Step S13 Perform network model training according to the decoder training set and encoder training set of each task, and update the network parameters of the conditional neural network model of each task and the parameters of the similarity network layer according to the minimized loss function;
- Step S14 When the network model training reaches the number of model training iterations, return to perform step S12 to step S14 in sequence; until the maximum number of algorithm iterations is reached, a multi-task conditional neural network model is obtained.
- step S13 further includes: updating the network parameters of the conditional neural network model and the parameters of the similarity network layer of each task according to the minimized loss function and using the Adam optimizer back propagation algorithm.
- the generation process of the observation set is as follows: First, the objective function of each task is generated according to the objective function of each task and the number of initial values, and the corresponding objective function value is obtained according to the initial value and objective of each task.
- the function value is the observation value of each task.
- the screening step is: calculating the collection function EI of each task according to the value of the objective function for each task at each point in the unknown area, and using the collection function EI of each task as the fitness of the particle swarm algorithm Function, and then select the point with the largest function value of the collection function EI according to the particle swarm algorithm as the candidate point for each task.
- An electronic device including a memory, a processor, and a computer program stored on the memory and running on the processor.
- the processor implements a deep neural network as one of the objectives of the present invention when the processor executes the program Steps of hyperparameter optimization method.
- a computer-readable storage medium has a computer program stored thereon, and when the computer program is executed by a processor, the steps of a deep neural network hyperparameter optimization method adopted as one of the objectives of the present invention are realized.
- the present invention has the following beneficial effects:
- the present invention implements the training of the model by using the conditional neural network model instead of the Gaussian model, and combines it with multi-task learning to form a multi-task conditional neural network model, and applies the multi-task conditional neural network model to the Bayesian algorithm framework, In turn, multi-task hyperparameter optimization is realized, so that the calculation cost and running time cost are reduced in the process of hyperparameter optimization.
- FIG. 1 is a flowchart of a method for optimizing hyperparameters of a deep neural network provided by the present invention
- Figure 2 is a network structure diagram of the conditional neural network model provided by the present invention.
- Figure 3 is a network structure diagram of a multi-task conditional neural network model provided by the present invention.
- Figure 4 is a network training diagram of a multi-task conditional neural network model provided by the present invention.
- FIG. 5 is a schematic diagram of the prediction of the prediction point x * in the multi-task conditional neural network model provided by the present invention.
- Fig. 6 is a flowchart of a hyperparameter optimization method for applying a multi-task conditional neural network model to a Bayesian decision algorithm provided by the present invention
- Fig. 7 is a schematic diagram of the extended data set provided by the present invention.
- the present invention proposes a method that uses Conditional Neural Processes (CNPs) to replace Gaussian Processes (GPs), and CNPs combine the characteristics of random processes and neural networks, and are inspired and used by the flexibility of Gaussian processes Gradient descent is used for neural network training to realize the hyperparameter optimization method.
- CNPs Conditional Neural Processes
- GPs Gaussian Processes
- the conditional neural network process is parameterized by the neural network when learning the known observation data.
- the conditional neural network model is trained by randomly sampling the data set and following the gradient steps to maximize the random number given a random observation set. Set the conditional likelihood.
- the general method of hyperparameter optimization in the prior art is to use the Gaussian process for model training, and this patent uses a conditional neural network model with stronger learning ability to replace the traditional Gaussian process, avoiding the calculation of complex covariance functions .
- a multi-task conditional neural network process Multi-task Conditional Neural Processes in One-to-many case, OMc-MTCNPs
- Bayesian decision algorithm Bayesian
- the multi-task conditional neural network model can make full use of the known small amount of data to enhance the generalization performance of each single task.
- Applying the multi-task conditional neural network model to the framework of the Bayesian decision algorithm can find a better approximate optimal solution in the hyperparameter optimization process, and solve the long running time and high computational cost of the existing hyperparameter optimization processing. problem.
- each task is a deep network model, and each model has a deep network hyperparameter optimization training set.
- each task has multiple hyperparameters, and the multi-task conditional neural network model can iteratively find a better hyperparameter combination corresponding to each task according to the respective training set and training results of each task.
- the optimal hyperparameter combination is found, the unknown point can be accurately predicted under the condition that the optimal hyperparameter exists.
- the present invention first introduces the model training process of the conditional neural network model and the model training process of the multi-task conditional neural network model.
- CNPs model For conditional neural network model (CNPs model):
- the conditional neural network model uses two stages of learning data to alleviate the problem that a small amount of data cannot be effectively trained: the first stage is to learn the statistical distribution information of the training data, the second stage only uses part of the training data and the first stage to learn The distribution information of the joint fits a specific function.
- Figure 2 shows the process structure diagram of the conditional neural network.
- the conditional neural network is composed of two parts, namely the encoder h and the decoder g, both of which are composed of neural networks.
- the encoder h mainly learns the mapping relationship between the data, that is, parameterized conditional probability;
- the decoder g uses the learned information to perform predictive regression, that is, calculates the conditional probability.
- the second stage fitting function p: X ⁇ Y
- the decoder g finally makes f and p as close as possible.
- conditional neural network model is that there is no need to set a Gaussian prior, but to obtain distribution information from the data directly through the neural network. If we have a set of m unobserved data CNPs in observations The above parameterization f(T
- O,T) P ⁇ (f(T)
- O,T') P ⁇ (f(T)
- T ⁇ x1,x2,x3,...,xn ⁇ . Therefore, P ⁇ (f(T)
- O , T) ⁇ x ⁇ T P ⁇ (f(T)
- h ⁇ : X ⁇ Y ⁇ R d and g ⁇ : X ⁇ R d ⁇ R e are neural networks.
- the encoder h in formula (2) extracts the known data information to obtain r, and then uses the formula ( 3) It is integrated and used as one of the decision factors for the encoder g to predict each x.
- N ⁇ uniform[1,...,n] is usually set, that is, the first N elements of the data set O are selected to obtain the subset Use it as the encoder h for model training.
- Step A1 First set the learning rate, the number of model training iterations, and the maximum number of algorithm iterations.
- these hyperparameters can be given based on historical experience to facilitate the start of model training.
- the learning rate is also a hyperparameter, which is mainly used to set the neural network model, how to learn, through the optimization of the hyperparameter, set it to the neural network model to achieve model training.
- Step A2 according to a preset ratio O randomly selected subset of the data set from O, N.
- N Uniform[1,...,n]. That is, the first N data are selected from the data set O as the subset ON .
- Step A3 Use the data set O as the training set of the decoder g, and the subset ON as the training set of the encoder h.
- Step A4 Use the minimized negative conditional logarithmic probability as the minimized loss function, that is: among them,
- the minimized loss function is obtained by minimizing the conditional logarithmic probability, and its calculation formula is a conventional formula, and the present invention does not improve the calculation formula.
- Step A5 Use the Adam optimizer update algorithm to update the parameters of the multi-condition network model according to the minimized loss function; return to step A2 to step A5 after each model training iteration number, continue to select subsets, model training and parameters Update, and so on.
- Step A6 When the maximum number of iterations of the algorithm is reached, the training ends, and the conditional neural network model of the task is obtained.
- the unknown point x * into the decoder g of the conditional neural network model, and the unknown point x * can be predicted by the conditional neural network model to obtain the objective function value, such as the objective function The predicted mean and variance of the value.
- the Gaussian process is suitable for occasions with small sample size, but it has the disadvantage of high computational complexity of the covariance function; and the above-mentioned conditional neural network model can only perform model training on a single task.
- the present invention expands the conditional neural network model and expands it into a multi-task conditional neural network model (Multi-task Conditional Neural Processes in One-to-many case, OMC-MTCNPs)
- OMC-MTCNPs Multi-task Conditional Neural Processes in One-to-many case
- the present invention combines multi-task learning with the strong learning and predictive capabilities of the conditional neural network model, proposes a multi-task conditional neural network model, and applies the multi-task conditional neural network model to Bayesian decision making
- the optimization framework of the algorithm is used to optimize the hyperparameters, and then obtain the optimal hyperparameters.
- Both the conditional neural network model and the multi-task conditional neural network model have strong scalability. For some simple low-dimensional data, a fully connected layer can achieve better performance, and for some complex high-dimensional data problems, we can use Convolutional layer to efficiently extract data information.
- the present invention provides the network structure of the multi-task conditional neural network model, as shown in Fig. 3.
- each task samples the same set X at the same time, and Y l represents the set of objective function values of the point set X on the l-th task.
- the multi-task conditional neural network model combines the conditional neural network models of multiple tasks, and combines the output layer of each conditional neural network model with the similarity network layer k to form the multi-task conditional neural network.
- Network model combines the conditional neural network models of multiple tasks, and combines the output layer of each conditional neural network model with the similarity network layer k to form the multi-task conditional neural network.
- the similarity network layer k is composed of a fully connected network, it mainly learns the similarity measure between each task by updating parameters, and combines the similarity information to perform a certain linear combination of the prediction results of each task to obtain the final the result of.
- the mean matrix of each conditional neural network model output by the similarity network layer k Sum variance matrix Connected by a neural network
- the mean matrix m and variance matrix v output by the entire model are respectively determined by with Said.
- w m and b m respectively represent the weight matrix and the bias matrix of the mean connection term
- w v and b v represent the weight matrix and the bias matrix of the variance connection term, respectively.
- the final output of each task is related to other tasks after passing through the similarity network layer k.
- the multi-task conditional neural network model uses the experience learned from similar related tasks to effectively spread the information of one task to the network model of other tasks through an appropriate loss function, and realize the information sharing between tasks. Due to the characteristics of the multi-task model, the scale of observation data is increased by M times, and each task can use the information shared by other tasks, avoiding the task to learn from scratch from its own data, and it can also update these experiences from the network. Integration in the process of dealing with similar problems. Figure 3 only draws the addition of a fully connected layer to represent the relevance of tasks, but if you deal with complex problems, the network structure or the number of layers in the similarity network layer k can be flexibly changed according to different tasks to meet requirements.
- the similarity network layer k can learn the mathematical distribution characteristics of all data. These characteristics are represented in hidden nodes. When the network is forwarded, it can be Activate these nodes to encourage tasks to learn from each other.
- Q l ⁇ be the probability distribution on the function f l :X l ⁇ Y l , namely f l represents the objective function of the lth task
- O l ⁇ X l
- L l is the loss function corresponding to the lth task
- ⁇ is All parameter vectors that define Q.
- conditional neural network model of each task of the multi-task conditional neural network model the mathematical structure is similar to the aforementioned formulas (2), (3) and (4).
- the training process of the multi-task conditional neural network model is as follows:
- the multi-task conditional neural network model prevents overfitting under the premise of fully learning the observation data, and at the same time guarantees the uncertainty of the model in the learning decision space X.
- a subset of the data set is randomly selected multiple times according to the set data ratio as the training data set of the encoder, which can make the generalization effect of the model better.
- w l is the weight of the loss function of the lth task. In order to balance the weight of each task, we set the weight to 1.
- Figure 4 shows the model training of the multi-task conditional neural network model.
- the conditional neural network model of each task will be trained accordingly, and the conditional neural network model of each task can be Independently learn related data set features for other tasks.
- Step B1 Set network parameters such as the number of model training iterations, the maximum number of algorithm iterations, and the minimized loss function.
- the minimized loss function is the sum of the loss functions of each task.
- Step B2 Select the corresponding decoder training set O and encoder training set O ⁇ n for each task according to the preset ratio for model training.
- Step B3 Update the parameters of the conditional neural network model of each task and the parameters of the similarity network layer k according to the minimized loss function and the Adam optimizer back propagation update algorithm, that is, the dashed line and the dashed box related parameters in FIG. 4.
- Step B4 When the model training reaches the number of model training iterations, return to step B2 to step B4, and randomly select the encoder training set again in turn Perform model training and parameter update, and so on.
- Step B5 When the maximum number of iterations of the algorithm is reached, the network training is completed, and the multi-task conditional neural network model is obtained.
- the multi-task conditional neural network model After the multi-task conditional neural network model is trained, it can predict any unknown point x *, and predict the objective function value of the unknown point on each task, such as predicting the mean and variance of the unknown point on each task.
- the decoder of the conditional neural network model of each task inputs the unknown point x * , and the corresponding training set is still input on the encoder training set. This part is to read the training set information and make subsequent predictions Help out. Since the information between tasks is shared through the similarity network layer k, the predicted value output by the conditional neural network model of each task is significantly better than the result predicted by a single task.
- the present invention provides a hyperparameter optimization method based on a multi-task conditional neural network model.
- the multi-task conditional neural network model is applied to the framework of the Bayesian decision algorithm to realize the hyperparameter optimization.
- the multi-task conditional neural network model has excellent predictive performance.
- it is applied as a proxy model to the framework of the Bayesian decision algorithm to optimize the hyperparameters .
- the method specifically includes the following steps:
- Step S2 Obtain the initial value set of each task according to the dimension of the parameters and the number of initial data of each task.
- the initial value set can be expressed as
- the dimension of the parameter is 2, that is, the two-dimensional parameter [x1,x2], where the range of x1 is [0,6], and the range of x2 is [6,10].
- 4 initial points are to be generated, that is, x1 takes 2,4; x2 takes 8,9.
- the initial points [2,8],[2,9],[4,8],[4,9] are obtained. That is to say, divide evenly in the decision space, and then find points in each small part, and finally get the data of the initial point, and generate the initial point set.
- Step S3 Input each initial value in the initial value set of each task into the objective function of the corresponding task, calculate the objective function value of the corresponding task, and then according to the initial value of each task and the corresponding objective function value Form the observation set for each task. That is, the objective function value is The observation set is
- Step S4 Select a subset of the corresponding task from the observation set of each task according to the preset ratio, use the observation set as the decoder training section, and use the subset as the encoder training set for network model training, and then train to obtain Multi-task conditional neural network model.
- the subset is
- Step S5 Predict the predicted value of any point x in each task in the unknown area according to the multi-task conditional neural network model, that is, input the arbitrary point into the conditional neural network model of each task to obtain the corresponding objective function value .
- each task corresponds to multiple unknown points and corresponding objective function values, where the unknown points are the combination of parameters of the task.
- Step S6 Use the particle swarm algorithm to find the optimal parameter combination for each task.
- the optimal parameter combination is Since each task is combined with multiple parameters, the unknown points are filtered according to all the unknown points and objective function values of each task, as well as the particle swarm algorithm, and then the best advantage of each task, that is, the optimal Parameter combination.
- the present invention also uses the collection function EI to screen. Since the collection function EI has the expected improvement characteristics of oil volume, the collection function EI is used as the fitness function of the particle swarm algorithm. It calculates the collection function EI of each task according to the value of the objective function of each task at each point in the unknown area, and uses the collection function EI of each task as the fitness function of the particle swarm algorithm, and then selects it according to the particle swarm algorithm The point with the largest function value of the collection function EI is used as the candidate point of each task, and the candidate point of each task is the optimal parameter combination of the corresponding task.
- Step S6 Input the optimal parameter combination of each task into the objective function of the corresponding task to calculate the true value to obtain the objective function value, and add the optimal parameter combination and objective function value of each task to the observation of the corresponding task Concentrate and form a new observation set corresponding to the task, and then continue to perform steps S3 to S6 according to the new observation set. For example, for the mth task, add the calculated point and the objective function value to the observation set Then perform model training, prediction, and screening.
- the optimal parameter combination of each task is actually evaluated, and then the objective function value corresponding to the optimal parameter combination of each task is calculated. That is to say, the optimal parameter combination of each task and the corresponding objective function value are added to the observation set of the corresponding task, and then the model training, prediction, screening, and evaluation operations are performed again.
- the real evaluation here refers to: after the decision variables are obtained, the decision variables are substituted into the test problem and recalculated, that is, the optimal parameter combination is input into the objective function of the corresponding task for prediction, and the corresponding objective function value is obtained .
- Step S7 By analogy, the above steps are executed in turn, when the maximum number of iterations of the algorithm T is reached, the new observation set formed by each task is to find the point corresponding to the maximum value of the objective function of each task, that is Is a combination of parameters: That is the optimal hyperparameter for each task.
- the multi-task conditional neural network model can accurately predict the objective function value of any point on each task, such as the mean value and the variance.
- the present invention also provides related experimental tests, which further illustrate the effectiveness and accuracy of the hyperparameter optimization of the present invention relative to the prior art.
- the specific data are referred to as follows:
- This experiment sets up two kinds of hyper-parameter problems: the first one sets two variable hyper-parameters of learning rate and bias, and the other hyper-parameters are set to default fixed values; the second one is to add the dropout value on the basis of the first one .
- the table use Q1 and Q2 to represent these two problems, and show the mean and standard deviation of the accuracy of the three tasks on the test set (data in parentheses).
- This experiment uses two proxy models, the conditional neural network model and the multi-task conditional neural network model, to optimize the hyperparameters in the framework of the Bayesian decision algorithm. That is, to find a hyperparameter combination that can make the network prediction accuracy high, and then can verify the effectiveness and accuracy of the multi-task conditional neural network model given in this paper for the hyperparameter optimization problem based on the experimental results.
- the neural network may get different results each time it is run, in order to eliminate errors, all accuracy values are the average of the results of ten runs.
- Table 1 The mean and standard deviation of the accuracy of MNIST tested by neural network
- the hyperparameter performance of the multi-task conditional neural network model is better than the hyperparameter optimization based on the conditional neural network model (CNPs) and the traditional hyperparameter optimization based on the Gaussian process. Therefore, the hyperparameter optimization method based on the multi-task conditional neural network model proposed by the present invention is better.
- the present invention adopts the conditional neural network model as the agent model in the Bayesian decision algorithm framework, replacing the traditional Gaussian process, realizing single-task hyperparameter optimization, and adopting the conditional neural network model on the basis of
- the similarity network layer is extended to a multi-task conditional neural network model, which realizes multi-task hyperparameter optimization.
- the present invention uses the conditional neural network model to replace the traditional Gaussian process to build a model for a single task, and the conditional neural network model is composed of a multi-layer special network structure, it can avoid the complex covariance of the traditional Gaussian process
- the problem of calculation, and the learning and fitting function of the conditional neural network model is also stronger than the traditional Gaussian process, and there is no need to set a suitable prior distribution in advance (because the Gaussian process needs to set the data set distribution model (Gaussian distribution ), and then use the data set to update the parameters in the Gaussian distribution, and it may not be suitable for the current data set.
- Conditional neural network model is to set a specific network structure, only need to update the parameters, it has strong adaptability) Therefore, the multi-task conditional neural network model formed by combining the conditional neural network model with multi-task learning greatly improves the learning ability of the model, thereby realizing hyperparameter optimization, and avoiding computational complexity.
- the present invention also provides an electronic device, which includes a memory, a processor, and a computer program that is stored in the memory and can be run on the processing.
- the processor implements the deep neural network described in the article when the program is executed. Steps of hyperparameter optimization method.
- the present invention also provides a computer-readable storage medium on which a computer program is stored.
- the computer program is executed by a processor, the steps of a deep neural network hyperparameter optimization method as described in the text are realized.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
本发明涉及超参数优化方法,尤其涉及一种深度神经网络超参数优化方法、电子设备及储存介质。The invention relates to a hyperparameter optimization method, in particular to a deep neural network hyperparameter optimization method, electronic equipment and storage medium.
目前,深度学习的网络模型中有许多不是在训练过程中学习到的参数,而是在训练开始之前就直接设置,这些参数被称为神经网络超参数。而超参数优化一直是限制网络模型性能提升的难题,目前神经网络模型越来越复杂,超参数的种类也越来越多,往往造成不能根据超参数之间的关系选择合适的超参数组合。目前比较常用的超参数优化方法为Bayesian优化方法,该方法可以在尽可能的真实评估次数下从决策空间中找到好的决策组合,主要思路是根据历史数据构建了整个问题过程的代理模型而不对真实问题进行评估,并通过代理模型预测的不确定性来决定下一步采样点,通过不断迭代后寻找一个近似最优解。目前Bayesian优化中主要的代理模型为高斯过程(Gaussian Process,GPs),但是一般每次只优化单个问题,或者以牺牲硬件为代价并行运行确保在一定时间内充分利用数据信息等问题。单任务学习单独地从零开始学习,忽略了其他相似任务的相关信息来深入研究数据特征,并且经常会遇到噪声大、数据维度较高或数据量偏小等对结果影响较大的问题。通常需要大量的观测数据来训练得到足够精确的单任务代理模型,但现实生活中很难达到要求,导致根据数据训练的模型有一定的局限性,也就造成了模型预测不够准确。随着数据量的增多,高斯过程中协方差函数计算复杂度呈指数增长,导致计算成本高、运行时间长等问题。At present, many of the network models of deep learning are not learned during the training process, but are set directly before the training starts. These parameters are called neural network hyperparameters. Hyperparameter optimization has always been a problem that limits the performance of network models. At present, neural network models are becoming more and more complex, and there are more and more types of hyperparameters, which often results in failure to select appropriate hyperparameter combinations according to the relationship between hyperparameters. At present, the most commonly used hyperparameter optimization method is the Bayesian optimization method. This method can find a good decision combination from the decision space under the actual evaluation times as much as possible. The main idea is to construct a proxy model of the entire problem process based on historical data. Real problems are evaluated, and the next sampling point is determined by the uncertainty of the proxy model prediction, and an approximate optimal solution is found after constant iteration. At present, the main proxy model in Bayesian optimization is Gaussian Process (GPs), but generally only a single problem is optimized at a time, or runs in parallel at the expense of hardware to ensure full use of data and information within a certain period of time. Single-task learning learns from scratch independently, ignoring the relevant information of other similar tasks to in-depth study of data characteristics, and often encounters problems such as large noise, high data dimensionality, or small amount of data that have a greater impact on the results. A large amount of observation data is usually required to train a sufficiently accurate single-task agent model, but it is difficult to meet the requirements in real life, which leads to certain limitations of the model trained on the data, which results in the model prediction is not accurate enough. As the amount of data increases, the computational complexity of the covariance function in the Gaussian process increases exponentially, leading to problems such as high computational cost and long running time.
发明内容Summary of the invention
为了克服现有技术的不足,本发明的目的之一在于提供一种深度神经网络超参数优化方法,其能够解决现有技术中对于超参数优化时计算成本高、运行时间长等问题。In order to overcome the shortcomings of the prior art, one of the objectives of the present invention is to provide a deep neural network hyperparameter optimization method, which can solve the problems of high calculation cost and long running time for hyperparameter optimization in the prior art.
本发明的目的之二在于提供一种电子设备,其能够解决现有技术中对于超参数优化时计算成本高、运行时间长等问题。The second object of the present invention is to provide an electronic device that can solve the problems of high calculation cost and long running time when optimizing hyperparameters in the prior art.
本发明的目的之三在于提供一种计算机存储介质,其能够解决现有技术中对于超参数优化时计算成本高、运行时间长等问题。The third object of the present invention is to provide a computer storage medium, which can solve the problems of high calculation cost and long running time when optimizing hyperparameters in the prior art.
本发明的目的之一采用如下技术方案实现:One of the objectives of the present invention is achieved by the following technical solutions:
一种深度神经网络超参数优化方法,包括以下步骤:A deep neural network hyperparameter optimization method, including the following steps:
参数设定步骤:设定训练的任务个数、每个任务的目标函数、每个任务的初始值个数、最大迭代次数;Parameter setting steps: set the number of training tasks, the objective function of each task, the number of initial values of each task, and the maximum number of iterations;
模型训练步骤:根据每个任务的初始值个数以及目标函数得出每个任务的观测集,并根据每个任务的观测集进行网络模型训练得出多任务条件神经网络模型;Model training steps: Obtain the observation set of each task according to the number of initial values of each task and the objective function, and train the network model according to the observation set of each task to obtain a multi-task conditional neural network model;
预测步骤:根据多任务条件神经网络模型对未知区域中的任意一点进行预测得出每个点对应每个任务的目标函数值;Prediction step: predict any point in the unknown area according to the multi-task conditional neural network model to obtain the objective function value of each point corresponding to each task;
筛选步骤:根据粒子群算法以及未知区域的所有点对每个任务的目标函数值进行筛选,得出每个任务的候选点;Screening step: Screen the objective function value of each task according to the particle swarm algorithm and all points in the unknown area to obtain candidate points for each task;
迭代步骤:将每个任务的候选点代入到多任务条件神经网络模型中进行真实值计算,并将每个任务的候选点以及计算出的真实值加入到每个任务的观测集中并形成每个任务的新的观测集,然后根据新的观测集依次执行模型训练步骤、预测步骤、筛选步骤以及迭代步骤;直到达到最大迭代次数,从最后一次 迭代所形成的观测集中选择响应值最大时所对应的参数组合作为超参数组合。Iterative steps: Substitute the candidate points of each task into the multi-task conditional neural network model to calculate the true value, and add the candidate points of each task and the calculated true value to the observation set of each task to form each The new observation set of the task, and then according to the new observation set, the model training step, the prediction step, the screening step and the iteration step are executed in sequence; until the maximum number of iterations is reached, the observation set formed by the last iteration is selected from the observation set corresponding to the largest response value The parameter combination of as a hyperparameter combination.
进一步地,所述多任务神经条件网络模型是通过多任务学习将多个任务在条件神经网络模型进行训练过程中进行相互学习,进而形成多任务条件神经网络模型。Further, the multi-task neural conditional network model learns from each other during the training of the conditional neural network model through multi-task learning, and then forms the multi-task conditional neural network model.
进一步地,多任务条件神经网络模型是将多个任务的条件神经网络模型的输出层用相似性网络层组合起来,形成多任务条件神经网络模型;相似性网络层是由全连接网络构成。Furthermore, the multi-task conditional neural network model combines the output layers of the conditional neural network model of multiple tasks with a similarity network layer to form a multi-task conditional neural network model; the similarity network layer is composed of a fully connected network.
进一步地,所述模型训练步骤具体为:Further, the model training step specifically includes:
步骤S11:设置模型训练迭代次数、算法最大迭代次数以及最小化损失函数;最小化损失函数为所有任务的最小化损失函数的总和;每个任务的最小化损失函数为:最小化负条件对数概率;Step S11: Set the number of model training iterations, the maximum number of algorithm iterations, and the minimized loss function; the minimized loss function is the sum of the minimized loss functions of all tasks; the minimized loss function of each task is: minimize the negative conditional logarithm Probability
步骤S12:按照预设比例分别从每个任务的观测集中随机选择出对应的初始值,并生成对应任务的解码器训练集,以及将每个任务的观测集作为对应任务的编码器训练集;Step S12: randomly select the corresponding initial value from the observation set of each task according to the preset ratio, generate the decoder training set of the corresponding task, and use the observation set of each task as the encoder training set of the corresponding task;
步骤S13:根据每个任务的解码器训练集、编码器训练集进行网络模型训练,同时根据最小化损失函数更新每个任务的条件神经网络模型的网络参数和相似性网络层的参数;Step S13: Perform network model training according to the decoder training set and encoder training set of each task, and update the network parameters of the conditional neural network model of each task and the parameters of the similarity network layer according to the minimized loss function;
步骤S14:当网络模型训练达到模型训练迭代次数后,返回依次执行步骤S12~步骤S14;直到达到算法最大迭代次数,得到多任务条件神经网络模型。Step S14: When the network model training reaches the number of model training iterations, return to perform step S12 to step S14 in sequence; until the maximum number of algorithm iterations is reached, a multi-task conditional neural network model is obtained.
进一步地,所述步骤S13还包括:根据最小化损失函数并利用Adam优化器反向传播算法更新每个任务的条件神经网络模型的网络参数和相似性网络层的参数。Further, the step S13 further includes: updating the network parameters of the conditional neural network model and the parameters of the similarity network layer of each task according to the minimized loss function and using the Adam optimizer back propagation algorithm.
进一步地,所述观测集的生成过程如下:首先根据每个任务的目标函数以 及初始值个数生成每个任务的目标函数中得出对应的目标函数值,根据每个任务的初始值以及目标函数值得出每个任务的观测值。Further, the generation process of the observation set is as follows: First, the objective function of each task is generated according to the objective function of each task and the number of initial values, and the corresponding objective function value is obtained according to the initial value and objective of each task. The function value is the observation value of each task.
进一步地,所述筛选步骤为:根据未知区域每个点对每个任务的目标函数值计算得出每个任务的采集函数EI,并将每个任务的采集函数EI作为粒子群算法的适应度函数,然后根据粒子群算法选取采集函数EI的函数值最大的点作为每个任务的候选点。Further, the screening step is: calculating the collection function EI of each task according to the value of the objective function for each task at each point in the unknown area, and using the collection function EI of each task as the fitness of the particle swarm algorithm Function, and then select the point with the largest function value of the collection function EI according to the particle swarm algorithm as the candidate point for each task.
本发明的目的之二采用如下技术方案实现:The second objective of the present invention is achieved by adopting the following technical solutions:
一种电子设备,包括存储器、处理器以及存储在存储器上并可在处理器上运行的计算机程序,所述处理器执行所述程序时实现如本发明目的之一所采用的一种深度神经网络超参数优化方法的步骤。An electronic device, including a memory, a processor, and a computer program stored on the memory and running on the processor. The processor implements a deep neural network as one of the objectives of the present invention when the processor executes the program Steps of hyperparameter optimization method.
本发明的目的之三采用如下技术方案实现:The third objective of the present invention is achieved by the following technical solutions:
一种计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现如本发明目的之一所采用的一种深度神经网络超参数优化方法的步骤。A computer-readable storage medium has a computer program stored thereon, and when the computer program is executed by a processor, the steps of a deep neural network hyperparameter optimization method adopted as one of the objectives of the present invention are realized.
相比现有技术,本发明的有益效果在于:Compared with the prior art, the present invention has the following beneficial effects:
本发明通过使用条件神经网络模型来代替高斯模型实现模型的训练,并将其与多任务学习结合一起组成多任务条件神经网络模型,将多任务条件神经网络模型应用到贝叶斯算法框架中,进而实现多任务的超参数优化,使得超参数优化的处理过程中,降低了计算成本和运行时间成本。The present invention implements the training of the model by using the conditional neural network model instead of the Gaussian model, and combines it with multi-task learning to form a multi-task conditional neural network model, and applies the multi-task conditional neural network model to the Bayesian algorithm framework, In turn, multi-task hyperparameter optimization is realized, so that the calculation cost and running time cost are reduced in the process of hyperparameter optimization.
图1为本发明提供的一种深度神经网络超参数优化方法的流程框图;FIG. 1 is a flowchart of a method for optimizing hyperparameters of a deep neural network provided by the present invention;
图2为本发明提供的条件神经网络模型网络结构图;Figure 2 is a network structure diagram of the conditional neural network model provided by the present invention;
图3为本发明提供的多任务条件神经网络模型网络结构图;Figure 3 is a network structure diagram of a multi-task conditional neural network model provided by the present invention;
图4为本发明提供的多任务条件神经网络模型网络训练图;Figure 4 is a network training diagram of a multi-task conditional neural network model provided by the present invention;
图5为本发明提供的多任务条件神经网络模型中预测点x *的预测示意图; FIG. 5 is a schematic diagram of the prediction of the prediction point x * in the multi-task conditional neural network model provided by the present invention;
图6为本发明提供的将多任务条件神经网络模型应用于贝叶斯决策算法中的超参数优化方法流程图;Fig. 6 is a flowchart of a hyperparameter optimization method for applying a multi-task conditional neural network model to a Bayesian decision algorithm provided by the present invention;
图7为本发明提供的扩充数据集示意图。Fig. 7 is a schematic diagram of the extended data set provided by the present invention.
下面,结合附图以及具体实施方式,对本发明做进一步描述,需要说明的是,在不相冲突的前提下,以下描述的各实施例之间或各技术特征之间可以任意组合形成新的实施例。Hereinafter, the present invention will be further described with reference to the drawings and specific implementations. It should be noted that, provided that there is no conflict, the following embodiments or technical features can be combined to form new embodiments. .
实施例一:Example one:
本发明提出了一种利用条件神经网络过程(Conditional Neural Processes,CNPs)来代替高斯过程(Gaussian Process,GPs),而CNPs结合了随机过程和神经网络的特点,受到高斯过程灵活性的启发并利用梯度下降进行神经网络训练,进而实现超参数优化方法。The present invention proposes a method that uses Conditional Neural Processes (CNPs) to replace Gaussian Processes (GPs), and CNPs combine the characteristics of random processes and neural networks, and are inspired and used by the flexibility of Gaussian processes Gradient descent is used for neural network training to realize the hyperparameter optimization method.
条件神经网络过程对已知观测数据的学习时由神经网络参数化通过对数据集进行随机采样并遵循梯度步骤来训练形成条件神经网络模型,以在给定随机观察集的情况下最大化随机子集的条件似然。The conditional neural network process is parameterized by the neural network when learning the known observation data. The conditional neural network model is trained by randomly sampling the data set and following the gradient steps to maximize the random number given a random observation set. Set the conditional likelihood.
现有技术中的超参数优化的一般方法都是采用高斯过程进行模型训练,而本专利利用了学习能力更强的条件神经网络模型来代替传统的高斯过程,避免了复杂的协方差函数的计算。通过结合当前多任务学习的优点,在条件神经网络模型的基础上提出了多任务条件神经网络过程(Multi-task Conditional Neural Processes in One-to-many case,OMc-MTCNPs),并将其应用到贝叶斯决策算法(Bayesian)中,实现对超参数进行优化。The general method of hyperparameter optimization in the prior art is to use the Gaussian process for model training, and this patent uses a conditional neural network model with stronger learning ability to replace the traditional Gaussian process, avoiding the calculation of complex covariance functions . By combining the advantages of current multi-task learning, a multi-task conditional neural network process (Multi-task Conditional Neural Processes in One-to-many case, OMc-MTCNPs) is proposed on the basis of the conditional neural network model and applied to In the Bayesian decision algorithm (Bayesian), the hyperparameters are optimized.
在多任务条件神经网络模型中,通过多任务条件神经网络模型可以充分利用已知的少量数据来增强每个单任务的泛化性能。将多任务条件神经网络模型应用在贝叶斯决策算法的框架中,可以对超参数优化过程中寻找到更好的近似最优解,解决现有超参数优化处理的运行时间长、计算成本高等问题。In the multi-task conditional neural network model, the multi-task conditional neural network model can make full use of the known small amount of data to enhance the generalization performance of each single task. Applying the multi-task conditional neural network model to the framework of the Bayesian decision algorithm can find a better approximate optimal solution in the hyperparameter optimization process, and solve the long running time and high computational cost of the existing hyperparameter optimization processing. problem.
例如,如图1所示,针对M个任务(也即是每个任务为一个深度网络模型,每个模型均有一个深度网络超参数优化训练集)。其中,每个任务都具有多个超参数,多任务条件神经网络模型可以根据每个任务各自的训练集以及训练结果迭代地找出每个任务对应的较好的超参数组合。当找到最优超参数组合时,就可以在该最优超参数存在的条件下,对未知点进行准确地预测。For example, as shown in Figure 1, for M tasks (that is, each task is a deep network model, and each model has a deep network hyperparameter optimization training set). Among them, each task has multiple hyperparameters, and the multi-task conditional neural network model can iteratively find a better hyperparameter combination corresponding to each task according to the respective training set and training results of each task. When the optimal hyperparameter combination is found, the unknown point can be accurately predicted under the condition that the optimal hyperparameter exists.
为了更好地说明多任务条件神经网络模型如何对超参数进行优化的,本发明首先介绍下条件神经网络模型的模型训练过程以及多任务条件神经网络模型的模型训练过程。In order to better explain how the multi-task conditional neural network model optimizes hyperparameters, the present invention first introduces the model training process of the conditional neural network model and the model training process of the multi-task conditional neural network model.
一、针对条件神经网络模型(CNPs模型):1. For conditional neural network model (CNPs model):
1、条件神经网络模型是通过两个阶段的学习数据以缓解少量数据无法有效训练的问题:第一阶段是学习训练数据的统计分布信息,第二阶段只利用部分训练数据和第一阶段学习到的分布信息共同拟合特定函数。1. The conditional neural network model uses two stages of learning data to alleviate the problem that a small amount of data cannot be effectively trained: the first stage is to learn the statistical distribution information of the training data, the second stage only uses part of the training data and the first stage to learn The distribution information of the joint fits a specific function.
例如假设我们有n个不同输入的数据集 有一个未知表达式的函数f:X→Y,将X输入到函数f中得到输出 其中y i=f(x i)。 For example, suppose we have n data sets with different inputs There is a function f:X→Y with an unknown expression, input X into the function f to get the output Where y i =f(x i ).
其中,第一阶段:如图2所示为条件神经网络过程结构图,条件神经网络由两部分组成,分别为编码器h和解码器g,它们都是由神经网络组成。其中,编码器h主要是学习数据之间的映射关系,即参数化条件概率;解码器g是利用学习到的信息进行预测回归,即计算条件概率。Among them, the first stage: Figure 2 shows the process structure diagram of the conditional neural network. The conditional neural network is composed of two parts, namely the encoder h and the decoder g, both of which are composed of neural networks. Among them, the encoder h mainly learns the mapping relationship between the data, that is, parameterized conditional probability; the decoder g uses the learned information to perform predictive regression, that is, calculates the conditional probability.
其中,第二阶段:拟合函数p:X→Y,解码器g最终就是使f和p尽可能的 接近。Among them, the second stage: fitting function p: X→Y, the decoder g finally makes f and p as close as possible.
条件神经网络模型的一个好处是不需要设定高斯先验,而是直接通过神经网络从数据中获取分布信息。假如我们有m个未观测数据的集合 CNPs在观测值 上参数化f(T|O,T),利用这种方式放弃了随机过程的数学计算,但增加了灵活性和扩展性。 One advantage of the conditional neural network model is that there is no need to set a Gaussian prior, but to obtain distribution information from the data directly through the neural network. If we have a set of m unobserved data CNPs in observations The above parameterization f(T|O,T), in this way, the mathematical calculation of the random process is abandoned, but the flexibility and scalability are increased.
2、条件神经网络模型的数学分析2. Mathematical analysis of conditional neural network model
设定P θ为一个在随机变量f(x),x∈T上的分布,其中θ是定义P θ的所有参数集合。则按一定循序排列已知观测集O和未知观测集T得到O'和T',也即是: Let P θ be a distribution on the random variable f(x),x∈T, where θ is the set of all parameters that define P θ . Then arrange the known observation set O and the unknown observation set T in a certain order to obtain O'and T', that is:
P θ(f(T)|O,T)=P θ(f(T)|O,T')=P θ(f(T)|O',T)。其中,T={x1,x2,x3,…,xn}。因此,P θ(f(T)|O ,T)=∏ x∈TP θ(f(T)|O ,x)(1),也即是将T中的每个x组合连乘变换,也即是可以将P θ分解并对每个输入x进行单独条件分布计算。 P θ (f(T)|O,T)=P θ (f(T)|O,T')=P θ (f(T)|O',T). Among them, T={x1,x2,x3,...,xn}. Therefore, P θ (f(T)|O , T)=∏ x∈T P θ (f(T)|O , x)(1), that is, to transform each combination of x in T, That is, it is possible to decompose P θ and perform a separate conditional distribution calculation for each input x.
如图2所示,其表示了条件神经网络模型的结构,其中r i=h θ(x i,y i) As shown in Figure 2, it represents the structure of the conditional neural network model, where r i =h θ (x i ,y i )
其中,h θ:X×Y→R d和g θ:X×R d→R e都是神经网络,公式(2)中的编码器h将已知数据信息提炼出来得到r,再利用公式(3)进行整合并作为编码器g预测每个x的决策因素之一。公式(4)为编码器g作出预测在数学上的表示。其中,符号 为整合所有学习到的信息并映射到单个x上。在大部分实验中 用来表示加权平均,即式子(2)可以等价于r=(r 1+r 2+…+r n-1+r n)/n。 Among them, h θ : X×Y→R d and g θ : X×R d →R e are neural networks. The encoder h in formula (2) extracts the known data information to obtain r, and then uses the formula ( 3) It is integrated and used as one of the decision factors for the encoder g to predict each x. Formula (4) is the mathematical representation of the prediction made by the encoder g. Among them, the symbol To integrate all learned information and map it to a single x. In most experiments Used to express the weighted average, that is, formula (2) can be equivalent to r=(r 1 +r 2 +...+r n-1 +r n )/n.
3、模型参数训练与预测3. Model parameter training and prediction
首先选取一数据集O,并且根据预设比例从该数据集O中随机选取一子集作为模型训练的编码器h,而将数据集O作为模型训练的解码器g。也即是通过数据集O中X的目标预测值与真实目标值的误差来更新训练P θ中的参数,进而得出的参数,能够使得预测的结果更准确。在选取子集时,一般来说,通常设置 N~uniform[1,...,n],也即是,选取数据集O的前N个元素得到子集 将其作为模型训练的编码器h。 First select a data set O, and randomly select a subset from the data set O according to a preset ratio as the encoder h for model training, and use the data set O as the decoder g for model training. That is, the parameters in the training P θ are updated by the error between the target predicted value of X in the data set O and the real target value, and the obtained parameters can make the predicted result more accurate. When selecting a subset, generally speaking, N~uniform[1,...,n] is usually set, that is, the first N elements of the data set O are selected to obtain the subset Use it as the encoder h for model training.
因此,基于数据集进行条件神经网络模型的训练,其步骤如下:Therefore, to train the conditional neural network model based on the data set, the steps are as follows:
步骤A1:首先设置学习率、模型训练迭代次数、算法最大迭代次数等参数。在一个模型训练之前,这些超参数可根据历史经验给出,便于开始模型训练。其中,学习率也是超参数,其主要是对神经网络模型进行设定,如何进行学习的,通过对该超参数的优化,将其设置给神经网络模型来实现模型训练。Step A1: First set the learning rate, the number of model training iterations, and the maximum number of algorithm iterations. Before a model is trained, these hyperparameters can be given based on historical experience to facilitate the start of model training. Among them, the learning rate is also a hyperparameter, which is mainly used to set the neural network model, how to learn, through the optimization of the hyperparameter, set it to the neural network model to achieve model training.
步骤A2:按照预设比例从数据集O中随机选取子集O N。一般来说,N~uniform[1,...,n]。也即是,从数据集O中选取前N个数据作为子集O N。 Step A2: according to a preset ratio O randomly selected subset of the data set from O, N. Generally speaking, N~uniform[1,...,n]. That is, the first N data are selected from the data set O as the subset ON .
步骤A3:将数据集O作为解码器g的训练集,子集O N作为编码器h的训练集。 Step A3: Use the data set O as the training set of the decoder g, and the subset ON as the training set of the encoder h.
步骤A4:将最小化负条件对数概率作为最小化损失函数,也即是: 其中, 另外,该最小化损失函数是通过最小化附条件对数概率得到的,其计算公式是常规公式,本发明不对计算公式进行改进。 Step A4: Use the minimized negative conditional logarithmic probability as the minimized loss function, that is: among them, In addition, the minimized loss function is obtained by minimizing the conditional logarithmic probability, and its calculation formula is a conventional formula, and the present invention does not improve the calculation formula.
步骤A5:根据该最小化损失函数利用Adam优化器更新算法更新多条件网络模型的参数;当每次模型训练迭代次数后返回步骤A2~步骤A5,继续进行子集的选取、模型的训练以及参数的更新,以此类推。Step A5: Use the Adam optimizer update algorithm to update the parameters of the multi-condition network model according to the minimized loss function; return to step A2 to step A5 after each model training iteration number, continue to select subsets, model training and parameters Update, and so on.
步骤A6:当达到算法最大迭代次数时,训练结束,进而得到任务的条件神经网络模型。Step A6: When the maximum number of iterations of the algorithm is reached, the training ends, and the conditional neural network model of the task is obtained.
模型训练完成后,向条件神经网络模型的解码器g中输入未知点x *,通过该 条件神经网络模型就可以对该未知点x *进行预测,得出其目标函数值,比如得出目标函数值的预测均值和方差。 After the model training is completed, input the unknown point x * into the decoder g of the conditional neural network model, and the unknown point x * can be predicted by the conditional neural network model to obtain the objective function value, such as the objective function The predicted mean and variance of the value.
二、多任务条件神经网络模型:2. Multi-task conditional neural network model:
高斯过程适合样本规模较小的场合,但是存在协方差函数计算复杂度性高的不足;而上述条件神经网络模型也只能对单一任务进行模型训练。而对于实现多任务的模型训练时,本发明将条件神经网络模型进行了扩展,将其扩展为多任务条件神经网络模型(Multi-task Conditional Neural Processes in One-to-many case,OMc-MTCNPs),通过多任务学习,不仅可以减轻数据不足的问题,还可以通过任务间相互学习提升单个任务的学习能力,也避免了复杂的协方差计算过程。The Gaussian process is suitable for occasions with small sample size, but it has the disadvantage of high computational complexity of the covariance function; and the above-mentioned conditional neural network model can only perform model training on a single task. For multi-task model training, the present invention expands the conditional neural network model and expands it into a multi-task conditional neural network model (Multi-task Conditional Neural Processes in One-to-many case, OMC-MTCNPs) Through multi-task learning, not only can the problem of insufficient data be alleviated, but also the learning ability of a single task can be improved through mutual learning between tasks, and the complicated covariance calculation process can also be avoided.
也即是说,本发明将多任务学习和条件神经网络模型的较强的学习预测能力结合在一起,提出了多任务条件神经网络模型,并将多任务条件神经网络模型应用到贝叶斯决策算法的优化框架来对超参数进行优化处理,进而得出最优的超参数。条件神经网络模型和多任务条件神经网络模型都具有很强的拓展性,对于一些简单低维数据,使用全连接层就能达到较好的性能,而对于一些复杂高维数据问题,我们可以使用卷积层来高效抽取数据信息。In other words, the present invention combines multi-task learning with the strong learning and predictive capabilities of the conditional neural network model, proposes a multi-task conditional neural network model, and applies the multi-task conditional neural network model to Bayesian decision making The optimization framework of the algorithm is used to optimize the hyperparameters, and then obtain the optimal hyperparameters. Both the conditional neural network model and the multi-task conditional neural network model have strong scalability. For some simple low-dimensional data, a fully connected layer can achieve better performance, and for some complex high-dimensional data problems, we can use Convolutional layer to efficiently extract data information.
1、首先本发明给出了多任务条件神经网络模型的网络结构,如图3所示。1. First, the present invention provides the network structure of the multi-task conditional neural network model, as shown in Fig. 3.
假设有M个任务, 代表训练集。其中,每个任务同时采样相同的集合X,Y l代表点集合X在第l个任务上的目标函数值的集合。 Suppose there are M tasks, Represents the training set. Among them, each task samples the same set X at the same time, and Y l represents the set of objective function values of the point set X on the l-th task.
也是说,多任务条件神经网络模型是将多个任务的条件神经网络模型组合在一起,并将每个条件神经网络模型的输出层用相似性网络层k组合起来,进而形成该多任务条件神经网络模型。In other words, the multi-task conditional neural network model combines the conditional neural network models of multiple tasks, and combines the output layer of each conditional neural network model with the similarity network layer k to form the multi-task conditional neural network. Network model.
而,对于相似性网络层k是由全连接网络构成,它主要通过更新参数学习每 个任务之间的相似性度量,并结合相似性信息对每个任务预测结果进行一定的线性组合得出最终的结果。也即是说,通过相似性网络层k将每个条件神经网络模型输出的均值矩阵 和方差矩阵 用神经网络连接起来,则整个模型输出的均值矩阵m和方差矩阵v分别由 和 表示。w m和b m分别表示均值连接项的权值矩阵和偏置矩阵,w v和b v分别表示方差连接项的权值矩阵和偏置矩阵。也就是说,每个任务最终输出经过相似性网络层k后都与其他任务相关。 However, for the similarity network layer k is composed of a fully connected network, it mainly learns the similarity measure between each task by updating parameters, and combines the similarity information to perform a certain linear combination of the prediction results of each task to obtain the final the result of. In other words, the mean matrix of each conditional neural network model output by the similarity network layer k Sum variance matrix Connected by a neural network, the mean matrix m and variance matrix v output by the entire model are respectively determined by with Said. w m and b m respectively represent the weight matrix and the bias matrix of the mean connection term, and w v and b v represent the weight matrix and the bias matrix of the variance connection term, respectively. In other words, the final output of each task is related to other tasks after passing through the similarity network layer k.
多任务条件神经网络模型利用从类似相关任务中学到的经验,通过适当的损失函数将一个任务的信息有效地传播到其他任务的网络模型中,实现任务之间的信息共享。由于多任务模型的特点,观测数据的规模增大M倍,而且每个任务都可以利用其他任务共享的信息,避免了任务只从本身数据从零开始学习,也可以通过更新网络将这些经验从处理类似问题的过程中整合起来。图3中只画出添加一个全连接层来表示任务的相关性,但是如果处理复杂的问题,相似性网络层k中网络结构或层数可以根据任务的不同灵活地改变以满足需求。当任务之间关系很难表达时,可通过添加更多的隐藏层,使相似性网络层k学习到所有数据的数学分布特征,这些特征被表征在隐藏节点中,当网络前向传播时可以激活这些节点以促使任务相互学习。The multi-task conditional neural network model uses the experience learned from similar related tasks to effectively spread the information of one task to the network model of other tasks through an appropriate loss function, and realize the information sharing between tasks. Due to the characteristics of the multi-task model, the scale of observation data is increased by M times, and each task can use the information shared by other tasks, avoiding the task to learn from scratch from its own data, and it can also update these experiences from the network. Integration in the process of dealing with similar problems. Figure 3 only draws the addition of a fully connected layer to represent the relevance of tasks, but if you deal with complex problems, the network structure or the number of layers in the similarity network layer k can be flexibly changed according to different tasks to meet requirements. When the relationship between tasks is difficult to express, by adding more hidden layers, the similarity network layer k can learn the mathematical distribution characteristics of all data. These characteristics are represented in hidden nodes. When the network is forwarded, it can be Activate these nodes to encourage tasks to learn from each other.
2、多任务条件神经网络模型的数学分析2. Mathematical analysis of multi-task conditional neural network model
假设我们测试M个任务,对于多任务条件神经网络模型中第l任务,有n个不同输入的观测数据 和未观测数据Tl。 Suppose we test M tasks, for the lth task in the multi-task conditional neural network model, there are n different input observation data And unobserved data Tl.
设Q lθ为函数f l:X l←Y l上的概率分布,即 f l表示第l任务的目标函数,O l={X l,Y l=f l(X l)}为第l个任务的观测集合,L l为对应第l个任务的损失函数,θ为所有定义Q的参数向量。 Let Q lθ be the probability distribution on the function f l :X l ←Y l , namely f l represents the objective function of the lth task, O l = {X l , Y l = f l (X l )} is the observation set of the lth task, L l is the loss function corresponding to the lth task, and θ is All parameter vectors that define Q.
在多任务条件神经网络模型的每个任务的条件神经网络模型中,数学结构与前述中的公式(2)、(3)和(4)相似。In the conditional neural network model of each task of the multi-task conditional neural network model, the mathematical structure is similar to the aforementioned formulas (2), (3) and (4).
对于第l任务相关性网络层k使用 结构, 代表 的参数,其含义与单任务的条件神经网络模型相似, 即为第l个任务的输出经过相似性网络层处理后的最终结果。 For the first task dependency network layer k use structure, representative The parameters of, whose meaning is similar to that of a single-task conditional neural network model, It is the final result after the output of the lth task is processed by the similarity network layer.
3、多任务条件神经网络模型的训练过程如下:3. The training process of the multi-task conditional neural network model is as follows:
在多任务条件神经网络模型中,每一个任务的输入观测数据都是相同的,即设置为X=X
1=X
2=…=X
M。由于先验知识是指保证输入每个任务的X保持一致,这样得出的相互结果才有关联性,否则会产生噪声,比如x1输入任务1,不同值x2输入任务2,得出的结果并没有一定的联系。因此,本申请中将每个任务的输入观测数据设置为相同的。
In the multi-task conditional neural network model, the input observation data of each task is the same, that is, it is set as X=X 1 =X 2 =...=X M. Since a priori knowledge means to ensure that the X input to each task is consistent, the mutual results obtained in this way are related, otherwise noise will be generated, such as
在训练中,我们会适当调整从训练的数据集中选取的编码器的训练数据比例σ。因此,多任务条件神经网络模型在充分学习观测数据的前提下防止过拟合,同时还能保证模型在学习决策空间X上的不确定性。在保证随机性的前提下,按照设定的数据比例多次随机从数据集中选取出子集,作为编码器的训练数据集,可使得模型的泛化效果更好。During training, we will appropriately adjust the training data ratio σ of the encoder selected from the training data set. Therefore, the multi-task conditional neural network model prevents overfitting under the premise of fully learning the observation data, and at the same time guarantees the uncertainty of the model in the learning decision space X. Under the premise of ensuring randomness, a subset of the data set is randomly selected multiple times according to the set data ratio as the training data set of the encoder, which can make the generalization effect of the model better.
同样的,在模型训练以及更新参数时,也需要用到最小化损失函数。Similarly, when training the model and updating the parameters, it is also necessary to minimize the loss function.
令 则最小化负条件对数概率: make Then minimize the negative conditional logarithmic probability:
其中w l为第l任务损失函数的权重。为了均衡每个任务的比重,我们将权重设置为1。 Where w l is the weight of the loss function of the lth task. In order to balance the weight of each task, we set the weight to 1.
如图4表示多任务条件神经网络模型的模型训练,在多任务条件神经网络模型进行训练时,每个任务的条件神经网络模型也会相应的训练,并且每个任务的条件神经网络模型都可以独立地学习其他任务的相关的数据集特征。Figure 4 shows the model training of the multi-task conditional neural network model. When the multi-task conditional neural network model is trained, the conditional neural network model of each task will be trained accordingly, and the conditional neural network model of each task can be Independently learn related data set features for other tasks.
也即是,首先对每个任务的条件神经网络模型进行训练,然后将各个任务组合成一个整体进行训练,同时更新每个任务的条件神经网络模型的参数,并且在最小化损失函数下反向更新每个任务的条件神经网络模型的网络参数,也即是将不同任务的信息分布到各个任务的条件神经网络模型的参数中。与此同时,也需要对相似性网络层k中的参数进行更新,并且每个参数趋向于表现出任务之间的相关性。That is, first train the conditional neural network model of each task, and then combine the tasks into a whole for training, and update the parameters of the conditional neural network model of each task at the same time, and reverse the loss function under the minimized loss function. To update the network parameters of the conditional neural network model of each task, that is, to distribute the information of different tasks to the parameters of the conditional neural network model of each task. At the same time, it is also necessary to update the parameters in the similarity network layer k, and each parameter tends to show the correlation between tasks.
多任务条件神经网络模型的训练具体步骤如下:The specific steps of multi-task conditional neural network model training are as follows:
步骤B1:设置模型训练迭代次数、算法最大迭代次数以及最小化损失函数等网络参数。该最小化损失函数为每个任务的损失函数的总和。Step B1: Set network parameters such as the number of model training iterations, the maximum number of algorithm iterations, and the minimized loss function. The minimized loss function is the sum of the loss functions of each task.
步骤B2:根据预设比例为每个任务选取对应的解码器训练集O和编码器训练集O σn进行模型训练。 Step B2: Select the corresponding decoder training set O and encoder training set O σn for each task according to the preset ratio for model training.
步骤B3:根据最小化损失函数以及Adam优化器反向传播更新算法更新每个任务的条件神经网络模型的参数和相似性网络层k的参数,即图4中虚线及虚线框相关参数。Step B3: Update the parameters of the conditional neural network model of each task and the parameters of the similarity network layer k according to the minimized loss function and the Adam optimizer back propagation update algorithm, that is, the dashed line and the dashed box related parameters in FIG. 4.
步骤B4:当模型训练达到模型训练迭代次数后,返回步骤B2~步骤B4,依次重新随机选取编码器训练集 进行模型训练以及参数更新等,以此类推。 Step B4: When the model training reaches the number of model training iterations, return to step B2 to step B4, and randomly select the encoder training set again in turn Perform model training and parameter update, and so on.
步骤B5:当达到算法最大迭代次数后,网络训练完毕,得到了多任务条件神经网络模型。Step B5: When the maximum number of iterations of the algorithm is reached, the network training is completed, and the multi-task conditional neural network model is obtained.
多任务条件神经网络模型训练完成后,即可对任意未知点x *进行预测,可以预测未知点在各个任务上的目标函数值,比如预测未知点在各个任务上的均值和方差。 After the multi-task conditional neural network model is trained, it can predict any unknown point x *, and predict the objective function value of the unknown point on each task, such as predicting the mean and variance of the unknown point on each task.
如图5,每个任务的条件神经网络模型的解码器输入未知点x *,在编码器训练集上仍然输入对应的训练集,这一部分是读取训练集的信息,并对后面的预 测做出帮助。由于通过相似性网络层k将任务之间的信息进行分享,应每个任务的条件神经网络模型输出的预测值要明显好于单个任务所预测的结果。 As shown in Figure 5, the decoder of the conditional neural network model of each task inputs the unknown point x * , and the corresponding training set is still input on the encoder training set. This part is to read the training set information and make subsequent predictions Help out. Since the information between tasks is shared through the similarity network layer k, the predicted value output by the conditional neural network model of each task is significantly better than the result predicted by a single task.
因此,为了实现超参数的优化,本发明提供了一种基于多任务条件神经网络模型的超参数优化方法,将多任务条件神经网络模型运用到贝叶斯决策算法的框架中,来实现超参数优化。如图6所示,根据前述可知,多任务条件神经网络模型具有优良的预测性能,本发明中将其作为代理模型应用到贝叶斯决策算法的框架中,实现对超参数的优化 。 Therefore, in order to realize the optimization of hyperparameters, the present invention provides a hyperparameter optimization method based on a multi-task conditional neural network model. The multi-task conditional neural network model is applied to the framework of the Bayesian decision algorithm to realize the hyperparameter optimization. As shown in FIG. 6, according to the foregoing, the multi-task conditional neural network model has excellent predictive performance. In the present invention, it is applied as a proxy model to the framework of the Bayesian decision algorithm to optimize the hyperparameters .
该方法具体包括以下步骤:The method specifically includes the following steps:
步骤S1:网络参数的设置。比如确定参数的维度d、每个任务的初始数据的个数n(通常n=11×d-1)、任务个数M、最大迭代次数T、以及第l个任务目标函数f l等等参数。 Step S1: Setting of network parameters. For example, determine the dimension d of the parameters, the number n of initial data for each task (usually n=11×d-1), the number of tasks M, the maximum number of iterations T, and the objective function f l of the lth task. .
步骤S2:根据参数的维度以及每个任务的初始数据的个数得出每个任务的初始值集合。比如初始值集合可表示为 Step S2: Obtain the initial value set of each task according to the dimension of the parameters and the number of initial data of each task. For example, the initial value set can be expressed as
例如:参数的维度为2,也即是两维的参数[x1,x2],其中x1的范围[0,6],x2的范围[6,10]。假设要产生4个初始点,即x1取2,4;x2取8,9。最后得出初始点[2,8],[2,9],[4,8],[4,9]。也即是说,在决策空间中均匀划分,然后在每个小部分中找点,最终得出初始点的数据,生成初始点集合。For example: the dimension of the parameter is 2, that is, the two-dimensional parameter [x1,x2], where the range of x1 is [0,6], and the range of x2 is [6,10]. Suppose that 4 initial points are to be generated, that is, x1 takes 2,4; x2 takes 8,9. Finally, the initial points [2,8],[2,9],[4,8],[4,9] are obtained. That is to say, divide evenly in the decision space, and then find points in each small part, and finally get the data of the initial point, and generate the initial point set.
步骤S3:将每个任务的初始值集合中的每个初始值输入到对应任务的目标函数中,计算得出对应任务的目标函数值,进而根据每个任务的初始值以及对应的目标函数值形成每个任务的观测集。也即是,目标函数值为 观测集为 Step S3: Input each initial value in the initial value set of each task into the objective function of the corresponding task, calculate the objective function value of the corresponding task, and then according to the initial value of each task and the corresponding objective function value Form the observation set for each task. That is, the objective function value is The observation set is
步骤S4:根据预设比例从每个任务的观测集中选取一对应任务的子集,并将观测集作为解码器训练节,将子集作为编码器的训练集进行网络模型训练, 进而训练得出多任务条件神经网络模型。其中,子集为 Step S4: Select a subset of the corresponding task from the observation set of each task according to the preset ratio, use the observation set as the decoder training section, and use the subset as the encoder training set for network model training, and then train to obtain Multi-task conditional neural network model. Among them, the subset is
步骤S5:根据多任务条件神经网络模型预测未知区域内任意点x在每个任务的预测值,也即是将该任意点输入到每个任务的条件神经网络模型中得出对应的目标函数值。比如均值 和方差σ l=1:M(x)。 Step S5: Predict the predicted value of any point x in each task in the unknown area according to the multi-task conditional neural network model, that is, input the arbitrary point into the conditional neural network model of each task to obtain the corresponding objective function value . Such as mean The sum variance σ l=1:M (x).
这样,每个任务均对应多个未知点以及对应目标函数值,其中未知点为任务的参数组合。In this way, each task corresponds to multiple unknown points and corresponding objective function values, where the unknown points are the combination of parameters of the task.
步骤S6:采用粒子群算法对上找到每个任务的最优参数组合。最优参数组合为 由于每个任务均在多个参数组合,因此根据每个任务的所有未知点与目标函数值,以及粒子群算法对未知点进行筛选,进而筛选出每个任务的最优点,也即是最优参数组合。 Step S6: Use the particle swarm algorithm to find the optimal parameter combination for each task. The optimal parameter combination is Since each task is combined with multiple parameters, the unknown points are filtered according to all the unknown points and objective function values of each task, as well as the particle swarm algorithm, and then the best advantage of each task, that is, the optimal Parameter combination.
进一步地,对未知点进行筛选,本发明还采用采集函数EI进行筛选。由于采集函数EI具有油量的期望提升特性,因此,将采集函数EI作为粒子群算法的适应度函数。是根据未知区域每个点对每个任务额目标函数值计算得出每个任务的采集函数EI,并将每个任务的采集函数EI作为粒子群算法的适应度函数,然后根据粒子群算法选取采集函数EI的函数值最大的点作为每个任务的候选点,每个任务的候选点即为对应任务的最优参数组合。Further, to screen unknown points, the present invention also uses the collection function EI to screen. Since the collection function EI has the expected improvement characteristics of oil volume, the collection function EI is used as the fitness function of the particle swarm algorithm. It calculates the collection function EI of each task according to the value of the objective function of each task at each point in the unknown area, and uses the collection function EI of each task as the fitness function of the particle swarm algorithm, and then selects it according to the particle swarm algorithm The point with the largest function value of the collection function EI is used as the candidate point of each task, and the candidate point of each task is the optimal parameter combination of the corresponding task.
步骤S6:将每个任务的最优参数组合输入到对应任务的目标函数中进行真实值计算得出目标函数值,并将每个任务的最优参数组合以及目标函数值加入到对应任务的观测集中,并形成对应任务的新的观测集,然后根据新的观测集继续执行步骤S3~S6。比如对于第m个任务,将计算后的点以及目标函数值加入到观测集 然后再进行模型训练、预测、筛选等。 Step S6: Input the optimal parameter combination of each task into the objective function of the corresponding task to calculate the true value to obtain the objective function value, and add the optimal parameter combination and objective function value of each task to the observation of the corresponding task Concentrate and form a new observation set corresponding to the task, and then continue to perform steps S3 to S6 according to the new observation set. For example, for the mth task, add the calculated point and the objective function value to the observation set Then perform model training, prediction, and screening.
如图7所示,当找出每个任务的最优参数组合后,将每个任务的最优参数 组合进行真实评价,进而计算得出每个任务的最优参数组合对应的目标函数值。也即是说,将每个任务的最优参数组合与对应的目标函数值加入到对应任务的观测集中,再重新进行模型训练、预测、筛选以及评价等操作。这里的真实评价是指:再得到决策变量之后,将决策变量代入到测试问题中重新进行计算,也即是将最优参数组合输入到对应任务的目标函数中进行预测,得到对应的目标函数值。As shown in Figure 7, when the optimal parameter combination of each task is found, the optimal parameter combination of each task is actually evaluated, and then the objective function value corresponding to the optimal parameter combination of each task is calculated. That is to say, the optimal parameter combination of each task and the corresponding objective function value are added to the observation set of the corresponding task, and then the model training, prediction, screening, and evaluation operations are performed again. The real evaluation here refers to: after the decision variables are obtained, the decision variables are substituted into the test problem and recalculated, that is, the optimal parameter combination is input into the objective function of the corresponding task for prediction, and the corresponding objective function value is obtained .
步骤S7:以此类推,依次循环执行上述步骤,当达到算法最大迭代次数T后,从每个任务所形成的新的观测集中找到每个任务的目标函数值最大时所对应的点,也即是参数组合: 即为每个任务的最优超参数。 Step S7: By analogy, the above steps are executed in turn, when the maximum number of iterations of the algorithm T is reached, the new observation set formed by each task is to find the point corresponding to the maximum value of the objective function of each task, that is Is a combination of parameters: That is the optimal hyperparameter for each task.
根据上述优化后的每个任务的超参数,通过该多任务条件神经网络模型就可以准确地预测任意点在每个任务上的目标函数值,比如均值和方差。According to the above optimized hyperparameters of each task, the multi-task conditional neural network model can accurately predict the objective function value of any point on each task, such as the mean value and the variance.
为了进一步说明上述超参数优化方法,本发明还给出了相关的实验测试,进而说明了本发明相对于现有技术的超参数优化的有效性、准确性,其具体数据参考如下:In order to further illustrate the above hyperparameter optimization method, the present invention also provides related experimental tests, which further illustrate the effectiveness and accuracy of the hyperparameter optimization of the present invention relative to the prior art. The specific data are referred to as follows:
在MNIST数据集错误!未找到引用源。和Fashion-MNIST数据集上测试Lenet-5网络多分类性能。Error in the MNIST data set! The reference source was not found. Test the multi-classification performance of the Lenet-5 network on the Fashion-MNIST dataset.
该实验设置了2种超参数问题:第一种设置学习率与偏置两个可变化超参数,其他超参数设置为默认定值;第二种是在第一种的基础上加上dropout值。在表格中使用Q1和Q2分别表示这两种问题,并展示三个任务在测试集上的准确率均值和标准差(括号中数据)。This experiment sets up two kinds of hyper-parameter problems: the first one sets two variable hyper-parameters of learning rate and bias, and the other hyper-parameters are set to default fixed values; the second one is to add the dropout value on the basis of the first one . In the table, use Q1 and Q2 to represent these two problems, and show the mean and standard deviation of the accuracy of the three tasks on the test set (data in parentheses).
本次实验通过采用条件神经网络模型和多任务条件神经网络模型这两种代理模型在贝叶斯决策算法框架中对于超参数优化的实验。也即是,寻找到一个 能够使网络预测准确率高的超参数组合,进而可根据实验结果来验证本文给出的多任务条件神经网络模型对于超参数优化问题的有效性、准确性。另外,由于神经网络每次运行可能会得到不同的结果,为了消除误差所有准确率数值都是十次运行结果的平均值。This experiment uses two proxy models, the conditional neural network model and the multi-task conditional neural network model, to optimize the hyperparameters in the framework of the Bayesian decision algorithm. That is, to find a hyperparameter combination that can make the network prediction accuracy high, and then can verify the effectiveness and accuracy of the multi-task conditional neural network model given in this paper for the hyperparameter optimization problem based on the experimental results. In addition, since the neural network may get different results each time it is run, in order to eliminate errors, all accuracy values are the average of the results of ten runs.
表1神经网络测试MNIST准确率的均值和标准差Table 1 The mean and standard deviation of the accuracy of MNIST tested by neural network
表2神经网络测试Fashion-MNIST准确率的均值和标准差Table 2 The mean and standard deviation of the accuracy of the neural network test Fashion-MNIST
从表1和表2中,多任务条件神经网络模型(OMc-MTCNPs模型)的超参数表现优于基于条件神经网络模型(CNPs)的超参数优化、以及传统的基于高斯过程的超参数优化。因此,本发明提出的基于多任务条件神经网络模型的超参数优化方法更好。From Table 1 and Table 2, the hyperparameter performance of the multi-task conditional neural network model (OMc-MTCNPs model) is better than the hyperparameter optimization based on the conditional neural network model (CNPs) and the traditional hyperparameter optimization based on the Gaussian process. Therefore, the hyperparameter optimization method based on the multi-task conditional neural network model proposed by the present invention is better.
也即是说,本发明采用条件神经网络模型为贝叶斯决策算法框架中的代理模型,取代了传统的高斯过程,实现了单任务的超参数优化,并在条件神经网络模型的基础上通过相似性网络层拓展为多任务条件神经网络模型,实现了多任务的超参数优化。通过在多任务学习中提出了相似性网络层,与一般的正则化约束项不同的是其由神经网络组成,可以参数化任务之间的相似性,并且共享各个任务学习到的不同信息以提高每个任务的泛化性能。That is to say, the present invention adopts the conditional neural network model as the agent model in the Bayesian decision algorithm framework, replacing the traditional Gaussian process, realizing single-task hyperparameter optimization, and adopting the conditional neural network model on the basis of The similarity network layer is extended to a multi-task conditional neural network model, which realizes multi-task hyperparameter optimization. By proposing a similarity network layer in multi-task learning, it is different from the general regularization constraint item in that it is composed of a neural network, which can parameterize the similarity between tasks and share different information learned by each task to improve Generalization performance of each task.
另外,由于本发明是通过利用条件神经网络模型代替传统的高斯过程对于单任务建立模型,而条件神经网络模型是由多层特殊的网络结构组成,可以避免传统的高斯过程存在的复杂的协方差计算的问题,同时条件神经网络模型的学习拟合函数能力也比传统的高斯过程更强,并且不需要提前设定合适的先验分布(由于高斯过程是需要设定数据集分布模型(高斯分布),然后利用数据集更新高斯分布中的参数,而且不一定适用于当前数据集。条件神经网络模型是设定特定的网络结构,只需要进行参数更新即可,具有很强的自适应性),因此将条件神经网络模型与多任务学习结合所形成的多任务条件神经网络模型,大大提高了模型的学习能力,进而实现超参数优化,同时也避免了计算的复杂度。In addition, because the present invention uses the conditional neural network model to replace the traditional Gaussian process to build a model for a single task, and the conditional neural network model is composed of a multi-layer special network structure, it can avoid the complex covariance of the traditional Gaussian process The problem of calculation, and the learning and fitting function of the conditional neural network model is also stronger than the traditional Gaussian process, and there is no need to set a suitable prior distribution in advance (because the Gaussian process needs to set the data set distribution model (Gaussian distribution ), and then use the data set to update the parameters in the Gaussian distribution, and it may not be suitable for the current data set. Conditional neural network model is to set a specific network structure, only need to update the parameters, it has strong adaptability) Therefore, the multi-task conditional neural network model formed by combining the conditional neural network model with multi-task learning greatly improves the learning ability of the model, thereby realizing hyperparameter optimization, and avoiding computational complexity.
本发明还提供了一种电子设备,其包括存储器、处理器以及存储在存储器上并可在处理上运行的计算机程序,所述处理器执行所述程序时实现文中所述的一种深度神经网络超参数优化方法的步骤。The present invention also provides an electronic device, which includes a memory, a processor, and a computer program that is stored in the memory and can be run on the processing. The processor implements the deep neural network described in the article when the program is executed. Steps of hyperparameter optimization method.
本发明还提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现如文中所述的一种深度神经网络超参数优化方法的步骤。The present invention also provides a computer-readable storage medium on which a computer program is stored. When the computer program is executed by a processor, the steps of a deep neural network hyperparameter optimization method as described in the text are realized.
上述实施方式仅为本发明的优选实施方式,不能以此来限定本发明保护的范围,本领域的技术人员在本发明的基础上所做的任何非实质性的变化及替换 均属于本发明所要求保护的范围。The foregoing embodiments are only preferred embodiments of the present invention, and cannot be used to limit the scope of protection of the present invention. Any insubstantial changes and substitutions made by those skilled in the art on the basis of the present invention belong to the present invention. The scope of protection required.
Claims (9)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2019/096379 WO2021007812A1 (en) | 2019-07-17 | 2019-07-17 | Deep neural network hyperparameter optimization method, electronic device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2019/096379 WO2021007812A1 (en) | 2019-07-17 | 2019-07-17 | Deep neural network hyperparameter optimization method, electronic device and storage medium |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2021007812A1 true WO2021007812A1 (en) | 2021-01-21 |
Family
ID=74209615
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2019/096379 Ceased WO2021007812A1 (en) | 2019-07-17 | 2019-07-17 | Deep neural network hyperparameter optimization method, electronic device and storage medium |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2021007812A1 (en) |
Cited By (230)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112765933A (en) * | 2021-01-29 | 2021-05-07 | 西安电子科技大学 | Integrated circuit design method based on population optimization algorithm |
| CN112766493A (en) * | 2021-01-19 | 2021-05-07 | 北京市商汤科技开发有限公司 | Training method and device for multitask neural network, electronic equipment and storage medium |
| CN112784919A (en) * | 2021-02-03 | 2021-05-11 | 华南理工大学 | Intelligent manufacturing multi-mode data oriented classification method |
| CN112784968A (en) * | 2021-01-29 | 2021-05-11 | 东南大学 | Hybrid pipeline parallel method for accelerating distributed deep neural network training |
| CN112784508A (en) * | 2021-02-12 | 2021-05-11 | 西北工业大学 | Deep learning-based airfoil flow field rapid prediction method |
| CN112861386A (en) * | 2021-03-23 | 2021-05-28 | 辽宁工程技术大学 | IACO-BRBP neural network-based Koch snowflake type ultra-wideband antenna reverse modeling method |
| CN112927013A (en) * | 2021-02-24 | 2021-06-08 | 国网电子商务有限公司 | Asset value prediction model construction method and asset value prediction method |
| CN112929306A (en) * | 2021-02-08 | 2021-06-08 | 西北工业大学 | Multicast frequency control array method based on deep learning channel estimation |
| CN112946784A (en) * | 2021-03-29 | 2021-06-11 | 杭州电子科技大学 | SuperDARN radar convection diagram short-term forecasting method based on deep learning |
| CN112989508A (en) * | 2021-02-01 | 2021-06-18 | 复旦大学 | Filter optimization design method based on deep learning algorithm |
| CN112990196A (en) * | 2021-03-16 | 2021-06-18 | 北京大学 | Scene character recognition method and system based on hyper-parameter search and two-stage training |
| CN113055111A (en) * | 2021-03-04 | 2021-06-29 | 武汉科技大学 | Channel modeling method and system based on Bayesian optimization |
| CN113076699A (en) * | 2021-04-22 | 2021-07-06 | 西安交通大学 | Antenna optimization method based on multi-output Gaussian process Bayesian optimization |
| CN113095477A (en) * | 2021-03-17 | 2021-07-09 | 西安理工大学 | Wind power prediction method based on DE-BP neural network |
| CN113113030A (en) * | 2021-03-22 | 2021-07-13 | 浙江大学 | High-dimensional damaged data wireless transmission method based on noise reduction self-encoder |
| CN113128124A (en) * | 2021-04-23 | 2021-07-16 | 东北大学 | Multi-grade C-Mn steel mechanical property prediction method based on improved neural network |
| CN113139624A (en) * | 2021-05-18 | 2021-07-20 | 南京大学 | Network user classification method based on machine learning |
| CN113159275A (en) * | 2021-03-05 | 2021-07-23 | 深圳市商汤科技有限公司 | Network training method, image processing method, device, equipment and storage medium |
| CN113189983A (en) * | 2021-04-13 | 2021-07-30 | 中国人民解放军国防科技大学 | Open scene-oriented multi-robot cooperative multi-target sampling method |
| CN113313152A (en) * | 2021-05-19 | 2021-08-27 | 北京大学 | Image classification method based on optimization-induced equilibrium neural network model |
| CN113407680A (en) * | 2021-06-30 | 2021-09-17 | 竹间智能科技(上海)有限公司 | Heterogeneous integrated model screening method and electronic equipment |
| CN113420508A (en) * | 2021-07-07 | 2021-09-21 | 华北电力大学 | Unit combination calculation method based on LSTM |
| CN113447021A (en) * | 2021-07-15 | 2021-09-28 | 北京理工大学 | MEMS inertial navigation system positioning enhancement method based on LSTM neural network model |
| CN113467766A (en) * | 2021-06-15 | 2021-10-01 | 江苏大学 | Method for determining optimal neural network input vector length |
| CN113469426A (en) * | 2021-06-23 | 2021-10-01 | 国网山东省电力公司东营供电公司 | Photovoltaic output power prediction method and system based on improved BP neural network |
| CN113470316A (en) * | 2021-05-24 | 2021-10-01 | 电子科技大学 | Debris flow monitoring and early warning method based on self-coding single classification model |
| CN113486923A (en) * | 2021-06-02 | 2021-10-08 | 长安大学 | Dynamic width-depth neural network for hyperspectral classification and learning method thereof |
| CN113505477A (en) * | 2021-06-29 | 2021-10-15 | 西北师范大学 | Process industry soft measurement data supplementing method based on SVAE-WGAN |
| CN113539517A (en) * | 2021-08-05 | 2021-10-22 | 浙江大学 | Prediction method of time sequence intervention effect |
| CN113569467A (en) * | 2021-07-12 | 2021-10-29 | 鞍钢股份有限公司 | Blast furnace tuyere raceway temperature soft measurement modeling method based on width learning |
| CN113571133A (en) * | 2021-09-14 | 2021-10-29 | 内蒙古农业大学 | A prediction method of lactic acid bacteria antimicrobial peptides based on graph neural network |
| CN113570040A (en) * | 2021-07-27 | 2021-10-29 | 南京航空航天大学 | A multi-domain action sequence intelligent optimization system and method based on evolution strategy |
| CN113612555A (en) * | 2021-08-19 | 2021-11-05 | 江西赛柏科技有限公司 | Intelligent calibration algorithm and system based on mobile terminal wireless radio frequency signal intensity |
| CN113640397A (en) * | 2020-07-13 | 2021-11-12 | 江南大学 | Remote sensing image change detection method and system based on depth network |
| CN113689928A (en) * | 2021-08-24 | 2021-11-23 | 平安国际智慧城市科技股份有限公司 | Recommendation method, device, equipment and storage medium for maintaining and preventing disease scheme |
| CN113721212A (en) * | 2021-07-16 | 2021-11-30 | 西安电子科技大学 | Radar scattering cross section confidence evaluation method based on neural network |
| CN113723692A (en) * | 2021-09-02 | 2021-11-30 | 深圳前海微众银行股份有限公司 | Data processing method, apparatus, device, medium, and program product |
| CN113779871A (en) * | 2021-08-26 | 2021-12-10 | 清华大学 | Electric heating coupling system scheduling method and device, electronic equipment and storage medium thereof |
| CN113780432A (en) * | 2021-09-15 | 2021-12-10 | 中国人民解放军国防大学联合作战学院 | An intelligent detection method for network information system operation and maintenance anomalies based on reinforcement learning |
| CN113890633A (en) * | 2021-11-11 | 2022-01-04 | 中国船舶重工集团公司第七一五研究所 | Underwater acoustic communication system self-adaptive selection method based on deep neural network |
| CN113887797A (en) * | 2021-09-29 | 2022-01-04 | 南京用能互联网研究院有限公司 | Building electric heating load prediction model establishing method, device and equipment |
| CN113919410A (en) * | 2021-08-25 | 2022-01-11 | 桂林电子科技大学 | An incremental image classification method and system |
| CN113915249A (en) * | 2021-10-29 | 2022-01-11 | 江苏大学 | Wet clutch state identification method based on fuzzy neural network |
| CN113947185A (en) * | 2021-09-30 | 2022-01-18 | 北京达佳互联信息技术有限公司 | Task processing network generation method, task processing device, electronic equipment and storage medium |
| CN113971992A (en) * | 2021-10-26 | 2022-01-25 | 中国科学技术大学 | Self-supervision pre-training method and system for molecular attribute prediction graph network |
| CN113988437A (en) * | 2021-11-01 | 2022-01-28 | 东南大学 | An Adaptive Interval Forecasting Method for Short-Term Residential Load |
| CN114048564A (en) * | 2021-11-05 | 2022-02-15 | 广东海洋大学 | Engineering forklift multi-target performance optimization method based on deep proxy model |
| CN114046815A (en) * | 2021-11-09 | 2022-02-15 | 上海精赋达传感技术有限公司 | Encoder self-correction method and device based on deep learning |
| CN114065635A (en) * | 2021-11-22 | 2022-02-18 | 中国民航大学 | Aircraft ground air conditioner energy consumption prediction method based on IALO-AM-BilSTM model and storage medium |
| CN114118339A (en) * | 2021-11-12 | 2022-03-01 | 吉林大学 | Radio modulation signal identification and classification method based on cuckoo algorithm improved ResNet |
| CN114120289A (en) * | 2022-01-25 | 2022-03-01 | 中科视语(北京)科技有限公司 | Method and system for identifying driving area and lane line |
| CN114114126A (en) * | 2021-11-22 | 2022-03-01 | 浙江华采科技有限公司 | Digital calibration method and system for low-permeability magnetic powder core current transformer |
| CN114118569A (en) * | 2021-11-24 | 2022-03-01 | 宁波力斗智能技术有限公司 | Wind power multi-step prediction method based on multi-mode multi-task Transformer network |
| CN114139720A (en) * | 2021-11-16 | 2022-03-04 | 广西中科曙光云计算有限公司 | Method and device for processing big data of government affairs based on machine learning |
| CN114169605A (en) * | 2021-12-06 | 2022-03-11 | 西安交通大学 | Physical field and performance collaborative prediction method suitable for pump and hydraulic turbine |
| CN114169240A (en) * | 2021-12-08 | 2022-03-11 | 中国石油大学(北京) | MMP prediction method and device based on conditional generative adversarial network |
| CN114190944A (en) * | 2021-12-23 | 2022-03-18 | 上海交通大学 | Robust emotion recognition method based on electroencephalogram signals |
| CN114217688A (en) * | 2021-12-24 | 2022-03-22 | 中国科学技术大学 | A system and method for NPU power consumption optimization based on neural network structure |
| CN114217944A (en) * | 2021-04-26 | 2022-03-22 | 无锡江南计算技术研究所 | A neural network-based dynamic load balancing method for model parallelism |
| CN114219069A (en) * | 2021-11-16 | 2022-03-22 | 北京工业大学 | Brain effect connection network learning method based on automatic variation self-encoder |
| CN114215706A (en) * | 2021-12-27 | 2022-03-22 | 南京邮电大学 | Wind turbine generator blade cracking fault early warning method and device |
| CN114236610A (en) * | 2021-12-24 | 2022-03-25 | 西安交通大学 | Iterative seismic data unmixing method and system based on depth convolution network prior |
| CN114268129A (en) * | 2021-12-21 | 2022-04-01 | 湖州越球电机有限公司 | Wind power frequency modulation gradual inertia control method based on deep neural network |
| CN114266953A (en) * | 2021-12-24 | 2022-04-01 | 福州大学 | Target detection process self-optimization method based on deep learning |
| CN114282725A (en) * | 2021-12-24 | 2022-04-05 | 山东大学 | Construction of transient oil reservoir agent model based on deep learning and oil reservoir prediction method |
| CN114280491A (en) * | 2021-12-23 | 2022-04-05 | 中山大学 | A method for estimating the remaining capacity of retired batteries based on active learning |
| CN114330456A (en) * | 2022-01-05 | 2022-04-12 | 北京邮电大学 | A Multi-task Learning-Based Optical Path Transmission Quality Prediction Method |
| CN114358278A (en) * | 2021-11-26 | 2022-04-15 | 北京邮电大学 | Training method and device for neural network model |
| CN114386364A (en) * | 2021-12-06 | 2022-04-22 | 西安交通大学 | Quality prediction and process parameter optimization method for electronic surface mount soldering |
| CN114399119A (en) * | 2022-01-18 | 2022-04-26 | 中国石油大学(北京) | MMP prediction method and device based on conditional convolution generative adversarial network |
| CN114417497A (en) * | 2021-12-23 | 2022-04-29 | 中国人民解放军军事科学院国防科技创新研究院 | Neural network initialization design method based on meta-learning embedded physical knowledge |
| CN114417011A (en) * | 2022-01-20 | 2022-04-29 | 北京北大英华科技有限公司 | Inspection service portrait knowledge fusion method and device, electronic equipment and storage medium |
| CN114462292A (en) * | 2021-11-22 | 2022-05-10 | 广州大学 | Circuit parameter optimization method based on SQP algorithm |
| CN114462309A (en) * | 2022-01-19 | 2022-05-10 | 苏州浪潮智能科技有限公司 | A physical simulation model training method, device, equipment and storage medium |
| CN114492279A (en) * | 2022-01-27 | 2022-05-13 | 西安交通大学 | A method and system for parameter optimization of an analog integrated circuit |
| CN114492941A (en) * | 2021-12-29 | 2022-05-13 | 东南大学 | Whole-county photovoltaic prediction method based on cluster division and data enhancement |
| CN114492624A (en) * | 2022-01-21 | 2022-05-13 | 中国电子科技集团公司第三十六研究所 | Small sample learning method and device based on meta-learning and electronic equipment |
| CN114519381A (en) * | 2021-12-31 | 2022-05-20 | 上海仙途智能科技有限公司 | Sensing method and device based on multitask learning network, storage medium and terminal |
| CN114565084A (en) * | 2022-03-02 | 2022-05-31 | 安徽大学 | Computer self-adaptive evaluation method based on learning question selection strategy |
| CN114584150A (en) * | 2022-03-04 | 2022-06-03 | 南京邮电大学 | Sensor network signal completion method realized by utilizing deep neural network |
| CN114611823A (en) * | 2022-03-23 | 2022-06-10 | 特斯联科技集团有限公司 | Optimized dispatching method and system for electricity-cold-heat-gas multi-energy-demand typical park |
| CN114630207A (en) * | 2022-03-15 | 2022-06-14 | 浙江大学 | Multi-sensor node sensing data collection method based on noise reduction autoencoder |
| CN114647950A (en) * | 2022-04-02 | 2022-06-21 | 西安交通大学 | Variable-density high-vacuum multi-layer heat insulation structure optimization method based on proxy model |
| CN114676639A (en) * | 2022-03-31 | 2022-06-28 | 上海大学 | Aircraft aerodynamic shape optimization method, device and medium based on neural network |
| CN114676909A (en) * | 2022-03-25 | 2022-06-28 | 东南大学 | A charging path planning method for unmanned vehicles based on deep reinforcement learning |
| CN114676522A (en) * | 2022-03-28 | 2022-06-28 | 西安交通大学 | Aerodynamic shape optimization design method, system and equipment integrating GAN and transfer learning |
| CN114692347A (en) * | 2022-03-16 | 2022-07-01 | 中国人民解放军军事科学院国防科技创新研究院 | Temperature field proxy model construction method based on neural network architecture search technology |
| CN114692504A (en) * | 2022-04-08 | 2022-07-01 | 东南大学 | An automatic synthesis method for variable spiral inductors on millimeter-wave chips |
| CN114707471A (en) * | 2022-06-06 | 2022-07-05 | 浙江大学 | Artificial intelligent courseware making method and device based on hyper-parameter evaluation graph algorithm |
| CN114742289A (en) * | 2022-03-31 | 2022-07-12 | 大连理工大学 | Gaussian process robust optimization method for production process parameters |
| CN114757096A (en) * | 2022-04-02 | 2022-07-15 | 华南理工大学 | Bridge temperature prediction method, device, equipment and medium based on NARX neural network |
| CN114819105A (en) * | 2022-05-23 | 2022-07-29 | 吉林大学 | A Neural Network-based Suspension Model Construction Method for Air Springs and Continuously Adjustable Shock Absorbers |
| CN114818477A (en) * | 2022-03-31 | 2022-07-29 | 上海大学 | Method, device and medium for predicting aerodynamic characteristics of aircraft |
| CN114842328A (en) * | 2022-03-22 | 2022-08-02 | 西北工业大学 | Hyperspectral change detection method based on cooperative analysis autonomous sensing network structure |
| CN114839439A (en) * | 2022-04-15 | 2022-08-02 | 厦门明翰电气股份有限公司 | Harmonic current detection method for brain emotion learning based on PID optimization |
| CN114912342A (en) * | 2022-03-28 | 2022-08-16 | 南京邮电大学 | Packaging lead bonding process parameter optimization method based on multiple quality parameters |
| CN114943182A (en) * | 2022-06-07 | 2022-08-26 | 清华大学深圳国际研究生院 | Robot cable shape control method and device based on graph neural network |
| CN114943162A (en) * | 2022-03-22 | 2022-08-26 | 北京交通大学 | Periodic super-barrier reverse design method based on deep learning |
| CN114966233A (en) * | 2022-05-16 | 2022-08-30 | 国网电力科学研究院武汉南瑞有限责任公司 | Lightning forecast system and method based on deep neural network |
| CN114966451A (en) * | 2022-06-06 | 2022-08-30 | 北京航空航天大学 | Power supply product residual service life online prediction method based on physical information neural network |
| CN114970819A (en) * | 2022-05-26 | 2022-08-30 | 哈尔滨工业大学 | A moving target search and tracking method and system based on intent reasoning and deep reinforcement learning |
| CN114997060A (en) * | 2022-06-10 | 2022-09-02 | 湖南大学 | A phononic crystal time-varying reliability testing method, computing device and storage medium |
| CN114997373A (en) * | 2022-06-14 | 2022-09-02 | 中国人民解放军国防科技大学 | Network depth dynamic adjustment method, apparatus, computer equipment and storage medium |
| CN114997491A (en) * | 2022-05-30 | 2022-09-02 | 武汉工程大学 | Method, device, equipment and storage medium for optimizing gasoline octane number loss |
| CN114997297A (en) * | 2022-05-26 | 2022-09-02 | 哈尔滨工业大学 | A method and system for inference of target motion intent based on multi-level region division |
| CN115018607A (en) * | 2022-07-01 | 2022-09-06 | 吉林工程技术师范学院 | Accounting data processing method and system based on artificial intelligence |
| CN115015869A (en) * | 2022-06-27 | 2022-09-06 | 清华大学 | Can learn low frequency broadband radar target parameter estimation method, equipment and program product |
| CN115018137A (en) * | 2022-05-23 | 2022-09-06 | 深圳市广汇源环境水务有限公司 | Water environment model parameter calibration method based on reinforcement learning |
| CN115018038A (en) * | 2022-03-14 | 2022-09-06 | 贵州大学 | Network interface flow prediction method based on sparrow optimization width learning system |
| CN115047894A (en) * | 2022-04-14 | 2022-09-13 | 中国民用航空总局第二研究所 | Unmanned aerial vehicle track measuring and calculating method, electronic equipment and storage medium |
| CN115049108A (en) * | 2022-05-20 | 2022-09-13 | 支付宝(杭州)信息技术有限公司 | Multitask model training method, multitask prediction method, related device and medium |
| CN115062501A (en) * | 2022-04-27 | 2022-09-16 | 南方科技大学 | Chip packaging design optimization method based on adaptive subproblem selection strategy |
| CN115099146A (en) * | 2022-06-27 | 2022-09-23 | 上海集成电路装备材料产业创新中心有限公司 | Circuit generation method, circuit generation device, electronic equipment and storage medium |
| CN115115052A (en) * | 2022-08-11 | 2022-09-27 | 杭州电子科技大学 | An Adaptive Distributed Parallel Training Method of Neural Network Based on Genetic Algorithm |
| CN115146536A (en) * | 2022-07-01 | 2022-10-04 | 纽威数控装备(苏州)股份有限公司 | Wavelet neural network and particle swarm algorithm-based heat source contribution analysis method |
| CN115205349A (en) * | 2022-08-12 | 2022-10-18 | 南京理工大学 | A Deep Neural Network-Based Interferogram Automatic Registration and Phase Unpacking Method |
| CN115222023A (en) * | 2022-08-15 | 2022-10-21 | 西北工业大学 | Non-cooperative target control strategy identification method and system based on neural network |
| CN115238602A (en) * | 2022-07-01 | 2022-10-25 | 中国海洋大学 | CNN-LSTM-based prediction method for contribution rate of transient liquefaction of wave-induced seabed to resuspension |
| CN115237878A (en) * | 2022-06-28 | 2022-10-25 | 华中科技大学 | Process database construction method and medium based on additive manufacturing |
| CN115240394A (en) * | 2022-09-22 | 2022-10-25 | 国网湖北省电力有限公司经济技术研究院 | Method and system for monitoring and early warning of water level in accident oil pool in substation |
| CN115412332A (en) * | 2022-08-25 | 2022-11-29 | 暨南大学 | Internet of things intrusion detection system and method based on hybrid neural network model optimization |
| CN115455782A (en) * | 2022-09-22 | 2022-12-09 | 吉林大学 | Bionic self-similar hierarchical structure performance prediction method based on machine learning |
| CN115468633A (en) * | 2022-09-01 | 2022-12-13 | 西安热工研究院有限公司 | Multi-point flowmeter correction method and device based on RBF neural network |
| CN115481790A (en) * | 2022-09-02 | 2022-12-16 | 广东省科学院生态环境与土壤研究所 | Carbon-based geological catalytic material fixed cadmium and methane emission reduction cooperative prediction method, device and medium |
| CN115512172A (en) * | 2022-09-30 | 2022-12-23 | 赵营鸽 | Uncertainty quantification method for multi-dimensional parameters in electrical impedance imaging technology |
| CN115549732A (en) * | 2022-09-13 | 2022-12-30 | 哈尔滨工业大学 | An Artificial Intelligence-Based Massive MIMO Hybrid Precoding Method |
| CN115542723A (en) * | 2022-10-08 | 2022-12-30 | 华南理工大学 | Parameter setting method for ankle exoskeleton PD controller for man-machine in-loop debugging |
| CN115579093A (en) * | 2022-12-08 | 2023-01-06 | 科丰兴泰(杭州)生物科技有限公司 | Method and system for designing nitrification inhibitor slow-release material based on deep learning |
| CN115600667A (en) * | 2022-09-22 | 2023-01-13 | 浙江大学(Cn) | A method and device for modeling multi-dimensional state change time series data of a system |
| CN115660786A (en) * | 2022-11-09 | 2023-01-31 | 中国工商银行股份有限公司 | Service recommendation method, apparatus, computer equipment, storage medium and program product |
| CN115688565A (en) * | 2022-10-13 | 2023-02-03 | 阿里巴巴(中国)有限公司 | Dynamic optimization method of network model, image processing method and equipment |
| CN115713099A (en) * | 2023-01-03 | 2023-02-24 | 税友信息技术有限公司 | Model design method, device, equipment and storage medium |
| CN115730509A (en) * | 2022-09-20 | 2023-03-03 | 中国人民解放军军事科学院国防科技创新研究院 | Spacecraft in-cabin temperature field reconstruction task research benchmark method based on machine learning |
| CN115759200A (en) * | 2022-11-22 | 2023-03-07 | 西北工业大学 | A key path construction method for deep network based on iterative optimization of gate parameters |
| CN115757384A (en) * | 2022-11-30 | 2023-03-07 | 安徽长正智库管理咨询有限公司 | A big data-based government data processing method |
| CN115758191A (en) * | 2022-10-31 | 2023-03-07 | 北京理工大学 | Knowledge service entity clustering number prediction method based on deep learning |
| CN115762674A (en) * | 2022-10-20 | 2023-03-07 | 西安交通大学 | Rapid solving method for inverse heat transfer problem of nano heat insulation composite material |
| CN115796346A (en) * | 2022-11-22 | 2023-03-14 | 烟台国工智能科技有限公司 | Yield optimization method and system and non-transitory computer readable storage medium |
| CN115809624A (en) * | 2023-02-01 | 2023-03-17 | 北京大学 | Automatic analysis design method for microstrip line transmission line of integrated circuit |
| CN115846852A (en) * | 2022-12-13 | 2023-03-28 | 大连交通大学 | FSW welding monitoring method based on DS evidence theory and neural network |
| CN115935626A (en) * | 2022-11-25 | 2023-04-07 | 河南大学 | Inversion method of river water-underground water vertical transient interaction water flux based on VAE-PINN algorithm |
| CN115935813A (en) * | 2022-11-28 | 2023-04-07 | 上海大学 | A device multi-stage degradation evolution prediction method, device and storage medium based on statistical model and deep learning |
| CN115988542A (en) * | 2021-10-13 | 2023-04-18 | 纬创资通股份有限公司 | Network construction system and method, recording medium, program product, and control system |
| CN116016223A (en) * | 2022-12-09 | 2023-04-25 | 国网湖北省电力有限公司信息通信公司 | Data transmission optimization method for data center network |
| CN116011337A (en) * | 2023-01-11 | 2023-04-25 | 西北工业大学 | An Efficient Engineering Design Method Against Numerical Noise |
| CN116150995A (en) * | 2023-02-21 | 2023-05-23 | 东南大学 | Rapid simulation method of switch arc model |
| CN116204998A (en) * | 2023-01-09 | 2023-06-02 | 华中科技大学 | SLM forming performance prediction and process parameter optimization method and system |
| CN116227671A (en) * | 2022-12-27 | 2023-06-06 | 山西省能源互联网研究院 | Optimization Method of Wind Power Joint Forecasting Model Based on Long Short-term Memory Neural Network |
| CN116306206A (en) * | 2022-11-06 | 2023-06-23 | 西北工业大学 | Airfoil transonic buffeting flow field rapid prediction method based on deep neural network |
| CN116321008A (en) * | 2023-04-13 | 2023-06-23 | 重庆邮电大学 | Indoor position fingerprint positioning method based on triplet self-encoder |
| CN116339136A (en) * | 2023-01-18 | 2023-06-27 | 浙江大学 | An actuator control method and controller based on online neural prediction |
| CN116343930A (en) * | 2023-01-10 | 2023-06-27 | 西北工业大学 | Metabolic dynamics and toxicity prediction method based on graph representation multitask learning |
| CN116380321A (en) * | 2023-02-24 | 2023-07-04 | 武汉科技大学 | Method for Decoupling Decoupling Model of Tire-Road Contact Triaxial Force Sensor |
| CN116449294A (en) * | 2023-03-27 | 2023-07-18 | 上海海事大学 | A real-time underwater acoustic positioning method based on IDE-BP neural network |
| CN116499607A (en) * | 2023-06-27 | 2023-07-28 | 之江实验室 | Optical fiber sensing signal noise reduction method and device and storage medium |
| CN116502566A (en) * | 2023-06-27 | 2023-07-28 | 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) | Multi-objective optimization method for performance of combustion chamber of gas turbine based on Bayesian optimization |
| CN116611349A (en) * | 2023-07-18 | 2023-08-18 | 华东交通大学 | Method and system for optimizing process parameters of drum silk drying based on neural network |
| CN116644662A (en) * | 2023-05-19 | 2023-08-25 | 之江实验室 | A well layout optimization method based on knowledge embedded neural network agent model |
| CN116680628A (en) * | 2023-05-11 | 2023-09-01 | 苏州大学 | A volatile liquid identification module and method based on SVM electronic nose system |
| CN116759031A (en) * | 2023-08-10 | 2023-09-15 | 内蒙古工业大学 | Design method of sludge ash concrete material mixing ratio based on ANN |
| CN116775687A (en) * | 2023-07-07 | 2023-09-19 | 哈尔滨工程大学 | Time sequence prediction closed-loop framework based on digital-analog symbiosis |
| CN116913435A (en) * | 2023-07-27 | 2023-10-20 | 常州威材新材料科技有限公司 | High-strength engineering plastic evaluation method and system based on component analysis |
| CN116992298A (en) * | 2023-09-28 | 2023-11-03 | 青岛东捷建设集团有限公司 | High-pressure jet grouting method and system based on seawater seepage |
| CN117004432A (en) * | 2022-04-26 | 2023-11-07 | 中国石油化工股份有限公司 | Data analysis methods, devices, media and equipment for feed oil hydrogenation units |
| CN117037971A (en) * | 2023-08-21 | 2023-11-10 | 电子科技大学 | Self-calibration design method for multilayer honeycomb wave-absorbing material based on neural network |
| CN117056847A (en) * | 2023-10-10 | 2023-11-14 | 中南大学 | Method, system, equipment and storage medium for detecting abnormality of stream data |
| CN117093871A (en) * | 2023-10-16 | 2023-11-21 | 之江实验室 | Deep learning-oriented distributed training evaluation method and system |
| CN117114141A (en) * | 2023-10-20 | 2023-11-24 | 安徽蔚来智驾科技有限公司 | Model training method, evaluation method, computer device and storage medium |
| CN117174219A (en) * | 2023-11-02 | 2023-12-05 | 济南大学 | Cement strength conditional probability distribution estimation method and system based on latent variable model |
| WO2023231244A1 (en) * | 2022-06-02 | 2023-12-07 | 大连理工大学 | Fsw tensile strength intelligent prediction method considering welding temperature and upsetting force |
| FR3136299A1 (en) | 2022-01-04 | 2023-12-08 | Alcom Technologies | Method for optimizing the hyperparameters of a machine learning model |
| CN117272778A (en) * | 2023-07-12 | 2023-12-22 | 上海交通大学 | Design methods and devices for microwave passive components |
| CN117272839A (en) * | 2023-11-20 | 2023-12-22 | 北京阿迈特医疗器械有限公司 | Support press-holding performance prediction method and device based on neural network |
| CN117331339A (en) * | 2023-12-01 | 2024-01-02 | 南京华视智能科技股份有限公司 | Coating machine die head motor control method and device based on time sequence neural network model |
| CN117432414A (en) * | 2023-12-20 | 2024-01-23 | 中煤科工开采研究院有限公司 | Method and system for regulating and controlling top plate frosted jet flow seam formation |
| CN117455320A (en) * | 2023-12-25 | 2024-01-26 | 江苏恒力化纤股份有限公司 | QGMVPN model-based multi-kettle polymerization process melt quality index online prediction method |
| CN117556877A (en) * | 2024-01-11 | 2024-02-13 | 西南交通大学 | Pulse neural network training method based on data pulse characteristic evaluation |
| CN117582185A (en) * | 2024-01-19 | 2024-02-23 | 季华实验室 | A pulse force level prediction method based on CLLSR hybrid model |
| CN117647367A (en) * | 2024-01-29 | 2024-03-05 | 四川航空股份有限公司 | Machine learning-based method and system for positioning leakage points of aircraft fuel tank |
| CN117875397A (en) * | 2024-03-11 | 2024-04-12 | 浪潮电子信息产业股份有限公司 | A method, device, computing device and storage medium for selecting parameters to be updated |
| CN117894397A (en) * | 2024-03-15 | 2024-04-16 | 北京科技大学 | A method for predicting viscosity of continuous casting mold slag based on machine learning |
| WO2024098522A1 (en) * | 2022-11-09 | 2024-05-16 | 广东海洋大学 | Machine learning-based multi-performance collaborative optimization high-entropy alloy component design method |
| CN118131689A (en) * | 2024-02-19 | 2024-06-04 | 哈尔滨理工大学 | A method for predicting machine tool processing energy consumption based on Hyperopt optimized neural network |
| CN118260840A (en) * | 2024-04-12 | 2024-06-28 | 湖南工程学院 | A method and system for generating a prefabricated area judgment model for an assembled building |
| CN118298792A (en) * | 2024-04-08 | 2024-07-05 | 哈尔滨工程大学 | Ship sonar platform self-noise suppression method based on deep learning |
| CN118333129A (en) * | 2024-06-07 | 2024-07-12 | 广东工业大学 | Identification model training method, nonlinear system identification method and system |
| CN118351556A (en) * | 2024-05-17 | 2024-07-16 | 优顶特技术有限公司 | Method, device, equipment and storage medium for classifying and identifying bills |
| CN118365183A (en) * | 2024-06-20 | 2024-07-19 | 西南石油大学 | A method and system for predicting oil well production based on hybrid neural network |
| CN118442745A (en) * | 2024-05-30 | 2024-08-06 | 华南农业大学 | Liquid nitrogen instant freezer temperature control method, device, equipment and medium |
| CN118469354A (en) * | 2024-07-15 | 2024-08-09 | 江西省赣地智慧科技有限公司 | Intelligent construction site data acquisition method and system for road construction |
| CN118538358A (en) * | 2024-06-07 | 2024-08-23 | 四川熠海瑞科技有限公司 | Immune system treatment information management platform |
| CN118569113A (en) * | 2024-08-05 | 2024-08-30 | 西安理工大学 | Optimization method and system for jet mechanism of impulse turbine |
| CN118747702A (en) * | 2024-06-06 | 2024-10-08 | 集美大学 | A mine inspection method and system based on reinforcement learning and machine vision |
| CN118778691A (en) * | 2024-09-10 | 2024-10-15 | 青岛哈尔滨工程大学创新发展中心 | A system and method for evaluating the hydrodynamic characteristics of a submarine-launched UAV during its entry and exit into water |
| CN118797354A (en) * | 2024-09-12 | 2024-10-18 | 青岛科技大学 | A method for determining parameters for improving the efficiency of methyl oleate epoxidation reaction |
| CN118798061A (en) * | 2024-09-11 | 2024-10-18 | 中国海洋大学 | Door sealing model parameter determination method, device, equipment, medium and program product |
| CN118915102A (en) * | 2024-08-22 | 2024-11-08 | 河南地矿职业学院 | Satellite clock error forecasting method and system |
| CN118919080A (en) * | 2024-10-11 | 2024-11-08 | 长春理工大学 | Noninvasive blood glucose concentration prediction method |
| CN118940612A (en) * | 2024-07-15 | 2024-11-12 | 华中科技大学 | Adapter Optimization Design Method Based on Parameter Optimization Neural Network |
| CN118963774A (en) * | 2024-10-18 | 2024-11-15 | 浙江大学 | A Sub-model-based Scheduling Method for AI Compilers |
| CN118966019A (en) * | 2024-10-14 | 2024-11-15 | 新雅电线电缆(深圳)有限公司 | A method for processing process data of insulated cables |
| CN118967670A (en) * | 2024-10-15 | 2024-11-15 | 浙江理工大学 | A parameter optimization method for electromagnetic induction detection based on multi-task learning model |
| CN118971947A (en) * | 2024-10-12 | 2024-11-15 | 中国人民解放军63921部队 | A UAV communication system based on Beidou messages |
| CN119025998A (en) * | 2024-10-25 | 2024-11-26 | 自然资源部第三海洋研究所 | A method for estuary water level time series prediction based on improved Informer deep learning framework |
| CN119025925A (en) * | 2024-10-28 | 2024-11-26 | 西南石油大学 | A multi-task reservoir parameter prediction method based on well logging data |
| CN119051981A (en) * | 2024-10-30 | 2024-11-29 | 海天地数码科技(北京)有限公司 | Online data service center of electronic reader based on deep learning |
| CN119066817A (en) * | 2024-11-05 | 2024-12-03 | 宁波隆跃科技有限公司 | Die casting parameter optimization and processing method and system based on orthogonal experiment |
| CN119109455A (en) * | 2024-08-30 | 2024-12-10 | 合肥工业大学 | A pipeline ADC background calibration method based on ant colony optimization time-delay neural network |
| CN119129437A (en) * | 2024-11-12 | 2024-12-13 | 宁波大学 | A fatigue life prediction method for automotive turbocharger pipelines |
| CN119127146A (en) * | 2024-09-19 | 2024-12-13 | 华南理工大学 | Liberty file generation method for SRAM memory compiler in 55nm CMOS process based on neural network |
| CN119155335A (en) * | 2024-11-19 | 2024-12-17 | 漳州职业技术学院 | Intelligent well lid data analysis method and device based on Internet of things and storage medium |
| CN119167211A (en) * | 2024-11-20 | 2024-12-20 | 济宁职业技术学院 | A method and system for managing people who take Mandarin proficiency tests |
| CN119168797A (en) * | 2024-08-30 | 2024-12-20 | 沈阳农业大学 | A method, system, device and medium for inversion of crop growth |
| CN119167786A (en) * | 2024-09-29 | 2024-12-20 | 浙江理工大学 | A modeling method for magnetorheological damper based on Gaussian process regression model |
| CN119227010A (en) * | 2024-12-02 | 2024-12-31 | 杭州市北京航空航天大学国际创新研究院(北京航空航天大学国际创新学院) | A cross-task dataset fusion method based on meta-Bayesian optimization |
| CN119316091A (en) * | 2024-09-30 | 2025-01-14 | 西北工业大学深圳研究院 | A semantic communication method and system based on LSTM network automatic training |
| CN119348220A (en) * | 2024-09-29 | 2025-01-24 | 武汉理工大学 | A stamping operation speed control method, device, electronic equipment and storage medium |
| CN119400308A (en) * | 2024-10-12 | 2025-02-07 | 中车株洲电力机车有限公司 | Battery material combination method, device and computer storage medium |
| CN119387758A (en) * | 2024-11-29 | 2025-02-07 | 徐州德高电动车科技有限公司 | Welding control method for electric tricycle manufacturing |
| CN119416362A (en) * | 2024-12-02 | 2025-02-11 | 北京航空航天大学 | A numerical and real fusion test method for reliability of aero-engine blade disk structure |
| CN119539804A (en) * | 2024-11-06 | 2025-02-28 | 孟约共梦区块链(杭州)有限责任公司 | A transaction method based on blockchain |
| CN119720906A (en) * | 2024-12-09 | 2025-03-28 | 西安电子科技大学 | A method, system, device and medium for optimizing analog equalizer design based on deep learning algorithm |
| CN119829955A (en) * | 2025-03-17 | 2025-04-15 | 中国海洋大学 | Ship nonlinear roll parameter identification method based on physical information neural network |
| CN119883576A (en) * | 2025-03-27 | 2025-04-25 | 吉林大学 | Task association degree-based multi-task learning auxiliary task selection method |
| CN119957986A (en) * | 2025-04-10 | 2025-05-09 | 山西太康高科节能股份有限公司 | A heat pump control method and system based on state data |
| CN120046813A (en) * | 2025-04-24 | 2025-05-27 | 杭州海亮数字科技有限公司 | Model optimization method and device suitable for special length prediction |
| CN120105109A (en) * | 2025-05-12 | 2025-06-06 | 山东永正产业技术研究院有限公司 | Training method and equipment for carbon emission prediction model in cement production |
| CN120256117A (en) * | 2025-03-26 | 2025-07-04 | 北京骏德智算科技有限公司 | A secure scheduling method for database storage resources based on artificial intelligence |
| CN120408251A (en) * | 2025-07-02 | 2025-08-01 | 山东大学日照研究院 | Multi-objective grinding decision-making method and system based on dynamic clustering and hyperparameter optimization |
| CN120468677A (en) * | 2025-07-10 | 2025-08-12 | 洛阳理工学院 | Lithium battery state of charge estimation method based on Bayes-TLCO optimized deep neural network |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20160328653A1 (en) * | 2013-05-30 | 2016-11-10 | Universite De Sherbrooke | Systems and methods for parallelizing bayesian optimization |
| CN108319928A (en) * | 2018-02-28 | 2018-07-24 | 天津大学 | A kind of deep learning model and application based on Multi-objective PSO optimization |
| CN109242105A (en) * | 2018-08-17 | 2019-01-18 | 第四范式(北京)技术有限公司 | Tuning method, apparatus, equipment and the medium of hyper parameter in machine learning model |
| CN109784355A (en) * | 2017-11-15 | 2019-05-21 | 财团法人资讯工业策进会 | Model generates server and its model generating method |
-
2019
- 2019-07-17 WO PCT/CN2019/096379 patent/WO2021007812A1/en not_active Ceased
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20160328653A1 (en) * | 2013-05-30 | 2016-11-10 | Universite De Sherbrooke | Systems and methods for parallelizing bayesian optimization |
| CN109784355A (en) * | 2017-11-15 | 2019-05-21 | 财团法人资讯工业策进会 | Model generates server and its model generating method |
| CN108319928A (en) * | 2018-02-28 | 2018-07-24 | 天津大学 | A kind of deep learning model and application based on Multi-objective PSO optimization |
| CN109242105A (en) * | 2018-08-17 | 2019-01-18 | 第四范式(北京)技术有限公司 | Tuning method, apparatus, equipment and the medium of hyper parameter in machine learning model |
Cited By (320)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113640397A (en) * | 2020-07-13 | 2021-11-12 | 江南大学 | Remote sensing image change detection method and system based on depth network |
| CN112766493A (en) * | 2021-01-19 | 2021-05-07 | 北京市商汤科技开发有限公司 | Training method and device for multitask neural network, electronic equipment and storage medium |
| CN112766493B (en) * | 2021-01-19 | 2023-04-07 | 北京市商汤科技开发有限公司 | Training method and device for multitask neural network, electronic equipment and storage medium |
| CN112765933B (en) * | 2021-01-29 | 2023-08-11 | 西安电子科技大学 | A Design Method of Integrated Circuit Based on Population Optimization Algorithm |
| CN112784968A (en) * | 2021-01-29 | 2021-05-11 | 东南大学 | Hybrid pipeline parallel method for accelerating distributed deep neural network training |
| CN112765933A (en) * | 2021-01-29 | 2021-05-07 | 西安电子科技大学 | Integrated circuit design method based on population optimization algorithm |
| CN112989508A (en) * | 2021-02-01 | 2021-06-18 | 复旦大学 | Filter optimization design method based on deep learning algorithm |
| CN112989508B (en) * | 2021-02-01 | 2022-05-20 | 复旦大学 | Filter optimization design method based on deep learning algorithm |
| CN112784919A (en) * | 2021-02-03 | 2021-05-11 | 华南理工大学 | Intelligent manufacturing multi-mode data oriented classification method |
| CN112784919B (en) * | 2021-02-03 | 2023-09-05 | 华南理工大学 | Classification method for intelligent manufacturing multi-mode data |
| CN112929306A (en) * | 2021-02-08 | 2021-06-08 | 西北工业大学 | Multicast frequency control array method based on deep learning channel estimation |
| CN112784508A (en) * | 2021-02-12 | 2021-05-11 | 西北工业大学 | Deep learning-based airfoil flow field rapid prediction method |
| CN112927013A (en) * | 2021-02-24 | 2021-06-08 | 国网电子商务有限公司 | Asset value prediction model construction method and asset value prediction method |
| CN112927013B (en) * | 2021-02-24 | 2023-11-10 | 国网数字科技控股有限公司 | An asset value prediction model construction method and asset value prediction method |
| CN113055111A (en) * | 2021-03-04 | 2021-06-29 | 武汉科技大学 | Channel modeling method and system based on Bayesian optimization |
| CN113159275A (en) * | 2021-03-05 | 2021-07-23 | 深圳市商汤科技有限公司 | Network training method, image processing method, device, equipment and storage medium |
| CN112990196A (en) * | 2021-03-16 | 2021-06-18 | 北京大学 | Scene character recognition method and system based on hyper-parameter search and two-stage training |
| CN112990196B (en) * | 2021-03-16 | 2023-10-24 | 北京大学 | Scene text recognition method and system based on super-parameter search and two-stage training |
| CN113095477A (en) * | 2021-03-17 | 2021-07-09 | 西安理工大学 | Wind power prediction method based on DE-BP neural network |
| CN113095477B (en) * | 2021-03-17 | 2024-02-09 | 西安理工大学 | Wind power prediction method based on DE-BP neural network |
| CN113113030A (en) * | 2021-03-22 | 2021-07-13 | 浙江大学 | High-dimensional damaged data wireless transmission method based on noise reduction self-encoder |
| CN113113030B (en) * | 2021-03-22 | 2022-03-22 | 浙江大学 | High-dimensional damaged data wireless transmission method based on noise reduction self-encoder |
| CN112861386A (en) * | 2021-03-23 | 2021-05-28 | 辽宁工程技术大学 | IACO-BRBP neural network-based Koch snowflake type ultra-wideband antenna reverse modeling method |
| CN112946784B (en) * | 2021-03-29 | 2022-09-16 | 杭州电子科技大学 | A Deep Learning-Based Short-Term Forecast Method for SuperDARN Radar Convective Maps |
| CN112946784A (en) * | 2021-03-29 | 2021-06-11 | 杭州电子科技大学 | SuperDARN radar convection diagram short-term forecasting method based on deep learning |
| CN113189983A (en) * | 2021-04-13 | 2021-07-30 | 中国人民解放军国防科技大学 | Open scene-oriented multi-robot cooperative multi-target sampling method |
| CN113189983B (en) * | 2021-04-13 | 2022-05-31 | 中国人民解放军国防科技大学 | Open scene-oriented multi-robot cooperative multi-target sampling method |
| CN113076699A (en) * | 2021-04-22 | 2021-07-06 | 西安交通大学 | Antenna optimization method based on multi-output Gaussian process Bayesian optimization |
| CN113128124A (en) * | 2021-04-23 | 2021-07-16 | 东北大学 | Multi-grade C-Mn steel mechanical property prediction method based on improved neural network |
| CN113128124B (en) * | 2021-04-23 | 2023-09-15 | 东北大学 | Multi-grade C-Mn steel mechanical property prediction method based on improved neural network |
| CN114217944A (en) * | 2021-04-26 | 2022-03-22 | 无锡江南计算技术研究所 | A neural network-based dynamic load balancing method for model parallelism |
| CN113139624A (en) * | 2021-05-18 | 2021-07-20 | 南京大学 | Network user classification method based on machine learning |
| CN113313152A (en) * | 2021-05-19 | 2021-08-27 | 北京大学 | Image classification method based on optimization-induced equilibrium neural network model |
| CN113313152B (en) * | 2021-05-19 | 2023-09-22 | 北京大学 | Image classification method based on balanced neural network model of optimization induction |
| CN113470316A (en) * | 2021-05-24 | 2021-10-01 | 电子科技大学 | Debris flow monitoring and early warning method based on self-coding single classification model |
| CN113470316B (en) * | 2021-05-24 | 2022-03-25 | 电子科技大学 | A Debris Flow Monitoring and Early Warning Method Based on Self-encoding Single Classification Model |
| CN113486923A (en) * | 2021-06-02 | 2021-10-08 | 长安大学 | Dynamic width-depth neural network for hyperspectral classification and learning method thereof |
| CN113467766A (en) * | 2021-06-15 | 2021-10-01 | 江苏大学 | Method for determining optimal neural network input vector length |
| CN113469426A (en) * | 2021-06-23 | 2021-10-01 | 国网山东省电力公司东营供电公司 | Photovoltaic output power prediction method and system based on improved BP neural network |
| CN113505477B (en) * | 2021-06-29 | 2022-05-20 | 西北师范大学 | SVAE-WGAN-based soft sensing data supplement method for process industry |
| CN113505477A (en) * | 2021-06-29 | 2021-10-15 | 西北师范大学 | Process industry soft measurement data supplementing method based on SVAE-WGAN |
| CN113407680A (en) * | 2021-06-30 | 2021-09-17 | 竹间智能科技(上海)有限公司 | Heterogeneous integrated model screening method and electronic equipment |
| CN113420508B (en) * | 2021-07-07 | 2024-02-27 | 华北电力大学 | Unit combination calculation method based on LSTM |
| CN113420508A (en) * | 2021-07-07 | 2021-09-21 | 华北电力大学 | Unit combination calculation method based on LSTM |
| CN113569467A (en) * | 2021-07-12 | 2021-10-29 | 鞍钢股份有限公司 | Blast furnace tuyere raceway temperature soft measurement modeling method based on width learning |
| CN113569467B (en) * | 2021-07-12 | 2023-11-10 | 鞍钢股份有限公司 | Soft measurement modeling method for temperature of blast furnace tuyere swirling zone based on width learning |
| CN113447021A (en) * | 2021-07-15 | 2021-09-28 | 北京理工大学 | MEMS inertial navigation system positioning enhancement method based on LSTM neural network model |
| CN113721212A (en) * | 2021-07-16 | 2021-11-30 | 西安电子科技大学 | Radar scattering cross section confidence evaluation method based on neural network |
| CN113721212B (en) * | 2021-07-16 | 2023-06-30 | 西安电子科技大学 | Radar scattering section confidence assessment method based on neural network |
| CN113570040A (en) * | 2021-07-27 | 2021-10-29 | 南京航空航天大学 | A multi-domain action sequence intelligent optimization system and method based on evolution strategy |
| CN113539517A (en) * | 2021-08-05 | 2021-10-22 | 浙江大学 | Prediction method of time sequence intervention effect |
| CN113539517B (en) * | 2021-08-05 | 2024-04-16 | 浙江大学 | Prediction Methods of Time-Sequence Intervention Effects |
| CN113612555A (en) * | 2021-08-19 | 2021-11-05 | 江西赛柏科技有限公司 | Intelligent calibration algorithm and system based on mobile terminal wireless radio frequency signal intensity |
| CN113689928A (en) * | 2021-08-24 | 2021-11-23 | 平安国际智慧城市科技股份有限公司 | Recommendation method, device, equipment and storage medium for maintaining and preventing disease scheme |
| CN113689928B (en) * | 2021-08-24 | 2023-06-20 | 深圳平安智慧医健科技有限公司 | Recommended method, apparatus, device and storage medium for maintenance and prevention of illness |
| CN113919410B (en) * | 2021-08-25 | 2024-05-24 | 桂林电子科技大学 | An incremental image classification method and system |
| CN113919410A (en) * | 2021-08-25 | 2022-01-11 | 桂林电子科技大学 | An incremental image classification method and system |
| CN113779871A (en) * | 2021-08-26 | 2021-12-10 | 清华大学 | Electric heating coupling system scheduling method and device, electronic equipment and storage medium thereof |
| CN113723692A (en) * | 2021-09-02 | 2021-11-30 | 深圳前海微众银行股份有限公司 | Data processing method, apparatus, device, medium, and program product |
| CN113571133A (en) * | 2021-09-14 | 2021-10-29 | 内蒙古农业大学 | A prediction method of lactic acid bacteria antimicrobial peptides based on graph neural network |
| CN113780432A (en) * | 2021-09-15 | 2021-12-10 | 中国人民解放军国防大学联合作战学院 | An intelligent detection method for network information system operation and maintenance anomalies based on reinforcement learning |
| CN113780432B (en) * | 2021-09-15 | 2022-10-28 | 中国人民解放军国防大学联合作战学院 | Intelligent detection method for operation and maintenance abnormity of network information system based on reinforcement learning |
| CN113887797A (en) * | 2021-09-29 | 2022-01-04 | 南京用能互联网研究院有限公司 | Building electric heating load prediction model establishing method, device and equipment |
| CN113947185A (en) * | 2021-09-30 | 2022-01-18 | 北京达佳互联信息技术有限公司 | Task processing network generation method, task processing device, electronic equipment and storage medium |
| CN115988542A (en) * | 2021-10-13 | 2023-04-18 | 纬创资通股份有限公司 | Network construction system and method, recording medium, program product, and control system |
| CN113971992B (en) * | 2021-10-26 | 2024-03-29 | 中国科学技术大学 | Self-supervised pre-training method and system for molecular attribute prediction graph network |
| CN113971992A (en) * | 2021-10-26 | 2022-01-25 | 中国科学技术大学 | Self-supervision pre-training method and system for molecular attribute prediction graph network |
| CN113915249B (en) * | 2021-10-29 | 2023-11-24 | 江苏大学 | Wet clutch status identification method based on fuzzy neural network |
| CN113915249A (en) * | 2021-10-29 | 2022-01-11 | 江苏大学 | Wet clutch state identification method based on fuzzy neural network |
| CN113988437A (en) * | 2021-11-01 | 2022-01-28 | 东南大学 | An Adaptive Interval Forecasting Method for Short-Term Residential Load |
| CN114048564A (en) * | 2021-11-05 | 2022-02-15 | 广东海洋大学 | Engineering forklift multi-target performance optimization method based on deep proxy model |
| CN114046815A (en) * | 2021-11-09 | 2022-02-15 | 上海精赋达传感技术有限公司 | Encoder self-correction method and device based on deep learning |
| CN114046815B (en) * | 2021-11-09 | 2024-02-09 | 上海精赋达传感技术有限公司 | Encoder self-correction method and device based on deep learning |
| CN113890633B (en) * | 2021-11-11 | 2023-10-20 | 中国船舶重工集团公司第七一五研究所 | Underwater acoustic communication system self-adaptive selection method based on deep neural network |
| CN113890633A (en) * | 2021-11-11 | 2022-01-04 | 中国船舶重工集团公司第七一五研究所 | Underwater acoustic communication system self-adaptive selection method based on deep neural network |
| CN114118339A (en) * | 2021-11-12 | 2022-03-01 | 吉林大学 | Radio modulation signal identification and classification method based on cuckoo algorithm improved ResNet |
| CN114118339B (en) * | 2021-11-12 | 2024-05-14 | 吉林大学 | Radio modulation signal identification and classification method based on improvement ResNet of cuckoo algorithm |
| CN114219069A (en) * | 2021-11-16 | 2022-03-22 | 北京工业大学 | Brain effect connection network learning method based on automatic variation self-encoder |
| CN114219069B (en) * | 2021-11-16 | 2024-04-26 | 北京工业大学 | Brain effect connection network learning method based on automatic variation self-encoder |
| CN114139720A (en) * | 2021-11-16 | 2022-03-04 | 广西中科曙光云计算有限公司 | Method and device for processing big data of government affairs based on machine learning |
| CN114065635A (en) * | 2021-11-22 | 2022-02-18 | 中国民航大学 | Aircraft ground air conditioner energy consumption prediction method based on IALO-AM-BilSTM model and storage medium |
| CN114114126A (en) * | 2021-11-22 | 2022-03-01 | 浙江华采科技有限公司 | Digital calibration method and system for low-permeability magnetic powder core current transformer |
| CN114462292A (en) * | 2021-11-22 | 2022-05-10 | 广州大学 | Circuit parameter optimization method based on SQP algorithm |
| CN114118569A (en) * | 2021-11-24 | 2022-03-01 | 宁波力斗智能技术有限公司 | Wind power multi-step prediction method based on multi-mode multi-task Transformer network |
| CN114358278A (en) * | 2021-11-26 | 2022-04-15 | 北京邮电大学 | Training method and device for neural network model |
| CN114386364A (en) * | 2021-12-06 | 2022-04-22 | 西安交通大学 | Quality prediction and process parameter optimization method for electronic surface mount soldering |
| CN114169605A (en) * | 2021-12-06 | 2022-03-11 | 西安交通大学 | Physical field and performance collaborative prediction method suitable for pump and hydraulic turbine |
| CN114386364B (en) * | 2021-12-06 | 2024-01-19 | 西安交通大学 | Electronic surface mounting welding quality prediction and process parameter optimization method |
| CN114169240A (en) * | 2021-12-08 | 2022-03-11 | 中国石油大学(北京) | MMP prediction method and device based on conditional generative adversarial network |
| CN114268129B (en) * | 2021-12-21 | 2024-02-02 | 湖州越球电机有限公司 | Wind power frequency modulation gradual inertia control method based on deep neural network |
| CN114268129A (en) * | 2021-12-21 | 2022-04-01 | 湖州越球电机有限公司 | Wind power frequency modulation gradual inertia control method based on deep neural network |
| CN114190944A (en) * | 2021-12-23 | 2022-03-18 | 上海交通大学 | Robust emotion recognition method based on electroencephalogram signals |
| CN114280491A (en) * | 2021-12-23 | 2022-04-05 | 中山大学 | A method for estimating the remaining capacity of retired batteries based on active learning |
| CN114417497A (en) * | 2021-12-23 | 2022-04-29 | 中国人民解放军军事科学院国防科技创新研究院 | Neural network initialization design method based on meta-learning embedded physical knowledge |
| CN114190944B (en) * | 2021-12-23 | 2023-08-22 | 上海交通大学 | Robust emotion recognition method based on electroencephalogram signals |
| CN114280491B (en) * | 2021-12-23 | 2024-01-05 | 中山大学 | Retired battery residual capacity estimation method based on active learning |
| CN114266953A (en) * | 2021-12-24 | 2022-04-01 | 福州大学 | Target detection process self-optimization method based on deep learning |
| CN114217688A (en) * | 2021-12-24 | 2022-03-22 | 中国科学技术大学 | A system and method for NPU power consumption optimization based on neural network structure |
| CN114236610A (en) * | 2021-12-24 | 2022-03-25 | 西安交通大学 | Iterative seismic data unmixing method and system based on depth convolution network prior |
| CN114282725A (en) * | 2021-12-24 | 2022-04-05 | 山东大学 | Construction of transient oil reservoir agent model based on deep learning and oil reservoir prediction method |
| CN114266953B (en) * | 2021-12-24 | 2024-06-11 | 福州大学 | Target detection flow self-optimization method based on deep learning |
| CN114217688B (en) * | 2021-12-24 | 2024-05-10 | 中国科学技术大学 | NPU power consumption optimization system and method based on neural network structure |
| CN114215706A (en) * | 2021-12-27 | 2022-03-22 | 南京邮电大学 | Wind turbine generator blade cracking fault early warning method and device |
| CN114215706B (en) * | 2021-12-27 | 2024-02-20 | 南京邮电大学 | Early warning method and device for cracking faults of wind turbine generator blades |
| CN114492941A (en) * | 2021-12-29 | 2022-05-13 | 东南大学 | Whole-county photovoltaic prediction method based on cluster division and data enhancement |
| CN114519381A (en) * | 2021-12-31 | 2022-05-20 | 上海仙途智能科技有限公司 | Sensing method and device based on multitask learning network, storage medium and terminal |
| FR3136299A1 (en) | 2022-01-04 | 2023-12-08 | Alcom Technologies | Method for optimizing the hyperparameters of a machine learning model |
| CN114330456A (en) * | 2022-01-05 | 2022-04-12 | 北京邮电大学 | A Multi-task Learning-Based Optical Path Transmission Quality Prediction Method |
| CN114399119A (en) * | 2022-01-18 | 2022-04-26 | 中国石油大学(北京) | MMP prediction method and device based on conditional convolution generative adversarial network |
| CN114462309B (en) * | 2022-01-19 | 2023-11-17 | 苏州浪潮智能科技有限公司 | A physical simulation model training method, device, equipment and storage medium |
| CN114462309A (en) * | 2022-01-19 | 2022-05-10 | 苏州浪潮智能科技有限公司 | A physical simulation model training method, device, equipment and storage medium |
| CN114417011A (en) * | 2022-01-20 | 2022-04-29 | 北京北大英华科技有限公司 | Inspection service portrait knowledge fusion method and device, electronic equipment and storage medium |
| CN114492624A (en) * | 2022-01-21 | 2022-05-13 | 中国电子科技集团公司第三十六研究所 | Small sample learning method and device based on meta-learning and electronic equipment |
| CN114120289B (en) * | 2022-01-25 | 2022-05-03 | 中科视语(北京)科技有限公司 | Method and system for identifying driving area and lane line |
| CN114120289A (en) * | 2022-01-25 | 2022-03-01 | 中科视语(北京)科技有限公司 | Method and system for identifying driving area and lane line |
| CN114492279B (en) * | 2022-01-27 | 2024-01-19 | 西安交通大学 | Parameter optimization method and system for analog integrated circuit |
| CN114492279A (en) * | 2022-01-27 | 2022-05-13 | 西安交通大学 | A method and system for parameter optimization of an analog integrated circuit |
| CN114565084A (en) * | 2022-03-02 | 2022-05-31 | 安徽大学 | Computer self-adaptive evaluation method based on learning question selection strategy |
| CN114584150A (en) * | 2022-03-04 | 2022-06-03 | 南京邮电大学 | Sensor network signal completion method realized by utilizing deep neural network |
| CN115018038B (en) * | 2022-03-14 | 2024-03-05 | 贵州大学 | A network interface traffic prediction method based on the Sparrow optimization width learning system |
| CN115018038A (en) * | 2022-03-14 | 2022-09-06 | 贵州大学 | Network interface flow prediction method based on sparrow optimization width learning system |
| CN114630207B (en) * | 2022-03-15 | 2023-04-07 | 浙江大学 | Multi-sensing-node sensing data collection method based on noise reduction self-encoder |
| CN114630207A (en) * | 2022-03-15 | 2022-06-14 | 浙江大学 | Multi-sensor node sensing data collection method based on noise reduction autoencoder |
| CN114692347A (en) * | 2022-03-16 | 2022-07-01 | 中国人民解放军军事科学院国防科技创新研究院 | Temperature field proxy model construction method based on neural network architecture search technology |
| CN114842328B (en) * | 2022-03-22 | 2024-03-22 | 西北工业大学 | Hyperspectral change detection method based on collaborative analysis of autonomous sensing network structure |
| CN114943162A (en) * | 2022-03-22 | 2022-08-26 | 北京交通大学 | Periodic super-barrier reverse design method based on deep learning |
| CN114842328A (en) * | 2022-03-22 | 2022-08-02 | 西北工业大学 | Hyperspectral change detection method based on cooperative analysis autonomous sensing network structure |
| CN114611823B (en) * | 2022-03-23 | 2022-11-08 | 特斯联科技集团有限公司 | Optimized dispatching method and system for electricity-cold-heat-gas multi-energy-demand typical park |
| CN114611823A (en) * | 2022-03-23 | 2022-06-10 | 特斯联科技集团有限公司 | Optimized dispatching method and system for electricity-cold-heat-gas multi-energy-demand typical park |
| CN114676909A (en) * | 2022-03-25 | 2022-06-28 | 东南大学 | A charging path planning method for unmanned vehicles based on deep reinforcement learning |
| CN114676909B (en) * | 2022-03-25 | 2024-04-09 | 东南大学 | Unmanned vehicle charging path planning method based on deep reinforcement learning |
| CN114912342A (en) * | 2022-03-28 | 2022-08-16 | 南京邮电大学 | Packaging lead bonding process parameter optimization method based on multiple quality parameters |
| CN114676522B (en) * | 2022-03-28 | 2024-03-29 | 西安交通大学 | Aerodynamic shape optimization design method, system and equipment integrating GAN and transfer learning |
| CN114676522A (en) * | 2022-03-28 | 2022-06-28 | 西安交通大学 | Aerodynamic shape optimization design method, system and equipment integrating GAN and transfer learning |
| CN114818477B (en) * | 2022-03-31 | 2024-12-06 | 上海大学 | Aircraft aerodynamic characteristics prediction method, device and medium |
| CN114676639A (en) * | 2022-03-31 | 2022-06-28 | 上海大学 | Aircraft aerodynamic shape optimization method, device and medium based on neural network |
| CN114742289A (en) * | 2022-03-31 | 2022-07-12 | 大连理工大学 | Gaussian process robust optimization method for production process parameters |
| CN114818477A (en) * | 2022-03-31 | 2022-07-29 | 上海大学 | Method, device and medium for predicting aerodynamic characteristics of aircraft |
| CN114647950A (en) * | 2022-04-02 | 2022-06-21 | 西安交通大学 | Variable-density high-vacuum multi-layer heat insulation structure optimization method based on proxy model |
| CN114757096A (en) * | 2022-04-02 | 2022-07-15 | 华南理工大学 | Bridge temperature prediction method, device, equipment and medium based on NARX neural network |
| CN114647950B (en) * | 2022-04-02 | 2024-02-20 | 西安交通大学 | An optimization method for variable density high vacuum multi-layer insulation structure based on proxy model |
| CN114757096B (en) * | 2022-04-02 | 2024-03-15 | 华南理工大学 | Bridge temperature prediction method, device, equipment and medium based on NARX neural network |
| CN114692504A (en) * | 2022-04-08 | 2022-07-01 | 东南大学 | An automatic synthesis method for variable spiral inductors on millimeter-wave chips |
| CN115047894B (en) * | 2022-04-14 | 2023-09-15 | 中国民用航空总局第二研究所 | Unmanned aerial vehicle track measuring and calculating method, electronic equipment and storage medium |
| CN115047894A (en) * | 2022-04-14 | 2022-09-13 | 中国民用航空总局第二研究所 | Unmanned aerial vehicle track measuring and calculating method, electronic equipment and storage medium |
| CN114839439A (en) * | 2022-04-15 | 2022-08-02 | 厦门明翰电气股份有限公司 | Harmonic current detection method for brain emotion learning based on PID optimization |
| CN117004432A (en) * | 2022-04-26 | 2023-11-07 | 中国石油化工股份有限公司 | Data analysis methods, devices, media and equipment for feed oil hydrogenation units |
| CN115062501A (en) * | 2022-04-27 | 2022-09-16 | 南方科技大学 | Chip packaging design optimization method based on adaptive subproblem selection strategy |
| CN114966233A (en) * | 2022-05-16 | 2022-08-30 | 国网电力科学研究院武汉南瑞有限责任公司 | Lightning forecast system and method based on deep neural network |
| CN115049108A (en) * | 2022-05-20 | 2022-09-13 | 支付宝(杭州)信息技术有限公司 | Multitask model training method, multitask prediction method, related device and medium |
| CN115018137A (en) * | 2022-05-23 | 2022-09-06 | 深圳市广汇源环境水务有限公司 | Water environment model parameter calibration method based on reinforcement learning |
| CN114819105A (en) * | 2022-05-23 | 2022-07-29 | 吉林大学 | A Neural Network-based Suspension Model Construction Method for Air Springs and Continuously Adjustable Shock Absorbers |
| CN115018137B (en) * | 2022-05-23 | 2024-06-11 | 深圳市广汇源环境水务有限公司 | Water environment model parameter calibration method based on reinforcement learning |
| CN114997297B (en) * | 2022-05-26 | 2024-05-03 | 哈尔滨工业大学 | Target movement intention reasoning method and system based on multi-level region division |
| CN114970819A (en) * | 2022-05-26 | 2022-08-30 | 哈尔滨工业大学 | A moving target search and tracking method and system based on intent reasoning and deep reinforcement learning |
| CN114970819B (en) * | 2022-05-26 | 2024-05-03 | 哈尔滨工业大学 | Moving target searching and tracking method and system based on intention reasoning and deep reinforcement learning |
| CN114997297A (en) * | 2022-05-26 | 2022-09-02 | 哈尔滨工业大学 | A method and system for inference of target motion intent based on multi-level region division |
| CN114997491A (en) * | 2022-05-30 | 2022-09-02 | 武汉工程大学 | Method, device, equipment and storage medium for optimizing gasoline octane number loss |
| WO2023231244A1 (en) * | 2022-06-02 | 2023-12-07 | 大连理工大学 | Fsw tensile strength intelligent prediction method considering welding temperature and upsetting force |
| CN114707471A (en) * | 2022-06-06 | 2022-07-05 | 浙江大学 | Artificial intelligent courseware making method and device based on hyper-parameter evaluation graph algorithm |
| CN114707471B (en) * | 2022-06-06 | 2022-09-09 | 浙江大学 | Artificial intelligence courseware making method and device based on hyperparameter evaluation graph algorithm |
| CN114966451B (en) * | 2022-06-06 | 2024-04-26 | 北京航空航天大学 | Power supply product residual service life online prediction method based on physical information neural network |
| CN114966451A (en) * | 2022-06-06 | 2022-08-30 | 北京航空航天大学 | Power supply product residual service life online prediction method based on physical information neural network |
| CN114943182A (en) * | 2022-06-07 | 2022-08-26 | 清华大学深圳国际研究生院 | Robot cable shape control method and device based on graph neural network |
| CN114943182B (en) * | 2022-06-07 | 2024-08-13 | 清华大学深圳国际研究生院 | Robot cable shape control method and equipment based on graph neural network |
| CN114997060A (en) * | 2022-06-10 | 2022-09-02 | 湖南大学 | A phononic crystal time-varying reliability testing method, computing device and storage medium |
| CN114997373A (en) * | 2022-06-14 | 2022-09-02 | 中国人民解放军国防科技大学 | Network depth dynamic adjustment method, apparatus, computer equipment and storage medium |
| CN114997373B (en) * | 2022-06-14 | 2024-10-15 | 中国人民解放军国防科技大学 | Network depth dynamic adjustment method, device, computer equipment and storage medium |
| CN115099146A (en) * | 2022-06-27 | 2022-09-23 | 上海集成电路装备材料产业创新中心有限公司 | Circuit generation method, circuit generation device, electronic equipment and storage medium |
| CN115015869A (en) * | 2022-06-27 | 2022-09-06 | 清华大学 | Can learn low frequency broadband radar target parameter estimation method, equipment and program product |
| CN115237878A (en) * | 2022-06-28 | 2022-10-25 | 华中科技大学 | Process database construction method and medium based on additive manufacturing |
| CN115018607B (en) * | 2022-07-01 | 2023-01-24 | 吉林工程技术师范学院 | Accounting data processing method and system based on artificial intelligence |
| CN115238602B (en) * | 2022-07-01 | 2023-07-11 | 中国海洋大学 | Prediction method of contribution rate of wave-induced seabed transient liquefaction to resuspension |
| CN115238602A (en) * | 2022-07-01 | 2022-10-25 | 中国海洋大学 | CNN-LSTM-based prediction method for contribution rate of transient liquefaction of wave-induced seabed to resuspension |
| CN115146536B (en) * | 2022-07-01 | 2024-10-18 | 纽威数控装备(苏州)股份有限公司 | Heat source contribution degree analysis method based on wavelet neural network and particle swarm algorithm |
| CN115018607A (en) * | 2022-07-01 | 2022-09-06 | 吉林工程技术师范学院 | Accounting data processing method and system based on artificial intelligence |
| CN115146536A (en) * | 2022-07-01 | 2022-10-04 | 纽威数控装备(苏州)股份有限公司 | Wavelet neural network and particle swarm algorithm-based heat source contribution analysis method |
| CN115115052A (en) * | 2022-08-11 | 2022-09-27 | 杭州电子科技大学 | An Adaptive Distributed Parallel Training Method of Neural Network Based on Genetic Algorithm |
| CN115205349A (en) * | 2022-08-12 | 2022-10-18 | 南京理工大学 | A Deep Neural Network-Based Interferogram Automatic Registration and Phase Unpacking Method |
| CN115222023A (en) * | 2022-08-15 | 2022-10-21 | 西北工业大学 | Non-cooperative target control strategy identification method and system based on neural network |
| CN115222023B (en) * | 2022-08-15 | 2025-09-05 | 西北工业大学 | Non-cooperative target control strategy identification method and system based on neural network |
| CN115412332A (en) * | 2022-08-25 | 2022-11-29 | 暨南大学 | Internet of things intrusion detection system and method based on hybrid neural network model optimization |
| CN115412332B (en) * | 2022-08-25 | 2024-06-07 | 暨南大学 | Internet of Things Intrusion Detection System and Method Based on Hybrid Neural Network Model Optimization |
| CN115468633B (en) * | 2022-09-01 | 2025-09-12 | 西安热工研究院有限公司 | Calibration method and device for multi-point flowmeter based on RBF neural network |
| CN115468633A (en) * | 2022-09-01 | 2022-12-13 | 西安热工研究院有限公司 | Multi-point flowmeter correction method and device based on RBF neural network |
| CN115481790A (en) * | 2022-09-02 | 2022-12-16 | 广东省科学院生态环境与土壤研究所 | Carbon-based geological catalytic material fixed cadmium and methane emission reduction cooperative prediction method, device and medium |
| CN115549732B (en) * | 2022-09-13 | 2024-05-28 | 哈尔滨工业大学 | A large-scale MIMO hybrid precoding method, device and storage medium based on artificial intelligence |
| CN115549732A (en) * | 2022-09-13 | 2022-12-30 | 哈尔滨工业大学 | An Artificial Intelligence-Based Massive MIMO Hybrid Precoding Method |
| CN115730509A (en) * | 2022-09-20 | 2023-03-03 | 中国人民解放军军事科学院国防科技创新研究院 | Spacecraft in-cabin temperature field reconstruction task research benchmark method based on machine learning |
| CN115455782A (en) * | 2022-09-22 | 2022-12-09 | 吉林大学 | Bionic self-similar hierarchical structure performance prediction method based on machine learning |
| CN115600667A (en) * | 2022-09-22 | 2023-01-13 | 浙江大学(Cn) | A method and device for modeling multi-dimensional state change time series data of a system |
| CN115240394A (en) * | 2022-09-22 | 2022-10-25 | 国网湖北省电力有限公司经济技术研究院 | Method and system for monitoring and early warning of water level in accident oil pool in substation |
| CN115512172A (en) * | 2022-09-30 | 2022-12-23 | 赵营鸽 | Uncertainty quantification method for multi-dimensional parameters in electrical impedance imaging technology |
| CN115512172B (en) * | 2022-09-30 | 2023-09-15 | 赵营鸽 | Uncertainty quantification method for multidimensional parameters in electrical impedance imaging technology |
| CN115542723A (en) * | 2022-10-08 | 2022-12-30 | 华南理工大学 | Parameter setting method for ankle exoskeleton PD controller for man-machine in-loop debugging |
| CN115688565A (en) * | 2022-10-13 | 2023-02-03 | 阿里巴巴(中国)有限公司 | Dynamic optimization method of network model, image processing method and equipment |
| CN115762674A (en) * | 2022-10-20 | 2023-03-07 | 西安交通大学 | Rapid solving method for inverse heat transfer problem of nano heat insulation composite material |
| CN115758191A (en) * | 2022-10-31 | 2023-03-07 | 北京理工大学 | Knowledge service entity clustering number prediction method based on deep learning |
| CN116306206A (en) * | 2022-11-06 | 2023-06-23 | 西北工业大学 | Airfoil transonic buffeting flow field rapid prediction method based on deep neural network |
| CN115660786A (en) * | 2022-11-09 | 2023-01-31 | 中国工商银行股份有限公司 | Service recommendation method, apparatus, computer equipment, storage medium and program product |
| WO2024098522A1 (en) * | 2022-11-09 | 2024-05-16 | 广东海洋大学 | Machine learning-based multi-performance collaborative optimization high-entropy alloy component design method |
| CN115796346A (en) * | 2022-11-22 | 2023-03-14 | 烟台国工智能科技有限公司 | Yield optimization method and system and non-transitory computer readable storage medium |
| CN115759200A (en) * | 2022-11-22 | 2023-03-07 | 西北工业大学 | A key path construction method for deep network based on iterative optimization of gate parameters |
| CN115935626B (en) * | 2022-11-25 | 2023-09-08 | 河南大学 | Inversion method of river water-groundwater vertical transient interactive flow flux |
| CN115935626A (en) * | 2022-11-25 | 2023-04-07 | 河南大学 | Inversion method of river water-underground water vertical transient interaction water flux based on VAE-PINN algorithm |
| CN115935813A (en) * | 2022-11-28 | 2023-04-07 | 上海大学 | A device multi-stage degradation evolution prediction method, device and storage medium based on statistical model and deep learning |
| CN115757384A (en) * | 2022-11-30 | 2023-03-07 | 安徽长正智库管理咨询有限公司 | A big data-based government data processing method |
| CN115579093A (en) * | 2022-12-08 | 2023-01-06 | 科丰兴泰(杭州)生物科技有限公司 | Method and system for designing nitrification inhibitor slow-release material based on deep learning |
| CN116016223A (en) * | 2022-12-09 | 2023-04-25 | 国网湖北省电力有限公司信息通信公司 | Data transmission optimization method for data center network |
| CN116016223B (en) * | 2022-12-09 | 2024-02-02 | 国网湖北省电力有限公司信息通信公司 | Data transmission optimization method for data center network |
| CN115846852A (en) * | 2022-12-13 | 2023-03-28 | 大连交通大学 | FSW welding monitoring method based on DS evidence theory and neural network |
| CN116227671A (en) * | 2022-12-27 | 2023-06-06 | 山西省能源互联网研究院 | Optimization Method of Wind Power Joint Forecasting Model Based on Long Short-term Memory Neural Network |
| CN115713099A (en) * | 2023-01-03 | 2023-02-24 | 税友信息技术有限公司 | Model design method, device, equipment and storage medium |
| CN115713099B (en) * | 2023-01-03 | 2023-05-09 | 税友信息技术有限公司 | Model design method, device, equipment and storage medium |
| CN116204998A (en) * | 2023-01-09 | 2023-06-02 | 华中科技大学 | SLM forming performance prediction and process parameter optimization method and system |
| CN116204998B (en) * | 2023-01-09 | 2025-07-29 | 华中科技大学 | SLM forming performance prediction and process parameter optimization method and system |
| CN116343930A (en) * | 2023-01-10 | 2023-06-27 | 西北工业大学 | Metabolic dynamics and toxicity prediction method based on graph representation multitask learning |
| CN116011337A (en) * | 2023-01-11 | 2023-04-25 | 西北工业大学 | An Efficient Engineering Design Method Against Numerical Noise |
| CN116339136A (en) * | 2023-01-18 | 2023-06-27 | 浙江大学 | An actuator control method and controller based on online neural prediction |
| CN115809624A (en) * | 2023-02-01 | 2023-03-17 | 北京大学 | Automatic analysis design method for microstrip line transmission line of integrated circuit |
| CN116150995B (en) * | 2023-02-21 | 2023-07-25 | 东南大学 | Rapid simulation method of switch arc model |
| CN116150995A (en) * | 2023-02-21 | 2023-05-23 | 东南大学 | Rapid simulation method of switch arc model |
| CN116380321A (en) * | 2023-02-24 | 2023-07-04 | 武汉科技大学 | Method for Decoupling Decoupling Model of Tire-Road Contact Triaxial Force Sensor |
| CN116449294A (en) * | 2023-03-27 | 2023-07-18 | 上海海事大学 | A real-time underwater acoustic positioning method based on IDE-BP neural network |
| CN116321008A (en) * | 2023-04-13 | 2023-06-23 | 重庆邮电大学 | Indoor position fingerprint positioning method based on triplet self-encoder |
| CN116680628A (en) * | 2023-05-11 | 2023-09-01 | 苏州大学 | A volatile liquid identification module and method based on SVM electronic nose system |
| CN116644662B (en) * | 2023-05-19 | 2024-03-29 | 之江实验室 | Well-arrangement optimization method based on knowledge embedded neural network proxy model |
| CN116644662A (en) * | 2023-05-19 | 2023-08-25 | 之江实验室 | A well layout optimization method based on knowledge embedded neural network agent model |
| CN116499607A (en) * | 2023-06-27 | 2023-07-28 | 之江实验室 | Optical fiber sensing signal noise reduction method and device and storage medium |
| CN116499607B (en) * | 2023-06-27 | 2023-09-08 | 之江实验室 | Optical fiber sensing signal noise reduction method and device and storage medium |
| CN116502566A (en) * | 2023-06-27 | 2023-07-28 | 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) | Multi-objective optimization method for performance of combustion chamber of gas turbine based on Bayesian optimization |
| CN116775687A (en) * | 2023-07-07 | 2023-09-19 | 哈尔滨工程大学 | Time sequence prediction closed-loop framework based on digital-analog symbiosis |
| CN117272778B (en) * | 2023-07-12 | 2024-03-12 | 上海交通大学 | Design method and device of microwave passive device |
| CN117272778A (en) * | 2023-07-12 | 2023-12-22 | 上海交通大学 | Design methods and devices for microwave passive components |
| CN116611349A (en) * | 2023-07-18 | 2023-08-18 | 华东交通大学 | Method and system for optimizing process parameters of drum silk drying based on neural network |
| CN116611349B (en) * | 2023-07-18 | 2023-10-10 | 华东交通大学 | Optimization method and system of drum drying process parameters based on neural network |
| CN116913435B (en) * | 2023-07-27 | 2024-01-26 | 常州威材新材料科技有限公司 | High-strength engineering plastic evaluation method and system based on component analysis |
| CN116913435A (en) * | 2023-07-27 | 2023-10-20 | 常州威材新材料科技有限公司 | High-strength engineering plastic evaluation method and system based on component analysis |
| CN116759031B (en) * | 2023-08-10 | 2024-01-26 | 内蒙古工业大学 | Design method of sludge ash concrete material mixing ratio based on ANN |
| CN116759031A (en) * | 2023-08-10 | 2023-09-15 | 内蒙古工业大学 | Design method of sludge ash concrete material mixing ratio based on ANN |
| CN117037971A (en) * | 2023-08-21 | 2023-11-10 | 电子科技大学 | Self-calibration design method for multilayer honeycomb wave-absorbing material based on neural network |
| CN116992298B (en) * | 2023-09-28 | 2023-12-22 | 青岛东捷建设集团有限公司 | High-pressure jet grouting method and system based on seawater seepage |
| CN116992298A (en) * | 2023-09-28 | 2023-11-03 | 青岛东捷建设集团有限公司 | High-pressure jet grouting method and system based on seawater seepage |
| CN117056847A (en) * | 2023-10-10 | 2023-11-14 | 中南大学 | Method, system, equipment and storage medium for detecting abnormality of stream data |
| CN117056847B (en) * | 2023-10-10 | 2024-01-30 | 中南大学 | Anomaly detection method, system, equipment and storage medium for streaming data |
| CN117093871B (en) * | 2023-10-16 | 2024-02-13 | 之江实验室 | Deep learning-oriented distributed training evaluation method and system |
| CN117093871A (en) * | 2023-10-16 | 2023-11-21 | 之江实验室 | Deep learning-oriented distributed training evaluation method and system |
| CN117114141B (en) * | 2023-10-20 | 2024-02-27 | 安徽蔚来智驾科技有限公司 | Model training method, evaluation method, computer device and storage medium |
| CN117114141A (en) * | 2023-10-20 | 2023-11-24 | 安徽蔚来智驾科技有限公司 | Model training method, evaluation method, computer device and storage medium |
| CN117174219B (en) * | 2023-11-02 | 2024-02-02 | 济南大学 | Cement strength conditional probability distribution estimation method and system based on latent variable model |
| CN117174219A (en) * | 2023-11-02 | 2023-12-05 | 济南大学 | Cement strength conditional probability distribution estimation method and system based on latent variable model |
| CN117272839B (en) * | 2023-11-20 | 2024-02-06 | 北京阿迈特医疗器械有限公司 | Support press-holding performance prediction method and device based on neural network |
| CN117272839A (en) * | 2023-11-20 | 2023-12-22 | 北京阿迈特医疗器械有限公司 | Support press-holding performance prediction method and device based on neural network |
| CN117331339B (en) * | 2023-12-01 | 2024-02-06 | 南京华视智能科技股份有限公司 | Coating machine die head motor control method and device based on time sequence neural network model |
| CN117331339A (en) * | 2023-12-01 | 2024-01-02 | 南京华视智能科技股份有限公司 | Coating machine die head motor control method and device based on time sequence neural network model |
| CN117432414B (en) * | 2023-12-20 | 2024-03-19 | 中煤科工开采研究院有限公司 | Method and system for regulating and controlling top plate frosted jet flow seam formation |
| CN117432414A (en) * | 2023-12-20 | 2024-01-23 | 中煤科工开采研究院有限公司 | Method and system for regulating and controlling top plate frosted jet flow seam formation |
| CN117455320A (en) * | 2023-12-25 | 2024-01-26 | 江苏恒力化纤股份有限公司 | QGMVPN model-based multi-kettle polymerization process melt quality index online prediction method |
| CN117455320B (en) * | 2023-12-25 | 2024-03-19 | 江苏恒力化纤股份有限公司 | QGMVPN model-based multi-kettle polymerization process melt quality index online prediction method |
| CN117556877A (en) * | 2024-01-11 | 2024-02-13 | 西南交通大学 | Pulse neural network training method based on data pulse characteristic evaluation |
| CN117556877B (en) * | 2024-01-11 | 2024-04-02 | 西南交通大学 | Impulse neural network training method based on data impulse feature evaluation |
| CN117582185B (en) * | 2024-01-19 | 2024-05-07 | 季华实验室 | A pulse strength level prediction method based on CLLSR hybrid model |
| CN117582185A (en) * | 2024-01-19 | 2024-02-23 | 季华实验室 | A pulse force level prediction method based on CLLSR hybrid model |
| CN117647367B (en) * | 2024-01-29 | 2024-04-16 | 四川航空股份有限公司 | Machine learning-based method and system for positioning leakage points of aircraft fuel tank |
| CN117647367A (en) * | 2024-01-29 | 2024-03-05 | 四川航空股份有限公司 | Machine learning-based method and system for positioning leakage points of aircraft fuel tank |
| CN118131689A (en) * | 2024-02-19 | 2024-06-04 | 哈尔滨理工大学 | A method for predicting machine tool processing energy consumption based on Hyperopt optimized neural network |
| CN117875397B (en) * | 2024-03-11 | 2024-06-04 | 浪潮电子信息产业股份有限公司 | A method, device, computing device and storage medium for selecting parameters to be updated |
| CN117875397A (en) * | 2024-03-11 | 2024-04-12 | 浪潮电子信息产业股份有限公司 | A method, device, computing device and storage medium for selecting parameters to be updated |
| CN117894397B (en) * | 2024-03-15 | 2024-05-28 | 北京科技大学 | A method for predicting viscosity of continuous casting mold slag based on machine learning |
| CN117894397A (en) * | 2024-03-15 | 2024-04-16 | 北京科技大学 | A method for predicting viscosity of continuous casting mold slag based on machine learning |
| CN118298792A (en) * | 2024-04-08 | 2024-07-05 | 哈尔滨工程大学 | Ship sonar platform self-noise suppression method based on deep learning |
| CN118260840A (en) * | 2024-04-12 | 2024-06-28 | 湖南工程学院 | A method and system for generating a prefabricated area judgment model for an assembled building |
| CN118351556A (en) * | 2024-05-17 | 2024-07-16 | 优顶特技术有限公司 | Method, device, equipment and storage medium for classifying and identifying bills |
| CN118442745A (en) * | 2024-05-30 | 2024-08-06 | 华南农业大学 | Liquid nitrogen instant freezer temperature control method, device, equipment and medium |
| CN118747702A (en) * | 2024-06-06 | 2024-10-08 | 集美大学 | A mine inspection method and system based on reinforcement learning and machine vision |
| CN118538358A (en) * | 2024-06-07 | 2024-08-23 | 四川熠海瑞科技有限公司 | Immune system treatment information management platform |
| CN118333129A (en) * | 2024-06-07 | 2024-07-12 | 广东工业大学 | Identification model training method, nonlinear system identification method and system |
| CN118365183A (en) * | 2024-06-20 | 2024-07-19 | 西南石油大学 | A method and system for predicting oil well production based on hybrid neural network |
| CN118940612A (en) * | 2024-07-15 | 2024-11-12 | 华中科技大学 | Adapter Optimization Design Method Based on Parameter Optimization Neural Network |
| CN118469354A (en) * | 2024-07-15 | 2024-08-09 | 江西省赣地智慧科技有限公司 | Intelligent construction site data acquisition method and system for road construction |
| CN118569113A (en) * | 2024-08-05 | 2024-08-30 | 西安理工大学 | Optimization method and system for jet mechanism of impulse turbine |
| CN118915102A (en) * | 2024-08-22 | 2024-11-08 | 河南地矿职业学院 | Satellite clock error forecasting method and system |
| CN119168797A (en) * | 2024-08-30 | 2024-12-20 | 沈阳农业大学 | A method, system, device and medium for inversion of crop growth |
| CN119109455A (en) * | 2024-08-30 | 2024-12-10 | 合肥工业大学 | A pipeline ADC background calibration method based on ant colony optimization time-delay neural network |
| CN118778691A (en) * | 2024-09-10 | 2024-10-15 | 青岛哈尔滨工程大学创新发展中心 | A system and method for evaluating the hydrodynamic characteristics of a submarine-launched UAV during its entry and exit into water |
| CN118798061A (en) * | 2024-09-11 | 2024-10-18 | 中国海洋大学 | Door sealing model parameter determination method, device, equipment, medium and program product |
| CN118797354A (en) * | 2024-09-12 | 2024-10-18 | 青岛科技大学 | A method for determining parameters for improving the efficiency of methyl oleate epoxidation reaction |
| CN119127146A (en) * | 2024-09-19 | 2024-12-13 | 华南理工大学 | Liberty file generation method for SRAM memory compiler in 55nm CMOS process based on neural network |
| CN119348220A (en) * | 2024-09-29 | 2025-01-24 | 武汉理工大学 | A stamping operation speed control method, device, electronic equipment and storage medium |
| CN119167786A (en) * | 2024-09-29 | 2024-12-20 | 浙江理工大学 | A modeling method for magnetorheological damper based on Gaussian process regression model |
| CN119316091A (en) * | 2024-09-30 | 2025-01-14 | 西北工业大学深圳研究院 | A semantic communication method and system based on LSTM network automatic training |
| CN118919080A (en) * | 2024-10-11 | 2024-11-08 | 长春理工大学 | Noninvasive blood glucose concentration prediction method |
| CN118971947A (en) * | 2024-10-12 | 2024-11-15 | 中国人民解放军63921部队 | A UAV communication system based on Beidou messages |
| CN119400308A (en) * | 2024-10-12 | 2025-02-07 | 中车株洲电力机车有限公司 | Battery material combination method, device and computer storage medium |
| CN118966019B (en) * | 2024-10-14 | 2024-12-27 | 新雅电线电缆(深圳)有限公司 | Insulated cable process data processing method |
| CN118966019A (en) * | 2024-10-14 | 2024-11-15 | 新雅电线电缆(深圳)有限公司 | A method for processing process data of insulated cables |
| CN118967670A (en) * | 2024-10-15 | 2024-11-15 | 浙江理工大学 | A parameter optimization method for electromagnetic induction detection based on multi-task learning model |
| CN118963774A (en) * | 2024-10-18 | 2024-11-15 | 浙江大学 | A Sub-model-based Scheduling Method for AI Compilers |
| CN119025998A (en) * | 2024-10-25 | 2024-11-26 | 自然资源部第三海洋研究所 | A method for estuary water level time series prediction based on improved Informer deep learning framework |
| CN119025925A (en) * | 2024-10-28 | 2024-11-26 | 西南石油大学 | A multi-task reservoir parameter prediction method based on well logging data |
| CN119051981A (en) * | 2024-10-30 | 2024-11-29 | 海天地数码科技(北京)有限公司 | Online data service center of electronic reader based on deep learning |
| CN119066817A (en) * | 2024-11-05 | 2024-12-03 | 宁波隆跃科技有限公司 | Die casting parameter optimization and processing method and system based on orthogonal experiment |
| CN119539804A (en) * | 2024-11-06 | 2025-02-28 | 孟约共梦区块链(杭州)有限责任公司 | A transaction method based on blockchain |
| CN119129437A (en) * | 2024-11-12 | 2024-12-13 | 宁波大学 | A fatigue life prediction method for automotive turbocharger pipelines |
| CN119155335A (en) * | 2024-11-19 | 2024-12-17 | 漳州职业技术学院 | Intelligent well lid data analysis method and device based on Internet of things and storage medium |
| CN119167211A (en) * | 2024-11-20 | 2024-12-20 | 济宁职业技术学院 | A method and system for managing people who take Mandarin proficiency tests |
| CN119387758A (en) * | 2024-11-29 | 2025-02-07 | 徐州德高电动车科技有限公司 | Welding control method for electric tricycle manufacturing |
| CN119227010B (en) * | 2024-12-02 | 2025-02-18 | 杭州市北京航空航天大学国际创新研究院(北京航空航天大学国际创新学院) | Cross-task data set fusion method based on meta-Bayesian optimization |
| CN119416362A (en) * | 2024-12-02 | 2025-02-11 | 北京航空航天大学 | A numerical and real fusion test method for reliability of aero-engine blade disk structure |
| CN119227010A (en) * | 2024-12-02 | 2024-12-31 | 杭州市北京航空航天大学国际创新研究院(北京航空航天大学国际创新学院) | A cross-task dataset fusion method based on meta-Bayesian optimization |
| CN119720906A (en) * | 2024-12-09 | 2025-03-28 | 西安电子科技大学 | A method, system, device and medium for optimizing analog equalizer design based on deep learning algorithm |
| CN119720906B (en) * | 2024-12-09 | 2025-08-26 | 西安电子科技大学 | A method, system, device and medium for optimizing analog equalizer design based on deep learning algorithm |
| CN119829955A (en) * | 2025-03-17 | 2025-04-15 | 中国海洋大学 | Ship nonlinear roll parameter identification method based on physical information neural network |
| CN120256117A (en) * | 2025-03-26 | 2025-07-04 | 北京骏德智算科技有限公司 | A secure scheduling method for database storage resources based on artificial intelligence |
| CN119883576A (en) * | 2025-03-27 | 2025-04-25 | 吉林大学 | Task association degree-based multi-task learning auxiliary task selection method |
| CN119957986A (en) * | 2025-04-10 | 2025-05-09 | 山西太康高科节能股份有限公司 | A heat pump control method and system based on state data |
| CN120046813A (en) * | 2025-04-24 | 2025-05-27 | 杭州海亮数字科技有限公司 | Model optimization method and device suitable for special length prediction |
| CN120105109A (en) * | 2025-05-12 | 2025-06-06 | 山东永正产业技术研究院有限公司 | Training method and equipment for carbon emission prediction model in cement production |
| CN120408251A (en) * | 2025-07-02 | 2025-08-01 | 山东大学日照研究院 | Multi-objective grinding decision-making method and system based on dynamic clustering and hyperparameter optimization |
| CN120468677A (en) * | 2025-07-10 | 2025-08-12 | 洛阳理工学院 | Lithium battery state of charge estimation method based on Bayes-TLCO optimized deep neural network |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| WO2021007812A1 (en) | Deep neural network hyperparameter optimization method, electronic device and storage medium | |
| CN110598842A (en) | Deep neural network hyper-parameter optimization method, electronic device and storage medium | |
| US12001957B2 (en) | Methods and systems for neural architecture search | |
| Dubois et al. | Data-driven predictions of the Lorenz system | |
| Uria et al. | A deep and tractable density estimator | |
| Seldin et al. | PAC-Bayesian inequalities for martingales | |
| Liang et al. | An equivalent measure of partial correlation coefficients for high-dimensional gaussian graphical models | |
| Karshenas et al. | Regularized continuous estimation of distribution algorithms | |
| KR20200092989A (en) | Production organism identification using unsupervised parameter learning for outlier detection | |
| US20230206036A1 (en) | Method for generating a decision support system and associated systems | |
| EP3895080A1 (en) | Regularization of recurrent machine-learned architectures | |
| Ghanbari et al. | Reconstruction of gene networks using prior knowledge | |
| US20230124258A1 (en) | Embedding optimization for machine learning models | |
| CN118414621A (en) | Supermarameter selection using budget-aware Bayesian optimization | |
| Rudner et al. | On the connection between neural processes and gaussian processes with deep kernels | |
| Hoffmann et al. | Minimising the expected posterior entropy yields optimal summary statistics | |
| Al Ali et al. | Enhancing financial distress prediction through integrated Chinese whisper clustering and federated learning | |
| CN114819295B (en) | Data analysis and prediction method, device, server, storage medium and program product | |
| Sacco et al. | On‐line machine‐learning forecast uncertainty estimation for sequential data assimilation | |
| Sun et al. | Variational rectification inference for learning with noisy labels | |
| Singh et al. | A review of online supervised learning | |
| Lapko et al. | Methods for rapid selection of kernel function blur coefficients in a nonparametric pattern recognition algorithm | |
| US20230092949A1 (en) | System and method for estimating model metrics without labels | |
| Zhang et al. | Self-paced deep clustering with learning loss | |
| Kim | Active label correction using robust parameter update and entropy propagation |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19937742 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 19937742 Country of ref document: EP Kind code of ref document: A1 |
|
| 32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 030822) |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 19937742 Country of ref document: EP Kind code of ref document: A1 |