Deep neural network hyper-parameter optimization method, electronic device and storage medium

A technology of deep neural network and optimization method, applied in the direction of neural learning method, biological neural network model, neural architecture, etc., can solve the problems of long running time and high computing cost, and achieve the effect of reducing computing cost and running time cost

Inactive Publication Date: 2019-12-20
SHENZHEN UNIV
View PDF0 Cites 20 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] In order to overcome the deficiencies in the prior art, one of the purposes of the present invention is to provide a deep neural network hyperparameter optimization method, which can solve the problems of high computational cost and long running time in the prior art for hyperparameter optimization
[0004] The second object of the present invention is to p...

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Deep neural network hyper-parameter optimization method, electronic device and storage medium
  • Deep neural network hyper-parameter optimization method, electronic device and storage medium
  • Deep neural network hyper-parameter optimization method, electronic device and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0038] The present invention proposes a conditional neural network process (Conditional Neural Processes, CNPs) to replace the Gaussian Process (Gaussian Process, GPs), and the CNPs combines the characteristics of the random process and the neural network, inspired by the flexibility of the Gaussian process and using Gradient descent is used to train the neural network, and then realize the hyperparameter optimization method.

[0039] The conditional neural network process is parameterized by the neural network when learning from known observation data. The conditional neural network model is trained by randomly sampling the data set and following the gradient step to maximize the random sub- The conditional likelihood of the set.

[0040] The general method of hyperparameter optimization in the prior art is to use Gaussian process for model training, but this patent uses a conditional neural network model with stronger learning ability to replace the traditional Gaussian proc...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a deep neural network hyper-parameter optimization method. The method comprises the steps of using a multi-task conditional neural network model for replacing an existing Gaussian process model to realize a data processing process of hyper-parameter optimization; determining the parameters, sequentially selecting the training sets of a decoder and an encoder according to the set parameters to perform network training, prediction, screening and evaluation; adding the candidate point of each task obtained after evaluation and a target function value into an observation set of the corresponding task; and carrying out the network training, prediction, screening and evaluation again, repeating the above steps until the maximum iterative frequency, and finding out the point, namely the hyper-parameter, when the target function value in the observation set of each task is maximum, thereby solving the problems of complex covariance calculation and the like during the hyper-parameter optimization realized by adopting a Gaussian process in the prior art, and simultaneously realizing the hyper-parameter optimization of multiple tasks. The invention further provides anelectronic device and a storage medium.

Description

technical field [0001] The present invention relates to a hyperparameter optimization method, in particular to a deep neural network hyperparameter optimization method, electronic equipment and a storage medium. Background technique [0002] At present, there are many parameters in the network model of deep learning that are not learned during the training process, but are set directly before the training starts. These parameters are called neural network hyperparameters. However, hyperparameter optimization has always been a difficult problem that limits the performance improvement of network models. At present, neural network models are becoming more and more complex, and there are more and more types of hyperparameters, which often results in the inability to select the appropriate combination of hyperparameters according to the relationship between hyperparameters. At present, the commonly used hyperparameter optimization method is the Bayesian optimization method. This ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06N3/04G06N3/08
CPCG06N3/084G06N3/045
Inventor 骆剑平陈亮
Owner SHENZHEN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products