Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Trainable piecewise linear activation function generation method

An activation function and linear function technology, applied in the computer field, can solve the problems of slow hardware implementation, lack of computing resources, and no solution to nonlinear mapping, and achieve the effect of speeding up the operation speed and simple linear operation conditions.

Inactive Publication Date: 2018-10-26
NANJING UNIV
View PDF0 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] With the development of deep networks, a variety of activation functions are used in the network, but in terms of application, due to the huge network of machine learning, the implementation of general hardware is relatively slow, so it is inevitable to accelerate the deep network through hardware. Must do
In the process of hardware acceleration, due to the lack of computing resources, it is difficult to implement complex operations of nonlinear activation functions
The existing linear activation functions, such as ReLU function, PReLU function, etc., although the use of these linear functions simplifies the network, the use of simple linear functions still does not solve the problem of nonlinear mapping, and the output of the deep network is still stuck in the linear mapping. stage

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Trainable piecewise linear activation function generation method
  • Trainable piecewise linear activation function generation method
  • Trainable piecewise linear activation function generation method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0025] The present invention will be described in detail below in conjunction with the accompanying drawings.

[0026] This implementation case uses the approximation of the tanh function as an example, and discloses a method for generating a trainable piecewise linear activation function. The process is as follows figure 1 As shown, the steps are as follows:

[0027] Step 1) Determine that the non-linear function to be replaced is the tanh function.

[0028] Segment the selected nonlinear function. In this implementation case, the nonlinear activation function is divided into eight segments, and the negative semi-axis is divided into four segments, respectively (-∞,-6], (-6,-4], (-4 ,-2], (-2,0], four segments of the positive semi-axis, respectively (0,2], (2,4], (4,6], (6,+∞), positive and negative semi-axis with The y-axis is the center and is axisymmetric.

[0029] Step 2) Analyze the non-linear activation function divided into eight sections separately, and analyze the...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a trainable piecewise linear activation function generation method, used for simplifying the computing problem of a complex non-linear activation function in a hardware acceleration process. The invention uses the piecewise linear function to replace the non-linear activation function, and constantly updates the coefficient through self-learning to achieve the purpose of using the linear function to replace the non-linear activation function. Compared with the general linear activation function, the trainable piecewise linear activation function generation method is closer to the original non-linear activation function; the relative error is less, and the coefficient is updated in reverse propagation continuously; therefore, the convergence speed of learning is accelerated, and the problems of gradient explosion and gradient disappearance are eliminated as far as possible in a certain range.

Description

technical field [0001] The invention belongs to the field of computers, in particular to a method for generating a trainable segmented linear activation function. Background technique [0002] In recent years, machine learning has been practically used in many fields such as computers and the Internet, and has played a huge role, greatly improving the success rate of functions including image recognition and language recognition. In the neural network, the result of each layer of the network is processed by the activation function before it is used as the final output. The continuous development of the activation function is an important part of the continuous improvement of the deep network. The continuous improvement of the activation function makes the output of the neural network more accurate. for precision. [0003] The learning network that only includes the convolutional layer and the fully connected layer is obtained through multi-layer operations, and the results ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06N3/08G06N3/04
CPCG06N3/08G06N3/048
Inventor 潘红兵郭良蛟秦子迪李丽何书专李伟
Owner NANJING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products