Apparatus and method for realizing sparse neural network
A neural network, sparse technology, applied in neural learning methods, biological neural network models, neural architectures, etc., can solve problems such as limited acceleration, and achieve the effect of improving performance, reducing quantity, and reducing memory access.
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment Construction
[0031] Compression and Computation for Deep Neural Networks (DNNs)
[0032] The FC layer of DNN performs calculations as follows:
[0033] b=f(Wa+v) (3)
[0034] where a is the input excitation vector, b is the output excitation vector, υ is the bias, W is the weight matrix, and f is the non-linear function, a typical linear rectification unit (ReLU) in CNNs and some RNNs. Sometimes υ is combined with W by adding an extra to the vector a, so the bias is ignored in the following paragraphs.
[0035] For a typical FC layer, like VGG-16 or AlexNet's FC7, the excitation vector is 4k long, and the weight matrix is 4K×4K (16M weights). Weights are represented as single precision floating point numbers, so such a layer requires 64MB of storage. The output excitation of equation (3) is calculated element by element, as:
[0036]
[0037]The paper "Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding" by Han Song et al. int...
PUM
Abstract
Description
Claims
Application Information
- R&D Engineer
- R&D Manager
- IP Professional
- Industry Leading Data Capabilities
- Powerful AI technology
- Patent DNA Extraction
Browse by: Latest US Patents, China's latest patents, Technical Efficacy Thesaurus, Application Domain, Technology Topic, Popular Technical Reports.
© 2024 PatSnap. All rights reserved.Legal|Privacy policy|Modern Slavery Act Transparency Statement|Sitemap|About US| Contact US: help@patsnap.com