Compression method and system used for neural network language model (NN LM)

A language model and neural network technology, applied in the field of neural network language model compression, can solve problems such as complex architecture and negligible memory reduction rate

Inactive Publication Date: 2018-08-17
AISPEECH CO LTD
View PDF3 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

It solves the above two problems, but the memory reduction rate is negligible (20% when |V|=10K) with tiny vocabulary size
Also, the architecture is complicated because it calls an additional layer in the output layer

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Compression method and system used for neural network language model (NN LM)
  • Compression method and system used for neural network language model (NN LM)
  • Compression method and system used for neural network language model (NN LM)

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0019] In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments It is a part of embodiments of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of the present invention.

[0020] In the following, the embodiment of the present application will be introduced first, and then the experimental data will be used to verify the difference between the solution of the present application and the prior art, and what beneficial effects can be achieved.

[0021] Please refer to figure 1 , which shows a flow chart of an embodim...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a compression method and system used for a neural network language model (NN LM). The method includes: inputting training data into the neural network language model for pre-training; respectively carrying out base decomposition and clustering quantization operations on word vector matrices of input and / or output of the language model to compress the word vector matrices; and inputting the training data again into the neural network language model after compression of the word vector matrices to finely tune the language model. The invention provides a novel and efficient structured word embedding framework based on product quantization, the framework is used for compressing the input / output word vector matrices, and a significant memory reducing rate can be obtainedin a case of not damaging NN LM performance.

Description

technical field [0001] The invention belongs to the technical field of language model compression, in particular to a compression method and system for neural network language models. Background technique [0002] In Automatic Speech Recognition (ASR, Automatic Speech Recognition), the language model (LM, Language Model) is the core component that combines the syntactic and semantic constraints of a given language. Although the traditional N-gram backoff language model with smoothing has been widely used in ASR, its context length is limited, and the memory requirement for a large vocabulary is also large. Recently, neural network-based language models (NN LM, Neural Network Language Model) have attracted great interest due to their efficient encoding of word context history and memory efficiency. In neural network-based language models, both word contexts and target words are projected into a continuous space. Projections represented by transformation matrices are learned...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/22G06F17/30
CPCG06F16/35G06F40/12
Inventor 俞凯石开宇
Owner AISPEECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products