Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Pre-training framework with two-stage decoder for language understanding and generation

A decoder and pre-training technology, applied in the field of language understanding and generation of pre-training frameworks, can solve the problems of insufficient ability to generate text, not smooth, insufficient subsequent generation ability, etc., so as to achieve convenient access to context information and good pre-training quality. Effect

Pending Publication Date: 2021-09-28
AISPEECH CO LTD
View PDF8 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] In order to at least solve the problems in the prior art that there is no relevant pre-training for subsequent text generation, resulting in insufficient subsequent generation capabilities, and lack of explicit generation related understanding, resulting in insufficient and unsmooth subsequent text generation capabilities

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Pre-training framework with two-stage decoder for language understanding and generation
  • Pre-training framework with two-stage decoder for language understanding and generation
  • Pre-training framework with two-stage decoder for language understanding and generation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment approach

[0056] As an implementation manner, the condition generation task includes: a text summarization task;

[0057] Encoder, used to receive the text of the text summarization task, and encode to obtain the text sequence;

[0058] a decoder for receiving the text sequence output by said encoder,

[0059] In the first decoding stage, the text sequence is reconstructed to obtain a coherent text sequence intelligible after reconstruction, and a meaning representation of the text sequence is generated;

[0060] In a second decoding stage, based on the coherent text sequence and meaning representation, a text summary is generated.

[0061] In this embodiment, if Figure 4 In the summarization task shown in (b), x represents questions (in text summarization tasks, these questions refer to the overall text). After being encoded by the encoder, it is input to the decoder for two-stage decoding to obtain a coherent text sequence that can be reconstructed and understandable, as well as t...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention provides a pre-training framework with a two-stage decoder for language understanding and generation, comprising: an encoder for receiving a noisy text in a conditional generation task and encoding to obtain a noisy text sequence, the noisy text comprising a text with covered segments and / or disordered sentences; a decoder used for receiving the noisy text sequence output by the encoder, reconstructing the noisy text sequence in a first decoding stage to obtain a reconstructed text, and generating a significance representation corresponding to the noisy text sequence. In a second decoding stage, subsequent text is generated based on the reconstructed text and the meaning representation. The embodiment of the invention also provides a two-stage decoder. According to the embodiment of the invention, reconstruction and understanding are carried out in the first decoding stage, the understanding of the text is ensured in the second decoding stage, and the subsequent text can be generated, so that the pre-training quality is better. Explicit understanding can be applied to downstream tasks, contextual information can be conveniently obtained, and the application task range is wider.

Description

technical field [0001] The invention relates to the field of intelligent speech, in particular to a pre-training framework for language understanding and generation with a two-stage decoder. Background technique [0002] Self-supervised pre-training improves the state of the art for natural language generation (NLG) tasks, various language generation (conditional generation) tasks aim to generate natural language sentences given a document (conditional), such as context in task-oriented dialogue To response (Context-to-Response), session response generation, etc. [0003] To achieve pre-training, there are various types of pre-trained models for conditional text generation tasks. Most of these models fall into two categories. The first type of model, such as MASS (Masked Sequence to Sequence, covered sequence to sequence) and BART (Bidirectional and Auto-Regressive Transformers, two-way auto-regressive Transformer), decodes the masked part given the damaged text or reply ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/126G06F40/205G06F16/35G06N3/04G06N3/08
CPCG06F40/126G06F40/205G06F16/355G06N3/08G06N3/045
Inventor 俞凯陈露马达陈志
Owner AISPEECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products