Double-flow video generation method based on different feature spaces of text

A feature space, two-stream technology, applied in the fields of natural language processing, computer vision, and pattern recognition, can solve problems such as difficult high-quality generation, overestimation of the learning ability of a single model, and insufficient understanding of the model

Active Publication Date: 2019-07-05
PEKING UNIV SHENZHEN GRADUATE SCHOOL
View PDF5 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Most of the existing methods uniformly process the entire text, and directly generate the entire video clip from the extracted features. However, such processing overestimates the learning ability of a single model, which requires not only learning spatial features (appearance information), but also Learning temporal features (motion information) makes the model unable to fully understand both features, and it is difficult to effectively generate high-quality

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Double-flow video generation method based on different feature spaces of text
  • Double-flow video generation method based on different feature spaces of text
  • Double-flow video generation method based on different feature spaces of text

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0079] Below in conjunction with accompanying drawing, further describe the present invention through embodiment, but do not limit the scope of the present invention in any way.

[0080] The present invention provides a dual-stream video generation method based on different feature spaces of the text. By separating the spatial features and temporal features contained in the text and modeling these features in a dual-stream manner, the learning ability of the specified features is maximized. , and use the way of confrontation training to optimize the generation results.

[0081] The method provided by the present invention includes: a text feature extraction process, a dual-stream video generation process and an adversarial training process; figure 1 Shown is the flow process of method provided by the present invention, and concrete steps are as follows:

[0082] 1. Perform text feature extraction and separation, see steps 11)-13)

[0083] 11) Use the bidirectional long short...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a double-flow video generation method based on different feature spaces of a text. The double-flow video generation method comprises a text feature extraction process, a double-flow video generation process and an adversarial resistance training process. According to the method, the input text content is analyzed, the feature information obtained by using the attention mechanism is separated, and the information of the appearance feature space and the information of the motion feature space are extracted respectively; a double-flow video generation mode is adopted, so that the learning capability of a single model on specified features is utilized to the maximum extent. According to the method, the adversarial resistance training process is used, the model is guidedand trained from the two aspects of the quality of each frame and the coherence of overall motion, and meanwhile text features are added to strengthen the training process, so that the generated video can conform to input text description, and the generation accuracy is achieved. The method provided by the invention can realize intelligent generation, and has wide market demands and application prospects in the aspects of material accumulation and data set automatic generation.

Description

technical field [0001] The present invention relates to the technical fields of pattern recognition, natural language processing, computer vision, etc., and in particular to a video generation method based on different feature spaces of texts, intelligently generating A video clip that matches the text content. Background technique [0002] In recent years, with the rapid development of computer vision and the introduction of generative adversarial networks, the research on image generation has received more and more attention. It has very positive significance in material accumulation and automatic generation of data sets. Compared with images, videos are more vivid and difficult to generate, so the exploration of video generation is more meaningful. At the same time, if randomly generating videos like most image generation methods does not have much practical value, users will more likely want to generate based on some given information, for example, the user inputs "a pe...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06N3/045G06F18/213
Inventor 王文敏李炜棉黄钟毅
Owner PEKING UNIV SHENZHEN GRADUATE SCHOOL
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products