Unlock instant, AI-driven research and patent intelligence for your innovation.

Method and device for generating cover image, equipment and storage medium

A cover and image technology, which is applied in the fields of instruments, computing, electrical digital data processing, etc., can solve the problems of long-term refinement and low efficiency, and achieve the effect of improving the experience of watching videos, improving accuracy and efficiency

Pending Publication Date: 2021-12-21
TENCENT TECH (BEIJING) CO LTD
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this method requires video creators to have a high level of creation, and it takes a long time to refine the highlight content in the video, and its efficiency is low

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and device for generating cover image, equipment and storage medium
  • Method and device for generating cover image, equipment and storage medium
  • Method and device for generating cover image, equipment and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment approach 1

[0109] Implementation mode 1. Extraction mode.

[0110] Predicting first probabilities that at least one piece of text information is target description information based on the integrated text features, and then selecting at least one target description information from the at least one piece of text information based on the obtained first probabilities.

[0111] Specifically, the comprehensive text features are input into the encoder to obtain the first probability that at least one piece of text information is target description information, wherein the encoder can be ALBERT, BERT, Transformer-Encoder, etc. Sort the obtained first probabilities in order from large to small, and use the text information corresponding to the first probability greater than the screening threshold as the target description information, or the text information corresponding to the first probability ranked in the top N positions, As the target description information, N is a preset number.

[01...

Embodiment approach 2

[0116] Embodiment 2, generation method.

[0117] The integrated text features are encoded to obtain at least one second text feature corresponding to each of the text information. Decoding the second text feature corresponding to the at least one text information respectively, generating at least one candidate description information corresponding to the target video, and a second probability that each candidate description information is the target description information. Then, based on the obtained second probability, at least one target description information is selected from at least one candidate description information.

[0118] Specifically, an encoder is used to encode the comprehensive text feature to obtain at least one second text feature corresponding to each text information, wherein the encoder can be ALBERT, BERT, Transformer-Encoder, etc. A decoder is used to decode the second text feature corresponding to at least one text information respectively, and at l...

Embodiment approach 3

[0126] Embodiment 3: The extraction method is combined with the generation method.

[0127] In this embodiment of the present application, at least one target description information includes at least one first type of description information and at least one second type of description information, wherein the first type of description information is the target description information obtained by extraction, and the second description information is The target description information obtained by generating.

[0128] Based on the comprehensive text features, the process of predicting at least one target description information corresponding to the target video is as follows: Figure 7 shown, including the following steps:

[0129] Step S701, encoding the integrated text features to obtain at least one second text feature corresponding to each of the text information.

[0130] An encoder is used to encode the integrated text features to obtain at least one second text feature ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The embodiment of the invention provides a method and device for generating a cover image, equipment and a storage medium, and relates to the technical field of artificial intelligence, and the method comprises the steps: carrying out the feature extraction of each piece of text information, obtaining a first text feature of each piece of text information, and enabling each piece of text information to obtain a better feature representation. The first text features are fused, context joint modeling is carried out on the relation between the text information, the obtained comprehensive text features can more accurately represent the content of the target video, and then when at least one piece of target description information corresponding to the target video is predicted based on the comprehensive text features, the accuracy of predicting the target description information can be effectively improved. By combining the original cover image of the target video with the target description information, the target cover image capable of better representing the video bright spots is obtained, so that a user can quickly and intuitively find the video bright spots, and the video watching experience of the user is improved.

Description

technical field [0001] The embodiments of the present invention relate to the technical field of artificial intelligence, and in particular to a method, device, device and storage medium for generating a cover image. Background technique [0002] With the development of video technology, people's requirements for viewing experience are getting higher and higher. When browsing recommended videos, users may skip the current video if they cannot quickly discover the highlights of the video. [0003] In order to facilitate users to quickly and intuitively discover video highlights, video creators manually construct video highlights and synthesize them into the cover image before uploading the video to the video platform. However, this method requires video creators to have a high level of creation, and it takes a long time to refine the highlight content in the video, and its efficiency is low. Contents of the invention [0004] Embodiments of the present application provide...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F16/738G06F16/783G06F16/735
CPCG06F16/739G06F16/7844G06F16/735
Inventor 陈小帅
Owner TENCENT TECH (BEIJING) CO LTD