Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

End-to-end character image generation method guaranteeing consistent style

An image generation and style technology, applied in neural learning methods, editing/combining graphics or text, electronic digital data processing, etc., can solve problems such as easy to generate artifacts, unable to handle scenes with complex backgrounds, and difficult to ensure semantic coherence. Achieve the effect of reducing the loss of effect and performance

Pending Publication Date: 2021-10-26
HARBIN INST OF TECH
View PDF7 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0009] (1) It cannot handle scenes with more complex backgrounds, and artifacts are prone to occur in the process of reconstructing the background
[0010] (2) It is difficult to simulate complex deformation of text, such as perspective deformation, curved text, etc., and it may be difficult to ensure semantic coherence with the global image after reverse correction
[0011] (3) Unable to fully capture the style of the original text, such as fonts, shapes, shadows and other effects

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • End-to-end character image generation method guaranteeing consistent style
  • End-to-end character image generation method guaranteeing consistent style
  • End-to-end character image generation method guaranteeing consistent style

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0067] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be described in detail below in conjunction with natural scene images.

[0068] The system development platform is the Linux operating system CentOS7.2, and the GPU is an NVIDIA GeForce GTXTITAN X GPU. The recognition program is written in python3.7 and uses the PyTorch1.6 framework.

[0069] Since there is no group data after text replacement in reality, and there is no related data set, the training data adopts synthetic data.

[0070] 1. Training data synthesis

[0071] Collect data such as font files, corpus, and images without text to generate training data.

[0072] Collect Chinese and English thesaurus. An English thesaurus (more than 160,000 words) and a Chinese thesaurus THUOCL (more than 150,000 words) were collected from the Internet. In order to prevent the font file from rendering a certain Chinese character, delete words other than ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an end-to-end character image generation method guaranteeing consistent style. The method comprises the following steps: step 1, erasing characters in a source style image, supplementing background textures, and obtaining a background image without characters; step 2, rendering a target text into an image, embedding the image into the background image without characters output by a background reconstruction module, and migrating text styles of the source image into the target text under the same background. By adopting the end-to-end method, the work such as text style migration and character erasing is integrated into one network, so that the potential losses of effectiveness and performance caused by intermediate steps are reduced; by using a GAN algorithm, a result that is more real, more consistent in style, and more coherent in semantics can be generated.

Description

technical field [0001] The present invention relates to a text image generation method, in particular to an end-to-end image generation method capable of maintaining the text style in a complex background environment, which is mainly oriented to picture translation scenarios, and the target text is drawn according to the source style, and the background texture is preserved And the style of the original text (font, color, shape, etc.), to achieve high-fidelity replacement of the translated text. Background technique [0002] Style-consistent text-to-image generation aims to achieve high-fidelity text replacement, and has many practical applications, such as image translation, text detection and recognition tasks, movie poster text editing, etc. For image translation tasks, it can improve translation results and enhance user experience; for text detection and text recognition tasks, it can quickly expand data for specific scenarios; for designers, it can quickly design poster...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F40/103G06T11/60G06N3/04G06N3/08
CPCG06F40/103G06T11/60G06N3/08G06N3/045
Inventor 苏统华杨富祥王忠杰徐晓飞
Owner HARBIN INST OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products