Unsupervised image-to-image translation method based on content style separation

An unsupervised and style-based technology, applied in the field of computer vision and image processing, can solve the problems of feature attention that cannot be content, difficulty in expanding multiple images, and inability to effectively extract different styles of different objects, so as to reduce the impact and modify the range. , the effect of improving scalability

Active Publication Date: 2021-05-07
BEIHANG UNIV
View PDF4 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Therefore, the disadvantages of this type of method are: 1. It cannot guarantee that the features of the content can effectively focus on meaningful objects in the image
2. Style features often focus on the appearance of the entire image, and cannot effectively extract different styles of different objects
The disadvantage of this type of method is that different network architectures need to be designed for specific high-level visual tasks. If the corresponding high-level visual information cannot be provided on the new data, it is difficult to extend the same method to multiple image-to-image translation tasks.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Unsupervised image-to-image translation method based on content style separation
  • Unsupervised image-to-image translation method based on content style separation
  • Unsupervised image-to-image translation method based on content style separation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0018] Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. Although certain embodiments of the disclosure are shown in the drawings, it should be understood that the disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein. Rather, these examples are provided so that the understanding of this disclosure will be thorough and complete. It should be understood that the drawings and embodiments of the present disclosure are for exemplary purposes only, and are not intended to limit the protection scope of the present disclosure.

[0019] It should also be noted that, for the convenience of description, only the parts related to the related invention are shown in the drawings. In the case of no conflict, the embodiments in the present disclosure and the features in the embodiments can be combined with each other.

[0020] It should be noted that the ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The embodiment of the invention discloses an unsupervised image-to-image translation method. A specific embodiment of the method comprises the following steps: acquiring an initial image, and zooming the initial image to a specific size; performing spatial feature extraction on the initial image through an encoder to obtain feature information; inputting the feature information into a content style separation module to obtain content feature information and style feature information; generating reference style feature information of the reference image in response to the acquired reference image, and setting the reference style feature information as Gaussian noise consistent with the style feature information in shape in response to the fact that the reference image is not acquired; inputting the content feature information and the reference style feature information into a generator to obtain a target image translating the initial image into a reference image style; and zooming the target image to a size matched with the initial image to obtain a final target image. The implementation mode can be applied to various different advanced vision tasks, and the expandability of the whole system is improved.

Description

technical field [0001] Embodiments of the present disclosure relate to the technical fields of computer vision and image processing, and in particular to an unsupervised image-to-image translation method. Background technique [0002] Image-to-image translation has attracted much attention since it can learn the mapping between different visual domains. In the current social media or chat software, many chatting objects are converted into cute animals, and the converted animal expressions are consistent with the original object during the chatting process. Or change the style of the chat background while maintaining the spatial structure information of the background. Or it is necessary to exchange the virtual scene with the real scene in the virtual-real scene of the augmented reality application, etc., which are typical applications of image-to-image. Pix2pix (pixels to pixels, pixel-to-pixel conversion tool) is the first to use conditional generative adversarial network...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04
CPCG06N3/04G06V30/413G06N3/045G06F18/253G06T3/40G06V10/32G06V10/82G06V10/806Y02D10/00G06V10/7715G06V20/64G06V10/255G06V10/764G06T3/4046G06T5/002G06T5/20
Inventor 陆峰刘云飞
Owner BEIHANG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products