Image style migration model training method, system and device and storage medium

A training method and image technology, applied in computing models, character and pattern recognition, machine learning, etc., can solve the problems of deformation style, no obvious improvement, content leakage, etc., and achieve the effect of improving ability and enhancing fidelity

Active Publication Date: 2022-05-13
UNIV OF SCI & TECH OF CHINA
View PDF10 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Specifically, the method based on cycle consistency usually encodes the source domain image into a low-dimensional latent space, transforms the domain-related image information from the source domain to the target domain in the low-dimensional latent space, and uses the transformed image information to reconstruct the target Domain-style images. In this process, two pairs of generators and discriminators are used, and the images migrated to the target domain are required to be converted back to the source domain. Most current methods based on cycle consistency use pixel-level constraints, so , most of them have problems such as deformation and style disorder
Decoupling-based methods tend to decouple the source and target domain images into domain-invariant content features that remain unchanged during the transformation process and domain-specific style features that change during the transformation process, and the transformation is achieved by preserving the content features and replacing the style features. purpose, but there is a problem of content leakage
[0004] In the Chinese patent application with publication number CN113808011A "A Style Migration Method, Device and Related Components Based on Feature Fusion", the decoupling-based style migration method is used to encode the style and content of the image separately, and then the required The content features and style features of this method are fused, and the final migration result is output by the decoder, which can improve the quality of content details and the color similarity with the target domain. There is the problem of content leakage; in the Chinese patent application "Image Migration Method Based on Mean Standard Deviation" with the publication number CN113837926A, the features are normalized in different levels of feature spaces, and PSNR and SSIM are performed through feature maps and source images In contrast, it reduces the time required to train the model and reduces the distortion and artifacts of image features. However, this method focuses on improving the training efficiency and does not significantly improve the quality of the generated image; in the Chinese patent application with the publication number CN107705242A In "An Image Stylization Migration Method Combining Deep Learning and Depth Perception", the content loss, style loss and depth of field loss are calculated on the output feature maps of the perceptual loss network and the depth perceptual network respectively, which improves the three-dimensionality of the generated image. But there are still some situations like corresponding semantic content mismatch and object distortion, so overall, the transfer ability of this scheme is not good

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Image style migration model training method, system and device and storage medium
  • Image style migration model training method, system and device and storage medium
  • Image style migration model training method, system and device and storage medium

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0033] An embodiment of the present invention provides a training method for an image style transfer model, which is used to further improve the quality of image style transfer and improve the accuracy of downstream tasks. Aiming at the common problems of object structure deformation and semantic content mismatch in existing methods, the present invention adopts the mainstream encoder-decoder generator structure and the idea of ​​adversarial learning to establish a style transfer model, and uses the cycle consistency loss constraint model training process. At the same time, a new positive and negative sample selection method is proposed, which improves the fit between contrastive learning and style transfer tasks, and makes the contrastive learning method better applied to the transfer model. The category information on which new positive and negative samples are selected is determined by the image block classification results obtained by the weakly supervised semantic segment...

Embodiment approach

[0039] This section mainly calculates three types of losses, and the preferred implementation method for calculating each type of loss is as follows:

[0040] 1) Calculate the total confrontation loss: the first generator uses the input source domain image to generate the target domain image, and the first discriminator is used to judge whether the input image is the target domain image generated by the first generator. At this time, the first discriminator’s The input image includes the target domain image generated by the first generator and the target domain image acquired for training; the second generator uses the input target domain image to generate the source domain image, and the second discriminator is used to judge whether the input image is the second The source domain image generated by the generator. At this time, the input image of the second discriminator includes the source domain image generated by the second generator and the source domain image acquired for ...

Embodiment 2

[0111] The present invention also provides a training system for an image style transfer model, which is mainly implemented based on the method provided in the first embodiment above, as Figure 5 As shown, the system mainly includes:

[0112] The model construction and image data acquisition unit is used to construct an image style transfer model including two generators and two discriminators. A single generator and a single discriminator form an adversarial structure, which constitutes two adversarial structures, and obtains the used training source domain images and target domain images;

[0113] The total confrontation loss calculation unit is used to input both the source domain image and the target domain image to each confrontation structure, and calculate the total confrontation loss by using the output of the two confrontation structures;

[0114] The total cycle consistency loss calculation unit is used to input the output of the generator of the current confrontat...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a training method, system and device of an image style migration model and a storage medium. Three parts of losses are designed to train the image style migration model: 1) the optimal balance of a generator and a discriminator can be achieved through the total confrontation loss; the method has the advantages that the method is simple, the reconstruction effect of the generator can be guaranteed through the total circulation consistency loss, the fidelity degree of the output image of the generator can be improved through the comparison loss, the image style migration model is trained by integrating the three losses, the image style migration capability can be improved, and the better image after style migration can be obtained.

Description

technical field [0001] The present invention relates to the technical field of image style transfer, in particular to a training method, system, device and storage medium for an image style transfer model. Background technique [0002] With the development of computer science and the improvement of modeling capabilities, computers can simulate virtual scenes that are very similar to real scenes, thereby batch-producing simulated virtual images and labels that can be used for other computer vision tasks. However, due to the limitations of related technologies and the complexity of real scenes and many other factors, it is difficult for virtual images to be completely consistent with the style of real images, which will not only affect the user's visual experience, but also reduce its performance in many downstream tasks to a certain extent. applications, such as target detection, semantic segmentation, etc. Therefore, the style transfer task came into being, by preserving th...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06V10/764G06V10/774G06V10/40G06V10/26G06K9/62G06N20/00
CPCG06N20/00G06F18/241G06F18/214
Inventor 王子磊毛语实
Owner UNIV OF SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products