Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Image description generation method based on conditional generative adversarial network

A conditional generation and image description technology, applied in the field of computer vision, can solve the problems of monotonous generation description, insufficient description and high training data requirements.

Active Publication Date: 2019-09-27
ZHEJIANG UNIV OF TECH
View PDF4 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] In order to overcome the deficiencies of the existing image description generation technology, such as high requirements for training data, monotonous generated descriptions, and unrealistic descriptions, the present invention provides a condition-based adversarial training method with better robustness and lower requirements for training data. Image description generation method

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Image description generation method based on conditional generative adversarial network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0057] The present invention will be further described below in conjunction with the accompanying drawings.

[0058] refer to figure 1 , an image description generation method based on a conditional generative adversarial network, the method includes four processes: construction of a conditional generative adversarial training network, data set preprocessing, network training, and evaluation index testing.

[0059] The pictures in this implementation case come from the MSCOCO dataset, including training set, verification set and test set. Train the model on the training set, and verify the training results on the test set and validation set. The framework of image description generation method based on conditional generative confrontation network is as follows: figure 1 As shown, the operation steps include four processes of network construction, data set preprocessing, network training and image retrieval testing.

[0060] The image description generation method based on c...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

An image description generation method based on a conditional generative adversarial network comprises the following steps of 1, constructing a network, wherein a conditional generative adversarial network framework is composed of a generation model and a judgment model, the generation model and the judgment model are similar in structure, but the parameters are independently trained and updated; 2, preprocessing a data set; 3, performing network training, wherein the process comprises 3.1, initializing the parameters of a genearation model and a discrimination model by using the random weights; 3.2, training the generation model; 3.3, training a discrimination model; 3.4, minimizing a loss function by using an RMSprop descent algorithm; and 4, testing the precision, generating the description of the test picture through the operation of the above steps. The image description generation method based on the conditional generative adversarial training is better in robustness and lower in requirement for training data.

Description

technical field [0001] The present invention relates to multimedia big data processing and analysis in the field of computer vision, in particular to a method for generating image descriptions based on conditional generation confrontation, which spans two fields of computer vision and natural language processing. Background technique [0002] With the development of network sharing technology, more and more pictures on the network can be shared and received in real time. How to use a machine to understand the content represented by an image and output it as a coherent and semantically correct sentence has become a key research issue. In recent years, with the rapid development of deep learning methods, thanks to the accurate expression of image content by deep features, significant progress has been made in using machines to automatically generate descriptions. However, these methods have gradient disappearance and image features loss in the network during the training proc...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F16/583G06N3/04G06N3/08
CPCG06F16/583G06N3/08G06N3/045
Inventor 白琮黄远李宏凯陈胜勇
Owner ZHEJIANG UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products