Image captioning with weak supervision

A weak supervision and subtitle technology, applied in image communication, still image data retrieval, still image data indexing, etc., can solve problems such as poor work, difficult for users to search for specific images, and neglect of image details

Active Publication Date: 2017-07-21
ADOBE SYST INC
View PDF6 Cites 5 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this top-down approach does not work well in capturing fine details of an image, such as local objects, attributes, and regions that contribute to an accurate description of the image.
As such, it can be difficult to generate accurate and complex image captions using traditional methods, such as "person feeding a child holding a toy in a high chair"
Therefore, subtitles generated using conventional techniques may omit important image details, which makes it difficult for users to search for a specific image and fully understand the content of the image based on the associated subtitles

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Image captioning with weak supervision
  • Image captioning with weak supervision
  • Image captioning with weak supervision

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0018] overview

[0019] Traditional techniques for image processing do not support high-precision natural language subtitles and image search due to the limitations of traditional image tagging and search algorithms. This is because traditional techniques only associate labels with images, but do not define the relationship between labels or between labels and images themselves. Alternatively, traditional techniques may include using a top-down approach in which the entire "gist" of an image is first taken and refined into appropriate descriptive words and captions through language modeling and sentence generation. However, this top-down approach does not work well in capturing fine details of an image, such as local objects, attributes, and regions that contribute to an accurate description of the image.

[0020] This paper describes techniques for captioning images with weak supervision. In one or more implementations, weakly supervised data about the target image is ac...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

Techniques for image captioning with weak supervision are described herein. In implementations, weak supervision data regarding a target image is obtained and utilized to provide detail information that supplements global image concepts derived for image captioning. Weak supervision data refers to noisy data that is not closely curated and may include errors. Given a target image, weak supervision data for visually similar images may be collected from sources of weakly annotated images, such as online social networks. Generally, images posted online include ''weak'' annotations in the form of tags, titles, labels, and short descriptions added by users. Weak supervision data for the target image is generated by extracting keywords for visually similar images discovered in the different sources. The keywords included in the weak supervision data are then employed to modulate weights applied for probabilistic classifications during image captioning analysis.

Description

Background technique [0001] Automatically generated natural language descriptions of images continue to attract interest due to practical applications for image search, accessibility for the visually impaired, and management of image acquisition. Traditional techniques for image processing do not support high-precision natural language subtitles and image search due to the limitations of traditional image tagging and search algorithms. This is because conventional techniques only associate labels with images, but do not define the relationship between labels or between labels and images themselves. Alternatively, traditional techniques can include using a top-down approach, in which the entire "gist" of an image is first obtained and then refined into appropriate descriptive words or captions through language modeling and sentence generation. However, this top-down approach does not work well in capturing fine details of an image, such as local objects, attributes, and region...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): H04N5/278H04N21/431H04N21/488G06N3/08G06N3/04G06V10/764
CPCG06N3/08H04N5/278H04N21/4314H04N21/4882G06N3/044G06F16/583G06V20/10G06V10/454G06V20/70G06V10/82G06V10/764G06N3/045G06F16/51G06F16/334G06F18/2411G06F18/24143G06N7/01
Inventor 王兆闻尤全增金海琳方晨
Owner ADOBE SYST INC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products