Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Attention mechanism-based natural language generating method for remote sensing images

A remote sensing image and natural language technology, applied in the field of computer vision, can solve the problems of easy overfitting of models, neglect of remote sensing images, low density of rotation ambiguity, etc., to avoid complex parameter initialization process, enrich image information, and prevent network performance. falling effect

Inactive Publication Date: 2018-11-06
UNIV OF ELECTRONICS SCI & TECH OF CHINA
View PDF6 Cites 14 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0009] The purpose of the present invention is to solve the following problems: (1) remote sensing images are different from natural images, they contain a large amount of irrelevant information, and have the characteristics of scale fuzzy, classification ambiguity and rotation ambiguity, and low density, which may lead to remote sensing images. Objects show different semantics at different scales, and it is difficult to describe the fusion area in remote sensing images, which can easily lead to ambiguity in classification; (2) the data set of remote sensing images is limited and the trained model is easy to overfit; (3) Neglecting the low-density characteristics of the remote sensing image itself and the characteristics of multiple images in the same area at different times may result in only a small area containing a small effective area in a large remote sensing image, making the model unable to fully describe the effective area. information; the present invention provides a method for generating natural language based on attention mechanism

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Attention mechanism-based natural language generating method for remote sensing images
  • Attention mechanism-based natural language generating method for remote sensing images
  • Attention mechanism-based natural language generating method for remote sensing images

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0044]The present invention will be further described below in conjunction with the embodiments, and the described embodiments are only a part of the embodiments of the present invention, not all of them. Based on the embodiments of the present invention, other used embodiments obtained by persons of ordinary skill in the art without creative efforts all belong to the protection scope of the present invention.

[0045] A method for generating natural language from remote sensing images based on an attention mechanism, comprising the following steps:

[0046] Step 1: Slice the natural language corresponding to the RSICD remote sensing image into characters and number them. The type of the number is an integer, and establish a dictionary space corresponding to the number and the character; that is, establish a mapping relationship between the integer number and the dictionary, and generate Called in natural language.

[0047] Step 2: Build a system model, such as figure 1 As s...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an attention mechanism-based natural language generating method for remote sensing images, and belongs to the technical field of computer vision. A CNN part comprises a commonconvolutional pooling layer, a C-S layer and a full connection layer; an RNN part structurally comprises a multi-layer model unit, a GRU unit and an attention unit; the remote sensing images at different moments in the same area are input into the initialized CNN part to obtain eigenvectors; the eigenvectors are input into the GRU unit and the attention unit, the attention unit also receives hidden layer states from the GRU unit, and the attention unit maps the eigenvectors and the hidden layer states to low dimension for further compression and normalization to obtain eigenvectors subjected to weighted average; and the eigenvectors subjected to the weighted average and the hidden layer states of the GRU unit are integrated through the multi-layer model unit, and then maximum likelihood isperformed through a hidden layer and a normalization layer to obtain an integer sequence. The problem that a result of generating a natural language by the remote sensing images at present is not ideal is solved.

Description

technical field [0001] The invention belongs to the technical field of computer vision, and in particular relates to a method for generating natural language of remote sensing images based on an attention mechanism. Background technique [0002] With the development and popularization of artificial intelligence, image annotation, as the integration of computer vision and natural language processing, has become more and more important. It is used in many important fields such as image retrieval, children's education and life assistance for visually impaired people. have important applications. A method that combines deep convolutional neural networks and recurrent neural networks has achieved remarkable progress on the problem of image annotation. However, the achievements in the field of computer vision in some areas of remote sensing images, image annotation has not well realized its potential. For remote sensing images, the classic combination of CNN and RNN in the field...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06N3/04
CPCG06N3/045G06F18/24G06F18/214
Inventor 郭一菲高建彬
Owner UNIV OF ELECTRONICS SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products