Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Local self-attention image processing method and model based on deformable block division

An image processing and attention technology, applied in the field of computer vision, can solve the problems of impairing the representation ability of the model and reducing the performance of the model, so as to reduce the amount of calculation and improve the performance.

Active Publication Date: 2022-01-21
OBJECTEYE (BEIJING) TECH CO LTD
View PDF17 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, this kind of crude windowing for self-attention calculation prevents some adjacent blocks from participating in self-attention calculation, even if they have a high similarity, which damages the representation ability of the model and reduces the performance of the model. performance

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Local self-attention image processing method and model based on deformable block division
  • Local self-attention image processing method and model based on deformable block division
  • Local self-attention image processing method and model based on deformable block division

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0062] In order to make the purpose, technical solutions and advantages of the present invention clearer, the technical solutions in the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the present invention. Obviously, the described embodiments are part of the embodiments of the present invention , but not all examples. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts fall within the protection scope of the present invention.

[0063] Combine below Figure 1-Figure 6 Describe the image processing method of the present invention based on local self-attention of deformable blocks.

[0064] see figure 1 , figure 1 It is a schematic flowchart of an image processing method based on local self-attention of deformable blocks provided by the present invention, figure 1 The shown image processing method based on local se...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a local self-attention image processing method and model based on deformable block division. The method comprises the following steps: performing partitioning processing on a first feature map of an input image, predicting a first offset value of each block obtained by partitioning processing based on the first feature map, and correcting a range of each block obtained by partitioning processing based on the first offset value; performing feature extraction based on each corrected block in the first feature map to obtain a second feature map; performing windowing processing on the second feature map, predicting a second deviation value of each window obtained by the windowing processing based on the second feature map, and correcting a range of each window obtained by the windowing processing based on the second deviation value; and determining the self-attention of each block in the second feature map based on each corrected window, and taking the feature map after the self-attention is determined as a first feature map or a target feature map. According to the method, the performance of the model can be effectively improved under the condition of not increasing a large amount of calculation.

Description

technical field [0001] The invention relates to the technical field of computer vision, in particular to an image processing method and model based on local self-attention of deformable blocks. Background technique [0002] Transformer (self-attention model) is a mainstream model in various tasks in the field of natural language processing, and has also attracted widespread attention in the field of computer vision in recent years. Compared with traditional convolutional networks, Transformer has obvious advantages in establishing long-distance relationship models and fitting large-scale data sets. At present, Transformer has achieved performance beyond the traditional convolution model in the fields of image classification, target detection, and semantic segmentation. [0003] Transformer mainly includes two modules, self-attention module and fully connected forward module. The earliest Transformer evenly divides the input image into several small area blocks according to...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/11G06V10/26G06V10/40
CPCG06T7/11G06F18/24
Inventor 王金桥朱优松陈志扬赵朝阳
Owner OBJECTEYE (BEIJING) TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products