Scene text recognition method based on HRNet coding and double-branch decoding

A text recognition and branching technology, applied in neural learning methods, character and pattern recognition, instruments, etc., can solve problems such as the inability to meet the accuracy requirements of scene text recognition, the inaccuracy of the decoder to recognize the target sequence, and the loss of visual feature information. Improve the effect of upsampling, suppress the amount of information, and reduce the effect of feature loss

Pending Publication Date: 2022-03-04
HANGZHOU NORMAL UNIVERSITY
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, scene text images are usually disturbed by factors of different sources and varying degrees, such as complex backgrounds, text distortions, etc., which often cause information loss in the visual features extracted by the encoding network, which in turn causes the decoder to suffer from noisy decoding timestamps. Recognition of the target sequence is not accurate enough
Although the existing methods have a good effect on scene text recognition, they still cannot meet the accuracy requirements of scene text recognition.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Scene text recognition method based on HRNet coding and double-branch decoding
  • Scene text recognition method based on HRNet coding and double-branch decoding
  • Scene text recognition method based on HRNet coding and double-branch decoding

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0038] The present invention is further analyzed below in conjunction with specific embodiment.

[0039] A scene text recognition method based on HRNet encoding and dual-branch decoding framework, the adopted model includes modified network TPS, encoding module, super-segmentation branch and recognition branch. The encoding module includes HRNet network, supervised attention. The super-resolution branch includes transposed convolution (TransConv2D) upsampling. The recognition branch includes IndependentTransConv2D Layers for multi-scale fusion and attention-based decoding to obtain text characters. The encoding module is used to perform feature extraction on a single scene text image to obtain visual features, and four resolution feature maps are obtained; the super-separation branch is used for the highest resolution feature map output by the encoding module, and the transposed convolution is upsampled to generate super Resolution image; the recognition branch is used to ex...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a scene text recognition method based on HRNet coding and double-branch decoding. When a traditional deep learning method is used for scene text recognition, the recognition accuracy is reduced when the problems of text distortion, image blurring and low resolution are encountered. The method comprises the following steps of: performing random Gaussian blur on an original text image of a single scene to obtain a low-resolution image; and building a scene text recognition model based on HRNet coding and double-branch decoding, wherein the scene text recognition model based on HRNet coding and double-branch decoding comprises a correction network TPS, a coding module, a super-branch branch and a recognition branch. According to the method, HRNet coding and double-branch decoding are introduced, the recognition accuracy of the model for fuzzy and low-resolution images is improved, and the model parameter quantity and time consumption are reduced in the mode of abandoning super branches during testing.

Description

technical field [0001] The invention relates to the technical field of computer vision and image text recognition, in particular to a scene text recognition method based on HRNet coding and double-branch decoding. Background technique [0002] Scene text recognition aims to automatically identify text content in natural scene images. Different from regular document text, text in natural scene images has the characteristics of changeable shape, complex background, distorted text, and blurred images. Early scene text recognition models were usually based on time series feature classification, using deep convolutional network VGG to extract image feature sequences, using recurrent neural network RNN ​​to learn bidirectional dependencies of feature sequences, and predicting the probability of text character sequences, and finally transcribing through continuous time series classification layer, which transcribes the sequence of predicted character probabilities into text accordi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06V20/62G06V10/80G06V10/774G06K9/62G06N3/08
CPCG06N3/08G06F18/214G06F18/253
Inventor 李秀梅李美玲孙军梅
Owner HANGZHOU NORMAL UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products