Multi-instance natural scene text detection method based on relevance level residual errors

A text detection, natural scene technology, applied in the fields of deep learning, computer vision and text detection, can solve the problem of poor detection effect of small text area, loss function can not well evaluate the actual regression of text detection frame, etc., to reduce parameters The amount of calculation, the effect of improving text detection accuracy, and improving performance

Active Publication Date: 2020-09-29
XI AN JIAOTONG UNIV
View PDF2 Cites 7 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The purpose of the present invention is to provide a multi-instance natural scene text detection method based on the relevance level residual to solve the problem that the current text detection method is not effective in detecting small text areas, and the loss function commonly used in text detection is not very good The problem of evaluating the actual regression of text detection boxes

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-instance natural scene text detection method based on relevance level residual errors
  • Multi-instance natural scene text detection method based on relevance level residual errors
  • Multi-instance natural scene text detection method based on relevance level residual errors

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0052] Below in conjunction with accompanying drawing, the present invention is further described:

[0053] see figure 1 , the present invention comprises the following steps:

[0054] Step 101, using a camera to acquire image data or directly uploading image data as image input.

[0055] Step 102, use the feature extraction network based on the correlation residual to extract the features of the original input image, and obtain the combined sizes of the coarse-grained and fine-grained images as 1 / 32, 1 / 16, 1 / 8, 1 / 4 feature map f 1 , f 2 , f 3 , f 4 , these multi-scale feature maps respectively represent rich feature information from low-level to high-level.

[0056] Step 103, reverse step-by-step feature fusion from the feature map f 1 start, turn to f 1 , f 2 , f 3 , f 4 Upsampling and feature splicing are carried out, and finally a multi-scale fusion feature map with a size of 1 / 4 of the original input image is generated.

[0057] Step 104, perform text region...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a multi-instance natural scene text detection method based on relevance level residual errors. According to the method, a feature extraction network is adopted to extract coarse-grained and fine-grained combined multi-scale features by utilizing relevance level residual errors and reverse level-by-level feature fusion, and more accurate and more complete text information iscontained, so that the text detection accuracy is improved; secondly, text detection box regression loss used in the method is composed of two parts, namely CIoU Loss and angle loss; particularly, theuse of CIoU Loss considers the factors such as the overlapping area, the center distance and the length-width ratio between the prediction text detection box and the real text detection box, so thatthe actual regression condition of the text detection box can be evaluated more accurately, and the performance of the text detection method can be improved; then, the hardware calculation pressure isreduced by adopting a proper mode in multiple steps, and finally, the detection effect on a conventional text region and a small text region is very good.

Description

technical field [0001] The invention belongs to the fields of deep learning, computer vision and text detection, and in particular relates to a multi-instance natural scene text detection method based on relevance hierarchical residuals. Background technique [0002] As a main way of information transmission, text plays an indispensable role in our daily life. With the advent of the era of big data, how to obtain text information in massive images has become an urgent problem to be solved. . Therefore, based on the development of deep learning, natural scene text detection technology has become a very popular research direction in the field of computer vision, which is of great significance for image retrieval and scene understanding. [0003] At present, the advent of a large number of research results has made natural scene text detection widely used in various industries and fields. For example, many Internet companies have developed image retrieval, street view navigat...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/32G06K9/20G06K9/62G06N3/04G06N3/08
CPCG06N3/084G06V10/22G06V20/63G06N3/045G06F18/241G06F18/253
Inventor 田智强王春晖杜少毅兰旭光
Owner XI AN JIAOTONG UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products