Chinese environment-oriented complex scene text positioning method

A technology for complex scene and text positioning, applied in the field of image processing, can solve the problems of poor robustness of positioning methods, high false alarm rate, and large amount of calculation, so as to save time, improve accuracy, and enhance robustness.

Inactive Publication Date: 2011-08-24
XIDIAN UNIV
View PDF4 Cites 55 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The positioning method based on the connected domain is more accurate, especially for larger characters, and the speed is faster, but when the text background is complex, some text-like objects are difficult to be classified; at the same time, how to select an appropriate threshold for character connectivity Domain confirmation is also a difficult point
The positioning method based on edge detection has good positioning for images with clear text background or dense text, and the calculation speed is relatively fast; but when there are many strong edge objects in the image, it often produces a high false alarm rate , with low accuracy
[0005] The texture-based positioning method considers text as a special texture. This method can

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Chinese environment-oriented complex scene text positioning method
  • Chinese environment-oriented complex scene text positioning method
  • Chinese environment-oriented complex scene text positioning method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] The present invention will be described in detail below in conjunction with the accompanying drawings.

[0041] In the present invention, its input image can be the image that various image acquisition devices acquire, for example: digital camera DC, the mobile phone with camera function, the PDA with camera function or can be one of the video sequences from digital video camera DV frame etc. The image processed by the method of the present invention may be for various image coding formats, such as JPEG, BMP, and the like. In the following description, the library used for the parameter learning involved in the present invention is a self-built database. Since there is currently no public scene text database oriented to the Chinese environment, the present invention is dedicated to building a database with 5,000 to 10,000 pictures, which covers various types of complex scene text images, and the text in the images includes Chinese and English Characters, so this embod...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a Chinese environment-oriented complex scene text positioning method, which mainly solves the problem that scene text positioning in a complex background has a high false alarm rate in the prior art. In the method, the advantages of an edge-based text positioning method and a texture-based text positioning method are combined, and edge features and texture features are combined by stages. The method comprises the following steps of: detecting a text area by using the edge features, namely performing edge extraction and binarization processing on a down-sampled color image, connecting the edges of characters to form a block through morphological operation, extracting the features of each communicated domain, and removing a great number of non-character communicated domains by using a cascaded threshold classifier to acquire candidate character communicated domains; and extracting the texture features of the candidate character communicated domains, and further judging whether the candidate character communicated domains are character communicated domains or not by using a back propagation (BP) network classifier. The method has the advantages of high text positioning accuracy and high speed under the conditions of complex illumination and backgrounds, and can be used for text automatic extraction and identification in a complex scene.

Description

technical field [0001] The invention belongs to the technical field of image processing, relates to an image text positioning method, and can be used for automatic text extraction and recognition. Background technique [0002] Natural scenes contain rich text information, such as shop signs, road signs and various warning or prompt signs, etc. If people can extract these information and do further processing, it will have a broad application prospect. For example, the text can be recognized by the optical character recognition OCR module, which can realize the retrieval of images or videos; or the text-to-speech TTS module can be used to recover the text in the form of sound, which will bring great benefits to international tourists and the blind. Great convenience. Under various demands, extracting text information from images has become a very practical topic. The text positioning is the premise of correct text information extraction, and it is the key part of the whole ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/60G06K9/00
Inventor 卢朝阳刘晓佩李静汪文芳王伟
Owner XIDIAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products