Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Complex optical text sequence identification system based on convolution and recurrent neural network

A technology of recursive neural network and convolutional neural network, applied in biological neural network models, character and pattern recognition, neural architecture, etc., can solve the problems of character size and gap difference, cumbersome and time-consuming, loss of available information in pictures, etc.

Inactive Publication Date: 2016-06-15
成都数联铭品科技有限公司
View PDF3 Cites 32 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Conventional OCR methods include image segmentation, feature extraction, single character recognition and other processing processes, wherein image segmentation includes a large number of image preprocessing processes, such as tilt correction, background denoising, and single character extraction; these processes The process is not only cumbersome and time-consuming, but also may cause the image to lose a lot of available information; and when the image to be recognized contains a string of multiple characters, the traditional OCR method needs to divide the original string into several small images containing a single character for separation. Recognition, and the most commonly used method for text segmentation is the projection method, that is, after binarizing the image text, find the dividing line between the two texts through vertical projection, and separate the text according to the dividing line. This method The main problems are: when the image text sequence to be recognized contains background noise, character distortion, character bonding, etc., it will cause difficulty in text segmentation
Especially when the image text sequence to be recognized is mixed with Chinese characters, letters, numbers, and symbols with left and right radicals, or characters in half-width and full-width formats are mixed in the image text sequence to be recognized, the character size and gaps exist due to the difference in format. The difference is that the single character in the image text sequence to be recognized cannot be accurately segmented through the simple projection method
Once there is a problem with segmentation, it is difficult to obtain accurate recognition results.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Complex optical text sequence identification system based on convolution and recurrent neural network
  • Complex optical text sequence identification system based on convolution and recurrent neural network
  • Complex optical text sequence identification system based on convolution and recurrent neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0041] The present invention will be further described in detail below in conjunction with test examples and specific embodiments. However, it should not be understood that the scope of the above subject matter of the present invention is limited to the following embodiments, and all technologies realized based on the content of the present invention belong to the scope of the present invention.

[0042] The present invention provides as figure 1 The technical solution shown: complex optical character sequence recognition system based on convolution and recursive neural network, including image and character input module, sliding sampling module, convolutional neural network and recurrent neural network classifier,

[0043] Wherein the sliding sampling module comprises a sliding sampling frame, and the sliding sampling frame carries out sliding sampling to the image text sequence to be recognized input by the image text input module (scanner, digital camera or image tex...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to the image and text identification field, and specifically relates to a complex optical text sequence identification system based on convolution and recurrent neural network. The complex optical text sequence identification system includes an image and text input module, a sliding sampling module, a CNN and an RNN, wherein the image and text input module is a scanner, a digital camera or an image and text storage module; the sliding sampling module performs sliding sampling of an image and text sequence to be identified, and inputs the sampling sub images in the CNN; the CNN extracts the characteristics and outputs the characteristics to the RNN; and the RNN successively identifies the front part of each character, the back part of each character, numbers, letters, punctuation, or blank according to the CNN input signal and the output data of the CNN for the last moment. The complex optical text sequence identification system based on convolution and recurrent neural network can realize complex image and text sequence identification, can overcome the cutting problem, and can significantly improve the identification efficiency and accuracy for the complex image and text sequence.

Description

technical field [0001] The invention relates to the field of image text recognition, in particular to complex optical text sequence recognition based on convolutional and recursive neural networks Background technique [0002] With the development of society, there is a large demand for the digitization of ancient books, documents, bills, business cards and other paper media. The digitization here is not limited to "photographic" using scanners or cameras, but more importantly, the Files are converted into readable and editable documents for storage. This process requires image text recognition on scanned pictures, and traditional image text recognition is optical text recognition (OCR). [0003] Conventional OCR methods include image segmentation, feature extraction, single character recognition and other processing processes, wherein image segmentation includes a large number of image preprocessing processes, such as tilt correction, background denoising, and single char...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/20G06K9/62G06N3/04
CPCG06N3/04G06V10/22G06F18/285G06F18/214
Inventor 刘世林何宏靖陈炳章吴雨浓姚佳
Owner 成都数联铭品科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products