Character segmentation and recognition method based on CTC deep neural network

A deep neural network and text segmentation technology, which is applied in neural learning methods, character and pattern recognition, biological neural network models, etc., can solve the problems of increased difficulty in structuring, high cost of labeling, and time-consuming recognition modules. The effects of improved precision, unique ideas, and novel ideas

Pending Publication Date: 2020-08-07
北京深智恒际科技有限公司
View PDF8 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The text information in the image content is structured based on templates and rule logic. The existing frameworks have certain disadvantages. Under the first framework above, the detection needs to mark the position of each character, and the cost of labeling is extremely high. At the same time, the difficulty of structuring is greatly increased. Therefore, the specific goal of general detection tasks is to detect text lines, rather than to detect individual characters independently; under the second framework, the recognition module takes a lot of time

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Character segmentation and recognition method based on CTC deep neural network
  • Character segmentation and recognition method based on CTC deep neural network
  • Character segmentation and recognition method based on CTC deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0025] The following will clearly and completely describe the technical solutions in the embodiments of the present invention in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiment is only one, not all embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0026] Below, in conjunction with accompanying drawing and specific embodiment, the invention is further described:

[0027] Such as figure 1 As shown, the method for text segmentation and recognition based on CTC deep neural network includes the following steps:

[0028] a1. Use CNN to extract features from the input image;

[0029] a2. Carry out CELL segmentation on the features extracted in step a1, the height and width of CELL are fixed, and the number is ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a character segmentation and recognition method based on a CTC deep neural network. The method comprises the following steps: a1, extracting features of an input image by usinga CNN; a2, carrying out the CELL segmentation on the features extracted in a1, fixing the height and width of CELL, and determining the number of CELL by the length of the image; a3, directly segmenting and classifying each CELL of the determined features, and outputting segmentation signals; a4, calculating the loss between the real segmentation signal and the segmentation signal output by the model by using CTCLOSS, feeding back the loss condition and training the whole model; a5, segmenting the text by using the segmentation signal output in the step a3, carrying out CNN + softmax classification identification on a single character, mapping a real segmentation signal from the annotated text, and automatically solving the text alignment problem by using the CTCLOSS. According to the invention, the OCR recognition speed is improved, and the recognition optimization is targeted after the characters are cut into single characters, so the final precision is improved; a recognition framework is improved, and a recognition process is separated into character segmentation and single character recognition, so optimization can be separately carried out in a targeted manner.

Description

technical field [0001] The present invention relates to the technical field of text segmentation and recognition, in particular to a method for text segmentation and recognition based on a CTC deep neural network. Background technique [0002] OCR (Optical Character Recognition) is an image processing technology that detects, recognizes, and structures text in images. The current OCR technology is divided into three modules: detection, recognition, and structuring; there are two frameworks for detection and recognition, which are: 1. Single-character detection + single-character recognition framework, specifically expressed as the core task of the detection module is detection Each independent character area of ​​the image; the recognition module is responsible for text recognition for each cropped character area image. The basic framework of the existing recognition model is: CNN+softmax; 2. Text line detection + entire line recognition framework, specifically It is stated...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/34G06K9/62G06N3/04G06N3/08
CPCG06N3/08G06V30/153G06V30/10G06N3/045G06F18/24
Inventor 侯进黄贤俊
Owner 北京深智恒际科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products