Video caption text extraction and identification method

A text extraction and recognition method technology, applied in character and pattern recognition, instruments, computer parts and other directions, can solve problems such as poor adaptability, low resolution, character segmentation, character image noise, etc., to avoid uneven spacing Effect

Active Publication Date: 2012-01-25
北京中科阅深科技有限公司
View PDF2 Cites 49 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In general, the existing video subtitle text extraction and recognition technology is not perfect, mainly reflected in: for complex and changeable backgrounds, especially those with similar or even the same color as the foreground cannot be processed; for changeable fonts, Such as: Song, Hei, Kai, etc. are not very adaptable; character recognition adopts the conventional OCR method, and the character segmentation, character image noise, and low resolution are not considered enough

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video caption text extraction and identification method
  • Video caption text extraction and identification method
  • Video caption text extraction and identification method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0031] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be described in further detail below in conjunction with specific embodiments and with reference to the accompanying drawings.

[0032] The present invention can be implemented on computing devices such as personal computers and servers.

[0033] The technical solution adopted by the present invention is: divide the subtitle area in the video into lines, automatically judge the foreground color for each line and generate a binary image, segment and recognize the characters in the binary image, and obtain the final text recognition result. The positioning of subtitles in the video is not the content of the present invention, and it is assumed that the subtitle area has been obtained by positioning by other methods.

[0034] refer to figure 1 , the video subtitle text extraction and recognition method that the present invention proposes specifically...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a video caption text extraction and identification method which comprises the following steps of: inputting an image of a caption area in a video; converting the input image into a grayscale; judging the arrangement direction of characters in the caption area; counterclockwise rotating the caption area in which vertical arrangement is adopted 90 degrees to obtain a horizontal caption area; lining the caption area to obtain single-line caption images; automatically judging foreground colors of the single-line caption images to obtain binary single-line caption images; over-segmenting the binary single-line caption images to obtain character segment sequences; and performing text line identification on the over-segmented binary single-line caption images. By utilizing the method, horizontal and vertical video caption text lines can be effectively segmented, the foreground colors of the characters can be accurately judged, noises can be filtered, and accurate character segmentation and identification results can be obtained; and the method can be applicable to a plurality of purposes such as video and image content editing, indexing, retrieving and the like.

Description

technical field [0001] The invention belongs to the field of pattern recognition and computer vision, in particular to a processing method for text detection and recognition in video images. Background technique [0002] As one of the most popular forms of media, video is widely disseminated through the Internet and television. In order to make users find the video content they are interested in more conveniently and quickly, video retrieval and classification has gradually become a research hotspot in the field of pattern recognition and computer vision. Among them, the text information in the video, especially the subtitle information, has the most significant effect on video retrieval and classification. This is because: (1) the text information in the video is closely related to the current content of the video; (2) the characters in the video have very obvious visual features, which are easy to extract; (3) character recognition (OCR) technology is relatively Recognit...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/20G06K9/32G06K9/46
Inventor 刘成林白博殷飞
Owner 北京中科阅深科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products