Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Rapid and automatic table extraction method based on deep neural network

A deep neural network and automatic extraction technology, which is applied in the application field of neural network, can solve the problems of easy mis-extraction of image contrast, inconsistent text spacing, slow speed, etc., and achieve good extraction effect, high degree of automation and high efficiency.

Active Publication Date: 2021-06-01
GUIZHOU POWER GRID CO LTD
View PDF16 Cites 3 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

In response to this problem, some people proposed to use word segmentation to extract image tables. This method needs to split a long paragraph of text into words, which is slow and inefficient. It is affected by inconsistent text spacing, table tilt, image brightness, image Influenced by contrast, etc., it is also easy to mis-extract

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Rapid and automatic table extraction method based on deep neural network
  • Rapid and automatic table extraction method based on deep neural network
  • Rapid and automatic table extraction method based on deep neural network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0075] Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings. It should be understood that the preferred embodiments are only for illustrating the present invention, but not for limiting the protection scope of the present invention.

[0076] like figure 1 Shown, a kind of table automatic extraction method based on depth neural network of the present invention comprises the following steps:

[0077] Step S1: performing image form correction preprocessing; including establishing an image correction preprocessing model, and then automatically correcting the input image form, including performing brightness correction on the image form and geometric correction on the image form.

[0078] (1) Image table brightness correction

[0079] The original image is affected by the shooting lighting conditions, shooting sensor or other factors, and the brightness of the image will be inconsistent and the differe...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

Design data of an overhead transmission line generally appear in the forms of a picture table, a PDF table and the like, data formats and representation forms provided by different design units are different, table data cannot be directly read by a computer, manual input is needed, the workload is large, and the speed is low. In order to solve the problem, the invention discloses a fast and automatic table extraction method based on a deep neural network, and the method comprises the steps: S1, carrying out the correction preprocessing of an image table; S2, carrying out image table frame line extraction to obtain a cut cell picture Q (i); S3, carrying out cell character positioning; S4, recognizing characters of the cells; and S5, performing sequential merging on all the cell recognition results to realize automatic and rapid table extraction. Compared with an existing table extraction method based on manual visual interpretation or single character segmentation, the method is higher in speed, higher in automation degree, more accurate and complete in recognition, higher in generalization ability and higher in efficiency, can still keep a good extraction effect on multi-source heterogeneous and complex image table data, and acceptance item judgment and acceptance work management are facilitated.

Description

technical field [0001] The invention relates to the technical field of neural network applications, in particular to a method for quickly and automatically extracting tables based on a deep neural network. Background technique [0002] Overhead transmission line design materials exist in the form of picture tables, PDF tables, etc. The types of materials are complicated. The design handover materials provided by different design units are different in content, format, and form of expression. It is difficult for the design handover data to be directly read by computers. Effective analysis requires manual input, but the manual extraction method has a large workload, long cycle, slow speed, low efficiency, high cost, and error-prone, especially for large-scale tabular data is extremely limited. At the same time, due to the lack of standardized management of design transfer data, it is difficult to quickly locate and browse design data, and it is not convenient for the judgment ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/38G06N3/04G06N3/08
CPCG06N3/08G06V30/414G06V10/28G06N3/044G06N3/045
Inventor 李晓春彭赤徐梁刚时磊杨恒陈科羽赵建余江顺周振锋杨渊
Owner GUIZHOU POWER GRID CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products