Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

2991 results about "Text recognition" patented technology

Realization method of image searching system based on image recognition

The invention discloses a realization method of an image searching system based on image recognition; wherein the realization method comprises the following steps: carrying out preprocessing steps such as format transformation on images in adding images; carrying out recursion multilevel cutting to obtain all potential characters, portraits, objects and object subimage blocks based on uniform space, contrast grade, similarity in color and texture; obtaining marking texts of position and color; carrying out texture and material recognition, character recognition, outline object recognition and text marking on the subimage blocks; carrying out combination and word segmentation on the marking results; establishing index for the word segmentation result and related subimage block information and storing related data of the subimage blocks; in the situation of searching based on images, first identifying images and marking texts; then carrying out word segmentation on the text to obtain index data of a keyword sequence; combining and sequencing the index data of the keyword sequence to obtain related data of the subimage blocks and the images matching with the result set, and returning to users. With the system and method of the invention, the users can input characters and images to retrieve the content of the images.
Owner:程治永

Method and system for name-face/voice-role association

A method for providing name-face / voice-role association includes determining whether a closed captioned text accompanies a video sequence, providing one of text recognition and speech to text conversion to the video sequence to generate a role-name versus actor-name list from the video sequence, extracting face boxes from the video sequence and generating face models, searching a predetermined portion of text for an entry on the role-name versus actor-name list, searching video frames for face models / voice models that correspond to the text searched by using a time code so that the video frames correspond to portions of the text where role-names are detected, assigning an equal level of certainty for each of the face models found, using lip reading to eliminate face models found that pronounce a role-name corresponding to said entry on the role-name versus actor-name list, scanning a remaining portion of text provided and updating a level of certainty for said each of the face models previously found. Once a particular face model / voice model and role-name association has reached a threshold the role-name, actor name, and particular face model / voice model is stored in a database and can be displayed by a user when the threshold for the particular face model has been reached. Thus the user can query information by entry of role-name, actor name, face model, or even words spoken by the role-name as a basis for the association. A system provides hardware and software to perform these functions.
Owner:UNILOC 2017 LLC

PDF document table extraction method, device and equipment and computer readable storage medium

The invention relates to the technical field of artificial intelligence, and discloses a PDF document table extraction method and device, equipment and a computer readable storage medium. The method comprises the steps of obtaining a to-be-identified PDF document, and processing the to-be-identified PDF document; preprocessing the processed PDF document, inputting the preprocessed PDF document into a convolutional neural network, outputting a feature map, inputting the feature map into an RPN region candidate network, and determining a table region; carrying out preprocessing and feature extraction on the table area based on the OCR character recognition technology, obtaining a feature picture, carrying out character detection on the feature picture, determining a text area, carrying out character recognition on the text area, determining text informatio, wherein the text information comprises text position information and text content information; and determining structure informationof the table according to the text coordinate information, dividing each cell of the table based on the structure information, and filling each corresponding cell of the table with a text corresponding to the text content information. According to the method and the device, the accuracy of PDF document table extraction is improved.
Owner:PING AN TECH (SHENZHEN) CO LTD

Method and device for touching character segmentation in character recognition

The present invention discloses a method and a device for touching character segmentation in character recognition, and belongs to the character recognition field. The method comprises the steps of carrying out preprocessing to obtain connected domains, an average character width and an average character height of a row image; carrying out connected domain analysis, marking touching connected domains, carrying out stroke extraction for selected touching connected domains, carrying out segmentation point detection of the touching connected domains to obtain pre-segmentation points, and saving character blocks for non-touching connected domains; merging extracted strokes according to the pre-segmentation points to obtain the character blocks; saving the character blocks and turning to a next connected domain for carrying out touching determination, and outputting a character block sequence after all connected domains having been traversed; and merging the character blocks according to reference information and outputting recognition result. The method and the device provided in the invention merge the strokes according to the pre-segmentation points to obtain the character blocks, guaranteeing that segmentation points in a larger scope can be detected, and take the mode that contour information is used to predetect the segmentation points as a parameter in merging, avoiding merging mistakes caused by merging correct segmentation points.
Owner:HANVON CORP

Method for searching for characters displayed in screen and based on mobile terminal and mobile terminal

The invention provides a method for searching for characters displayed in a screen and based on a mobile terminal. The method includes setting a word fetching tool, wherein the window rank of the word fetching tool is higher than that of an application program of the mobile terminal; utilizing the word fetching tool to intercept picture information on the screen according to the gesture of a user when detecting a triggering command of the user; conducting image-to-text identification operation on the picture information to obtain a plurality of characters, conducting word segmentation on the plurality of characters to obtain a plurality of word groups; acquiring a key word list according to the position of the characters in the word groups on the screen and the position of the word fetching tool on the screen during the picture information intercepting and displaying the key word list; receiving a searching word of the user, conducting searching according to key words in the key word list and displaying a searching result for the user. The method improves user experience, increases the page view of the searched pages, and has quickness, high efficiency and usability. The mobile terminal is further disclosed.
Owner:BAIDU ONLINE NETWORK TECH (BEIJIBG) CO LTD

Complex image and text sequence identification method based on CNN-RNN

The invention relates to the image and text identification field, and specifically relates to a complex image and text sequence identification method based on CNN-RNN. The complex image and text sequence identification method includes the steps: utilizing a sliding sampling box to perform sliding sampling on an image and text sequence to be identified; extracting the characteristics from the sub images obtained through sampling by means of a CNN and outputting the characteristics to an RNN, wherein the RNN successively identifies the front part of each character, the back part of each character, numbers, letters, punctuation, or blank according to the input signal; and successively recording and integrating the identification results for the RNN at each moment and acquiring the complete identification result, wherein the input signal for each moment for the RNN also includes the output signal of a recursion neural network for the last moment and the vector data converted from the recursion neural network identification result for the last moment. The complex image and text sequence identification method based on CNN-RNN can overcome the cutting problem of a complex image and text sequence and the problem that the identification result relies on a language model, thus significantly improving the identification efficiency and accuracy for images and text.
Owner:成都数联铭品科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products