Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

48 results about "Text alignment" patented technology

Data entity relationship extraction method based on deep learning

The invention discloses a data entity relationship extraction method based on deep learning. The method comprises the following steps: 1, obtaining training data by adopting an open entity relationship extraction method; mapping the data entity relationship instances to a large number of texts in an entity knowledge base by means of a DBPedia, OpenCyc, YAGO or FreeBase entity knowledge base, obtaining training data through a text alignment method, and obtaining training corpora with noise annotations; re-annotating the noise annotation by adopting a supervised entity relationship extraction method, and training a machine learning model on the basis of annotated training data; and extracting a data entity relationship corresponding to the entity pair combination. According to the method, the data entity relationship is extracted by combining the open entity relationship extraction method and the supervised entity relationship extraction method. The training data acquisition efficiency of the open entity relationship extraction method is high. The training data acquired by the supervised entity relationship extraction method is high in accuracy. The extraction efficiency and the accuracy of the entity relationship are improved.
Owner:福建奇点时空数字科技有限公司

Differential description statement generation method and device, equipment and medium

ActiveCN114511860AImprove accuracyGuaranteed normal reasoning functionCharacter and pattern recognitionText alignmentNoise monitoring
The invention discloses a difference description statement generation method and device, equipment and a medium, and relates to the technical field of artificial intelligence, and the method comprises the steps: carrying out the feature splicing of image coding features and text coding features, inputting the spliced coding features into a preset image-text alignment unit constructed based on a preset self-attention mechanism to obtain spliced alignment features, using a preset noise monitoring unit constructed based on a preset self-attention mechanism and a preset cross-attention mechanism to process image alignment features and text alignment features obtained after splitting the text coding features and the spliced alignment features so as to extract difference signals; the difference description statement is generated by utilizing the preset difference description generation algorithm and based on the difference signal, and therefore, the part, which cannot be aligned with the image, in the human language text is positioned based on the preset cross-attention mechanism, and the corresponding interpretation description is given, so that the problem that a computer cannot perform normal reasoning due to human language errors is solved.
Owner:SUZHOU LANGCHAO INTELLIGENT TECH CO LTD

Voice data labeling method and system, electronic equipment and storage medium

The invention discloses a voice data labeling method and system, electronic equipment and a storage medium, and the method comprises the steps: firstly screening original voice data, and carrying outthe reading text matching of screened voice, and obtaining proofreading voice and proofreading text; performing word segmentation on the proofreading text to obtain a word segmentation text; performing noise reduction on the proofreading voice to obtain noise reduction voice, and inputting the voice features after feature extraction into a VAD model to obtain VAD effective voice duration of the noise reduction voice; carrying out voice forced alignment on the word segmentation text by adopting an acoustic model to obtain the word-level alignment time, word-level time intervals, segmented texts, the segmented text starting time, the ending time and the text alignment time; determining a speech speed, an effective time ratio and an error word number according to the plurality of times, and performing speech quality inspection; and segmenting the original voice according to the starting time and the ending time of the segmented text, and taking the segmented text and the segmented voice as voice annotation results. The voice annotation text with qualified quality can be automatically acquired.
Owner:北京智慧星光信息技术有限公司

Character segmentation and recognition method based on CTC deep neural network

The invention provides a character segmentation and recognition method based on a CTC deep neural network. The method comprises the following steps: a1, extracting features of an input image by usinga CNN; a2, carrying out the CELL segmentation on the features extracted in a1, fixing the height and width of CELL, and determining the number of CELL by the length of the image; a3, directly segmenting and classifying each CELL of the determined features, and outputting segmentation signals; a4, calculating the loss between the real segmentation signal and the segmentation signal output by the model by using CTCLOSS, feeding back the loss condition and training the whole model; a5, segmenting the text by using the segmentation signal output in the step a3, carrying out CNN + softmax classification identification on a single character, mapping a real segmentation signal from the annotated text, and automatically solving the text alignment problem by using the CTCLOSS. According to the invention, the OCR recognition speed is improved, and the recognition optimization is targeted after the characters are cut into single characters, so the final precision is improved; a recognition framework is improved, and a recognition process is separated into character segmentation and single character recognition, so optimization can be separately carried out in a targeted manner.
Owner:北京深智恒际科技有限公司

Sound and text realignment and information presentation method and device, electronic equipment and storage medium

The invention provides a sound and text realignment and information presentation method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a target audio, a pre-editing recognition text and a post-editing recognition text, wherein the pre-editing recognition text is a recognition text obtained through the automatic voice recognition of the target audio, and the edited text is a text obtained by editing the recognition text before editing; performing forced alignment on the target audio and the recognition text before editing to determine audio starting and ending time corresponding to each character in the recognition text before editing; performing text alignment on the pre-edited recognition text and the post-edited recognition text to determine a character corresponding to each character in the post-edited recognition text in the pre-edited recognition text; and for each character in the edited recognition text, determining the audio starting and ending time of the character corresponding to the character in the pre-edited recognition text as the audio starting and ending time of the character. According to the invention, high-precision sound and text re-alignment between the target audio and the edited recognition text is realized.
Owner:BEIJING ZITIAO NETWORK TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products