Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

82 results about "Word graph" patented technology

Chinese phonetic symbol keyword retrieving method based on feed forward neural network language model

The invention provides a Chinese phonetic symbol keyword retrieving method based on a feed forward neural network language model. The method comprises: (1), an input sample including historical words and target words are inputted into a feed forward neural network model; for each target word wi, a plurality of noise words with probability distribution q (wi) are added and an active output of a last hidden layer is transmitted to the target words and nodes where the noise words are located, and conversion matrixes between all layers are calculated based on an objective function; errors between an output of an output layer and the target words are calculated, all conversion matrixes are updated until the feed forward neural network model training is completed; (2), a target word probability of inputting a word history is calculated by using the feed forward neural network model; and (3), the target word probability is applied to a decoder and voice decoding is carried out by using the decoder to obtain word graphs of multiple candidate identification results, the word graphs are converted into a confusion network and an inverted index is generated; and a keyword is retrieved in the inverted index and a targeted key word and occurrence time are returned.
Owner:INST OF ACOUSTICS CHINESE ACAD OF SCI +1

Information recommendation method and device based on user portrait, equipment and storage medium

The invention relates to the technical field of big data, and discloses a user portrait-based information recommendation method. The method comprises the following steps: receiving a man-machine conversation chat record in a first scene in real time through a log collection system Flume; performing desensitization processing on the chat record to obtain first data; performing stop word removal processing on the first data to obtain second data; extracting a keyword of the second data through a preset word graph; obtaining a first label set according to the keyword; performing duplicate removalon the first label set data to obtain a second label set; generating a user interest portrait based on the second label set, and storing the user interest portrait in a database; receiving a recommendation instruction, and obtaining a user interest portrait according to the recommendation instruction; and obtaining to-be-recommended information corresponding to the recommendation instruction according to the user interest portrait. The invention also provides an information recommendation device and equipment based on the user portrait, and a storage medium. The information recommendation method based on the user portrait provided by the invention improves the accuracy of the user portrait.
Owner:CHINA PING AN LIFE INSURANCE CO LTD

Text information extraction method and device, server and storage medium

PendingCN109408826AReflect the relationshipAccurately and fully determineNatural language data processingSpecial data processing applicationsAlgorithmTheoretical computer science
The embodiment of the invention provides a text information extraction method and device, a server and a storage medium. The method comprises the steps that word vectors of candidate words in a text are determined through a Word2Vec model, and similarity values between different word vectors are determined; taking the word vectors as nodes, and constructing edges between the nodes according to thesimilarity values between the word vectors to obtain a candidate word graph set; determining candidate word weights according to the candidate word atlas through a TextRank algorithm; and determiningkeywords of the text according to the weights of the candidate words. The method comprises the following steps of: converting candidate words into word vectors by adopting a Word2Vec model; Accordingto the method, the candidate words can be represented through the low-dimensional vectors, the processing efficiency is improved, the association relationship between the candidate words can be vividly reflected through similarity value calculation and image set construction, and finally the weight values of the candidate words are calculated through the TextRank algorithm, so that the keywords of the text are more accurately and comprehensively determined.
Owner:RUN TECH CO LTD BEIJING

Audio data labeling method, device and system

The invention discloses an audio data labeling method, device and system. The method comprises the following steps: performing voice recognition on to-be-labeled audio data by using a voice recognition engine to obtain a reference labeled text; searching an optimal identification path with the shortest editing distance from the reference labeled text in a word graph network obtained by decoding the to-be-labeled audio data; calculating the confidence coefficient of each word on the optimal recognition path, comparing the confidence coefficient of each word with a preset first confidence coefficient condition, and outputting a target word meeting the first confidence coefficient condition on the optimal recognition path; and aligning the target word according to the time parameter of each word in the word graph network to form an labeled text of the to-be-labeled audio data. The words in the word graph network of the to-be-labeled audio data are distinguished according to the confidencecoefficients, the words with high confidence coefficients are extracted to form the annotation text of the to-be-labeled audio data, the words with low confidence coefficients are annotated, audio data annotation is automatically completed, annotation efficiency is improved, and annotation accuracy is improved.
Owner:SUNING CLOUD COMPUTING CO LTD

Voice recognition method and system based on deep neural network acoustic model

The invention discloses a voice recognition method and system based on a deep neural network acoustic model. The method comprises steps of carrying out the sliding windowing preprocessing operation of a to-be-recognized voice, and extracting acoustic features; constructing and training a deep neural network acoustic model; calculating a likelihood probability corresponding to the extracted acoustic features by using the deep neural network acoustic model; constructing a static decoding graph, a decoder constructing a directed acyclic graph containing all recognition results as a decoding network through the static decoding graph and the likelihood probability on the basis of a viterbi algorithm of dynamic programming, and a word graph of the state level being obtained from the decoding network and being determined to obtain the word graph of the word level; and obtaining an optimal cost path word graph of the word-level word graph, obtaining a word sequence corresponding to an optimal state sequence of the word graph, taking the word sequence as a final recognition result, and completing voice recognition. According to the method, gradient dispersion and gradient explosion caused by a complex structure network model can be solved, the word error rate is reduced while the decoding speed is ensured, and recognition accuracy is improved.
Owner:XI AN JIAOTONG UNIV

Keyword detection method and system, mobile terminal and storage medium

The invention provides a keyword detection method and system, a mobile terminal and a storage medium. The method comprises the steps of: acquiring text corpus and a transliteration text to perform model training on a language model; performing model training on a chain model according to acoustic features in a training set, and combining the chain model with the language model to obtain a speech recognition model; inputting a to-be-detected speech segment into the speech recognition model for analysis to obtain a word graph, and performing inverted indexing on the word graph; converting the index result into a factor converter, and inputting a preset keyword into the factor converter for retrieval to obtain a keyword retrieval result; and calculating the occurrence probability of the preset keyword according to the keyword retrieval result, and when the occurrence probability is greater than a probability threshold, judging that the preset keyword occurs in the to-be-detected speech segment. According to the invention, the speech recognition model is controlled to decode the to-be-detected speech segment to generate the word graph, so that the situation of keyword detection errorscaused by speech recognition errors is avoided, and the accuracy of keyword detection is improved.
Owner:XIAMEN KUAISHANGTONG TECH CORP LTD

Voice recognition method and system based on incremental word graph re-scoring

The invention discloses a voice recognition method and system based on incremental word graph re-scoring. The method comprises the steps that: a to-be-recognized voice signal is obtained and acousticfeatures are extracted; a likelihood probability corresponding to the acoustic features is calculated by using a trained acoustic model; a decoder constructs a corresponding decoding network, obtainsa word graph of a state level from the decoding network and obtains a word graph of a word level by updating the word graph and determining the word graph; state-level word graphs of remaining decoding networks are determined, and the determined state-level word graphs are combined with the obtained word-level word graphs to generate a decoded word graph; a target word graph is obtained through afinite-state transcriber merging algorithm according to a re-scoring language model obtained through one-time decoded word graph and small corpus training; and an optimal cost path word graph of the target word graph is obtained, then a corresponding word sequence is obtained, and the word sequence is taken as a final recognition result. According to the invention, the calculation amount of determination after the decoding of a common decoder is finished is reduced, and the decoding speed is accelerated; the word error rate of speech recognition in a specific scene is reduced, and the accuracyis improved.
Owner:XI AN JIAOTONG UNIV

Entity identification method, terminal equipment and storage medium

The invention relates to an entity recognition method, terminal equipment and a storage medium. The method comprises the steps of S1, constructing a word graph containing domain entities correspondingto a to-be-recognized text; S2, expressing each word in the to-be-recognized text as a vocabulary tensor through a word vector embedding layer; S3, extracting candidate entities corresponding to theto-be-recognized text from the constructed word graph through a graph neural network module according to all vocabulary tensors of the to-be-recognized text, wherein the graph neural network module comprises a graph attention network layer and a bidirectional graph convolutional network layer; S4, converting a vocabulary tensor and a candidate entity of the text to be recognized into an intermediate calculation tensor containing context information through a bidirectional recurrent neural network layer; and S5, inputting the intermediate calculation tensor into a CRF decoding layer for decoding to obtain an entity contained in the finally recognized text to be recognized. According to the method, the secondary graph structure of the entity boundary is modeled, and the relationship betweenthe entity boundary and the graph neural network is analyzed, so that influence of insufficient judgment of the entity boundary on the result accuracy is reduced.
Owner:厦门渊亭信息科技有限公司

Voice recognition implementing method and system based on confidence coefficient

ActiveCN106782513APhoneme synchronization decoding is accurateProbability Confusion Network Competition Probability AccuracySpeech recognitionSpeech identificationLanguage model
The invention relates to a voice recognition implementing method and system based on a confidence coefficient. The method comprises the following steps: carrying out voice recognition of phoneme synchronous decoding from voice of a use to obtain a word graph acoustic information structure which is synchronous to decoding information generated phoneme, generating a confusion network on the basis of the word graph acoustic information structure so as to establish a competitive relation between voice recognition candidate results, namely a competition probability of the confusion network; meanwhile, establishing full search space of voice recognition by using an auxiliary search network based on a linguistic model; calculating to obtain complete lossless full search space probability; recording a search process of the generated full search space by voice recognition of phoneme synchronous decoding; carrying out path recalling by a whole search history so as to obtain the full search space probability; and finally, fusing the competition probability of the confusion network and the full search space probability to obtain a judgment result of voice recognition. On the one hand, correct confidence coefficient can be given for a voice recognition result so as to improve user experiences of voice recognition; and on the other hand, consumption of calculation and memory resources of voice recognition confidence coefficient algorithm can be reduced remarkably.
Owner:AISPEECH CO LTD

Multi-factor fused textrank keyword extraction algorithm

PendingCN110728136ASmall local feature influenceGlobal features have a significant impactNatural language data processingAlgorithmTheoretical computer science
The invention relates to the technical field of natural language processing, and especially relates to a multi-factor fused textrank keyword extraction algorithm. Influence factors of the keyword extraction algorithm TextRank include five factors including word coverage, word position, word frequency, word length, word span and the like. 1, global factors are greater than local factors in a keyword extraction process; 2, the word coverage, the word length, the word frequency, the word span and the word position influence weight are gradually increased; 3, the influence weights of the word coverage and the word length are basically equivalent, the word span and the word frequency influence weight are basically equivalent when the keyword of the text is extracted by using the TextRank algorithm, only two factors of word positions and word spans can be considered; wherein the ratio of the two factors is 7: 3; 3, because the text needs to be traversed again on the basis of establishing a word graph when the word span is calculated, a certain running time needs to be consumed, if the requirement on the running speed of the algorithm is strict, the word span can be replaced by the word frequency, and the extraction effect is slightly influenced, but is also good.
Owner:YANAN UNIV

Speech identification method and device, computer equipment and storage medium

The application belongs to the technical field of the artificial intelligence, and relates to a speech identification method and device, computer equipment, and a storage medium. The method comprisesthe following steps of acquiring to-be-identified speech information; inputting the to-be-identified speech information into a local first word graph model to perform decoding search to obtain a firstsearch result, wherein the first research result comprises a first path and a corresponding first path score, the first word graph model comprises an acoustic model, a pronunciation dictionary and afirst word graph space; inputting the first search result into a local second word and graph model to search so as to obtain a second search result, wherein the second search result comprises a secondpath and a corresponding second path score, the second word and graph model comprises a second word and graph space, and the first word and graph space is the sub-word and graph space of the second word and graph space; and selecting the corresponding second path to output according to the second path score. The search dimension is lowered, the amount of the word and graph search is reduced, andthe search time is shortened, and the speech identification speed is improved.
Owner:PING AN TECH (SHENZHEN) CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products