Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

927 results about "Word embedding" patented technology

Word embedding is the collective name for a set of language modeling and feature learning techniques in natural language processing (NLP) where words or phrases from the vocabulary are mapped to vectors of real numbers. Conceptually it involves a mathematical embedding from a space with many dimensions per word to a continuous vector space with a much lower dimension.

Natural language processing-based multi-language analysis method and device

The invention discloses a natural language processing-based multi-language analysis method and device. The method comprises the following steps of: selecting to input a natural language text information language category through a language detection training model; obtaining word embedding expression information of corresponding words which can be recognized by a computer through a trained word vector model, and extracting a keyword of the obtained word embedding expression information through a TF-IDF manner; calculating an article vector and a category vector of each preset category according to the keyword and a keyword weight, and calculating a similarity between an article of natural language text information and each preset category so as to determine a text classification result ofthe natural language text information; and inputting the word embedding expression information of the natural language text information into a trained convolutional neural network and a parallel-framework text emotion analysis model of a bidirectional gate circulation unit, and obtaining a final emotion tendency value through calculation. According to the method and device, the problem that traditional multi-language analysis method needs to know domain knowledges of related linguistics and needs plenty of manpower to carry out operation is solved.
Owner:北京百分点科技集团股份有限公司

A model method based on paragraph internal reasoning and joint question answer matching

The invention discloses a reading understanding model method based on paragraph internal reasoning and joint question answer matching, and the method comprises the following steps: S1, constructing avector for each candidate answer, the vector representing the interaction of a paragraph with a question and an answer, and then enabling the vectors of all candidate answers to be used for selectinganswers; S2, carrying out experiment. According to the model provided by the invention, paragraphs are firstly segmented into blocks under multiple granularities; an encoder is used for summing the intra-block word embedding vectors by utilizing neural word bag expression; then, a relationship between blocks with different granularities where each word is located through a two-layer forward neuralnetwork is modeled to construct a gating function, so that the model has greater context information and captures paragraph internal reasoning at the same time. Compared with a baseline neural network model such as a Stanford AR model and a GA Reader, the accuracy of the model is improved by 9-10%. Compared with a recent model SurfaceLR, the accurcay is at least improved by 3% and is about 1% higher than that of a single model of the TriAN, and in addition, the model effect can also be improved through pre-training on an RACE data set.
Owner:SICHUAN UNIV

A fine-grained emotion polarity prediction method based on a hybrid attention network

ActiveCN109948165AAccurate predictionMake up for the shortcoming that it is difficult to obtain global structural informationSpecial data processing applicationsSelf attentionAlgorithm
The invention discloses a fine-grained emotion polarity prediction method based on a hybrid attention network, and aims to overcome the problems of lack of flexibility, insufficient precision, difficulty in obtaining global structure information, low training speed, single attention information and the like in the prior art. The method comprises the following steps: 1, determining a text context sequence and a specific aspect target word sequence according to a comment text sentence; 2, mapping the sequence into two multi-dimensional continuous word vector matrixes through log word embedding;3, performing multiple different linear transformations on the two matrixes to obtain corresponding transformation matrixes; 4, calculating a text context self-attention matrix and a specific aspect target word vector attention matrix by using the transformation matrix, and splicing the two matrixes to obtain a double-attention matrix; 5, splicing the double attention matrixes subjected to different times of linear change, and then performing linear change again to obtain a final attention representation matrix; and 6, through an average pooling operation, inputting the emotion polarity into asoftmax classifier through full connection layer thickness to obtain an emotion polarity prediction result.
Owner:JILIN UNIV

Chinese image semantic description method combined with multilayer GRU based on residual error connection Inception network

The invention discloses a Chinese image semantic description method combined with multilayer GRU based on a residual error connection Inception network, and belongs to the field of computer vision andnatural language processing. The method comprises the steps: carrying out the preprocessing of an AI Challenger image Chinese description training set and an estimation set through an open source tensorflow to generate a file at the tfrecord format for training; pre-training an ImageNet data set through an Inception_ResNet_v2 network to obtain a convolution network pre-training model; loading a pre-training parameter to the Inception_ResNet_v2 network, and carrying out the extraction of an image feature descriptor of the AI Challenger image set; building a single-hidden-layer neural network model and mapping the image feature descriptor to a word embedding space; taking a word embedding characteristic matrix and the image feature descriptor after secondary characteristic mapping as the input of a double-layer GRU network; inputting an original image into a description model to generate a Chinese description sentence; employing an evaluation data set for estimation through employing the trained model and taking a Perplexity index as an evaluation standard. The method achieves the solving of a technical problem of describing an image in Chinese, and improves the continuity and readability of sentences.
Owner:HARBIN UNIV OF SCI & TECH

Deep learning model-based image Chinese description method

The invention discloses a deep learning model-based image Chinese description method and belongs to the field of computer vision and natural language processing. The method comprises the steps of preparing an ImageNet image data set and an AI Challenger image Chinese description data set; pre-training the ImageNet image data set by utilizing a DCNN to obtain a pre-trained DCNN model; performing image feature extraction and image feature mapping on the AI Challenger image Chinese description data set, and transmitting image features to a GRU threshold recursive network recurrent neural network;performing word coding matrix construction on an AI Challenger image mark set in the AI Challenger image Chinese description data set; extracting word embedding features by utilizing an NNLM, and finishing text feature mapping; taking the GRU threshold recursive network recurrent neural network as a language generation model, and finishing image description model building; and generating a Chinese description statement. According to the method, the blank of image Chinese description is filled up; a function of automatically generating the image Chinese description is realized; the accuracy ofdescription contents is well improved; and a foundation is laid for development of Chinese NLP and computer vision.
Owner:HARBIN UNIV OF SCI & TECH

Multi-source and multi-label text classification method and system based on improved seq2seq model

The invention belongs to the technical field of natural language processing text classification, in particular to a multi-source multi-label text classification method based on an improved seq2seq model and a system thereof. The method comprises the following steps: data input and pretreatment, word embedding, encoding, encoding and splicing, decoding, model optimization and prediction output. Themethod of the invention has the following beneficial effects: adopting a seq2seq depth learning framework, constructing a plurality of encoders, and combining the attention mechanism to be used for atext classification task, so as to maximize the use of multi-source corpus information and improve the classification accuracy of the multi-label; In the error feedback process of decoding step, according to the characteristics of multi-label text, an intervention mechanism is added to avoid the influence of label sorting, which is more in line with the essence of multi-label classification problem. The encoder adopts the circulating neural network, which can learn according to the time step effectively. The decoding layer adopts one-way loop neural network and adds attention mechanism to highlight the learning focus.
Owner:广州语义科技有限公司

Answer generation method based on multi-layer Transformer aggregation encoder

The invention discloses an answer generation method based on a multilayer Transformer aggregation encoder, and the method comprises the steps: receiving input information which comprises paragraph article information and question information; converting the input information through a character embedding layer and a word embedding layer to obtain corresponding character vectors and word vectors; splicing the character vector and the word vector to obtain a spliced word vector; performing addition splicing on the spliced word vector and the position coding vector to obtain an input sequence; inputting the input sequence into a multi-layer Transformer aggregation encoder to obtain higher-level semantic information; inputting higher-level semantic information into a context-question attentionlayer, and learning question and answer information; inputting a learning result into an encoding layer comprising three multi-layer Transformer aggregation encoders, and obtaining a starting position and an ending position through a softmax function; and taking the content determined by the starting position and the ending position as a target answer. By applying the embodiment of the invention,the problems of information loss and insufficient performance in the prior art are solved.
Owner:SHANGHAI MARITIME UNIVERSITY

Relation extraction method based on Bi-LSTM input information enhancement

The invention provides a relation extraction method based on Bi-LSTM input information enhancement and belongs to the field of artificial intelligence natural language processing of computers. The method comprises the steps that by applying a strategy annotation dataset of an indeterminate label, a redundancy encoding technology is used for conducting character-level encoding on each word to generate a word form encoding vector; the word form encoding vector and a word embedding vector are spliced to generate a word vector used for capturing word form and word meaning information; Bi-LSTM of input information enhancement is used as a model encoding layer, the word vector is input to an encoding layer, and the encoding vector is output; the encoding vector is input into a decoding layer, and a decoding vector is obtained; by applying three layers of NN, an entity label, a relation type and entity number information are extracted from the decoding vector; finally, the gradient is calculated, the weight is updated, and a model is trained through a maximum target function. By means of the relation extraction method, the robustness of the system is improved, interference information caused by non-entity words is reduced, and the accuracy rate and recall rate of relation extraction are effectively increased.
Owner:BEIJING INSTITUTE OF TECHNOLOGYGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products