Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

124 results about "Self attention" patented technology

Personalized recommendation method based on deep learning

The invention discloses a personalized recommendation method based on deep learning. The method comprises the steps of according to the viewing time sequence behavior sequence of the user, predictingthe next movie that the user will watch, including three stages of preprocessing the historical behavior characteristic data of the user watching the movie, modeling a personalized recommendation model, and performing model training and testing by using the user time sequence behavior characteristic sequence; at the historical behavior characteristic data preprocessing stage when the user watchesthe movie, using the implicit feedback of interaction between the user and the movie to sort the interaction data of each user and the movie according to the timestamp, and obtaining a corresponding movie watching time sequence; and then encoding and representing the movie data,wherein the personalized recommendation model modeling comprises the embedded layer design, the one-dimensional convolutional network layer design, a self-attention mechanism, a classification output layer and the loss function design. According to the method, the one-dimensional convolutional neural network technologyand the self-attention mechanism are combined, so that the training efficiency is higher, and the number of parameters is relatively small.
Owner:SOUTH CHINA UNIV OF TECH

A fine-grained emotion polarity prediction method based on a hybrid attention network

ActiveCN109948165AAccurate predictionMake up for the shortcoming that it is difficult to obtain global structural informationSpecial data processing applicationsSelf attentionAlgorithm
The invention discloses a fine-grained emotion polarity prediction method based on a hybrid attention network, and aims to overcome the problems of lack of flexibility, insufficient precision, difficulty in obtaining global structure information, low training speed, single attention information and the like in the prior art. The method comprises the following steps: 1, determining a text context sequence and a specific aspect target word sequence according to a comment text sentence; 2, mapping the sequence into two multi-dimensional continuous word vector matrixes through log word embedding;3, performing multiple different linear transformations on the two matrixes to obtain corresponding transformation matrixes; 4, calculating a text context self-attention matrix and a specific aspect target word vector attention matrix by using the transformation matrix, and splicing the two matrixes to obtain a double-attention matrix; 5, splicing the double attention matrixes subjected to different times of linear change, and then performing linear change again to obtain a final attention representation matrix; and 6, through an average pooling operation, inputting the emotion polarity into asoftmax classifier through full connection layer thickness to obtain an emotion polarity prediction result.
Owner:JILIN UNIV

A pedestrian searching method and device based on structural perception self-attention and online instance aggregation matching

The invention discloses a pedestrian searching method and device based on structural perception self-attention and online instance aggregation matching, and belongs to the technical field of computervision technology processing. The method comprises the following steps: firstly, in a training phase, combining a convolutional neural network with a non-local layer; carrying out feature extraction on an input whole scene image to obtain feature representation of the scene image, designing structure-perceived anchor points for a special object of a pedestrian, improving the performance of a detection framework, framing the detected pedestrian into the same size, then sending the pedestrian into a pedestrian re-identification network, and carrying out training, storage, optimization and updating of pedestrian features with tags. In the model testing stage, the trained non-local convolutional neural network is used for carrying out pedestrian detection on an input scene image, and after a pedestrian frame is detected, a target pedestrian image is used for carrying out special similarity matching sorting and retrieval. Pedestrian detection and re-identification can be carried out on large-scale real scene images at the same time, and the method plays an important role in the security and protection fields of urban monitoring and the like.
Owner:CHINA UNIV OF MINING & TECH

Sentence classification method based on LSTM and combining part-of-speech and multi-attention mechanism

The invention discloses a sentence classification method based on LSTM and combining part-of-speech and a multi-attention mechanism. The method comprises steps: each sentence is converted into a semantic word vector matrix and a part-of-speech word vector matrix which are based on continuity and density in an input layer; learning context information of words or part-of-speech in the sentences ina shared bidirectional LSTM layer respectively, connecting learning results of each step in series, and outputting the learning results; a self-attention mechanism and a point multiplication functionare adopted in the self-attention layer to learn important local features at all positions in the sentence from the semantic word vector sequence and the part-of-speech word vector sequence respectively, corresponding semantic attention vectors and part-of-speech attention vectors are obtained, and the semantic attention vectors and the part-of-speech attention vectors are constrained through theKL distance; carrying out weighted summation on the output sequence of the bidirectional LSTM layer in the merging layer by utilizing the obtained semantic attention vector and part-of-speech attention vector to obtain semantic representation and part-of-speech representation of the sentence, and obtaining final sentence semantic representation; And finally, prediction and classified output are carried out through an MLP output layer.
Owner:SOUTH CHINA UNIV OF TECH

Log sequence anomaly detection framework based on nLSTM (Non-Log Sequence Transfer Module)-self attention

PendingCN111209168AAvoid complex feature extraction stepsPreserve and control contextual informationHardware monitoringNeural architecturesSelf attentionAlgorithm
The invention relates to a log sequence anomaly detection framework based on nLSTM-self attention, and the framework comprises a training model and an anomaly detection model. The training model comprises: assuming that one log file contains k log templates E = {e1, e2L ek}, wherein the input of the training model is a sequence of the log template, the log sequence lt-h,...lt-2, lt-1 with the length of h comprises a log template li belongs to E, t-h < = i < = t-1, and the log template number | lt-h,...lt-2, lt-1 | in one sequence is equal to m < = h; enabling each log template to correspond toone template number, generating a log template dictionary, generating an input sequence from a normal log template sequence, and feeding the input sequence and target data into an anomaly detection model for training. The detection stage comprises the following steps: the data input method is the same as the training stage, anomaly detection is carried out by using the model generated in the training stage, the model output is a probability vector P = (p1, p2L pk), pi represents the probability that the target log template is ei, if the actual target data is within the prediction value, it isjudged that the log sequence is normal, otherwise it is judged that the log sequence is abnormal.
Owner:中国人民解放军陆军炮兵防空兵学院郑州校区

Machine reading comprehension answer obtaining method based on multi-round attention mechanism

The invention discloses a machine reading comprehension answer obtaining method based on a multi-round attention mechanism. The method comprises steps of performing word segmentation processing and vectorization processing on the questions and the texts corresponding to the questions respectively to obtain feature vectors, selecting a bidirectional long-short time memory network to encode contextsemantic information of the feature vectors, and performing modeling between the questions and the texts by using an attention mechanism to effectively capture information interaction between the questions and the texts. Attention of an article about a question is calculated through multiple rounds, context semantic information is fused, then BLSTM is used for coding the context semantic information, the above processes are repeated for multiple times, so that an nth text semantic vector is obtained, and a Self-Attention mechanism is used for obtaining a vector representation of the question;by calculating the similarity between the semantic vectors of the questions and the similarity of the semantic vectors, namely one representation of each word in the article in the question space, theaccuracy of predicting answers can be effectively improved, BLSTM and Attention are effectively combined, and the matching accuracy of the questions and the answers returned by text extraction can beimproved.
Owner:XI AN JIAOTONG UNIV

An image coloring method based on a self-attention generative adversarial network

The invention discloses an image coloring method based on a self-attention generative adversarial network. The method comprises the following steps: step 1, training a grayscale image coloring model;2, inputting the gray level images in the training data set into an adversarial network to execute a feature extraction stage, a feature fusion stage, a deconvolution calculation stage and a self-attention learning stage to reconstruct corresponding color images; step 3, comparing the color image reconstructed after the self-attention learning with the corresponding original color image, and calculating a penalty function shown in the specification; and step 4, calculating a penalty function according to the formula shown in the specification, 4, taking the loss function as the optimization loss of the GAN on the basis of the formula shown in the specification, wherein the formula shown in the specification is shown in the specification; and step 5, dividing the training process into a plurality of preset sub-training periods, and sequentially training the sub-training periods by adopting a step-by-step growth strategy to obtain the generator network. According to the method, the colorimage conforming to human subjective visual preferences is reconstructed from a black-white or gray level image by adopting an adversarial generation network, so that the color image is more realistic.
Owner:福建帝视信息科技有限公司

Joint extraction method for viewpoints and viewpoint holders based on self-attention

ActiveCN108628828AAvoid Situations That Don't Include OpinionsAvoid the influence of the extraction effectSemantic analysisSpecial data processing applicationsViewpointsSelf attention
The invention is based on a joint extraction method for viewpoints and viewpoint holders based on self-attention. The method comprises the steps of S1, constructing a corpus for extracting the viewpoints and the viewpoint holders; S2, identifying statements containing the viewpoints; S3, conducting joint extraction on the viewpoints and the viewpoint holders. The method has the advantages that thesituation that extracted sentences do not contain the viewpoints is avoided through a text classification model; a joint extraction model for the viewpoints and the viewpoint holders is free from natural language processing links such as part-of-speech tagging, named entity recognition and syntactic dependency analysis, avoids the influence of errors in the links on the extraction effect of the model, and has high flexibility and coverage; the method comprises the steps of constructing the corpus for extracting the viewpoints and the viewpoint holders, identifying the statements containing the viewpoints and conducting joint extraction on the viewpoints and the viewpoint holders; self-attention is used on the basis of two-way LSTM, the advantages of the self-attention and the two-way LSTMare effectively combined, the representation semantics of word sequences is more abundant, and the accuracy of the trained model is higher.
Owner:NAT COMP NETWORK & INFORMATION SECURITY MANAGEMENT CENT +1
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products