Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

561 results about "Task learning" patented technology

Scene and target identification method and device based on multi-task learning

InactiveCN106845549ARealize integrated identificationImprove single-task recognition accuracyCharacter and pattern recognitionNeural architecturesTask networkGoal recognition
The invention relates to a scene and target identification method and device based on multi-task learning. The method comprises the steps that pictures containing different scenes and targets are collected as image sample data; the image sample data is subjected to manual label marking, and target class labels and scene class labels are obtained; a multi-layer convolutional neural network model is built, and network initialization is conducted; the image sample data and the corresponding target class labels are adopted for pre-training the built model till convergence, and a target identification model is obtained; based on a multi-task learning technology, network branches are added into a specific layer of the target identification model, random initialization is conducted, and a multi-task network is obtained; the image sample data and the corresponding scene class labels and target class labels are adopted for e-training the multi-task network till convergence, and a multi-task learning model is obtained; new image data is input to the multi-task learning model, and classification results of scene and target identification of images are obtained. Accordingly, the single task identification precision is improved.
Owner:珠海习悦信息技术有限公司

Multi-task named entity recognition and confrontation training method for medical field

The invention discloses a multi-task named entity recognition and confrontation training method for medical field. The method includes the following steps of (1) collecting and processing data sets, so that each row is composed of a word and a label; (2) using a convolutional neural network to encode the information at the word character level, obtaining character vectors, and then stitching withword vectors to form input feature vectors; (3) constructing a sharing layer, and using a bidirection long-short-term memory nerve network to conduct modeling on input feature vectors of each word ina sentence to learn the common features of each task; (4) constructing a task layer, and conducting model on the input feature vectors and the output information in (3) through a bidirection long-short-term network to learn private features of each task; (5) using conditional random fields to decode labels of the outputs of (3) and (4); (6) using the information of the sharing layer to train a confrontation network to reduce the private features mixed into the sharing layer. According to the method, multi-task learning is performed on the data sets of multiple disease domains, confrontation training is introduced to make the features of the sharing layer and task layer more independent, and the task of training multiple named entity recognition simultaneously in a specific domain is accomplished quickly and efficiently.
Owner:ZHEJIANG UNIV

Dialog strategy online realization method based on multi-task learning

The invention discloses a dialog strategy online realization method based on multi-task learning. According to the method, corpus information of a man-machine dialog is acquired in real time, current user state features and user action features are extracted, and construction is performed to obtain training input; then a single accumulated reward value in a dialog strategy learning process is split into a dialog round number reward value and a dialog success reward value to serve as training annotations, and two different value models are optimized at the same time through the multi-task learning technology in an online training process; and finally the two reward values are merged, and a dialog strategy is updated. Through the method, a learning reinforcement framework is adopted, dialog strategy optimization is performed through online learning, it is not needed to manually design rules and strategies according to domains, and the method can adapt to domain information structures with different degrees of complexity and data of different scales; and an original optimal single accumulated reward value task is split, simultaneous optimization is performed by use of multi-task learning, therefore, a better network structure is learned, and the variance in the training process is lowered.
Owner:AISPEECH CO LTD

Multi-modal fusion emotion recognition system and method based on multi-task learning and attention mechanism and experimental evaluation method

The invention relates to a multi-modal fusion emotion recognition system and method based on multi-task learning and an attention mechanism and an experimental evaluation method, and aims to solve the problems that in the prior art, a multi-modal emotion recognition process without introducing a multi-modal fusion mechanism is low in efficiency and accuracy. The invention belongs to the field of human-computer interaction, and provides a multi-modal fusion emotion recognition method based on combination of multi-task learning and an attention mechanism, and compared with single-modal emotion recognition work, the multi-modal fusion emotion recognition method based on combination of multi-task learning and the attention mechanism is wider in application. Multi-task learning is utilized to introduce an auxiliary task, so that the emotion representation of each mode can be more efficiently learned, and an interactive attention mechanism can enable the emotion representations among the modes to mutually learn and complement each other, so that the recognition accuracy of the multi-mode emotion is improved; experiments are carried out on the multi-modal data sets CMU-MOSI and CMU-MOSEI, the accuracy and the F1 value are both improved, and meanwhile the accuracy and efficiency of emotion information recognition are improved.
Owner:HARBIN UNIV OF SCI & TECH

Multi-task learning method for real-time target detection and semantic segmentation based on lightweight network

The invention relates to a multi-task learning method for real-time target detection and semantic segmentation based on a lightweight network. The system comprises a feature extraction module, a semantic segmentation module, a target detection module and a multi-scale receptive field module. The feature extraction module selects a lightweight convolutional neural network MobileNet; features are extracted through a MobileNet network and sent to a semantic segmentation module to complete segmentation of a drivable area and a selectable driving area of a road, and meanwhile the features are sentto a target detection module to complete object detection appearing in a road scene. A multi-scale receptive field module is used for increasing the receptive domain of a feature map, convolution of different scales is used for solving the multi-scale problem, finally, weighted summation is carried out on a loss function of a semantic segmentation module and a loss function of a target detection module, and a total module is optimized. Compared with the prior art, the method provided by the invention has the advantage that two common unmanned driving perception tasks of road object detection and road driving area segmentation are completed more quickly and accurately.
Owner:SUN YAT SEN UNIV

Video recommendation method based on multi-modal video content and multi-task learning

The invention discloses a video recommendation method based on multi-modal video content and multi-task learning. The method comprises the following steps: extracting visual, audio and text features of a short video through a pre-trained model; fusing the multi-modal features of the video by adopting an attention mechanism method; learning feature representation of the social relationship of the user by adopting a deep walk method; proposing a deep neural network model based on an attention mechanism to learn multi-domain feature representation; embedding the features generated based on the above steps into a sharing layer as a multi-task model, and generating prediction results through a multi-layer perceptron. According to the method, the attention mechanism is combined with the user features to fuse the video multi-modal features, so that the whole recommendation is richer and more personalized; meanwhile, because of multi-domain features and with consideration of the importance ofinteraction features in recommendation learning, a deep neural network model based on an attention mechanism is provided, so that learning of high-order features is enriched, and more accurate personalized video recommendation is provided for users.
Owner:SOUTH CHINA UNIV OF TECH +1

Robot sequence task learning method based on visual simulation

The invention provides a robot sequence task learning method based on visual simulation. The robot sequence task learning method based on the visual simulation is used for guiding a robot to simulateand execute human actions from a video containing the human actions. The robot sequence task learning method comprises the following steps of (1) identifying object types and masks by using a region-based mask convolutional neural network according to an input image; (2) calculating actual plane physical coordinates (x, y) of objects according to the masks; (3) identifying atomic actions in a target video; (4) converting an atomic action sequence and the identified object types into a one-dimensional vector; (5) inputting the one-dimensional vector into a task planner, and outputting a task description vector capable of guiding the robot; and (6) controlling the robot to simulate a sequence task in the target video by combining the task description vector and the object coordinates. According to the robot sequence task learning method based on the visual simulation, the video and the image serve as input, the objects are recognized, a task sequence is deduced, the robot is guided to simulate the target video, and meanwhile, the generalization performance is high, so that the simulation of the task can still be completed under different environments or object types.
Owner:BEIHANG UNIV

Alzheimer's disease classification and prediction system based on multi-task learning

The invention discloses an Alzheimer's disease classification and prediction system based on multi-task learning, and relates to an Alzheimer's disease classification and prediction system. The objective of the invention is to solve the problem that an existing Alzheimer's disease classification system cannot judge whether a mild cognitive impairment individual will be transformed into Alzheimer'sdisease or not. The system comprises a an image processing main module, a clinical index processing main module, a neural network main module, a training main module and a detection main module; wherein the image processing main module is used for acquiring a head image, preprocessing the acquired head image to obtain a preprocessed image, and inputting the preprocessed image into the training main module and the detection main module; wherein the clinical index processing main module is used for selecting clinical indexes, acquiring feature vectors of the clinical indexes and inputting the feature vectors of the clinical indexes into the training main module and the detection main module; and the neural network main module is used for building an Alzheimer's disease classification and prediction model. The system is applied to the technical field of intelligent medical detection.
Owner:HARBIN INST OF TECH

Natural language relation extraction method based on multi-task learning mechanism

The invention discloses a natural language relation extraction method based on a multi-task learning mechanism. The method comprises the following steps: introducing information implied by different tasks by utilizing a plurality of auxiliary tasks to improve a relation extraction effect; introducing knowledge distillation to enhance the effect of assisting a task to guide and train a multi-task model, introducing a teacher annealing algorithm for relation and extraction based on multi-task learning to enable the effect of the multi-task model to serve as a single-task model of a guide task inan ultra-far mode, and finally improving the accuracy of relation extraction is improved. The method comprises the following steps: firstly, training on different auxiliary tasks to obtain a multi-task model for guiding training; then, a model learned by an auxiliary task and a real label are used as supervision information to guide the learning of a multi-task model at the same time; finally, evaluation is carried out on a SemEval2010 task-8 data set, and the performance of the model is superior to that of a model independently using improved BERT for relation extraction and is also superiorto that of a mainstream model based on deep learning relation extraction.
Owner:EAST CHINA NORMAL UNIV

Part surface roughness and tool wear prediction method based on multi-task learning

The invention belongs to the technical field of machining and provides a part surface roughness and tool wear prediction method based on multi-task learning. The method is characterized in that firstly, vibration signals in the machining process are collected, next, the surface roughness of a part and the abrasion condition of a cutter are measured, and the measured results are made to correspondto vibration signals respectively; secondly, sample expansion is carried out, and features are extracted and normalized; then, a multi-task prediction model based on a deep belief network is constructed, the surface roughness of the part and the cutter abrasion condition serve as model output, features are extracted as input, and a multi-task DBN network prediction model is established; and finally, test verification is performed, a vibration signal is inputted into the multi-task prediction model, and the surface roughness and the cutter wear condition are predicted. The method is mainly advantaged in that online prediction of the part surface roughness and the tool wear condition is achieved through one-time modeling, the hidden information contained in monitoring data is fully utilized,and the workload and model building cost are reduced.
Owner:DALIAN UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products