Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Scene understanding method based on multi-task learning

A multi-task learning and scene understanding technology, applied in neural learning methods, complex mathematical operations, instruments, etc., can solve problems such as unsatisfactory efficiency and prediction accuracy

Inactive Publication Date: 2017-12-08
SHENZHEN WEITESHI TECH
View PDF0 Cites 27 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] For the problem of unsatisfactory efficiency and prediction accuracy, the purpose of the present invention is to provide a scene understanding method based on multi-task learning, first perform the weighted linear sum of the losses of each individual task, learn the optimal task weight, and then derive A multi-task loss function that defines a probabilistic model that defines likelihood as a Gaussian function of the mean given by the model output, and finally builds models that learn regression and classification outputs at the pixel level, including semantic segmentation, instance segmentation, and deep regression

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Scene understanding method based on multi-task learning
  • Scene understanding method based on multi-task learning
  • Scene understanding method based on multi-task learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0054] It should be noted that, in the case of no conflict, the embodiments in the present application and the features in the embodiments can be combined with each other. The present invention will be further described in detail below in conjunction with the drawings and specific embodiments.

[0055] figure 1 It is a system framework diagram of a scene understanding method based on multi-task learning in the present invention. It mainly includes multi-task learning with homoscedastic uncertainty, multi-task likelihood function, and scene understanding model.

[0056] Multi-task learning with homoscedastic uncertainty, multi-task learning involves the problem of optimizing a model for multiple objectives; a way to combine multi-objective losses is to perform a weighted linear sum of the losses for each individual task:

[0057]

[0058] But the model performance is not as good as the weight w i The choice of w is very sensitive; at some optimal weights, the joint network...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a scene understanding method based on multi-task learning. The method includes: multi-task learning having unstable homoscedasticity, multi-task likelihood function, and scene understanding model. The method includes the following steps: first executing weighted linear sum on the loss of each individual task, learning the optimal task weight, inducting a multi-task loss function, defining a probability model, defining the probability as the gaussian function of the average values that are output by the model, eventually modeling pixel-level learning regression and classified output, comprising semantic division, incidence division and depth regression. According to the invention, the scene understanding model can learn multi-task weight, is more advantageous than models that independently train each task in that the method herein reduces computing amount, increases learning efficiency and prediction precision and can be real-timely operated.

Description

technical field [0001] The invention relates to the field of scene understanding, in particular to a scene understanding method based on multi-task learning. Background technique [0002] Scene understanding can effectively help and improve the computer's ability to analyze and recognize complex and changeable indoor and outdoor scenes, and is one of the research hotspots in the field of computer vision. Scene understanding can usually be divided into two categories: local scene understanding and global scene understanding. The former focuses on the analysis and description of the distribution and category of the local area of ​​the scene, such as the identification and positioning of various types of local targets in the scene; the latter focuses on understanding the global attributes of the scene, such as scene classification. Both can deepen the computer's cognition and grasp of unknown scenes from different cognitive levels, and have broad application prospects in the f...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06F17/14G06F17/18G06N3/08
CPCG06F17/14G06F17/18G06N3/08G06F18/23G06F18/2321
Inventor 夏春秋
Owner SHENZHEN WEITESHI TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products