Unlock instant, AI-driven research and patent intelligence for your innovation.

Behavior recognition and location method based on multi-task joint learning

A technology of behavior positioning and positioning methods, which is applied in biometric recognition, character and pattern recognition, instruments, etc., to achieve the effects of saving costs, expanding diversity, and solving the problems of labeling data sets

Active Publication Date: 2020-11-10
四川瞳知科技有限公司
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] The purpose of the present invention is to solve the above-mentioned problems in the prior art, and propose a behavior recognition and positioning method based on multi-task joint learning, which uses the combination of convolutional neural network in deep learning and multi-task joint learning to replace a single task Convolutional neural network algorithm to meet the needs of human behavior recognition and behavior positioning in video clips

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Behavior recognition and location method based on multi-task joint learning
  • Behavior recognition and location method based on multi-task joint learning
  • Behavior recognition and location method based on multi-task joint learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0070] Exemplary embodiments of the present invention will now be described in detail with reference to the accompanying drawings. It should be understood that the implementations shown and described in the drawings are only exemplary, intended to explain the principle and spirit of the present invention, rather than limit the scope of the present invention.

[0071] Embodiments of the present invention provide a behavior recognition and positioning method based on multi-task joint learning, such as figure 1 As shown, including the following steps S1-S5:

[0072] S1. Construct a multi-channel combined behavior recognition convolutional neural network.

[0073] Such as figure 2 As shown, in the embodiment of the present invention, the behavior recognition convolutional neural network includes an optical flow channel and an image channel, and the optical flow channel and the image channel respectively include independent first-layer networks, second-layer networks, third-laye...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a behavior recognition and positioning method based on multi-task joint learning, which combines the convolutional neural network in deep learning with multi-task joint learning to replace the single-task convolutional neural network algorithm, so as to realize human behavior in video Targets for identification and behavioral targeting. The invention improves the object detection depth network in the faster rcnn, and combines the behavior recognition depth network on this basis, so that the combined network can achieve the ability of multi-task joint learning, so that the two tasks can promote each other and enhance the robustness of the recognition algorithm and accuracy; at the same time, the present invention combines the video data set and the picture data set, which enhances the information diversity of the training set; in addition, if it will consume a huge amount of energy to mark the position of the human body in the video data set, the present invention can pass the algorithm The autonomous learning omits the labeling work of the data set, which can greatly reduce the labeling workload.

Description

technical field [0001] The invention belongs to the technical fields of computer vision, machine learning and deep learning, and specifically relates to the design of a behavior recognition and positioning method based on multi-task joint learning. Background technique [0002] In the field of security, there is a great demand for human behavior detection and positioning, such as the detection of violent behavior. If the violent behavior that endangers society and others can be detected in real time and measures taken, it will be of great significance to social stability. If it is possible to locate the specific thugs who committed violent acts in the video, combined with the application of face recognition, it will be of great value to quickly solve the case. However, the current video surveillance system is mainly based on manpower and supplemented by computers. It mainly identifies the content of the surveillance video manually. The workload is huge, and as the surveillan...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/00G06K9/62G06N3/04
CPCG06V40/20G06V40/10G06N3/045G06F18/241
Inventor 郝宗波
Owner 四川瞳知科技有限公司