Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Multi-person multi-mode perception data automatic marking and mutual learning method

A technology for automatic labeling and data perception, applied in the field of cross-domain perception, can solve the problems of limited perception ability of single modal data and difficulty of manual labeling data, and achieve the effect of improving recognition accuracy, improving ability, and increasing the number of categories

Pending Publication Date: 2021-01-12
德清阿尔法创新研究院
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The purpose of the present invention is to provide a method for automatic marking and mutual learning of multi-modal sensing data, which can automatically segment, align, and mark the sensing data stream, and then learn from each other to achieve higher-precision sensing and solve the difficulty of manually marking data. , the problem of limited awareness of single modality data

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Multi-person multi-mode perception data automatic marking and mutual learning method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0017]The following describes the technical solutions in the embodiments of the present invention clearly and completely with reference to the specific content of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, rather than all the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of the present invention. The content that is not described in detail in the embodiments of the present invention belongs to the prior art known to those skilled in the art.

[0018]Such asfigure 1As shown, the embodiment of the present invention provides a method for automatic marking and mutual learning of multi-person and multi-modal data, which can automatically segment, mark, and then learn from each other in multi-modal data streams, including:

[0019]Step 1. Data preprocessing: including cloc...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an automatic marking and mutual learning method for multi-person multi-mode perception data. The method comprises the following steps: step 1, carrying out clock alignment, denoising and other preprocessing on the data; 2, fusing the results of the modal models to segment the data stream; 3, when the scene contains multi-person information, data entity alignment is carriedout by utilizing the static state information and the motion state information of the persons at the same time; 4, automatically marking the data segments by utilizing a designed multi-modal data prediction fusion mechanism; and 5, improving the model understanding capability of each mode by utilizing the automatically marked data. In order to ensure the capability of the updated model, the quality of the automatically marked data is evaluated, i.e., only the data with high quality is selected to update the model. According to the method, automatic segmentation and labeling of the multi-modaldata can be realized by utilizing intrinsic correlation of the multi-modal data, labeled data segments are obtained, the model capability is further improved, and promotion of data understanding related research is facilitated.

Description

Technical field[0001]The invention relates to the field of cross-domain perception, in particular to a method for automatically marking and mutual learning of multi-person multi-modal perception data.Background technique[0002]The wide application of sensing devices (such as smart phones, wearable devices, cameras, and wireless access points, etc.) and the rapid growth of sensing data make intelligent perception a hot research topic. In recent years, there has been quite a lot of research on using machine learning technology to understand various modal data (such as video, audio, motion sensor data, and wireless data). But most of these studies are for single modal data. Monomodal data can only obtain information about the current part of the scene, which may result in low perception accuracy. At the same time, most of the existing models require labeled training data to train. The labeled data itself is a time-consuming and laborious task, and the trained model can only recognize th...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06K9/62G06N20/00
CPCG06N20/00G06V10/751G06F18/25G06F18/241
Inventor 张兰李向阳郑达人宋年卉
Owner 德清阿尔法创新研究院
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products