Action recognition method and device based on first person view angle

A first-person and action recognition technology, applied in the field of action recognition, can solve problems such as large amount of calculation, interference, and lack of original information, and achieve strong robustness and the effect of getting rid of dependence

Active Publication Date: 2021-08-27
GUANGDONG UNIV OF TECH
View PDF4 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0003] Based on the method of first-person perspective action recognition, the current methods mainly include three types: (1) use sensors such as LeapMtion to collect information about hand joints in demonstration videos, and then assist action recognition. This method requires hardware support and requires operation. Personnel need to demonstrate actions in a specific environment; (2) For the demonstration video, use dense trajectories to represent motion features, and use HOG to collect gesture features. This method is often disturbed by background and camera movement, and the calculation is relatively heavy (3) Segment the hands of the operator in the demonstration video and input them into the deep neural network for recognition. Although this method can effectively reduce background interference, most of the original information is missing
Obviously, the existing action recognition methods based on the first-person perspective have certain defects.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Action recognition method and device based on first person view angle
  • Action recognition method and device based on first person view angle
  • Action recognition method and device based on first person view angle

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0057] Example 1, please refer to figure 1 , figure 1 It is a flow chart of steps in Embodiment 1 of an action recognition method based on a first-person perspective in the present invention, including:

[0058] S101. Obtain an RGB video frame to be processed; the RGB video frame to be processed includes hand movement image information based on a first-person perspective;

[0059] S102, inputting all the RGB video frames to be processed into the pre-trained HOPE-Net deep neural network to obtain corresponding hand joint point position information;

[0060] S103, selecting a predetermined number of target RGB video frames from all the RGB video frames to be processed, and inputting them into the I3D model for identification to obtain corresponding video frame features;

[0061] S104. Input the position information of the hand joint points into the AGCN model to obtain corresponding position information features;

[0062] S105, merging the video frame features and the positio...

Embodiment 2

[0064] Example 2, please refer to figure 2 , figure 2 It is a flow chart of steps in Embodiment 2 of an action recognition method based on a first-person perspective in the present invention, specifically including:

[0065] Step S201, acquiring the video to be processed; the video to be processed contains hand movement image information based on the first-person perspective;

[0066] Step S202, converting the hand movement image information into the RGB video frame to be processed through OpenCV;

[0067] In the embodiment of the present invention, first the video to be processed is converted into a plurality of RGB video frames to be processed using OpenCV.

[0068] It should be noted that OpenCV is a cross-platform computer vision and machine learning software library that can run on Linux, Windows, Android and Mac OS operating systems. OpenCV has light-weight and high-efficiency features—consists of a series of C functions and a small number of C++ classes, and provid...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses an action recognition method and device based on a first person view angle. The method comprises the following steps: acquiring a to-be-processed RGB video frame, wherein the to-be-processed RGB video frame comprises hand action image information based on a first person view angle; inputting all the RGB video frames to be processed into a pre-trained HOPE-Net deep neural network to obtain corresponding hand joint point position information; selecting a predetermined number of target RGB video frames from all the to-be-processed RGB video frames, and inputting the selected target RGB video frames into the I3D model for identification to obtain corresponding video frame features; inputting the hand joint point position information into an AGCN model to obtain corresponding position information features; and fusing the video frame features and the position information features in a one-to-one correspondence manner to obtain the probability of identifying the action instruction. The video frame is sequentially subjected to hand skeleton joint extraction and RGB and skeleton action feature extraction, and finally feature fusion is performed to obtain the action instruction probability, so that dependence on external hardware equipment is eliminated, and the method has strong robustness for illumination and scene change.

Description

technical field [0001] The present invention relates to the technical field of motion recognition, in particular to a method and device for motion recognition based on a first-person perspective. Background technique [0002] Although robots can understand human behavior intentions well and learn human behavior autonomously by learning actions from human demonstration videos, in practical applications, the learning of human behavior by robots requires a detailed understanding process, especially It is particularly challenging for robots to learn behaviors derived from daily activities. For example, in the first-person video taken by a wearable camera, the robot can only obtain the human hand's operation from a single angle. In this case, it is full of things such as fast hand movement and occlusion when the hand is operated, resulting in a large degree of unpredictability. Therefore, the recognition of subtle differences in human actions by robots, and the process of learni...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/00G06K9/46G06K9/62G06N3/04
CPCG06N3/04G06V40/28G06V10/44G06F18/2415G06F18/253Y02T10/40
Inventor 刘文印田文浩陈俊洪
Owner GUANGDONG UNIV OF TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products