A Video Classification Method Based on 3D Convolutional Neural Network

A neural network and three-dimensional convolution technology, which is applied in the field of video processing, can solve the problems of reducing the learning complexity of three-dimensional convolutional neural network and insufficient video data resources, and achieve the goal of improving classification performance, reducing network complexity and saving recognition time. Effect

Active Publication Date: 2018-05-11
山东管理学院
View PDF4 Cites 1 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The purpose of the present invention is to provide a video classification method based on a small-scale video data set and a relatively low-configuration hardware condition. The method uses data set amplification and a distributed parallel computing method of a three-dimensional convolutional neural network to classify videos in multiple ways. The problem is transformed into a binary classification problem, which not only solves the problem of insufficient video data resources, but also greatly reduces the complexity of 3D convolutional neural network learning

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Video Classification Method Based on 3D Convolutional Neural Network
  • A Video Classification Method Based on 3D Convolutional Neural Network
  • A Video Classification Method Based on 3D Convolutional Neural Network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] Below in conjunction with accompanying drawing, invention is further described:

[0041] According to the present invention, a video classification method is provided. Firstly, the video in the video library is read, and the video frame is grayscaled; secondly, the grayscaled video is sampled into a video with a fixed number of frames by sampling at equal intervals. segment; for each type of video, use the video segment as a unit to formulate different training and test data sets, and set labels for each video segment. The tags are divided into two types: belonging to this category and not belonging to this category; Initialize a 3D CNN network for class video, and train the network with the training samples corresponding to the class, so that the 3D CNN can perform two-category classification on multiple video segments inside and outside the class; connect multiple trained 3D CNNs in parallel, and then The classification mechanism is set at the end, and the category of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a video classification method based on a three-dimensional convolutional neural network, belonging to the technical field of video processing. This method samples the video into multiple video segments at equal intervals, expands the video database, directly inputs the 3D video segment into the 3D CNN, and extracts the temporal and spatial features of the video at the same time, which improves the traditional video classification method due to manual selection of video. Limitations of feature and video modeling methods. The parallel distributed 3D CNN multi-classification model reduces the complexity of 3D CNN learning, and at the same time makes it easier for the classification system to implement distributed parallel computing. The 3D CNN-based multi-classification system can achieve a high recognition rate with only a small number of video segments, and can classify videos that do not belong to any category into new categories, avoiding classification errors for new categories.

Description

technical field [0001] The invention relates to a video classification method and belongs to the technical field of video processing. Background technique [0002] With the development of multimedia technology and Internet technology, people can now easily obtain a lot of video data from various channels, but because these massive video data are too large, how to classify these video data, so that people can more conveniently Obtaining the data you are interested in has become one of the very important and challenging research hotspots in the field of computer vision. [0003] The classification technology of video mainly includes three methods of video-based visual information, text information and audio information. As the most important information in video, visual information also contains the most video information and can best represent the video category, so it is also the most worthy. Research. The traditional video classification technology based on visual informa...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06K9/62G06N3/04
CPCG06N3/045G06F18/2413
Inventor 李静
Owner 山东管理学院
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products