Three-dimensional convolutional neural network based video classifying method

A neural network and three-dimensional convolution technology, applied in the field of video processing, can solve the problems of insufficient video data resources, reduce the learning complexity of three-dimensional convolutional neural network, etc., and achieve the effects of reducing network complexity, improving classification performance, and increasing speed

Active Publication Date: 2015-10-07
山东管理学院
View PDF4 Cites 54 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The purpose of the present invention is to provide a video classification method based on a small-scale video data set and a relatively low-configuration hardware condition. The method uses data set amplification and a distributed parallel computing method of a three-dimensional convolutional neural network to classify videos in multiple ways. The problem is transformed into a binary classification problem, which not only solves the problem of insufficient video data resources, but also greatly reduces the complexity of 3D convolutional neural network learning

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Three-dimensional convolutional neural network based video classifying method
  • Three-dimensional convolutional neural network based video classifying method
  • Three-dimensional convolutional neural network based video classifying method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] Below in conjunction with accompanying drawing, invention is further described:

[0041] According to the present invention, a video classification method is provided. Firstly, the video in the video library is read, and the video frame is grayscaled; secondly, the grayscaled video is sampled into a video with a fixed number of frames by sampling at equal intervals. segment; for each type of video, use the video segment as a unit to formulate different training and test data sets, and set labels for each video segment. The tags are divided into two types: belonging to this category and not belonging to this category; Initialize a 3D CNN network for class video, and train the network with the training samples corresponding to the class, so that the 3D CNN can perform two-category classification on multiple video segments inside and outside the class; connect multiple trained 3D CNNs in parallel, and then The classification mechanism is set at the end, and the category of...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a three-dimensional convolutional neural network (3D CNN) based video classifying method and belongs to the technical field of video processing. According to the method, a video is sampled at equal intervals to obtain a plurality of video segments, a video database is amplified, three-dimensional video segments are directly input into a 3D CNN, and time domain and space domain characteristics of the video are extracted, so that the limitation of a conventional video classifying method in manually selecting video characteristics and video modeling modes is improved. A parallel distributed 3D CNN multi-classification model lowers the complexity in learning the 3D CNN and enables a classification system to realize distributed parallel computation more conveniently. Relatively high identification rate can be achieved with only fewer video segments based on a 3D CNN multi-classification system, and videos not belonging to any type can be classified into new type, so that the classification error of the new type is avoided.

Description

technical field [0001] The invention relates to a video classification method and belongs to the technical field of video processing. Background technique [0002] With the development of multimedia technology and Internet technology, people can now easily obtain a lot of video data from various channels, but because these massive video data are too large, how to classify these video data, so that people can more conveniently Obtaining the data you are interested in has become one of the very important and challenging research hotspots in the field of computer vision. [0003] The classification technology of video mainly includes three methods of video-based visual information, text information and audio information. As the most important information in video, visual information also contains the most video information and can best represent the video category, so it is also the most worthy. Research. The traditional video classification technology based on visual informa...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06N3/04
CPCG06N3/045G06F18/2413
Inventor 孙建德赵冬李静
Owner 山东管理学院
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products