Video classification method and model training method and device thereof, and electronic equipment

A technology for video classification and training methods, applied in the field of image processing, can solve the problems of difficulty in mining high-level semantic features, large amount of parameters in 3D convolutional neural network, and shallow 3D convolutional neural network layers.

Active Publication Date: 2019-07-30
BEIJING KINGSOFT CLOUD NETWORK TECH CO LTD +1
View PDF5 Cites 50 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] In related technologies, videos can be classified through a three-dimensional convolutional neural network, and the spatiotemporal features of the video can be extracted through three-dimensional convolution. However, the network parameters of the three-dimensional convolutional neural network are relativel

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Video classification method and model training method and device thereof, and electronic equipment
  • Video classification method and model training method and device thereof, and electronic equipment
  • Video classification method and model training method and device thereof, and electronic equipment

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0046] In order to make the purpose, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions of the present invention will be clearly and completely described below in conjunction with the accompanying drawings. Obviously, the described embodiments are part of the embodiments of the present invention, not all of them. the embodiment. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0047] Considering the problems of high computing cost, high time overhead and low video classification accuracy in video classification by three-dimensional convolutional neural network, embodiments of the present invention provide a video classification method and its model training method, device and electronic equipment ; This technology can be widely used in the classification o...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a video classification method, a model training method and device thereof, and electronic equipment. The training method comprises the following steps: extracting initial features of a plurality of video frames through a convolutional neural network; extracting final features of the plurality of video frames from the initial features through a recurrent neural network; inputting the final feature into an output network, and outputting a prediction result of the multi-frame video frame; determining a loss value of the prediction result through a preset loss prediction function; and training the initial model according to the loss value until parameters in the initial model converge to obtain a video classification model. According to the method, the convolutional neural network and the recurrent neural network are combined, so that the operand can be greatly reduced, and the model training and recognition efficiency is improved; and meanwhile, the association information between the video frames can be considered in the feature extraction process, so that the extracted features can accurately represent the video types, and the accuracy of video classificationis improved.

Description

technical field [0001] The present invention relates to the technical field of image processing, in particular to a video classification method and its model training method, device and electronic equipment. Background technique [0002] In related technologies, videos can be classified through a three-dimensional convolutional neural network, and the spatiotemporal features of the video can be extracted through three-dimensional convolution. However, the network parameters of the three-dimensional convolutional neural network are relatively large, resulting in high computational costs in the network training process and recognition process. , the time overhead is high; in addition, the layers of the three-dimensional convolutional neural network are shallow, and it is difficult to mine high-level semantic features, resulting in low video classification accuracy. Contents of the invention [0003] In view of this, the purpose of the present invention is to provide a video ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06K9/00G06K9/62G06N3/04
CPCG06V20/46G06N3/045G06F18/214G06F18/2415
Inventor 苏驰李凯陈宜航刘弘也
Owner BEIJING KINGSOFT CLOUD NETWORK TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products