Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Short video classification method

A classification method and short video technology, applied in the field of computer vision, can solve problems such as suboptimal motion information, inability to solve long-sequence problems, long training time, etc., and achieve the effect of solving long-sequence problems, efficient and effective learning

Inactive Publication Date: 2020-01-03
HANGZHOU QUWEI SCI & TECH
View PDF2 Cites 6 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

There are the following problems in video classification training using the above methods: (1) The calculation of optical flow in advance requires additional GPU computing time and storage space, which has become the bottleneck of the two-stream algorithm; (2) the traditional optical flow calculation method is completely independent of two -stream framework, not end-to-end training, the motion information in advance is not optimal; (3) cannot solve the long-term sequence problem
Using the above method for video classification training has the following problems: the amount of parameters is huge, the training time is long, it is easy to overfit, and the performance on various public data sets generally requires a large amount of calculation and cannot be applied in real time.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Short video classification method
  • Short video classification method
  • Short video classification method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0023] The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are some of the embodiments of the present invention, but not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0024] refer to figure 1 , the invention discloses a short video classification method, comprising the following steps:

[0025] Network training, including choosing the BN-Inception building block because it has a better balance between accuracy and efficiency. During the learning process, Batch Normalization will estimate the activation mean and variance within each batch and use them to convert these activation values ​​into a standard Gaussia...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a short video classification method. The method comprises the following steps: network training; network test which comprises the following steps: performing equal-interval frame interception on to-be-classified short videos, and intercepting a certain number of frames from each video; enabling intercepted frames to be subjected image processing, taking out a picture with acertain size and inputting into the model for prediction; fusing the intercepted frame and the prediction scores of different streams before Softmax; and obtaing a final result. The model is obtainedin an manner that an input video is divided into K segments, one segment is randomly sampled from the corresponding segment, the category scores of different segments are fused by using a segment consensus function to generate a segment consensus, a video-level prediction is obtained, and the prediction of all modes are fused to generate a final prediction result.

Description

technical field [0001] The invention belongs to the technical field of computer vision, and in particular relates to a short video classification method. Background technique [0002] In the prior art, video classification training is based on Two-Stream and its derivative series. The basic principle is to train two convolutional networks to model video frame images (spatial) and dense optical flow (temporal). The two networks The structure is the same, both are two-dimensional convolution (2D ConvNets), such as figure 1 shown. The networks of the two streams respectively judge the category of the video to obtain a class score, and then perform fusion of the scores to obtain the final classification result. There are the following problems in video classification training using the above methods: (1) The calculation of optical flow in advance requires additional GPU computing time and storage space, which has become the bottleneck of the two-stream algorithm; (2) the tradi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06K9/62G06K9/00G06F16/75
CPCG06F16/75G06V20/41G06F18/25G06F18/241
Inventor 魏陈超范俊
Owner HANGZHOU QUWEI SCI & TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products