Coordination-model-based method for obtaining video abstract

A technology for acquiring videos and models, applied in special data processing applications, instruments, electrical digital data processing, etc., can solve the problem of not ensuring that adjacent frames have similar sparse expression coefficients, the influence of sparse expression accuracy, and the lack of ignoring the dispersion of video frames. And other issues

Active Publication Date: 2016-10-19
XI'AN INST OF OPTICS & FINE MECHANICS - CHINESE ACAD OF SCI
View PDF3 Cites 9 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] Although these methods achieved good results, they did not consider the relationship between adjacent frames when performing dictionary learning.
This will not only not guarantee that similar adjacent frames have similar sparse expression coefficients, but also cannot guarantee that the sparse expression coefficients of dissimilar adjacent frames must be different
Therefore, the accuracy of sparse representation will suffer
In addition, although these methods focus on the expressiveness of video frames, they ignore the dispersion of video frames, which leads to the selection of key frames that tend to contain redundant information and cannot effectively cover all important content.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Coordination-model-based method for obtaining video abstract
  • Coordination-model-based method for obtaining video abstract
  • Coordination-model-based method for obtaining video abstract

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0043] refer to figure 1 , the steps that the present invention realizes are as follows:

[0044] Step 1, generate feature representations for video frames.

[0045] (1a) extracting the bottom layer image feature operation frame by frame from the input video containing n frames, and obtaining the bottom layer feature set of the input video;

[0046] (1b) Use the BoW (Bag-of-Word) model on this feature set to obtain the feature description vector x of each frame of the video, thereby obtaining the feature expression matrix X of the video = [x 1 ,x 2 ,...,x n ];

[0047] Step 2, comprehensively evaluate the importance of video frames through the collaborative model.

[0048] (2a) Carry out dictionary learning on the obtained video feature expression matrix, and measure the expressivity of the video frame by calculating the reconstruction error of the sparse expression coefficients. A frame with a smaller reconstruction error indicates better expressivity. In order to achieve...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention, which belongs to the video processing technology, particularly relates to a coordination-model-based method for obtaining a video abstract. The method comprises: (1), generating a feature expression matrix of a video frame; (2), evaluating significance of the video frame comprehensively by a coordination mode; and (3), extracting a key frame to form a video abstract. According to the invention, with construction of the coordination model, expressivity and dispersity of the video frame are considered, so that accuracy of the video abstract is improved. Constraint information of the geometric structure of the video frame is added into the dictionary learning process, so that sparse representation becomes accurate and expressivity of the video frame can be measured accurately. With a dispersity measuring standard based on similarity measurement, redundant frame extraction can be avoided, thereby improving conciseness of the video abstract.

Description

technical field [0001] The invention belongs to video processing technology, and in particular relates to a method for acquiring video summaries based on a collaborative model, which can be used in the fields of public security monitoring management, military reconnaissance, and large-scale video data management. Background technique [0002] In recent years, with the increasing popularity of low-cost and large-capacity digital video recording equipment, video data is growing explosively at an alarming rate. Take YouTube, the largest video website in the world, as an example. As of January 2015, the total duration of videos uploaded every minute is 100 hours. This makes understanding and obtaining the main content of the video in a manual viewing method need to consume a huge amount of manpower and time. Therefore, people urgently need a technology that can efficiently analyze massive video data. Several video frames (called key frames) containing the main content of the o...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): H04N21/8549G06F17/30
Inventor 李学龙卢孝强陈潇
Owner XI'AN INST OF OPTICS & FINE MECHANICS - CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products