Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A video deshaking method based on deep learning

A technology of video de-shaking and deep learning, applied in the field of computer vision and video de-shaking, which can solve the problems of limitations and less research on video de-shaking, and achieve the effects of increasing stability, improving precision, and ensuring rationality and accuracy

Active Publication Date: 2021-05-07
UNIV OF SCI & TECH OF CHINA
View PDF12 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, its use of a single homography matrix for the transformation from a shaking perspective to a stable perspective limits its ability to solve 3D video
[0004] Over the past few years, CNN has achieved great success in solving traditional computer vision problems, but there are few studies on video deshaking using deep networks.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A video deshaking method based on deep learning
  • A video deshaking method based on deep learning
  • A video deshaking method based on deep learning

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0059] The present invention will be described in detail below in conjunction with the accompanying drawings and embodiments.

[0060] Such as figure 1 Shown, the inventive method comprises the steps:

[0061] Step1: In the training phase, the continuous shaking frame sequence is used as the input of the network, and the stable frame is used as the output of the network for supervised training to generate a weighted deep network;

[0062] Step2: In the test phase, the continuous jitter frame sequence is used as the input of the network to generate a pixel-level map;

[0063] Step3: During the test phase, the shaking frame is mapped point by point through the map generated by Step2 to generate a stable frame.

[0064] In the training step of the deep network, the continuous shaking frame sequence is used as the input of the network, and the stable frame is used as the output of the network for supervised training to generate a weighted deep network;

[0065] In the test pha...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a video deshaking method based on deep learning, which learns a point-by-point map by designing a fully convolutional network, and maps the pixels of the jittering view point to the stable view point point by point according to the map. The method proposed by the present invention is no longer based on the traditional feature matching strategy and homography matrix estimation, but instead performs pixel-level mapping relationship estimation. Such a transformation can solve the problem that the same homography matrix cannot be used locally due to discontinuous depth changes. problem, thus achieving better results in real videos. At the same time, the deep network trained by this method has better robustness, especially when dealing with low-quality videos (such as blurred videos, night videos, and watermarked videos), it has better results than traditional methods. With the help of GPU parallel processing characteristics, the present invention achieves faster processing speed than traditional methods, and can realize online real-time video deshaking.

Description

technical field [0001] The invention relates to a video debounce method based on deep learning, and belongs to the technical fields of computer vision and video debounce. Background technique [0002] In recent years, more and more cameras have been used in various scenes in real life, including a large number of portable and movable camera equipment. The quality of the video recorded by hand-held devices has been greatly reduced due to artificial shaking. The vibration caused by the human sensory discomfort. [0003] Ordinary handheld device recording video deshaking methods are roughly divided into three categories, 2D, 2.5D and 3D methods. 2D methods typically use a sequence of inter-frame matrices to model camera motion followed by smoothing [1]. The 3D method has a better effect on the processing of parallax, by using the motion recovery structure (Structure from Motion, SfM) to estimate the camera path [2], and then using the content-preserving warping (content-prese...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06T5/50H04N5/21
CPCG06T5/50G06T2207/10016G06T2207/20081G06T2207/20084H04N5/21
Inventor 凌强赵敏达李峰
Owner UNIV OF SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products