Unlock instant, AI-driven research and patent intelligence for your innovation.

Training method of depth image/video compression network

A video compression, depth image technology, applied in image communication, biological neural network model, digital video signal modification, etc., can solve the problems of network coding performance degradation, training-test inconsistency, and limiting the flexibility of the compression network.

Active Publication Date: 2021-07-06
UNIV OF SCI & TECH OF CHINA
View PDF11 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, in the actual encoding and decoding, the entire quantization layer will use the method of direct rounding to obtain discrete hidden layer variables, which causes the problem of inconsistency between training and testing, which in turn makes the encoding performance of the network drop greatly.
At the same time, since the additive mean noise usually ranges from -0.5 to 0.5, the quantization step size is limited, resulting in only integer quantization of hidden layer variables, which in turn limits the flexibility of the compression network.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Training method of depth image/video compression network
  • Training method of depth image/video compression network
  • Training method of depth image/video compression network

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0015] The technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some of the embodiments of the present invention, not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

[0016] The embodiment of the present invention provides a training method for a deep image / video compression network, which mainly includes the improvement of two quantization layers, one is to train the fine-tuning decoder by implementing accurate rounding and hard quantization, so as to ensure the consistency of training and testing . The other is to flexibly control the quantization granularity of the encoding network by pre-generating data cont...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a training method of a depth image / video compression network, which can eliminate quantization errors caused by additive mean noise on one hand, realize the consistency of training tests through two-stage fine tuning of a decoder, and obviously improve the rate distortion performance of the whole compression network on the other hand. On the other hand, the quantization granularity of the coding network is flexibly controlled by pre-generating the quantization step size adaptive to the data content from the marginal information, which is a new spatial domain code rate allocation strategy, so that the network can adaptively control the bit allocation on the spatial domain according to the image content during quantization. In addition, the two-stage adaptive quantization strategy can be effective for all coding networks adopting additive mean noise, and meanwhile, the stability of coding network training can be remarkably ensured. In conclusion, video / image compression based on the deep neural network can be more universal, flexible and efficient.

Description

technical field [0001] The invention relates to image / video compression coding, in particular to a training method for a deep image / video compression network. Background technique [0002] Image and video compression coding is an important technology in the electronic information age, which helps to reduce the transmission bandwidth and storage consumption of image and video. Image / video compression essentially realizes the effective representation of information by controlling the relationship between the bit rate (the size of the binary data required to characterize the image and video) and the distortion (the difference between the restored image and the original image). [0003] The existing image / video compression based on variational neural network contains a quantization layer, so it will lead to the problem that the direct training gradient cannot pass through the quantization layer. Specifically, taking the image compression algorithm as an example, the nonlinear t...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): H04N19/172H04N19/124H04N19/147H04N19/42H04N19/44H04N19/91G06N3/04
CPCH04N19/172H04N19/124H04N19/147H04N19/91H04N19/42H04N19/44G06N3/045
Inventor 陈志波郭宗昱
Owner UNIV OF SCI & TECH OF CHINA
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More