GPU (Graphics Processing Unit)-based HEVC (High Efficiency Video Coding) parallel decoding method

A decoding method and parallel algorithm technology, applied in the field of HEVC parallel decoding based on GPU, to achieve the effect of improving decoding speed and decoding efficiency

Inactive Publication Date: 2014-10-29
SUN YAT SEN UNIV
View PDF9 Cites 23 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, there is currently no GPU-based

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • GPU (Graphics Processing Unit)-based HEVC (High Efficiency Video Coding) parallel decoding method
  • GPU (Graphics Processing Unit)-based HEVC (High Efficiency Video Coding) parallel decoding method
  • GPU (Graphics Processing Unit)-based HEVC (High Efficiency Video Coding) parallel decoding method

Examples

Experimental program
Comparison scheme
Effect test

no. 1 example

[0064] HEVC decoding framework such as Figure 5 shown. The HEVC decoding process is the reverse of the encoding process. The decoder reads the code stream file and obtains the bit stream from the NAL (Network Abstraction Layer). The decoding is performed in a frame-by-frame order, and a frame of image is divided into several maximum encoding Unit LCU, in the order of raster scanning, entropy decoding is performed with LCU as the basic unit, and then reordered to obtain the residual coefficient of the corresponding coding unit; then the residual coefficient is dequantized and inversely transformed to obtain the image residual data. At the same time, the decoder generates a prediction block based on the header information decoded from the code stream: if it is an inter-frame prediction mode, it will generate a corresponding prediction block based on the motion vector and the reference frame; if it is an intra-frame prediction mode, it will generate a prediction block from the ...

Embodiment 2

[0068] This embodiment describes the HEVC inverse transformation parallel algorithm process of the present invention.

[0069] The inverse transform module is a process of converting the transform coefficient matrix of the current block into a residual sample matrix, and prepares for subsequent reconstruction. The inverse transformation is performed after the inverse quantization process, and the TU transformation unit is also used as the basic unit for processing, and the source data used is the result of inverse quantization. When the HEVC decoder of the present invention performs two-dimensional IDCT inverse transform, it first performs one-dimensional IDCT inverse discrete cosine transform in the horizontal direction, then performs one-dimensional IDCT inverse discrete cosine transform in the vertical direction, and finally multiplies the matrix to obtain The transformation coefficient matrix is ​​converted into a residual data matrix of the same size, thereby completing t...

Embodiment 3

[0079] This embodiment describes the HEVC motion compensation parallel algorithm process of the present invention.

[0080]The implementation principle of inter-frame motion compensation, simply put, is that the motion vector obtained by parsing the code stream is obtained according to the position pointed to on the reference frame to obtain the predicted value, and the position pointed to by the whole pixel point of the reference frame is directly read. If it is a sub-pixel position, it is necessary to obtain the sub-pixel predicted value through pixel interpolation, and then add the predicted value to the image residual value obtained through inverse quantization and inverse transformation to obtain the image reconstruction value. In the motion compensation module, the calculation of pixel interpolation and filtering takes up about 70% of the computational load. Therefore, the realization of the motion compensation of the present invention on the GPU is mainly to perform pix...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a GPU-based HEVC parallel decoding method. The method includes that a GPU performs entropy decoding, re-ordering and inverse quantization on a read code stream file to obtain a transformation coefficient matrix, and the GPU parses the obtained code stream file to a obtain motion vector and a reference frame; the GPU processes the transformation coefficient matrix through an HEVC inverse transformation parallel algorithm to obtain residual data of an image, and the GPU uses an HEVC motion compensation parallel algorithm to obtain a predicted pixel value of the image according to the reference frame position which the motion vector points to; the GPU sequentially performs summing, deblocking filter and sample self-adaption compensation on the residual data and the predicted pixel value of the image to obtain a reconstructed image, and a pixel value of the reconstructed image is copied to a memory of the CPU. The GPU-based HEVC parallel decoding method effectively improves the decoding speed and efficiency and can be widely used in the video coding and decoding field.

Description

technical field [0001] The invention relates to the field of video encoding and decoding, in particular to a GPU-based HEVC parallel decoding method. Background technique [0002] With the rapid development of Internet and mobile communication technology, digital video is moving towards the direction of high definition, high frame rate, and high compression rate. The format of video has developed from 720P to 1080P. In some occasions, 4Kx2K and 8Kx4K have even appeared. ultra-high-definition digital video. In video applications, transmission bandwidth and storage space are undoubtedly the core resources. How to realize high-definition video storage in a limited space and achieve good transmission in a network environment with bandwidth bottlenecks is a big problem. High-definition video can bring people a higher quality of life, but it will inevitably have a huge amount of data. For example, for 1080P high-definition video, the pixels are 1920X1080, the format is 4:2:0, an...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): H04N19/44H04N19/436H04N19/80
Inventor 梁凡罗林
Owner SUN YAT SEN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products