Unlock instant, AI-driven research and patent intelligence for your innovation.

High-efficiency video coding intra-frame CTU division method and system

A high-efficiency video and encoding frame technology, applied in neural learning methods, digital video signal modification, biological neural network models, etc., can solve the problems of shortening encoding time, high computational complexity, and long encoding time, and achieve shortening encoding time , reduce computational complexity, and ensure accuracy

Pending Publication Date: 2022-07-29
GUANGDONG UNIV OF TECH
View PDF0 Cites 2 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] In order to solve the problem of high computational complexity and long encoding time of the current CTU division method in HEVC, the present invention proposes a method and system for high-efficiency video coding intra-frame CTU division, based on the correlation between texture complexity and deep learning, using Divide the threshold and neural network to judge CU segmentation, reduce the computational complexity of HEVC intra-frame coding, and shorten the coding time

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • High-efficiency video coding intra-frame CTU division method and system
  • High-efficiency video coding intra-frame CTU division method and system
  • High-efficiency video coding intra-frame CTU division method and system

Examples

Experimental program
Comparison scheme
Effect test

Embodiment 1

[0068] figure 1 A schematic flowchart showing a method for dividing CTUs in a high-efficiency video coding frame proposed in an embodiment of the present invention, including the following steps:

[0069] S1. Collect an image data set, and use the image data set to make video sets of different resolutions;

[0070] There are many methods for making a video set. In this embodiment, the collected image data set is a RAISE ultra-high-definition image set, and four YUV format video sets with different resolutions are produced from the RAISE ultra-high-definition image set. The resolutions are Select several 4928x3264, 2560x1600, 1536x1024, 704x576; 4928x3264 ultra-high-definition images, and then downsample some photos to three resolutions of 2560x1600, 1536x1024, 704x576, and randomly divide the YUV format video set of each resolution into 90 % of the training set, 5% of the validation set and 5% of the test set; for the HEVC reference software HM16.20, use the All_intra configu...

Embodiment 2

[0093] image 3 Represents the structure diagram of the A neural network proposed in the embodiment of the present invention, such as image 3 As shown, A neural network includes a first convolution block, a second convolution block, a third convolution block, a fourth convolution block, a fifth convolution block, a fully connected layer and an output layer connected in sequence; the first volume The accumulation block consists of one convolutional layer, and each of the second convolutional block, the third convolutional block, the fourth convolutional block, and the fifth convolutional block consists of two convolutional layers, and the two convolutional The buildup parameters are set the same; see Figure 4 , the first convolution block is set to 64 convolution kernels, and the size of the convolution kernel is 7x7; the second convolution block is set to 64 convolution kernels, and the size of the convolution kernel is 3x3; the third convolution block is It is set to 128 ...

Embodiment 3

[0099] like Figure 5 As shown, the B neural network includes the first convolutional layer, the second convolutional layer, the pooling layer, the third convolutional layer, the fourth convolutional layer, the fully connected layer and the output layer; the first layer of volume The convolution layer is set to 32 convolution kernels, and the size of the convolution kernel is 3x3; the second convolution layer is set to 64 convolution kernels, and the size of the convolution kernel is 3x3; the pooling layer is set to AvgPool operation , the pooling kernel size is 2x2; the third convolution layer is set to 64 convolution kernels, and the size of the convolution kernel is 2x2; the fourth convolution layer is set to 128 convolution kernels, and the size is 2x2;

[0100] The fully connected layer includes two hidden layers, which are randomly lost between the second hidden layer and the output layer with a probability of 50%: the output layer is activated by the Sigmoid function, ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention provides a high-efficiency video coding intra-frame CTU division method and a high-efficiency video coding intra-frame CTU division system, relates to the technical field of video coding unit division, solves the problems of high calculation complexity and long coding time of a current intra-frame CTU division method in HEVC, and makes video sets with different resolutions based on an image data set. The method comprises the following steps of: establishing a data set, extracting size division conditions of an intra-frame coding unit (CU) under different quantization parameter (QP) settings, obtaining a division threshold value T based on the established data set, determining 64 * 64 CU division according to a size relationship between an average absolute deviation M and T, training two neural networks through a training set, respectively judging whether a 32 * 32 CU and a 16 * 16 CU continue to be divided or not, and determining whether the 32 * 32 CU and the 16 * 16 CU continue to be divided or not. And finally, the obtained final CU size is determined to be subjected to subsequent coding processing, so that the calculation complexity of HEVC intra-frame coding is reduced, and the coding time is shortened.

Description

technical field [0001] The present invention relates to the technical field of video coding unit division, and in particular, to a method and system for high-efficiency video coding intra-frame CTU division. Background technique [0002] The original Coding Tree Unit (CTU) division algorithm in High Efficiency Video Coding (HEVC) is as follows: a CTU contains one or more coding units (Coding Unit, CU), and the CU has four sizes, respectively. 64x64, 32x32, 16x16, 8x8. The quadtree traversal algorithm of the CTU can be used to find the best partitioning method of each CTU and determine the best CU partition, but it greatly increases the coding complexity. As people's requirements for video quality are getting higher and higher, it brings challenges to the video coding and decoding of mobile devices. Therefore, it is necessary to reduce the coding complexity through algorithm optimization, so that mobile devices can apply HEVC with a lower threshold. [0003] There are three...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): H04N19/119H04N19/124H04N19/159H04N19/70G06N3/04G06N3/08
CPCH04N19/119H04N19/124H04N19/159H04N19/70G06N3/08G06N3/045
Inventor 庞贵杰原玲晏陈旭王耀葛文瑞森
Owner GUANGDONG UNIV OF TECH