Unlock instant, AI-driven research and patent intelligence for your innovation.

Architecture and algorithms for data compression

a data compression and architecture technology, applied in the field of graphics processing, can solve the problems of limiting the overall system speed and efficiency, increasing the use of valuable board area, and adding more memory circuits, so as to save bandwidth and save dram bandwidth

Active Publication Date: 2019-07-02
NVIDIA CORP
View PDF4 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

The patent describes a way to make computer graphics processing units (GPUs) more efficient by adding compression capabilities to certain parts of the GPU. This allows for better compression of data transfers between the GPU and the main computer memory, saving bandwidth and reducing internal data transfer requests. Overall, this approach results in improved performance and efficiency for graphics processing.

Problems solved by technology

Consequently, transfers between on-chip processors and external memory may present a bottleneck that limits the overall system speed and efficiency.
Adding more memory circuits corresponding to the increase in memory transfers may improve memory capacity, but adding more memory circuits incurs increased cost and increased use of valuable board area.
The cache is typically limited in capacity, and so the GPU frequently executes transfers to and from a main memory, such as a system memory.
In sum, present architectures implement compression with only about forty percent of memory transfers, while the remaining memory transfers impose a burden on memory bandwidth usage.
Further, as memory transfers are likely to increase with developing processor technology, the memory bandwidth usage may be expected to also increase, resulting in an increase in the bottleneck that further would further limit processing efficiency.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Architecture and algorithms for data compression
  • Architecture and algorithms for data compression
  • Architecture and algorithms for data compression

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0016]In the following description, numerous specific details are set forth to provide a more thorough understanding of the present invention. However, it will be apparent to one of skill in the art that the present invention may be practiced without one or more of these specific details.

System Overview

[0017]FIG. 1 is a block diagram illustrating a computer system 100 configured to implement one or more aspects of the present invention. As shown, computer system 100 includes, without limitation, a central processing unit (CPU) 102 and a system memory 104 coupled to a parallel processing subsystem 112 via a memory bridge 105 and a communication path 113. Memory bridge 105 is further coupled to an I / O (input / output) bridge 107 via a communication path 106, and I / O bridge 107 is, in turn, coupled to a switch 116.

[0018]In operation, I / O bridge 107 is configured to receive user input information from input devices 108, such as a keyboard or a mouse, and forward the input information to C...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A system architecture conserves memory bandwidth by including compression utility to process data transfers from the cache into external memory. The cache decompresses transfers from external memory and transfers full format data to naive clients that lack decompression capability and directly transfers compressed data to savvy clients that include decompression capability. An improved compression algorithm includes software that computes the difference between the current data word and each of a number of prior data words. Software selects the prior data word with the smallest difference as the nearest match and encodes the bit width of the difference to this data word. Software then encodes the difference between the current stride and the closest previous stride. Software combines the stride, bit width, and difference to yield final encoded data word. Software may encode the stride of one data word as a value relative to the stride of a previous data word.

Description

BACKGROUND OF THE INVENTIONField of the Invention[0001]Embodiments of the present invention relate generally to graphics processing and, more specifically, to architecture and algorithms for data compression.Description of the Related Art[0002]In computer systems, in general, and in graphics processing units (GPUs), in particular, evolving technology affords manufacturers the ability to produce processors with increasing density that process data with increasing speed and efficiency. Higher data rates associated with streaming video, among other types of data transmission, magnify the importance of conserving memory bandwidth. Accordingly, efforts to reduce the memory bandwidth utilization are increasingly important.[0003]However, present technology has advanced in the areas of processor speed and efficiency at a faster rate than corresponding advances in memory access technology. Consequently, transfers between on-chip processors and external memory may present a bottleneck that li...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(United States)
IPC IPC(8): H03M7/00G06F3/06H03M7/30G06T1/20H03M7/40
CPCG06F3/061G06F3/0647G06F3/0661G06T1/20H03M7/40H03M7/60G06F3/0673G06F2212/302
Inventor DIMITROV, ROUSLANPOOL, JEFFKRISHNAMURTHY, PRAVEENAMSINCK, CHRISMEHRA, KARANCUTLER, SCOTT
Owner NVIDIA CORP