Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Data size-based shuffle switch matrix compression method

A technology of data granularity and switch matrix, applied in machine execution devices, concurrent instruction execution, etc., can solve problems such as large proportion of complex operations and differences in data granularity, and achieve the effects of compressing capacity, reducing power consumption, and improving utilization

Active Publication Date: 2013-03-27
NAT UNIV OF DEFENSE TECH
View PDF0 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0008] Programs of different applications (or different program segments of the same application) often have different data widths. For example, in video algorithms, data widths are mostly bytes or halfwords. In wireless communication algorithms, complex operations (data widths of word or double word) accounted for a large proportion, the above phenomenon also leads to a large difference in the data granularity of the shuffling requests of different programs (or different program segments of the same application program), which seeks a more efficient The switch matrix compression method offers the possibility of

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Data size-based shuffle switch matrix compression method
  • Data size-based shuffle switch matrix compression method
  • Data size-based shuffle switch matrix compression method

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0040] The present invention will be further described in detail below in conjunction with the accompanying drawings and specific embodiments.

[0041] In the preprocessing stage of the present invention, for each shuffling request, the final shuffling data granularity is determined step by step, and the switch matrix is ​​compressed according to the data granularity determined step by step, and finally further compressed according to the current granularity index method, so that Complete the final compression of the switch matrix.

[0042]Assume that the total width of the data path of the processor is W bits, the minimum data granularity is G bits, the size of the Crossbar is N*N, and the width of each port is G bits (where W and G are integer powers of 2, And N=W / G), then the shuffling mode of each shuffling request is initially an N*N switch matrix, let it be X, and the value of each element in it is X i,j (1≤i≤N-1, 1≤j≤N-1), assuming that the maximum data granularity sup...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a data size-based shuffle switch matrix compression method, which comprises the following steps of: step 1, when performing data size-based shuffle switch matrix compression according to a shuffle request, setting the data size of the current shuffle request to be the smallest data size first; step 2, judging whether the data size of the current shuffle request can be doubled, and if the data size of the current shuffle request cannot be doubled, skipping to the step 3, otherwise, skipping to the step 4; step 3, doubling the current data size, compressing the current shuffle switch matrix according to the doubled data size, and skipping to the step 2; step 4, further compressing the current switch matrix by using the current size indexing method; and step 5, judging whether the current shuffle request is the last shuffle request in the procedure, if the current shuffle request is not the last shuffle request, processing the next shuffle request, and skipping to the step 1, otherwise, ending the algorithm. The data size-based shuffle switch matrix compression method is simple in principle, is easy and convenient to operate, and can effectively improve the utilization ratio and the compression efficiency of a shuffle mode memory.

Description

technical field [0001] The present invention mainly relates to the field of general-purpose processors or digital signal processors (Digital Signal Processor, DSP) using Single Instruction Stream Multiple Datastreams (Single Instruction Stream Multiple Datastreams, SIMD) technology, and specifically refers to a kind of processor used in SIMD or DSP. Shuffle switch matrix compression method. Background technique [0002] SIMD technology, represented by subword parallel and vector technology, has great advantages in solving data-intensive applications such as media processing and wireless communication, so it has been widely used and developed, and is favored by processors, especially DSP. . The shuffling unit is mainly responsible for the data interaction between the registers of each processing unit of the SIMD component, and is a key component that restricts the performance of parallel processing units. [0003] The hardware main structure of the shuffling unit generally ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F9/38
Inventor 陈书明刘胜万江华陈海燕陈跃跃孙海燕张凯王耀华刘祥远李振涛
Owner NAT UNIV OF DEFENSE TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products