Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Dynamic caching method for solving storage bandwidth efficiency of universal AI processor

A storage bandwidth and dynamic cache technology, applied in memory systems, electrical digital data processing, instruments, etc., can solve problems such as the inability to meet the needs of new general-purpose AI processors, and achieve the effects of reducing memory overhead, improving efficiency, and increasing utilization.

Active Publication Date: 2021-04-23
轸谷科技(南京)有限公司 +1
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0031] In view of the high-bandwidth data flow requirements of general-purpose AI processors, existing data caches need to transmit data in the basic unit of cache lines, and the data exchanged each time is the size of a single cache line, which obviously cannot meet the needs of new general-purpose AI processing. device needs

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Dynamic caching method for solving storage bandwidth efficiency of universal AI processor
  • Dynamic caching method for solving storage bandwidth efficiency of universal AI processor
  • Dynamic caching method for solving storage bandwidth efficiency of universal AI processor

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0054] The present invention will be further described now in conjunction with accompanying drawing.

[0055] Such as Figure 1 to Figure 7 As shown, the dynamic caching method used to solve the storage bandwidth efficiency of a general-purpose AI processor adds a bit flag C in the cache line to form a data segment cache, and the data segment is continuously stored in several conventional cache line data Medium; follow the steps below to read data when reading cached data:

[0056] S1. According to the read instruction given by the CPU, it is judged whether it is conventional data reading or data segment reading. If it is conventional data, it will be read according to the conventional data reading steps. If it is data segment reading, it will enter S2;

[0057] S2. Determine the position of the data segment in the cache according to the index field in the start address of the data segment;

[0058] S3, comparing the tag of the start address of the data segment with the tag ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a dynamic caching method for solving the storage bandwidth efficiency of a unisersal AI processor, which comprises adding a bit flag bit C in a caching line, continuously storing a data segment in conventional caching line data, and reading the data according to the following steps: S1, judging according to a CPU (Central Processing Unit) reading instruction, and if the data segment is read, entering S2; s2, determining the position of the data segment according to the index domain; s3, comparing the tag of the data segment with the tag in the cache, and if the tag of the data segment is consistent with the tag in the cache, entering S4; s4, checking a V bit, and if the V bit is 1, entering S5; s5, checking the C bit, and if the C bit is 1, entering S6; s6, according to the length of the data needing to be read by the reading instruction, checking the C bits in the cache behind the initial address, if the C bits of the continuously stored data are all set to be 1 and the data length meets the data reading length, performing continuous reading, and otherwise, performing cache miss. And the operation requirements of the CPU and the AI processor can be met at the same time.

Description

technical field [0001] The invention relates to a data cache, in particular to a dynamic cache method for solving storage bandwidth efficiency of a general AI processor. Background technique [0002] At present, artificial intelligence is widely used in various fields. Deep neural network technology has become a representative algorithm in the field of artificial intelligence. Key technologies such as character recognition, image classification, or speech recognition based on deep neural network technology have been widely used in search engines and intelligence. products such as mobile phones. [0003] The core computing unit of deep neural network technology is multiply-accumulate operation. Multiply-accumulator array is often used in matrix multiplication operation. Therefore, MAC (multiply-accumulator) array is the core of AI calculation. More and more general-purpose computing chips cater to AI For computing needs, a MAC array is specially added to improve computing po...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F12/0886G06F12/0868
CPCG06F12/0868G06F12/0886
Inventor 蔡浩田沈亚明葛悦飞
Owner 轸谷科技(南京)有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products