Tensor cache and access structure and method thereof

A cache and access method technology, applied in memory systems, register devices, instruments, etc., can solve problems such as inconsistent data access delays, complex memory access scheduling, and memory bank access conflicts.

Active Publication Date: 2021-06-08
杭州慧芯达科技有限公司
View PDF7 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] However, in the case of shared Cache, using multiple computing engines for processing is prone to problems such as complex memory access scheduling and storage bank access conflicts.
However, when using private cache / separate scratch pad memory, since there are multiple copies of the same data on the chip, it will lead to waste of limited on-chip storage space and inconsistent data access delays.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Tensor cache and access structure and method thereof
  • Tensor cache and access structure and method thereof
  • Tensor cache and access structure and method thereof

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0039] The present invention will be described in further detail below in conjunction with the accompanying drawings and specific embodiments.

[0040] This embodiment discloses a Tensor cache and access structure and method thereof, which is mainly aimed at a data cache, access and parallel processing scheme for two-dimensional / three-dimensional regular data access and processing.

[0041] Such as figure 1 As shown, this solution includes a Tensor cache 1 with multi-dimensional distribution of data. Specifically, the Tensor cache 1 is divided into multiple Block cache blocks in three dimensions, and each cache block has multiple Slice caches distributed in a two-dimensional array Each cache slice has multiple memory banks distributed in a two-dimensional array.

[0042] The parameters of the Tensor structure variable tensor are as follows:

[0043] Length on the third dimension: DIM;

[0044] Block cache block array contained in two dimensions: BLOCK={BLOCK_ROW,BLOCK_COLUM...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention provides a Tensor cache, an access structure and a method thereof. The method comprises the following steps: S1, acquiring a mapping relationship between Tensor structure variables of the Tensor cache and an incidence matrix; S2, executing a) or b) according to the mapping relationship: a) loading the data of the incidence matrix in the main memory into a Tensor cache; and b) writing the data in the Tensor cache into the incidence matrix in the main memory. According to the scheme, the data of the incidence matrix in the main memory is mapped and extracted into the Tensor cache chip, and the data of the incidence matrix is cut and respectively placed in different cache chips, so that subsequent data conversion and parallel operation are facilitated, and meanwhile, the utilization rate of a limited on-chip memory can be improved; distributing the data of the Tensor cache in a multi-dimensional manner, and processing the data in a multi-dimensional manner, so that the data processing efficiency can be effectively improved, and the problem of memory bank access conflicts can be effectively avoided; realizing cache access and parallel processing of data, and avoiding the problems of waste of on-chip storage space and inconsistent data access delay.

Description

technical field [0001] The invention belongs to the technical field of data cache and access, and in particular relates to a Tensor cache and an access method. Background technique [0002] Algorithms such as deep learning and image processing have the characteristics of memory access and regular calculation. At present, there are three main ways to cache and access data for algorithms such as deep learning and image processing: [0003] 1) Private Cache, shared address space [0004] 2) Independent sticky note memory [0005] 3) Shared Cache [0006] However, in the case of shared Cache, using multiple computing engines for processing is likely to cause problems such as complex memory access scheduling and storage bank access conflicts. However, when using a private cache / separate scratch pad memory, since there are multiple copies of the same data on-chip, it will lead to waste of limited on-chip storage space and inconsistent data access delays. Contents of the inven...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F12/0862G06F9/30G06F3/06
CPCG06F12/0862G06F9/30098G06F3/064Y02D10/00
Inventor 黄权李兆亮王辰琛
Owner 杭州慧芯达科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products