Hybrid precision arithmetic unit for FP16 floating point data and INT8 integer data operation

A FP16, floating-point data technology, applied in the direction of electrical digital data processing, digital data processing components, instruments, etc., can solve the problems of calculation accuracy loss, accuracy loss, huge, etc., and achieve high-precision results

Active Publication Date: 2021-05-28
厦门壹普智慧科技有限公司
View PDF6 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, when using 16-bit or 8-bit integer precision calculations, it often brings a certain loss of calculation accuracy
Especially under 8-bit integer precision, if no additional techniques such as quantization training are used to compensate, the loss of accuracy caused by 8-bit integer precision calculations will often be huge

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Hybrid precision arithmetic unit for FP16 floating point data and INT8 integer data operation
  • Hybrid precision arithmetic unit for FP16 floating point data and INT8 integer data operation
  • Hybrid precision arithmetic unit for FP16 floating point data and INT8 integer data operation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0030] To further illustrate the various embodiments, the present invention is provided with accompanying drawings. These drawings are a part of the disclosure of the present invention, which are mainly used to illustrate the embodiments, and can be combined with related descriptions in the specification to explain the operating principles of the embodiments. With reference to these contents, those skilled in the art should understand other possible implementations and advantages of the present invention. Components in the figures are not drawn to scale, and similar component symbols are generally used to denote similar components.

[0031] The present invention will be further described in conjunction with the accompanying drawings and specific embodiments.

[0032] The present invention proposes a mixed-precision arithmetic unit for FP16 floating-point data and INT8 integer data operations, and applies it to the data flow calculation engine of the neural network tensor proc...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a hybrid precision arithmetic unit for FP16 floating point data and INT8 integer data operation. The mixed precision arithmetic unit comprises a precision conversion module, an arithmetic unit, two input data and one, two or four output data. The input data and the output data are expressed in an EM16 format, and the EM16 format expression is 16 digits and comprises FP16 floating point expression, INT8 integer expression and INT8 double integer expression; the two input data are respectively feature data and parameter data in neural network calculation; the precision conversion module is used for executing precision conversion among expressions of the EM16 format of the feature data according to the external configuration information; and the arithmetic unit is used for executing additive operation or multiplication operation between two pieces of data which are both FP16 floating point data or both INT8 integer data according to external configuration information. According to the hybrid precision arithmetic unit, FP16 floating point data and INT8 integer data can be used in a mixed mode in the calculation process, so that hybrid precision calculation can have the precision of the FP16 and the speed of the INT8 in a neural network calculation task.

Description

technical field [0001] The invention relates to the field of neural network tensor processor architecture, in particular to a mixed-precision arithmetic unit used for FP16 floating-point data and INT8 integer data operations. Background technique [0002] Traditional neural network computing hardware uses fixed computing precision. For example, DSPs usually use 16-bit fixed-point calculations, GPUs usually use 32-bit floating-point or 16-bit floating-point calculations, and AI accelerators usually use 16-bit floating-point or 8-bit integer calculations. [0003] Neural network algorithms usually use 32-bit floating-point precision for model training, and the parameter data obtained from training usually use 32-bit floating-point precision. On the other hand, when using 32-bit floating point or 16-bit floating point precision for inference calculations, you can usually achieve comparable accuracy to training. However, when calculating with 16-bit or 8-bit integer precision, ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F7/483G06F7/575
CPCG06F7/483G06F7/575Y02D10/00
Inventor 罗闳訚何日辉周志新
Owner 厦门壹普智慧科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products