Method, device, computing device, and computer storage medium for allocating display memory

A device and data storage technology, applied in the computer field, can solve problems such as low utilization rate, high implementation cost, and the inability of the graphics card to load multiple deep learning algorithm models, so as to achieve the effect of reducing implementation cost and video memory resources

Active Publication Date: 2019-03-08
杭州比智科技有限公司
View PDF9 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

One of the more typical problems is as follows: most open source deep learning frameworks are mainly developed for academic research, but the memory utilization rate of graphics cards has not been very high, and general graphics card devices cannot afford to feature ultra-large-scale deep neural networks such as ResNet Memory Requirements of Deep Learning Algorithms for Extractive Networks
Moreover, in actual industrial use, if you use an open source deep learning framework, such as Caffe, which is a general deep learning framework, its hardware cost is very expensive, which will cause a graphics card to be unable to load multiple deep learning algorithm models.
[0003] Therefore, a solution suitable for industrial production is needed to overcome the problems of low memory utilization and high implementation cost of existing deep learning algorithm models during operation

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method, device, computing device, and computer storage medium for allocating display memory
  • Method, device, computing device, and computer storage medium for allocating display memory
  • Method, device, computing device, and computer storage medium for allocating display memory

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0026] Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. Although exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited by the embodiments set forth herein. Rather, these embodiments are provided for more thorough understanding of the present disclosure and to fully convey the scope of the present disclosure to those skilled in the art.

[0027] The deep learning algorithm model is a highly modular and modular algorithm model. Currently commonly used deep learning frameworks such as Caffe, Mxnet, tensorflow, etc., their design concept is to abstract the deep learning algorithm model into a calculation roadmap ( It is a defined directed acyclic graph), and then control the input data flow of the algorithm model, and calculate layer by layer, and each layer will have a...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a method, a device, a computing device and a computer storage medium for allocating a display memory, wherein the method comprises the following steps of analyzing the data flow paths of one or more depth learning algorithm models loaded in the device to obtain an analysis result; according to the parsing result, obtaining the first number of display memory blocks needed inthe data flow process of each depth learning algorithm model; determining an allocation rule of a first number of display blocks in a data flow process of each depth learning algorithm model; allocating display memory blocks to the depth learning algorithm model according to allocation rules corresponding to each depth learning algorithm model. In accordance with that invention, according to theanalysis results of the data flow path, the number of the display memory blocks and the allocation rules of the number of the display memory blocks needed by the depth learning algorithm model are determined, so that the display memory in the device can play the role of cache to the maximum extent, and the project implementation cost of the depth learning algorithm model can be reduced by improving the utilization rate of the display memory.

Description

technical field [0001] The invention relates to the technical field of computers, in particular to a video memory allocation method, device, computing equipment and computer storage medium. Background technique [0002] With the rapid development of hardware technology, the mainstream hardware devices currently on the market can already meet the basic operation of deep learning algorithms, but there are many problems hidden in actual industrial production. One of the more typical problems is as follows: most open source deep learning frameworks are mainly developed for academic research, but the memory utilization rate of graphics cards has not been very high, and general graphics card devices cannot afford to feature ultra-large-scale deep neural networks such as ResNet Extract the memory requirements of deep learning algorithms for networks. Moreover, in actual industrial use, if you use an open source deep learning framework, such as Caffe, a general deep learning framew...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06N3/063
CPCG06N3/063
Inventor 洪伟
Owner 杭州比智科技有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products