Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Concurrent calculation system based on Spark and GPU (Graphics Processing Unit)

A parallel computing and distributed computing technology, applied in computing, computers, digital computer components, etc., can solve problems such as inability to perceive GPU resources, inability to distinguish between CPU tasks and GPU tasks, and task execution failures

Inactive Publication Date: 2017-09-15
FUDAN UNIV
View PDF4 Cites 36 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Since Spark cannot perceive GPU computing tasks, it cannot distinguish between CPU tasks and GPU tasks. When scheduling task execution, GPU tasks may be started on nodes without GPU devices, resulting in task execution failures.
In addition, in the YARN resource manager, it only supports the scheduling of CPU and memory resources, and cannot perceive GPU resources. It cannot provide the allocation and scheduling of GPU resources to the upper-level Spark framework.
Due to the reasons of YARN and Spark framework itself, the traditional method of performing GPU computing in Spark cannot adapt to the heterogeneous cluster environment

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Concurrent calculation system based on Spark and GPU (Graphics Processing Unit)
  • Concurrent calculation system based on Spark and GPU (Graphics Processing Unit)
  • Concurrent calculation system based on Spark and GPU (Graphics Processing Unit)

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0054] The technical scheme of the present invention will be further described below in conjunction with the accompanying drawings.

[0055] figure 1 It is a block diagram of model training and image recognition, mainly including:

[0056] 1. In terms of resource representation, you can first customize the number of GPU devices included in the node, and modify the resource representation protocol to increase the representation of GPU resources. When the node is started, the node manager initializes the resource list, and reports the resource information of the node to the resource manager through the heartbeat mechanism.

[0057] 2. In terms of resource scheduling, the present invention adds GPU, CPU, and memory resources to the hierarchical management queue of the resource management platform.

[0058] 3. The resource manager sends the resource list information to be released to the corresponding node manager through the heartbeat mechanism. When the node manager detects th...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention belongs to the technical field of concurrent calculation and concretely relates to a concurrent calculation frame system based on Spark and a GPU (Graphics Processing Unit). Based on YARN resource management platform, a resource manager and a node manager are improved so that the concurrent calculation frame system can effectively sense GPU resources of a heterogeneous cluster, thereby supporting management and scheduling of the cluster GPU resource; and then in a YARN deployment mode, a job scheduling mechanism and a task execution mechanism of the Spark are improved so that the concurrent calculation frame system supports scheduling and execution for a GPU task. Identification for the GPU resource is introduced in stages such as resource application, resource distribution, DAG (Directed Acyclic Graph) generation, stage division and task execution so that an execution engine can sense the GPU task and effectively executes the GPU task in the heterogeneous cluster; and meanwhile, an effective programming model under the framework is provided by use of the characteristics of efficient memory calculation of Spark in combination with the advantages of multi-core concurrent calculation of the GPU. The system can effectively process data-intensive and compute-intensive jobs and greatly improves the job processing efficiency.

Description

technical field [0001] The invention belongs to the technical field of parallel computing, in particular to a parallel computing framework system based on Spark and GPU. Background technique [0002] In today's society, the scale of data that needs to be processed by various industries has shown a trend of massive amounts, and big data has attracted widespread attention from various industries in society. Undoubtedly, big data contains a wealth of useful information, and if it can be properly mined and used, it will greatly promote scientific research and social economy. Because the information contained in big data can assist business decision-making and scientific research, it has been rapidly developed and applied in many industries. In the era of big data, everything is centered on data, and a lot of effective information that cannot be obtained through other methods can be mined and analyzed from massive historical data, so as to improve the accuracy of decision-making...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F9/48G06F9/50G06F15/173
CPCG06F9/4881G06F9/5038G06F9/505G06F9/5083G06F15/17318G06F15/17331
Inventor 郑健杜姗姗冯瑞金城薛向阳
Owner FUDAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products