Supercharge Your Innovation With Domain-Expert AI Agents!

Task configuration method for heterogeneous distributed machine learning cluster

A machine learning and task configuration technology, applied in the physical field, can solve the problems of high time overhead, high configuration task time overhead, low utilization of node resources, etc., and achieve the effect of strong adaptability and high model accuracy

Pending Publication Date: 2021-11-02
XIDIAN UNIV
View PDF1 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0005] The purpose of the present invention is to address the above-mentioned deficiencies in the prior art, and propose a task configuration method based on heterogeneous distributed machine learning clusters, which is used to solve the problem of low resource utilization of heterogeneous distributed machine learning cluster nodes and heterogeneous parameter servers. The problem of large time overhead for each node configuration task in a distributed machine learning cluster
The present invention dynamically configures tasks for each node by replacing the characteristic parameters of the node resources before the change with the characteristic parameters of the changed node resources. Compared with the traditional method, it does not need to count the number of tasks completed by each node training, and solves the parameter server problem. The problem of high time overhead for configuring tasks for each node in a heterogeneous distributed machine learning cluster

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Task configuration method for heterogeneous distributed machine learning cluster
  • Task configuration method for heterogeneous distributed machine learning cluster
  • Task configuration method for heterogeneous distributed machine learning cluster

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0042] By the following figure 1 Further description of the invention.

[0043] Refer figure 1 Specific steps of the present invention achieves further described.

[0044] Step 1, to build a heterogeneous distributed machine learning cluster.

[0045] A parameter server and at least four nodes a heterogeneous distributed machine learning cluster.

[0046] Refer figure 2 , Heterogeneous distributed learning machine parameter consists of a server cluster and eight nodes constructed according to the embodiment of the present invention is further described.

[0047] Step 2, generating a training set and the prediction set.

[0048] Selecting at least 10,000 server parameters set image composed of images, each image comprising at least one target.

[0049] Embodiments of the present invention, an image data set derived from the open source cifar10, 20000 were selected image. Each image contains an image of an airplane.

[0050] Each aircraft image for each image to mark, label and gene...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a task configuration method for a heterogeneous distributed machine learning cluster. The method comprises the following steps: constructing the heterogeneous distributed machine learning cluster; generating a training set and a prediction set; pre-training the convolutional neural network; generating a random forest training sample subset of the parameter server; constructing a random forest model; generating an inferred training time for each node; configuring a task for each node; updating the pre-trained convolutional neural network; training a convolutional neural network; and when the number of training times of the convolutional neural network corresponding to each node reaches the maximum number of times, reconfiguring node tasks for node resource feature parameter changes existing in the distributed machine learning cluster. According to the method, the matching degree of the task configured for each node in the heterogeneous distributed machine learning cluster by the parameter server and the resource of the node is improved, and the task can be dynamically reconfigured for each node according to the resource change of each node.

Description

Technical field [0001] The present invention belongs to the field of physical technology, and more is a further involve a task-oriented configuration method for a shared machine learning cluster in the field of distributed machine learning. The present invention can be used in a large-scale heterogeneous distributed machine learning cluster through reasonable task configuration to make full use of the cluster node to calculate resources. Background technique [0002] With the arrival of the big data era, the data set for training machine learning models presents the trend of explosion growth. Fast training, dynamic flexible distributed machine learning clusters have become an inevitable trend of large-scale machine learning development. Distributed Machine Learning Through the parameter server to put the model training task to a large number of computing performance cluster nodes, shorten the time of the entire model training. However, most existing node resources are inconsisten...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): G06F9/50G06K9/62G06N3/04G06N3/08G06N20/00
CPCG06F9/5088G06N3/08G06N20/00G06N3/045G06F18/214G06F18/24323
Inventor 姬文浩顾华玺李竟蔚余晓杉任泽昂李硕
Owner XIDIAN UNIV
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More