Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

96 results about "GPU cluster" patented technology

A GPU cluster is a computer cluster in which each node is equipped with a Graphics Processing Unit (GPU). By harnessing the computational power of modern GPUs via General-Purpose Computing on Graphics Processing Units (GPGPU), very fast calculations can be performed with a GPU cluster.

Resource management method and system facing GPU (Graphic Processing Unit) cluster

The invention discloses a resource management method facing a GPU (Graphic Processing Unit) cluster, which comprises the following steps: a main management node establishes two charts (a resource information chart and a task information chart); the main management node receives a new task; the main management node judges whether the task is a CPU (Central Processing Unit) task or a GPU task; the main management node seeks free resource meeting the requirement of the task; if the task is a CPU task, a secondary management node conducts pretreatment on the data of the task, and dispensing pieces of the data to all nodes managed by the secondary management node for calculation, the main management node reclaims CPU resource related to all the nodes managed by the secondary management node according to the number of the task after calculation; if the task is a GPU task, the main management node reclaims the GPU resource related to all the nodes managed by the secondary management node according to the number of the task when GPU calculation is detected to be finished; meanwhile, the CPUs of all the nodes managed by the secondary management node are used for post-processing of a result, and the post-processing is finished. According to the invention, CPU resource and the GPU resource are treated differently; through the detection of the task, free GPU resource can be reclaimed fast.
Owner:HUAZHONG UNIV OF SCI & TECH

A GPU cluster deep learning edge computing system oriented to sensing information processing

The invention relates to a GPU cluster deep learning edge computing system oriented to sensing information processing. pre-feature extraction is carried out on sensing information by using weak computing power of front-end intelligent sensing equipment; the quantity of original data information is greatly compressed, then the rest processing tasks are transmitted to a GPU cluster for large-scale sensing data feature clustering set processing, the computing power of front-end intelligent sensing equipment can be dynamically adapted through task splitting processing, and the cost pressure of theconsistency requirement of the front-end sensing equipment and hardware versions is reduced; The communication pressure of the edge computing network is reduced, so that the cost of constructing theedge computing network is greatly reduced; Network data feature transmission hides user privacy;, the SPMD advantages of the GPU are brought into play through the clustering operation according to thedata transmitted in the network and the stored data core characteristics, the parallel computing efficiency of edge computing is improved, and meanwhile, the large-scale parallel computing capacity of the GPU cluster and the advantages of low cost and high reliability are effectively brought into play.
Owner:UNIV OF SHANGHAI FOR SCI & TECH

GPU cluster deep learning task parallelization method, device and electronic equipment

The embodiment of the invention provides a GPU cluster deep learning task parallelization method, a GPU cluster deep learning task parallelization device and electronic equipment, and relates to the technical field of internet. The method comprises the following steps: firstly, analyzing the similarity between a to-be-processed deep learning task and each computing node of a GPU cluster; determining a target computing node of the to-be-processed deep learning task in the GPU cluster, reducing the possibility of computing node resource contention, and improving the utilization rate and the execution efficiency of deep learning task system resources; according to the number of GPUs required by the to-be-processed deep learning task, obtaining a to-be-processed deep learning task; dividing the to-be-processed deep learning task into a plurality of target sub-tasks, analyzing the interference level and the communication cost of the target subtask; determining the target GPU of the target sub-task in the target computing node, and avoiding unbalanced resource allocation on the GPU in the computing node. The high parallelization of the deep learning task is realized. The resource utilization rate of the GPU cluster is improved. Meanwhile, the execution efficiency of the deep learning task is improved.
Owner:BEIJING UNIV OF POSTS & TELECOMM

Heterogeneous network perception model division and task placement method in pipelined distributed deep learning

The invention provides a heterogeneous network perception model division and task placement method in assembly line distributed deep learning, which mainly comprises three parts, namely deep learningmodel description, model division and task placement and assembly line distributed training. According to the method, firstly, for resource requirements of deep learning application in the GPU training process, corresponding indexes such as calculation time, intermediate result communication quantity and parameter synchronization quantity in the training execution process of the deep learning application are described and serve as input of model division and task placement; then indexes and heterogeneous network connection topology of the GPU cluster are obtained according to model description, a dynamic programming algorithm based on min-max is designed to execute model division and task placement, and the purpose is to minimize the maximum value of task execution time of each stage afterdivision so as to ensure load balance. And finally, according to a division placement result, performing distributed training by using assembly line time-sharing injection data on the basis of modelparallelism, thereby realizing effective guarantee of training speed and precision.
Owner:SOUTHEAST UNIV

WPA shared key cracking system based on GPU cluster

The invention relates to the technical field of password cracking, and discloses a WPA shared key cracking system based on a GPU cluster. The WPA shared key cracking system based on the GPU cluster specifically comprises a control node and a plurality of GPU computational nodes. The control node intercepts and obtains a WPA data package, a characteristic value extracting module extracts cracking characteristic values, and a user interaction module of the control node receives a password cracking range defined by a user; a password space partition module partitions password sections of certain ranges to all GPU computational nodes, and sends the password sections to all the corresponding GPU computational nodes; the GPU computational nodes calculate and obtain temporary verification parameters MIC_TMP, a breaking password passphrase is regarded as the shared key when the temporary verification parameters MIC_TMP are identical with a verification parameter MIC value, and password cracking is completed. According to the WPA shared key cracking system based on the GPU cluster, the GPU cluster is adopted to crack a password of a WPA/WPA2-PSK, multi-node GPU clusters are supported, expansion can be carried out properly according to needs, and cracking performance is improved well. Meanwhile, oriented to the heterogeneous characteristic of the GPU clusters, a reliable task dispatching system is designed, load balancing is achieved, and cracking speed is improved.
Owner:NO 30 INST OF CHINA ELECTRONIC TECH GRP CORP

Deep learning-oriented multi-type GPU cluster resource management scheduling method and system

The invention discloses a deep learning-oriented multi-type GPU cluster resource management and scheduling method and system. The method comprises the following steps: dividing a GPU cluster into a plurality of GPU groups according to the model of a GPU, counting the idle operational capability of each GPU group, obtaining all users accessing the GPU cluster, and recording the minimum operationalcapability requirement of each user; and periodically accessing the job queue, obtaining the job to be processed with the highest priority in the job queue, and scheduling GPU cluster resources according to the job to be processed. According to the invention, GPUs of different brands and models are uniformly managed as one cluster for deep learning, the number of maintained GPU clusters is reduced, and the GPU cluster management complexity is simplified; the requirements of different users in deep learning can be met; reasonable user attributes are set according to user requirements, users donot need to be familiar with and care about GPU cluster environments, resource scheduling is carried out according to operational capability requirements and priorities of the users, resources meetingthe requirements can be automatically allocated through the scheduling method, and the resource utilization rate of different GPU type groups is increased.
Owner:HANGZHOU EBOYLAMP ELECTRONICS CO LTD

Method for checking model training notifications and training logs at mobile terminal

The invention provides a method for dynamically receiving an artificial intelligence model training notification at a mobile terminal in a cross-platform manner. During artificial intelligence technology development, a large number of deep learning model tasks need to use a GPU cluster at the cloud to complete training tasks. The deep learning model training state message notification can be checked in real time by using the mobile terminal device, the experience of checking the model training state and checking the training state feedback in real time by the user can be greatly improved, andthe process of participating in the competition becomes more and more convenient and efficient. Therefore, according to a technical scheme, a competition training task instruction can be sent to a GPUserver for training on equipment such as a mobile terminal HTML5 page or a local terminal PC; according to the method for checking the training notice and the training process in the mobile terminalequipment at any time, the single use scene limitation that a user checks the training process and the notice in the competition training process is solved. A user can check the competition training process and the result notification in real time through the mobile terminal equipment in any scene, and the user experience and the competition training time efficiency of participating in competitiontraining are improved.
Owner:北京智能工场科技有限公司

GPU cluster shared video memory system, method, device and equipment

PendingCN113674133AImprove the performance of shared video memoryHelp with integrationResource allocationProcessor architectures/configurationVideo memoryComputer architecture
The invention discloses a GPU cluster video memory sharing method, device and system and equipment. The method comprises the following steps: determining GPU cluster global video memory address mapping information of a target application according to a GPU cluster global virtual video memory address space of the target application running on a first computing node; when page missing abnormity occurs when the target application accesses the GPU video memory, determining a second computing node where the target page data is located according to global video memory address mapping information of the target application; and calling the target page data in the second computing node into the GPU video memory of the first computing node, and reading the target page data from the GPU video memory of the first computing node by the target application. By the adoption of the processing mode, the video memory resources are aggregated from the GPU cluster system level, a unified GPU video memory address space and a single programming view are provided for a distributed GPU cluster in the face of large loads with high video memory resource requirements, explicit management data migration and communication are avoided, and GPU cluster system programming is simplified.
Owner:ALIBABA SINGAPORE HLDG PTE LTD

GPU cluster service management system and method

The invention belongs to the field of computer management, and particularly relates to a GPU cluster service management system and method. The GPU cluster service management system comprises a resource monitoring module used for monitoring GPU cluster resources, generating cluster resource data and sending the cluster resource data, a resource allocation module used for acquiring task informationand the cluster resource data and allocating task resources according to the task information and the cluster resource data, a checking module used for obtaining the cluster resource data sent by theresource monitoring module, checking the GPU cluster resource state according to the cluster resource data, generating a checking result and sending the checking result, and an isolation module used for acquiring the inspection result and isolating abnormal resources according to the inspection result. According to the GPU cluster service management system and method, all resource states in the GPU cluster can be monitored in real time, and it is ensured that resources are efficiently utilized; and abnormal resources can be automatically checked out and isolated, so that normal operation of the GPU cluster is ensured, and the processing efficiency of the GPU cluster is improved.
Owner:北京中科云脑智能技术有限公司 +1

Fluorescent diffusion optical cross-sectional image reestablishing method based on dfMC model

The invention discloses a fluorescent diffusion optical cross-sectional image reestablishing method based on a decoupling fluorescence Monte Carlo (dfMC) model, and belongs to the technical field of biomedical engineering. The method includes the steps of firstly, determining a detecting area, selecting a plurality of scanning points in the detecting area, and obtaining fluorescent intensity distribution on a detector; secondly, establishing a three-dimensional digital model for depicting a tissue optical parameter space structure, conducting forward-direction white Monte Carlo simulation of stimulating photons according to the scanning positions and directions of a light source, tracking the stimulating photons, and recording the corresponding physical quantities of the photons reaching the detector on a path; thirdly, calculating the weight of fluorescent photons through a dfMC method, and calculating a fluorescent Jacobi matrix; fourthly, calculating the positions and absorbing coefficients of fluorophores in tissue through iterative reconstruction of GPU clusters. The method has the advantage of providing an accurate and rapid reestablishing method for a three-dimensional fluorescence tomography system through the high-precision dfMC model on the basis of the accelerated iterative reconstruction process of the GPU clusters.
Owner:HUAZHONG UNIV OF SCI & TECH

Hybrid pipeline parallel method for accelerating distributed deep neural network training

The invention provides a hybrid pipeline parallel method for accelerating distributed deep neural network training, and mainly solves the problems that resources are not fully utilized and efficient distributed training cannot be realized in a traditional GPU cluster distributed training process. The core mechanism of the method mainly comprises three parts, namely deep learning model description, model hybrid division and hybrid pipeline parallel division. The method comprises: firstly, for the resource requirements of deep learning application in a GPU training process, describing corresponding indexes such as a calculated amount, an intermediate result communication amount and a parameter synchronization amount in the training process, and using the indexes as input of model hybrid division and task placement; and then, according to a model description result and an environment of a GPU cluster, designing two division algorithms based on dynamic programming to realize model hybrid division and hybrid pipeline parallel division so as to minimize the maximum value of task execution time of each stage after division, so that load balance is ensured, and efficient distributed training of a deep neural network is realized.
Owner:SOUTHEAST UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products