Distributed computing

a distributed computing and computing technology, applied in the field of high-performance computing, can solve the problems of low priority background jobs and restricted utilization to overnight idle periods, and achieve the effects of increasing integration and deployment, on-going savings in maintenance and administrative overhead, and enormous savings

Inactive Publication Date: 2006-08-31
BERNARDIN JAMES +1
View PDF5 Cites 222 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

[0027] The invention provides an off-the-shelf product solution to target the specific needs of commercial users with naturally parallel applications. A top-level, public API provides a simple “compute server” or “task farm” model that dramatically accelerates integration and deployment. By providing built-in, turnkey support for enterprise features like fault-tolerant scheduling, fail-over, load balancing, and remote, central administration, the invention eliminates the need for customized middleware and yields enormous, on-going savings in maintenance and administrative overhead.
[0028] Behind the public API is a layered, peer-to-peer (P2P) messaging implementation that provides tremendous flexibility to configure data transport and overcome bottlenecks, and a powerful underlying SDK based on pluggable components and equipped with a run-time XML scripting facility that provides a robust migration path for future enhancements.
[0029] Utilizing the techniques described in detail below, the invention supports effectively unlimited scaling over commoditized resource pools, so that end-users can add resources as needed, with no incremental development cost. The invention seamlessly incorporates both dedicated and intermittently idle resources on multiple platforms (Windows™, Unix, Linux, etc.). And it provides true idle detection and automatic fault-tolerant rescheduling, thereby harnessing discrete pockets of idle capacity without sacrificing guaranteed service levels. (In contrast, previous efforts to harness idle capacity have run low-priority background jobs, restricted utilization to overnight idle periods, or imposed intrusive measures, such as checkpointing.) The invention provides a system that can operate on user desktops during peak business hours without degrading performance or intruding on the user experience in any way.

Problems solved by technology

(In contrast, previous efforts to harness idle capacity have run low-priority background jobs, restricted utilization to overnight idle periods, or imposed intrusive measures, such as checkpointing.)

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Distributed computing
  • Distributed computing
  • Distributed computing

Examples

Experimental program
Comparison scheme
Effect test

example 1

[0263] Discriminators may be attached to Tasks or Jobs to ensure that they are assigned to Engines that are equipped with specific resources. Common examples include: Tasks that must run under a particular operating system or subset of operating systems; Tasks that must have at least a specified minimum of memory or disk space to run; Tasks that must run on a specific subset of Engines because of administrative or security restrictions, possibly including database or file system access restrictions; etc. Discriminators may also be used to impose analogous restrictions in order to optimize performance. For example, they may restrict very long-running Tasks to run only on processors meeting minimum performance requirements or on dedicated (as opposed to interruptible) Engines.

example 2

[0264] Scoring Discriminators may be used in connection with distributed caching to maximize reuse of distributed resources, such as objects or data. In this case, the Engines maintain a local cache, and update their associated property list whenever entries are added or removed from the cache. The Scoring Discriminator associated with each Task returns a score for each Engine based on the overlap between the resources that are available in the Engine's local cache and the resources specifically required by the Task.

[0265] The simplest examples of this kind are those in which each Task requires a single object, for example, each Task within a portfolio pricing application might look for a single deal or subportfolio. In this case, the Scoring Discriminator might return 1 in case the relevant deal or subportfolio is present in the cache, and 0 otherwise. A slightly more complex example would be one in which the Scoring Discriminator examines the Engine property list for multiple key...

example 3

[0266] Engine Discriminators may be used to dedicate some number, N, of Engines to a specific Job. This is accomplished by having the first N Tasks within the Job install an Engine Discriminator that imposes two requirements: (1) The Job Id must match the Job Id for the given Task, and (2) the Task Id must be greater than N.

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

A distributed computing system manages execution of jobs and their associated tasks. Multiple scheduling strategies respect job priority preferences. A graphical user interface allows viewing of job status information and on-the-fly modification of job priorities.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS [0001] This application is a continuation of U.S. patent application Ser. No. 10 / 306,689 (“Adaptive Scheduling”), filed on Nov. 27, 2002 by the inventors herein.FIELD OF THE INVENTION [0002] The present invention relates generally to the field of high-performance computing (“HPC”); more specifically, to systems and techniques for distributed and / or parallel processing. BACKGROUND OF THE INVENTION [0003] HPC has long been a focus of both academic research and commercial development, and the field presents a bewildering array of standards, products, tools, and consortia. Any attempt at comparative analysis is complicated by the fact that many of these interrelate not as mutually exclusive alternatives, but as complementary component or overlapping standards. [0004] Probably the most familiar, and certainly the oldest, approach is based on dedicated supercomputing hardware. The earliest supercomputers included vector-based array processors, whose...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(United States)
IPC IPC(8): G06F15/16G06F9/50
CPCG06F9/505G06F2209/503
Inventor BERNARDIN, JAMESLEE, PETER
Owner BERNARDIN JAMES
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products