Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Virtual machine task management system

a virtual machine and task management technology, applied in the field of network switches, can solve the problems of over-expensive solutions, failure of any given switch, and destruction of the integrity of the entire switching system,

Inactive Publication Date: 2006-02-09
RAPTOR NETWORKS TECH
View PDF11 Cites 217 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

"The present invention provides a system that can switch at both the distributed edge nodes and within a distributed core. The system acts as a single switch through encapsulation of information using a special header that is added by the system upon ingress, and removed by the system upon egress. The routing header includes destination elements and destination port addresses, and can also include destination cluster addresses and source cluster addresses. The ingress and egress elements support connectionless media with stateful connections, and can communicate at high speeds. The switch includes management protocols for discovering which elements are connected, constructing appropriate connection tables, designating a master element, and resolving failures and off-line conditions among the switches. The system can use Strict Ring Topology (SRT) and can be geographically separated by at least one kilometer, and in some cases by over 150 kilometers. The technical effects of the invention include improved efficiency, reliability, and resilience of network switching, as well as improved scalability and flexibility for network deployment and management."

Problems solved by technology

One problem with simple embodiments of the prior art design of FIG. 1 is that failure of any given switch destroys integrity of the entire switching system.
That solution, however, is overly expensive because an entire backup must be deployed for each working system.
The solution is also problematic in that the redundant system must be engaged upon failure of substantially any component within the working system.
But that solution is problematic because all the components are situated locally to one another.
A fire, earthquake or other catastrophe will still terminally disrupt the functionality of the entire system.
Despite numerous potential advantages, Beshai's solution in the '546 patent has significant drawbacks.
A second drawback is that the optical core is contemplated to be entirely passive.
A third, related disadvantage, is that Beshai's concept only supports a single channel from one module to another.
All of those deficiencies reduce functionality.
Thus, even though the '522 application incorporates by reference Beshai's Ser. No. 09 / 244824 application regarding High-Capacity Packet Switch (issued as U.S. Pat. No. 6,721,271 in April 2004), the '522 application still fails to teach, suggest, or motivate one of ordinary skill to provide a fully distributed network (edge and core) that acts as a single switch.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Virtual machine task management system
  • Virtual machine task management system
  • Virtual machine task management system

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0028] In FIG. 2 a switching system 200 generally includes ingress elements 210A-C, egress elements 230A-C, core switching elements 220A-C and connector elements 240A-C. The ingress elements encapsulate incoming packets with a routing header (see FIG. 3), and perform initial switching. The encapsulated packets then enter the core elements for further switching. The intermediate elements facilitate communication between core elements. The egress elements remove the header, and deliver the packets to a sink or final destination.

[0029] Those skilled in the art will appreciate that switching (encapsulation) header must, at a bare minimum, include at least a destination element address. In preferred embodiments the header also includes destination port ID, and where elements are clustered and optional destination cluster ID. Also optional are fields for source cluster, source element, and source port IDs. As used herein an “ID” is something that is the same as, or can be resolved into a...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A switch encapsulates incoming information using a header, and removes the header upon egress. The header is used by both distributed ingress nodes and within a distributed core to facilitate switching. The ingress and egress elements preferably support Ethernet or other protocol providing connectionless media with a stateful connection. Preferred switches include management protocols for discovering which elements are connected, for constructing appropriate connection tables, for designating a master element, and for resolving failures and off-line conditions among the switches. Secure data protocol (SDP), port to port (PTP) protocol, and active / active protection service (AAPS) are all preferably implemented. Systems and methods contemplated herein can advantageously use Strict Ring Topology (SRT), and conf configure the topology automatically. Components of a distributed switching fabric can be geographically separated by at least one kilometer, and in some cases by over 150 kilometers.

Description

[0001] This application claims priority to provisional application number 60 / 511,145 filed Oct. 14, 2003; provisional application number 60 / 511,144 filed Oct. 14, 2003; provisional application number 60 / 511,143 filed Oct. 14, 2003; provisional application number 60 / 511,142 filed Oct. 14, 2003; provisional application number 60 / 511,141 filed Oct. 14, 2003; provisional application number 60 / 511,140 filed Oct. 14, 2003; provisional application number 60 / 511,139 filed Oct. 14, 2003; provisional application number 60 / 511,138 filed Oct. 14, 2003; provisional application number 60 / 511,021 filed Oct. 14, 2003; and provisional application number 60 / 563,262 filed Apr. 16, 2004, all of which are incorporated herein by reference in their entirety.FIELD OF THE INVENTION [0002] The field of the invention is network switches. BACKGROUND [0003] Modem computer networks typically communicate using discrete packets or frames of data according to predefined protocols. There are multiple such standards,...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(United States)
IPC IPC(8): H04L12/50H04Q11/00G06FH04L12/56
CPCH04L49/102H04L49/552H04L49/351H04L49/3009
Inventor PERERA, ANANDAHOFFMAN, EDWIN
Owner RAPTOR NETWORKS TECH
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products