Tcp/ip processor and engine using rdma

a processor and engine technology, applied in the field of storage networking semiconductors, can solve the problems of lower line rate, lower performance host processors, and prohibitive software stack overhead, and achieve the effects of low cost, low cost, and low cos

Inactive Publication Date: 2008-10-16
MEMORY ACCESS TECH LLC
View PDF6 Cites 99 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

[0006]I describe a high performance hardware processor that sharply reduces the TCP / IP protocol stack overhead from a host processor and enables a high line rate storage and data transport solution based on IP.
[0007]Traditionally, TCP / IP networking stack is implemented inside the operating system kernel as a software stack. The software TCP / IP stack implementation consumes, as mentioned above, more than 50% of the processing cycles available in a 1 GHz processor when serving a 1 Gbps network. The overhead comes from various aspects of the software TCP / IP stack including checksum calculation, memory buffer copy, processor interrupts on packet arrival, session establishment, session tear down and other reliable transport services. The software stack overhead becomes prohibitive at higher lines rates. Similar issues occur in networks with lower line rates, like wireless networks, that use lower performance host processors. A hardware implementation can remove the overhead from the host processor.

Problems solved by technology

The overhead comes from various aspects of the software TCP / IP stack including checksum calculation, memory buffer copy, processor interrupts on packet arrival, session establishment, session tear down and other reliable transport services.
The software stack overhead becomes prohibitive at higher lines rates.
Similar issues occur in networks with lower line rates, like wireless networks, that use lower performance host processors.
This is an expensive compute intensive operation performed on each packet involving each received byte in the packet.
When the line rates increase to above 1 Gbps, the memory size overhead and memory speed bottleneck resulting from these add significant cost to the network cards and also cause huge performance overhead.
Another function that consumes a lot of processor resources is the copying of the data to / from the network card buffers, kernel buffers and the application buffers.
Interrupting these processors on arrival of small packets will cause severe performance degradation due to context switching overhead, pipeline flushes and refilling of the pipelines.
When the block storage traffic is transported over TCP / IP networks, these performance issues become critical, severely impacting the throughput and the latency of the storage traffic.
However, there usually isn't alignment relationship between the TCP segments and the protocol data units that are encapsulated by TCP packets.
This becomes an issue when the packets arrive out of order, which is a very frequent event in today's networks.
This can be expensive from the size of the memory storage required and also the performance that the memory subsystem is expected to support, particularly at line rates above 1 Gbps.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Tcp/ip processor and engine using rdma
  • Tcp/ip processor and engine using rdma
  • Tcp/ip processor and engine using rdma

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0067]I provide a new high performance and low latency way of implementing a TCP / IP stack in hardware to relieve the host processor of the severe performance impact of a software TCP / IP stack. This hardware TCP / IP stack is then interfaced with additional processing elements to enable high performance and low latency IP based storage applications. This can be implemented in a variety of forms to provide benefits of TCP / IP termination, high performance and low latency IP storage capabilities, remote DMA (RDMA) capabilities, security capabilities, programmable classification and policy processing features and the like. Following are some of the embodiments that can implement this:

[0068]Server

[0069]The described architecture may be embodied in a high performance server environment providing hardware based TCP / IP functions that relieve the host server processor or processors of TCP / IP software and performance overhead. The IP processor may be a companion processor to a server chipset, pr...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

A TCP / IP processor and data processing engines for use in the TCP / IP processor is disclosed. The TCP / IP processor can transport data payloads of Internet Protocol (IP) data packets using an architecture that provides capabilities to transport and process Internet Protocol (IP) packets from Layer 2 through transport protocol layer and may also provide packet inspection through Layer 7. The engines may perform pass-through packet classification, policy processing and / or security processing enabling packet streaming through the architecture at nearly the full line rate. A scheduler schedules packets to packet processors for processing. An internal memory or local session database cache stores a TCP / IP session information database and may also store a storage information session database for a certain number of active sessions. The session information that is not in the internal memory is stored and retrieved to / from an additional memory. An application running on an initiator or target can in certain instantiations register a region of memory, which is made available to its peer(s) for access directly without substantial host intervention through RDMA data transfer.

Description

CROSS REFERENCE TO RELATED APPLICATIONS[0001]This application is a divisional of U.S. patent application Ser. No. 10 / 459,349 filed on Jun. 10, 2003 titled TCP / IP Processor and Engine Using RDMA and allowed, which claims the benefit of U.S. Provisional Patent Application Ser. No. 60 / 388,407, filed on Jun. 11, 2002 titled A High Performance IP Storage Processor and now expired, both of which are hereby incorporated by reference in their entirety.[0002]This application is related to the following non-provisional applications, each filed on Jun. 10, 2003 claiming the benefit of U.S. Provisional Patent Application Ser. No. 60 / 388,407, and each of which is hereby incorporated by reference for all purposes: 10 / 458,844 titled Data Processing System Using Internet Protocols, which is Pending, 10 / 458,855 titled Data Processing System Using Internet Protocols and RDMA, which is Pending, 10 / 459,019 titled Memory System for a High Performance IP Processor, which is Pending, 10 / 459,297 titled Hig...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(United States)
IPC IPC(8): H04J3/16G06F15/173G06F15/177H04L29/06H04L29/08
CPCH04L29/06H04L67/1097H04L67/34H04L69/16H04L69/166H04L69/18H04L69/161H04L69/163H04L69/10H04L69/12H04L69/165H04L69/32H04L69/329H04L69/326H04L69/323H04L9/40
Inventor PANDYA, ASHISH A.
Owner MEMORY ACCESS TECH LLC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products