Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Content-based information retrieval architecture

a content-based information and architecture technology, applied in the field of architectures for fast information retrieval, can solve problems such as complicated reconstruction of tables, and achieve the effects of convenient manufacture, convenient hardware implementation, and convenient construction

Inactive Publication Date: 2005-08-11
NEC CORP
View PDF6 Cites 32 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

[0007] In accordance with another embodiment of another aspect of the invention, the tables can be segmented into a plurality of banks, each bank associated with one of the plurality of hashed values generated by the hashing module. This advantageously improves the access speed and obviates any need for multi-ported memories. In addition, the different parts of the information retrieval architecture, e.g., the hashing module and the table of encoded values, can be pipelined into different stages, thereby allowing implementation using conventional random access memory chips. The architecture can use stream-based data flow and can achieve very high throughputs via deep pipelining.
[0008] In accordance with another embodiment of the invention, the table of lookup values and the table of associated input values can be made smaller than the table of encoded values, so that the width of the encoded table is at least log(n), where n is the number of lookup values. This advantageously reduces the memory consumed by the tables. The table of encoded values preferably should be constructed using sequential address generation for the table of lookup values.
[0009] In accordance with another embodiment of the invention, a filtering module can be provided which performs pre-filtering on input values before passing an input value to the hashing module. The filtering module forwards an input value to the hashing module only if the input value is not a member of some filtered set of input values. For example, and without limitation, the filtered set of input values can be those input values that are recognized as not being members of the lookup set. This can result in significant power savings, since the tables in the information retrieval architecture are accessed only if the input value is part of the lookup set. Values can be added to the filtered set of input values, for example, when input values are recognized through use of the third table as not being part of the lookup set. The filtering module can also be used for other purposes. For example, the filtering module can be configured so as to remap certain input values into other more advantageous values that are forwarded to the hashing module. The filtered set of input values can be selected so as to facilitate construction of the first table of encoded values. For example, where an input value ends up generating a plurality of hashed values that correspond to not one singleton location in the first table of encoded values, this can complicate reconstruction of the table. It can be advantageous to add this value to the filtered set of input values and handle that value separately in a spillover table.
[0011] One advantageous application of the present architecture is in network routers. A network router can be readily constructed using such an architecture to implement a plurality of filters which can perform longest prefix matching on a packet header, where the input value is a prefix of a pre-specified length and the lookup value is the forwarding information. An advantageous design is to utilize a plurality of filters, one for each prefix length operating in parallel. Where several of the filters signal a match, a priority encoder can be used to select the forwarding information from the filter with the longest prefix match. The filters can be implemented in pairs, where an update filter can be updated off-line and swapped for the filter that is performing lookups.
[0012] The content-based information retrieval architecture disclosed herein can be readily implemented in hardware or software or an advantageous combination of both. An implementation advantageously can be embodied in a single chip solution, with embedded memory, or a multi chip solution, with external SRAM / DRAM. The above-described architecture has a number of key advantages over related technologies. As discussed above, the design can use standard inexpensive memory components such as SRAM or DRAM, thereby facilitating ease of manufacture. The design is capable of high speeds, as it uses streaming data with no hardware dependencies and may be deeply pipelined to obtain high throughput. Additionally, the design has the potential to consume significantly less power than equivalent TCAM components. These and other advantages of the invention will be apparent to those of ordinary skill in the art by reference to the following detailed description and the accompanying drawings.

Problems solved by technology

For example, where an input value ends up generating a plurality of hashed values that correspond to not one singleton location in the first table of encoded values, this can complicate reconstruction of the table.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Content-based information retrieval architecture
  • Content-based information retrieval architecture
  • Content-based information retrieval architecture

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0025]FIG. 1 is a diagram of a content-based information retrieval architecture, in accordance with an embodiment of the invention. As depicted in FIG. 1, an input value 101 is received and processed by the architecture. Based on the processing of the input value 101, what is output is either a lookup value 102 retrieved from a table 120 or a signal 103, 104 that the input value 101 is not a member of the lookup set.

[0026] In FIG. 1, there are three tables 110, 120, 170 the construction and operation of which are described in further detail herein. Each table in FIG. 1 has the same number of entries, although, as further described herein and in accordance with another embodiment of the invention, the tables can have different numbers of entries. Table 120 is referred to herein as a lookup table and comprises a plurality of entries, each storing a lookup value. For example, in the case of a router application, the lookup values can be next hop information associated with different d...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A content-based information retrieval architecture is herein disclosed that can achieve correct and predictable high speed lookups while taking advantage of inexpensive conventional memory components.

Description

[0001] This Utility patent application is a Non-Provisional of and claims the benefit of U.S. Provisional Patent Application Ser. No. 60 / 541,983 entitled “INEXPENSIVE AND FAST CONTENT ADDRESSABLE MEMORY” filed on Feb. 5, 2004, the contents of which are incorporated by reference herein.BACKGROUND OF THE INVENTION [0002] The present invention relates to architectures for fast information retrieval. [0003] Content addressable memories (CAMs) are devices that are queried using data in contrast to conventional memories that are queried using addresses. CAMs search through their contents to check if the data exists, and if so, provide information associated with the data. Such memories find applications in Internet routers, processor caches, databases, and other situations where fast lookups are required. Most commercial routers employ CAM devices called Ternary-CAMs (TCAMs) which are specialized memory devices that check if a query matches every entry in parallel. TCAMs are more expensiv...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(United States)
IPC IPC(8): G06F17/30H03M7/00H04L12/56
CPCG06F17/30949H04L45/745H04L45/00G06F16/9014H04L45/7452
Inventor CADAMBI, SRIHARIKILIAN, JOSEPHASHAR, PRANAVCHAKRADHAR, SRIMAT T.
Owner NEC CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products