Method and device using vector instruction to process file index in parallel mode

A file indexing and parallel processing technology, applied in electrical digital data processing, special data processing applications, instruments, etc., can solve the problem that the binary search method is not suitable for parallel processing, cannot use parallel computing power to improve efficiency, etc., and reaches the peak of improvement. Computational power, exploiting data parallelism, reducing performance penalty

Inactive Publication Date: 2015-02-04
INST OF AUTOMATION CHINESE ACAD OF SCI
View PDF4 Cites 30 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, the binary search method is not suitable for parallel processing, and cannot use the parallel computing capabilities of modern superscalar processors to improve efficiency

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Method and device using vector instruction to process file index in parallel mode
  • Method and device using vector instruction to process file index in parallel mode
  • Method and device using vector instruction to process file index in parallel mode

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0033] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with specific embodiments and with reference to the accompanying drawings.

[0034] The following symbols are used to describe the process:

[0035] N: Number of index items to be processed B: Data width of index items R: Data width of vector registers n: B+ tree leaf node capacity α: Filling factor of leaf nodes (the ratio of index data to leaf node capacity when first constructing leaf nodes )

[0036] g: degree of B+ tree

[0037] h: the height of the B+ tree

[0038] In the following specific implementation manners, B+ tree is taken as an example for illustration, and the method described in the present invention can also be used for other tree indexes commonly used in file indexes.

[0039] figure 1 It is a structural block diagram of the file index data parallel processing system propose...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a method and a device using a vector instruction to process a file index in a parallel mode. The method comprises the following steps of calculating and sorting index key values of data in a data table, distributing needed storing space of all leaf nodes, using an SIMD (single instruction multiple data) instruction to load a plurality of index key values once, sequentially filling the data in all the leaf nodes, distributing the needed storing space of the internal nodes of the upper layer, building the internal nodes, continuously building the internal nodes of the upper layer until the node number of the upper layer is one, regarding the node as a root node, and finishing the building of an index tree; building a buffer pool in internal storage, using an inquiring queue to temporarily store all received inquiring requests, taking the inquiring requests out of the buffering pool when the length of the inquiring queue reaches a preset threshold value, sorting the inquiring requests, and forming a sorted to-be-inquired queue; sequentially obtaining a to-be-inquired term in the sorted inquiring queues, sequentially searching from the root node of the built index tree, and using the SIMD method to search the to-be-inquired term in a parallel mode inside the internal nodes and the leaf nodes.

Description

technical field [0001] The invention belongs to the field of computer application technology, and specifically relates to a method for parallel processing of file indexes by using vector instructions, which is closely related to file system structure and processor instruction system. Background technique [0002] In the era of cloud computing, information is growing explosively. In the face of massive data, storage systems need to meet the requirements of high performance, large capacity, high reliability, and high scalability. The distributed file storage system is an effective way to obtain large-capacity storage. The storage capacity is arranged on a large number of distributed computer nodes, and multiple copies of each data are stored. The data access speed is improved through concurrent access, and the data redundancy is improved. Data reliability. [0003] The current distributed file system mainly manages by storing file data and its index data (metadata) separately...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30
CPCG06F16/24532G06F16/2246G06F16/2282
Inventor 王东琳陈易龙蒿杰林建海舒琳宋亚芳穆敬彬梁小龙
Owner INST OF AUTOMATION CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products