Cache Management Method for Distributed Memory Columnar Database

A cache management, in-memory columnar technology, applied in the field of cache management of distributed in-memory columnar databases, achieves the effect of improving query efficiency, saving query time and storage space, and reducing the calculation of repetitive tasks

Active Publication Date: 2019-03-19
UNIV OF ELECTRONICS SCI & TECH OF CHINA
View PDF2 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0004] The query requests in the database are highly semantically related, so some query results are likely to have appeared in historical queries

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Cache Management Method for Distributed Memory Columnar Database
  • Cache Management Method for Distributed Memory Columnar Database
  • Cache Management Method for Distributed Memory Columnar Database

Examples

Experimental program
Comparison scheme
Effect test

Embodiment

[0030] This embodiment provides a cache management method for a distributed in-memory columnar database. For the structural diagram of the cache management system for a distributed in-memory columnar database, please refer to figure 1 , including a query execution engine, a cache master node, a standby node, and at least one cache slave node.

[0031] When a query request comes, the query execution engine parses the SQL statement into a physical execution plan represented by DAG. Each node in the physical execution plan represents a physical task, and the physical tasks are divided into GetColumn, Join, Filter, Group, BuildRow, etc. Each edge represents the transmission relationship of calculation results between two physical tasks. The physical execution plan of a typical query statement (SELECT A.id FROM A,B WHERE A.id=B.id AND A.idfigure 1 shown. In a cache management system, the granularity of cached data is the calculation result of a single physical task. When the cach...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a cache management method of a distributed memory columnar database, comprising: establishing a cache queue in a cache master control node; taking each physical task as a root node to cut its physical execution plan to obtain each physical task Corresponding cache calculation track; according to the cache calculation track corresponding to each physical task, a cache feature tree is built in the cache master node; when a query request arrives, the query execution engine parses the SQL statement into a physical execution plan; from the physical execution plan The root node starts to traverse each node in the physical execution plan hierarchically, and judges whether the cache calculation trajectory corresponding to each physical task matches the cache feature tree; if it matches, directly reads the cached actual data of the physical task from the cache slave node, otherwise Compute the physics task. The cache management method of the distributed memory column database provided by the present invention quickly detects whether the cache is hit through an efficient cache matching algorithm, thereby improving query efficiency.

Description

technical field [0001] The invention relates to the technical field of computer software, in particular to a cache management method of a distributed memory column database. Background technique [0002] With the development of the information age, the scale of data has shown explosive growth. How to extract valuable information from these massive data is a huge challenge facing the current society. The On-Line Analytical Processing (OLAP, On-Line Analytical Processing) system has demonstrated its powerful data analysis capabilities, and it has been widely used in commercial fields such as banking, telecommunications, and stock exchanges. [0003] The distributed in-memory columnar database that supports OLAP systems allows users to extract and analyze valuable information from massive data in multiple dimensions. This information may be a simple report or a complex analysis result. As the complexity of query statements increases, the time required for query operations will...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Patents(China)
IPC IPC(8): G06F16/2455
CPCG06F16/24552
Inventor 段翰聪闵革勇张建郑松詹文翰
Owner UNIV OF ELECTRONICS SCI & TECH OF CHINA
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products