CART-based decision-making tree construction method in cognitive computation

A construction method and decision tree technology, applied in the field of big data processing, can solve problems such as not meeting the needs of data processing, calculating optimal features, etc., to achieve the effect of reducing the possibility and improving the accuracy

Inactive Publication Date: 2016-08-10
NANJING UNIV OF POSTS & TELECOMM
View PDF4 Cites 10 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

However, since it is impossible to use infinite data sets to calculate the selection of optimal features, only known data can be used to form sample sets to train decision trees, which is far from meeting the needs of current data processing.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • CART-based decision-making tree construction method in cognitive computation
  • CART-based decision-making tree construction method in cognitive computation
  • CART-based decision-making tree construction method in cognitive computation

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0019] In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention.

[0020] CART first creates a root node N 0 , during the learning process, for each specified node N a There is a subset S of the specified training set S a Corresponding. For the root node, it corresponds to the training set S. When the subsets corresponding to a node belong to the same category, the node is set as a leaf node, which means that the training of the node has been completed. If the subsets corresponding to a node do not all belong to the same classification, the algorithm will continue to iterate until all the training subsets correspond to the corresponding leaf nodes.

[0021] F...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

The invention discloses a CART-based decision-making tree construction method in cognitive computation. The method comprises the following steps: initializing a root node of a decision-making tree and a characteristic attribute set corresponding to the root node according to establishment rules of the decision-making tree; carrying out sorting of data in a training set; judging whether all sample data in a node is of the same type or not; calculating the optimum characteristic attributes and the suboptimum characteristic attributes of the node to be split; judging whether the node satisfies the selected optimum splitting attributes and splitting conditions of an interruption mechanism or not; if the node satisfies the selected optimum splitting attributes and the splitting conditions of the interruption mechanism, carrying out splitting through the optimum splitting attributes, then iteratively replacing a current node with the node split through the characteristic attributes, and adding two new leaf nodes respectively from a left branch and a right branch, so as to achieve automatic splitting of the decision-making tree; and if the node does not satisfy the selected optimum splitting attributes or the splitting conditions of the interruption mechanism, waiting for data stream inputs, carrying out sample updating, and continuing to calculate node splitting. The method provided by the invention has the advantages that the accuracy of data stream processing is further improved, so that the possibility of system blocking is reduced.

Description

technical field [0001] The invention relates to the technical field of big data processing, in particular to a CART-based decision tree construction method in cognitive computing. Background technique [0002] The rapid development of emerging technologies such as cloud computing and the Internet of Things has caused the scale of data to grow at an unprecedented rate, and the era of big data has begun. Decision tree is a commonly used data processing model in data mining. Common decision tree construction algorithms include ID3, C4.5, and CART. However, decision tree construction algorithms such as ID3, C4.5, and CART are all for static data sets. By design, they cannot be directly applied to the processing of data streams, which are endless. In addition, the data flow will continuously flow into the system at a very fast speed, which also brings great challenges to the training of decision trees. Existing research also has some solutions to the processing of data streams....

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
Patent Type & Authority Applications(China)
IPC IPC(8): G06F17/30
CPCG06F16/285G06F16/2246G06F16/24568
Inventor 王堃陆恒张明翔岳东孙雁飞吴蒙亓晋陈思光
Owner NANJING UNIV OF POSTS & TELECOMM
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products