Check patentability & draft patents in minutes with Patsnap Eureka AI!

A Reference Point K-Nearest Neighbor Classification Method Based on MPI Parallelization

A classification method and k-nearest neighbor technology, applied in the direction of multi-program device, inter-program communication, program control design, etc., can solve the problems of high time complexity, increased time consumption, slow classification speed, etc., and achieve high classification accuracy Effect

Active Publication Date: 2021-07-20
CHONGQING UNIV OF POSTS & TELECOMM
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

The disadvantage of this type of algorithm is that as the dimension of the data set increases, the performance of these tree-structured algorithms gradually deteriorates. The reason is that high-dimensional data sets will increase the complexity of the tree structure, which will lead to the establishment of tree structures and search for neighbors. Points and increased time taken to calculate distances
The kNN algorithm is a commonly used data mining algorithm, but its time complexity is high and the classification speed is slow.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A Reference Point K-Nearest Neighbor Classification Method Based on MPI Parallelization
  • A Reference Point K-Nearest Neighbor Classification Method Based on MPI Parallelization
  • A Reference Point K-Nearest Neighbor Classification Method Based on MPI Parallelization

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0034] The preferred embodiments of the present invention will be described in detail below with reference to the accompanying drawings.

[0035] The k-nearest neighbor algorithm based on reference points uses the distance between sample points and several reference points to measure the position difference between each other. The idea is to set several reference points, calculate the similarity between training samples and reference points and generate an ordered similarity sequence, Then according to the similarity between the test sample and the reference point, the approximate neighbor samples in the training set are searched from the ordered sequence, and the exact similarity with the test sample is calculated from these approximate neighbor samples, so as to find k nearest neighbor samples and judge the category. Its core strategy is to greatly reduce the scope of searching for training set samples with the help of reference points.

[0036] Define the Location Differenc...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention relates to a reference point k-nearest neighbor classification method based on MPI parallelization, and belongs to the field of data classification. The method includes steps: S1: The k-nearest neighbor algorithm based on the reference point uses the distance between the sample point and several reference points to measure the position difference between each other, defines the position difference factor, calculates the similarity between the training sample and the reference point, and generates an ordered S2: Search the approximate neighbor samples in the training set from the ordered sequence according to the similarity between the test sample and the reference point; S3: Calculate the exact similarity with the test sample from the searched approximate neighbor samples, so as to find k neighbor samples and judge the category. The present invention accelerates the search of k-nearest neighbors by means of reference points and uses MPI technology to implement parallelization, thereby accelerating the classification speed of large-scale and high-dimensional data.

Description

technical field [0001] The invention belongs to the field of data classification and relates to a reference point k nearest neighbor classification method based on MPI parallelization. Background technique [0002] Classification is an important technology in the field of data mining. Its purpose is to construct a classification model (also known as classification function, classifier) ​​according to the characteristics of the data set, which can map samples of unknown categories to certain categories in a given category. one or several. The k-nearest neighbor algorithm was originally proposed by Cover and Hart in 1968. It is a non-parametric classification technique that has the advantages of robustness, clear concept, and easy implementation, and can achieve high classification accuracy for unknown and non-normal distributions. Rate. [0003] The traditional k-nearest neighbor algorithm has a high time complexity. At present, many scholars have proposed many improved alg...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): G06F9/54G06K9/00G06K9/62
CPCG06F9/546G06V10/94G06F18/24147
Inventor 陈子忠梁聪夏书银
Owner CHONGQING UNIV OF POSTS & TELECOMM
Features
  • R&D
  • Intellectual Property
  • Life Sciences
  • Materials
  • Tech Scout
Why Patsnap Eureka
  • Unparalleled Data Quality
  • Higher Quality Content
  • 60% Fewer Hallucinations
Social media
Patsnap Eureka Blog
Learn More