Methods for multi-class cost-sensitive learning

Inactive Publication Date: 2005-12-29
IBM CORP
View PDF2 Cites 36 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

[0009] One of the methods of invention, which is based on the idea 1) above, works by repeatedly sampling from the expanded data set, which is obtained by enhancing each example in the original data set with as many data points as there are possible labels for any single instance. It then repeatedly draws sub-sample from this expanded data set using weighted sampling according to a certain example weighting scheme, in which each labeled example is given the weight specified as the difference between the maximum possible misclassification cost for the instance in question and the misclassification associated with the label in the particular labeled example. The example weighting remains constant throughout the iterative sampling procedure. It then finally outputs a classifier hypothesis which is the average of all the hypotheses output in the respective iterations.
[0010] Another one of the methods of invention, which is based on the idea 2) above, works by iteratively applying weighted sampling from the same expanded data set, using a different weighting scheme. The weighting scheme

Problems solved by technology

The real world is messier.
The third category concerns methods that modify the distribution of training examples before applying the classifier learning method, so that the classifier learned from the modified distribution is cost-sensitive.
Unfortunately, current methods in this category suffer from a major limitation: they are well-understood only for

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Methods for multi-class cost-sensitive learning
  • Methods for multi-class cost-sensitive learning
  • Methods for multi-class cost-sensitive learning

Examples

Experimental program
Comparison scheme
Effect test

Example

[0018] We begin by introducing some general concepts and notation we use in the rest of the description.

Cost-Sensitive Learning and Related Problems

[0019] A popular formulation of the cost-sensitive learning problem is via the use of a cost matrix. A cost matrix, C(y1, y2), specifies how much cost is incurred when misclassifying an example labeled y2 as y1, and the goal of a cost-sensitive learning method is to minimize the expected cost. Zadrozny and Elkan (B. Zadrozny and C. Elkan, “Learning and making decisions when costs and probabilities are both unknown”, Proceedings of the seventh International Conference on Knowledge Discovery and Data Mining, pp. 204-213, ACM Press, 2001) noted that this formulation is not applicable in situations in which misclassification costs depend on particular instances, and proposed a more general form of cost function, C(x, y1 , y2), that allows dependence on the instance x. Here we adopt this general formulation, but note that in the reasonable ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

No PUM Login to view more

Abstract

Methods for multi-class cost-sensitive learning are based on iterative example weighting schemes and solve multi-class cost-sensitive learning problems using a binary classification algorithm. One of the methods works by iteratively applying weighted sampling from an expanded data set, which is obtained by enhancing each example in the original data set with as many data points as there are possible labels for any single instance, using a weighting scheme which gives each labeled example the weight specified as the difference between the average cost on that instance by the averaged hypotheses from the iterations so far and the misclassification cost associated with the label in the labeled example in question. It then calls the component classification algorithm on a modified binary classification problem in which each example is itself already a labeled pair, and its (meta) label is 1 or 0 depending on whether the example weight in the above weighting scheme is positive or negative, respectively. It then finally outputs a classifier hypothesis which is the average of all the hypotheses output in the respective iterations.

Description

BACKGROUND OF THE INVENTION [0001] 1. Field of the Invention [0002] The present invention generally relates to the field of cost-sensitive learning in the areas of machine learning and data mining and, more particularly, to methods for solving multi-class cost-sensitive learning problems using a binary classification algorithm. This algorithm is based on techniques of data space expansion and gradient boosting with stochastic ensembles. [0003] 2. Background Description [0004] Classification in the presence of varying costs associated with different types of misclassification is important for practical applications, including many data mining applications, such as targeted marketing, fraud and intrusion detection, among others. Classification is often idealized as a problem where every example is equally important, and the cost of misclassification is always the same. The real world is messier. Typically, some examples are much more important than others, and the cost of misclassifyi...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Application Information

Patent Timeline
no application Login to view more
IPC IPC(8): G06N20/00G06F9/44G06K9/62G06N3/08G06N7/02G06N7/06
CPCG06N99/005G06K9/6256G06N20/00G06F18/214Y10S706/932
Inventor ABE, NAOKIZADROZNY, BIANCA
Owner IBM CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products