Approaches to classification
Decision trees
bayesian classification
classification rules
neural networks
k-nearest nieghbours
SVM
Classification Requirements
Definition of classification
Given a collection of labels and one of data objects labelled, find a descripitve model that will enable to distinguish an object of one class from another.
Criteria to evaluate classification techniques
Decision Trees, Hunt’s Algo
Dt -> set of training records that reach a node t.
Steps:

Decision tree induction
Adopts a greedy strategy -> not global optimum
Issues:
Selection of best attribute in decision tree
Attributes with homogeneous class distribution are preferred, degree of impurity must be low.
Node impurity needs a measure:
Different algortithms rely on different measures of impurity.
GINI index

Computing quality of a split using GINI
This technique is used in CART, SLIQ, SPRINT
When a node p is split into k partitions (children), the quality of the split is.
Categorical attribute:
Continuos Attribute:

Entropy impurity measure (INFO)
INFO
Information gain:
Gain Ratio:

Classification error impurity measure
Meausures classification error made by current node.
Same maximum and mimum of GINI index.

Stopping criteria for tree induction
Stop expanding node when:
Overfitting vs Underfitting,
Noise overfitting
Underfitting: model is too simple, training and test set errors are high.

Addressing overfitting
Pre-pruning (early stopping):
Post-pruning:
Issues of decisions tree
Tree decision boundary and Oblique decision trees
Border line between two neighboring regions of different classes is known as decision boundary
Decision boundary is parallel ot axes because test condition involves a single attribute at a time.
In oblique decision trees, test condition may involve multiple attributes, more expressive representation.
Finding optimal test condition is computationally expensive.

Pros and cons of decision tree based classification
Pros:
Cons:
Rule Based classifier
Classify records by using a collection of if … then …
Rule: (Condition) → y
A rule r covers an instance x if the attributes of the instance satisfy the condition of the rule.
Example: R1: (Give birth = no) ⋂ (Can fly = yes) → Birds
Characteristics of rules:
Advantages of rule-based classifier:
Rules semplification: problems and solutions
From:
(Refund=No) ⋂ (Status=Married) → No
To:
Simplified Rule: (Status=Married) → No
Problems:
Ordered rule set
Rules are rank ordered according to their priority.
When a test record is presentd to the classifier it is assigned to the class lable of the highest ranked rule it has triggered. If no rules were triggered, it is assigned to the default class.

Methods to build classification rules
Direct method:
Indirect Method:
Associative classification
The classification models defined by means of associaltion rule: (Condition) → y
rule body is an itemset.
Model generation:
Pros and cons of associative classification
Strong points:
Weak points: