- Accuracy (Essential for classification), Weighted Accuracy = Weighted Kappa
- Logarithmic Loss (Depends on soft predictions probabilities)
- Area under Receiver Operating Curve (Considers ordering of objects, tries all threshold to convert soft predictions to hard labels)
- Kappa (Similar to R Squared)
N - Number of objects
L - Number of classes
y - Ground truth
yi - Predictions
[a = b] - indicator function
- Soft labels (soft predictions) are classifier's scores - Probabilities of objects
- Hard Labels (hard predictions) - argmax fi(x), [f(x)>b], b - threshold for binary classification, Predict label, maximum value from soft prediction and set class for prediction label. Function of soft label
- Most referred measure of classifier quality
- Higher is better
- Need hard predictions
- Number of correctly guessed objects
- Argmax of soft predictions
- Work with soft predictions
- Make classifier output posterior probabilities
- Penalises for wrong answers
- Set constant to frequencies of each class
- Based on threshold decide percentage of above / below the threshold
- Metric tries all possible ones and aggregate scores
- Depends on order of objects
- Compute TruePositive, FalsePositive
- AUC max value 1
- Fraction of correctly ordered pairs
= 1 - (Fraction of incorrectly ordered pairs / total number of pairs)
Cohen's Kappa
- Score = 1- ((1-accuracy)/(1-baseline))
- Baselines different for each data
- Similar to R squared
- Here R predictions for dataset used as baseline
- Error = (1- Accuracy)
- Weighted Error Score = Confusion matrix * Weight matrix and sum their results
- Weighted Kappa = 1 - ((weighted error)/(weighted baseline error))
- Useful for medical applications
No comments:
Post a Comment