Paper - Review

10.1038/nmeth.3945

DOI: 10.1038/nmeth.3945

❓ how to evaluate classifier performance?
→ using 1⃣ a numeric metric 2⃣ a graphical representation of performance

∴ Examine some common classifier metrics & Discuss the pitfalls of relying on a single metrics

❗ Metrics
→ how a classifier performs
→ critical for evaluating reports

Simulate → a hypothetical diagnostic test
→ on the basis of multiple clinical factors

❗ Confusion matrix
Relevance (← of each of these 4 quantities)
← depend on 1⃣ the purpose of the classifier 2⃣ motivate the choice of metric

High accuracy
→ does NOT ❌ necessarily characterize → a good classifier

Predicting will be negative offers high accuracy
← if the disease is rare
→ is NOT ❌ useful for diagnosis

NO ❌ single metric
→ can distinguish → all (the strengths & weaknesses) ← of a classifier

Important factor
← in interpreting classification results
→ is (class balance)

Imbalance makes
→ understanding (FPs & FNs) → more important
→ Extra care is required → to interpret the results

To evaluate the classifier
→ without having to select ← a specific threshold
∴ Can be visualized → using the ROC curve

ROC curve
→ A good classifier should aim → to reach as close to the top-left corner as possible

Class imbalance can cause ROC curves
→ to be poor visualizations of classifier performance