05 - More on evaluations
Overview
- Description:: continuing last lession on evaluations
Most common measures for verification
- 
FAR: False Acceptance Rate 
- 
FRR: False Rejection Rate 
- 
EER: Equal Error Rate 
- 
DET: Detection Error trade-off 
- 
ROC: Receiving Operating Curve 
- 
all such measures depend on the adopted acceptance threshold 
- 
topMatch(p_j, identity returns the best match between pj and the templates associated to the claimed identity in the gallery s(t1, t2) returns the similarity between t1 and t2 - it can be more than a result
 
Scoring
- 
a score is said genuine (authentic) if it results from matching two samples of the biometric trait of a same enrolled individual; 
- 
it is said impostor if it results from matching the sample of a non-enrolled individual. 
Acceptance threshold is crucial and depends on our application needs!
A too low threshold causes many type 1: FRR (rejected genuine)
A too high threshold causes many type 2: FAR (accepted impostors)
Therefore, a common good choice is FRR (ratio) = ERR: FAR

Open set
An open set problem is comparing someone against templates, without a claim on behalf of the person.
For example: terrorism checks at the airport
- 
rank(pj) = the position in the list where the first template for the correct identity is returned 
- 
DIR (at rank k) (Detection and Identication Rate): probability of correct identification at rank k (the correct subject is returned at position k) 
- 
FAR or more specifically FPIR (False Acceptance Rate or False Positive Identification Rate) or False Alarm Rate (Watch List): the probability of false acceptance/alarm 
- 
EER (Equal Error Rate): the point where the two probability errors are equal