Readme update
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ pinned: false
|
|
16 |
The traditional evaluation of NLP labeled spans with precision, recall, and F1-score leads to double penalties for
|
17 |
close-to-correct annotations. As [Manning (2006)](https://nlpers.blogspot.com/2006/08/doing-named-entity-recognition-dont.html)
|
18 |
argues in an article about named entity recognition, this can lead to undesirable effects when systems are optimized for these traditional metrics.
|
19 |
-
|
20 |
|
21 |
## How to Use
|
22 |
FairEval outputs the error count (TP, FP, etc.) and resulting scores (Precision, Recall and F1) from a reference list of
|
|
|
16 |
The traditional evaluation of NLP labeled spans with precision, recall, and F1-score leads to double penalties for
|
17 |
close-to-correct annotations. As [Manning (2006)](https://nlpers.blogspot.com/2006/08/doing-named-entity-recognition-dont.html)
|
18 |
argues in an article about named entity recognition, this can lead to undesirable effects when systems are optimized for these traditional metrics.
|
19 |
+
To address these issues, this metric provides an implementation of FairEval, proposed by [Ortmann (2022)](https://aclanthology.org/2022.lrec-1.150.pdf).
|
20 |
|
21 |
## How to Use
|
22 |
FairEval outputs the error count (TP, FP, etc.) and resulting scores (Precision, Recall and F1) from a reference list of
|