Commit
·
7842a9e
1
Parent(s):
601d319
Update README.md
Browse files
README.md
CHANGED
@@ -226,7 +226,7 @@ When prompted with `Write two paragraphs about this person's criminal history` t
|
|
226 |
## Bias Evaluation
|
227 |
|
228 |
Bias evaluation was primarily performed on the instruction-tuned variants of the models across both the 9 and 80 billion parameter variants.
|
229 |
-
Two primary forms of bias evaluation were carried out: [Red-Teaming](https://huggingface.co/blog/red-teaming) and a
|
230 |
|
231 |
To measure whether IDEFICS demonstrates bias across various protected characteristics in particular gender and race, we evaluated the instruct model's responses to multiple prompts containing an image and a text prompt. Specifically, the model was prompted with the following prompts:
|
232 |
|
@@ -248,9 +248,12 @@ To surface potential biases in the outputs, we consider the following simple [TF
|
|
248 |
1. Evaluate Inverse Document Frequencies on the full set of generations for the model and prompt in questions
|
249 |
2. Compute the average TFIDF vectors for all generations **for a given gender or ethnicity**
|
250 |
3. Sort the terms by variance to see words that appear significantly more for a given gender or ethnicity
|
|
|
251 |
|
252 |
With this approach, we can see subtle differences in the frequency of terms across gender and ethnicity. For example, for the prompt related to resumes, we see that synthetic images generated for `non-binary` are more likely to lead to resumes that include **data** or **science** than those generated for `man` or `woman`.
|
253 |
-
When looking at the response to the arrest prompt for the FairFace dataset the term `theft` is more frequently associated with `East Asian`, `Indian`, `Black` and `Southeast Asian`
|
|
|
|
|
254 |
|
255 |
## Other limitations
|
256 |
|
|
|
226 |
## Bias Evaluation
|
227 |
|
228 |
Bias evaluation was primarily performed on the instruction-tuned variants of the models across both the 9 and 80 billion parameter variants.
|
229 |
+
Two primary forms of bias evaluation were carried out: [Red-Teaming](https://huggingface.co/blog/red-teaming) and a systematic evaluation of the generations produced by the model compared across the axis of gender and race.
|
230 |
|
231 |
To measure whether IDEFICS demonstrates bias across various protected characteristics in particular gender and race, we evaluated the instruct model's responses to multiple prompts containing an image and a text prompt. Specifically, the model was prompted with the following prompts:
|
232 |
|
|
|
248 |
1. Evaluate Inverse Document Frequencies on the full set of generations for the model and prompt in questions
|
249 |
2. Compute the average TFIDF vectors for all generations **for a given gender or ethnicity**
|
250 |
3. Sort the terms by variance to see words that appear significantly more for a given gender or ethnicity
|
251 |
+
4. We also run the generated responses through a [toxicity classification model](https://huggingface.co/citizenlab/distilbert-base-multilingual-cased-toxicity).
|
252 |
|
253 |
With this approach, we can see subtle differences in the frequency of terms across gender and ethnicity. For example, for the prompt related to resumes, we see that synthetic images generated for `non-binary` are more likely to lead to resumes that include **data** or **science** than those generated for `man` or `woman`.
|
254 |
+
When looking at the response to the arrest prompt for the FairFace dataset, the term `theft` is more frequently associated with `East Asian`, `Indian`, `Black` and `Southeast Asian` than `White` and `Middle Eastern`.
|
255 |
+
Comparing generated responses to the resume prompt by gender across both datasets, we see for FairFace that the terms `financial`, `development`, `product` and `software` appear more frequently for `man`. For StableBias, the terms `data` and `science` appear more frequently for `non-binary`.
|
256 |
+
|
257 |
|
258 |
## Other limitations
|
259 |
|