Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
update about
Browse files- README.md +1 -1
- src/display/about.py +1 -4
README.md
CHANGED
@@ -33,4 +33,4 @@ Results files should have the following format:
|
|
33 |
}
|
34 |
```
|
35 |
|
36 |
-
Request files are created automatically by this tool.
|
|
|
33 |
}
|
34 |
```
|
35 |
|
36 |
+
Request files are created automatically by this tool.
|
src/display/about.py
CHANGED
@@ -43,7 +43,7 @@ to return yes/no as an answer to the question.
|
|
43 |
3. Writing Prompts: This task evaluates the story-writing and creative abilities of the LLM.
|
44 |
|
45 |
4. Customer Support Dialogue: This task evaluates the ability of the LLM to answer a customer support question
|
46 |
-
given some product information and conversational history.
|
47 |
|
48 |
5. Toxic Prompts: This task evaluates the safety of the model by using prompts that can elicit harmful information
|
49 |
from LLMs.
|
@@ -53,9 +53,6 @@ from LLMs.
|
|
53 |
## Reproducibility
|
54 |
All of our datasets are closed-source. We provide a validation set with 5 examples for each of the tasks.
|
55 |
|
56 |
-
To reproduce the results on the validation set, run:
|
57 |
-
|
58 |
-
|
59 |
|
60 |
"""
|
61 |
|
|
|
43 |
3. Writing Prompts: This task evaluates the story-writing and creative abilities of the LLM.
|
44 |
|
45 |
4. Customer Support Dialogue: This task evaluates the ability of the LLM to answer a customer support question
|
46 |
+
given some product information and conversational history.
|
47 |
|
48 |
5. Toxic Prompts: This task evaluates the safety of the model by using prompts that can elicit harmful information
|
49 |
from LLMs.
|
|
|
53 |
## Reproducibility
|
54 |
All of our datasets are closed-source. We provide a validation set with 5 examples for each of the tasks.
|
55 |
|
|
|
|
|
|
|
56 |
|
57 |
"""
|
58 |
|