Update README.md
Browse files
README.md
CHANGED
@@ -16,10 +16,10 @@ BLING models are fine-tuned with high-quality custom instruct datasets, designed
|
|
16 |
Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
|
17 |
Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
|
18 |
|
19 |
-
--**Accuracy Score**: **86.
|
20 |
--Not Found Classification: 85.0%
|
21 |
-
--Boolean:
|
22 |
-
--Math/Logic: 37.
|
23 |
--Complex Questions (1-5): 3 (Medium-High: multiple choice, table reading, causal)
|
24 |
--Summarization Quality (1-5): 3 (Coherent, extractive)
|
25 |
--Hallucinations: No hallucinations observed in test runs.
|
@@ -34,7 +34,7 @@ For test run results (and good indicator of target use cases), please see the fi
|
|
34 |
- **Model type:** TinyLlama
|
35 |
- **Language(s) (NLP):** English
|
36 |
- **License:** Apache 2.0
|
37 |
-
- **Finetuned from model:** TinyLlama-1.1b
|
38 |
|
39 |
## Uses
|
40 |
|
|
|
16 |
Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
|
17 |
Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
|
18 |
|
19 |
+
--**Accuracy Score**: **86.5** correct out of 100
|
20 |
--Not Found Classification: 85.0%
|
21 |
+
--Boolean: 82.50%
|
22 |
+
--Math/Logic: 37.50%
|
23 |
--Complex Questions (1-5): 3 (Medium-High: multiple choice, table reading, causal)
|
24 |
--Summarization Quality (1-5): 3 (Coherent, extractive)
|
25 |
--Hallucinations: No hallucinations observed in test runs.
|
|
|
34 |
- **Model type:** TinyLlama
|
35 |
- **Language(s) (NLP):** English
|
36 |
- **License:** Apache 2.0
|
37 |
+
- **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
|
38 |
|
39 |
## Uses
|
40 |
|