doberst commited on
Commit
d7582af
1 Parent(s): 11dacd8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -18,14 +18,17 @@ without using any advanced quantization optimizations.
18
  Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
19
  Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
20
 
21
- --**Accuracy Score**: **80.25** correct out of 100
22
- --Not Found Classification: 40.0%
23
- --Boolean: 41.5%
24
- --Math/Logic: 7.5%
25
  --Complex Questions (1-5): 1 (Low)
26
  --Summarization Quality (1-5): 3 (Coherent, extractive)
27
  --Hallucinations: No hallucinations observed in test runs.
28
 
 
 
 
29
  For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
30
 
31
 
 
18
  Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
19
  Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
20
 
21
+ --**Accuracy Score**: **89.0** correct out of 100
22
+ --Not Found Classification: 57.5%
23
+ --Boolean: 57.5%
24
+ --Math/Logic: 25%
25
  --Complex Questions (1-5): 1 (Low)
26
  --Summarization Quality (1-5): 3 (Coherent, extractive)
27
  --Hallucinations: No hallucinations observed in test runs.
28
 
29
+ Please note that these scoring results have been updated from the original (upward), as we corrected a small bug in the original test inference script for this model.
30
+ The corrected test results are in the files repo, and have been generated with the test scripts in the repo.
31
+
32
  For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
33
 
34