rshacter commited on
Commit
c00127a
·
verified ·
1 Parent(s): 2a6422c

Create README.md

Browse files

This is the first submission for the Uplimit Fine-tuning LLMs course: Evaluating a model

Files changed (1) hide show
  1. README.md +21 -0
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model Card: Uplimit Project 1 part 1
2
+ Model Description:
3
+ This is a model to test run publishing models. It has no real model assessment value.
4
+
5
+ This is a Large Language Model (LLM) trained on a dataset of DIBT/10k_prompts_ranked.
6
+ It was evaluated using using Eleuther Evaluation Harness
7
+
8
+ Hellaswag
9
+ Passed argument batch_size = auto:4.0. Detecting largest batch size
10
+ Determined largest batch size: 64
11
+ Passed argument batch_size = auto:4.0. Detecting largest batch size
12
+ Determined largest batch size: 64
13
+ hf (pretrained=EleutherAI/pythia-160m,revision=step100000,dtype=float), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto:4 (64,64,64,64,64)
14
+
15
+ | Tasks |Version|Filter|n-shot| Metric | |Value | |Stderr|
16
+ |---------|------:|------|-----:|--------|---|-----:|---|-----:|
17
+ |hellaswag| 1|none | 0|acc |↑ |0.2872|± |0.0045|
18
+ | | |none | 0|acc_norm|↑ |0.3082|± |0.0046|
19
+
20
+ How to Use
21
+ To use this model, simply download the checkpoint and load it into your preferred deep learning framework.