Solshine commited on
Commit
f154e60
1 Parent(s): 01eb803

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -22,8 +22,10 @@ datasets:
22
  - **Finetuned from model :** unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
23
 
24
  Inspired by and featuring the Reflection Tuning technique pioneered by Matt Shumer (possibly earlier innovated by the team at Anthropic.)
 
25
  *To the authors' knowledge, this is the first "reflection tuned" Llama 3.1 8B LLM*
26
 
 
27
  **As per the inspiring model "mattshumer/Reflection-Llama-3.1-70B" (this mode was not used in the training process nor as a foundational model, but only served as inspiration) :**
28
  '''
29
  During sampling, the model will start by outputting reasoning inside <thinking> and </thinking> tags, and then once it is satisfied with its reasoning, it will output the final answer inside <output> and </output> tags. Each of these tags are special tokens, trained into the model.
 
22
  - **Finetuned from model :** unsloth/meta-llama-3.1-8b-instruct-bnb-4bit
23
 
24
  Inspired by and featuring the Reflection Tuning technique pioneered by Matt Shumer (possibly earlier innovated by the team at Anthropic.)
25
+
26
  *To the authors' knowledge, this is the first "reflection tuned" Llama 3.1 8B LLM*
27
 
28
+
29
  **As per the inspiring model "mattshumer/Reflection-Llama-3.1-70B" (this mode was not used in the training process nor as a foundational model, but only served as inspiration) :**
30
  '''
31
  During sampling, the model will start by outputting reasoning inside <thinking> and </thinking> tags, and then once it is satisfied with its reasoning, it will output the final answer inside <output> and </output> tags. Each of these tags are special tokens, trained into the model.