Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,40 @@ license: apache-2.0
|
|
10 |
language:
|
11 |
- en
|
12 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
# Uploaded model
|
15 |
|
|
|
10 |
language:
|
11 |
- en
|
12 |
---
|
13 |
+
## Model Introduction
|
14 |
+
Early experimental model uses unique advance form of supervised tuning. This training program loads the model, and than loads the data from dataset. It will provide data in inference time. Than it trains the LLM.
|
15 |
+
During inference and than checks if it reaches the answer or goal. If not, it will keep training until it reaches the answer or solution.
|
16 |
+
|
17 |
+
## Installation
|
18 |
+
Update latest transformers
|
19 |
+
```python
|
20 |
+
pip install -U transformers
|
21 |
+
```
|
22 |
+
|
23 |
+
System prompt suggested for math:
|
24 |
+
```python
|
25 |
+
system_prompt="<problem>...</problem><solution>...</solution>"
|
26 |
+
```
|
27 |
+
|
28 |
+
|
29 |
+
Inference
|
30 |
+
```python
|
31 |
+
from transformers import pipeline
|
32 |
+
model_id = "EpistemeAI/OpenReasoner-Llama-3.2-1B-rs1"
|
33 |
+
pipe = pipeline(
|
34 |
+
"text-generation",
|
35 |
+
model=model_id,
|
36 |
+
torch_dtype=torch.bfloat16,
|
37 |
+
device_map="auto"
|
38 |
+
)
|
39 |
+
print(pipe("What is larger 9.9 or 9.11?"))
|
40 |
+
```
|
41 |
+
|
42 |
+
## Reference
|
43 |
+
Thank you so much to Hugging Face H4 and the dataset: [Math-500](https://huggingface.co/datasets/HuggingFaceH4/MATH-500)
|
44 |
+
|
45 |
+
We use this as evaluator.
|
46 |
+
|
47 |
|
48 |
# Uploaded model
|
49 |
|