artificialguybr commited on
Commit
320d399
1 Parent(s): 4dbd193

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: llama3.2
4
+ base_model: meta-llama/Llama-3.2-1B
5
+ tags:
6
+ - generated_from_trainer
7
+ - fine-tuned
8
+ - instruction-tuned
9
+ model-index:
10
+ - name: OpenHermes-Llama-3.2-1B
11
+ results: []
12
+ ---
13
+
14
+ # OpenHermes-Llama-3.2-1B
15
+
16
+ ## Model Description
17
+
18
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on the [OpenHermes-2.5](https://huggingface.co/datasets/teknium/OpenHermes-2.5) dataset. It is based on the Llama 3.2 architecture, which is an optimized transformer model designed for multilingual dialogue use cases, including agentic retrieval and summarization tasks.
19
+
20
+ ### Key Features:
21
+ - **Base Model**: Meta Llama 3.2 1B
22
+ - **Fine-tuning Dataset**: OpenHermes 2.5
23
+ - **Architecture**: Auto-regressive language model with optimized transformer architecture
24
+ - **Params**: 1.23B
25
+ - **Context Length**: 128k
26
+ - **Input/Output Modalities**: Multilingual Text and code
27
+ - **Supported Languages**: Primarily English, with potential for other languages supported by Llama 3.2
28
+
29
+ ## Intended Uses & Limitations
30
+
31
+ This model is intended for commercial and research use in multiple languages, particularly suited for assistant-like chat and agentic applications such as knowledge retrieval, summarization, and query rewriting. It inherits the capabilities and limitations of both the Llama 3.2 1B base model and the OpenHermes 2.5 dataset.
32
+
33
+ ### Out of Scope:
34
+ - Use that violates applicable laws or regulations
35
+ - Use prohibited by the Acceptable Use Policy and Llama 3.2 Community License
36
+ - Use in unsupported languages without proper evaluation and safety measures
37
+
38
+ ## Training Procedure
39
+
40
+ ### Training Data
41
+ The model was fine-tuned on the OpenHermes 2.5 dataset, which contains 1M primarily synthetically generated instruction and chat samples. This dataset is a compilation of various open-source datasets and custom-created synthetic datasets, designed to enhance the model's performance in instruction-following and chat scenarios.
42
+
43
+
44
+ ### Training Results
45
+
46
+ | Training Loss | Epoch | Step | Validation Loss |
47
+ |:-------------:|:-----:|:----:|:---------------:|
48
+ | 1.1101 | 0.0003| 1 | 0.9499 |
49
+ | 0.7977 | 0.5000| 1438 | 0.8729 |
50
+ | 0.8338 | 1.0000| 2876 | 0.8647 |
51
+ | 0.7714 | 1.4981| 4314 | 0.8637 |
52
+ | 0.8305 | 1.9983| 5752 | 0.8612 |
53
+ | 0.6801 | 2.4963| 7190 | 0.8631 |
54
+
55
+ ## Evaluation Results
56
+
57
+ The model achieves a final validation loss of 0.8631.
58
+
59
+ ## Ethical Considerations and Limitations
60
+
61
+ Users should be aware of potential biases in the training data and exercise caution when deploying the model, especially in sensitive applications. The model's outputs should be carefully monitored and filtered for inappropriate content.
62
+
63
+
64
+ For more information on the base Llama 3.2 model, please refer to the [official Llama 3.2 model card](https://huggingface.co/meta-llama/Llama-3.2-1B).