prithivMLmods commited on
Commit
f539049
·
verified ·
1 Parent(s): a26e0eb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -1
README.md CHANGED
@@ -9,4 +9,60 @@ tags:
9
  - trl
10
  - llama3.2
11
  - Reinforcement learning
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - trl
10
  - llama3.2
11
  - Reinforcement learning
12
+ ---
13
+ # **Bellatrix-Tiny-3B-R1**
14
+
15
+ Bellatrix is based on a reasoning-based model designed for the **DeepSeek-R1** synthetic dataset entries. The pipeline's instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks. These models outperform many of the available open-source options. Bellatrix is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions utilize supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF).
16
+
17
+ ## **Use with transformers**
18
+
19
+ Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.
20
+
21
+ Make sure to update your transformers installation via:
22
+
23
+ ```sh
24
+ pip install --upgrade transformers
25
+ ```
26
+
27
+ ```python
28
+ import torch
29
+ from transformers import pipeline
30
+
31
+ model_id = "prithivMLmods/Bellatrix-Tiny-3B-R1"
32
+ pipe = pipeline(
33
+ "text-generation",
34
+ model=model_id,
35
+ torch_dtype=torch.bfloat16,
36
+ device_map="auto",
37
+ )
38
+ messages = [
39
+ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"},
40
+ {"role": "user", "content": "Who are you?"},
41
+ ]
42
+ outputs = pipe(
43
+ messages,
44
+ max_new_tokens=256,
45
+ )
46
+ print(outputs[0]["generated_text"][-1])
47
+ ```
48
+
49
+ **Note:** You can also find detailed recipes on how to use the model locally, with `torch.compile()`, assisted generations, quantization, and more at [`huggingface-llama-recipes`](https://github.com/huggingface/huggingface-llama-recipes).
50
+
51
+ ## **Intended Use**
52
+
53
+ Bellatrix is designed for applications that require advanced reasoning and multilingual dialogue capabilities. It is particularly suitable for:
54
+
55
+ - **Agentic Retrieval**: Enabling intelligent retrieval of relevant information in a dialogue or query-response system.
56
+ - **Summarization Tasks**: Condensing large bodies of text into concise summaries for easier comprehension.
57
+ - **Multilingual Use Cases**: Supporting conversations in multiple languages with high accuracy and coherence.
58
+ - **Instruction-Based Applications**: Following complex, context-aware instructions to generate precise outputs in a variety of scenarios.
59
+
60
+ ## **Limitations**
61
+
62
+ Despite its capabilities, Bellatrix has some limitations:
63
+
64
+ 1. **Domain Specificity**: While it performs well on general tasks, its performance may degrade with highly specialized or niche datasets.
65
+ 2. **Dependence on Training Data**: It is only as good as the quality and diversity of its training data, which may lead to biases or inaccuracies.
66
+ 3. **Computational Resources**: The model’s optimized transformer architecture can be resource-intensive, requiring significant computational power for fine-tuning and inference.
67
+ 4. **Language Coverage**: While multilingual, some languages or dialects may have limited support or lower performance compared to widely used ones.
68
+ 5. **Real-World Contexts**: It may struggle with understanding nuanced or ambiguous real-world scenarios not covered during training.