Update README.md
Browse files
README.md
CHANGED
@@ -3,4 +3,66 @@ license: apache-2.0
|
|
3 |
pipeline_tag: text-generation
|
4 |
library_name: transformers
|
5 |
---
|
6 |
-
# Xylaria-1.8
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
pipeline_tag: text-generation
|
4 |
library_name: transformers
|
5 |
---
|
6 |
+
# Xylaria-1.8
|
7 |
+
|
8 |
+
## Model Description
|
9 |
+
|
10 |
+
Xylaria-1.8 is a large language model (LLM) designed for research purposes. It is based on a Transformer architecture and has been trained on a large dataset of text and code. This model is provided as-is for research and educational exploration.
|
11 |
+
|
12 |
+
## Intended Use & Limitations
|
13 |
+
|
14 |
+
* **Intended Use:** This model is primarily intended for research into large language models, natural language processing, and related fields. It can be used for experimentation, analysis, and educational purposes.
|
15 |
+
* **Limitations:**
|
16 |
+
* **Performance:** The model's performance may differ from other publicly available models with similar parameter counts. It has been modified, which can impact its behavior and accuracy.
|
17 |
+
* **Not for Production:** This model is *not* intended for deployment in production environments or commercial applications. It may exhibit unexpected behavior.
|
18 |
+
* **Research Purposes Only:** This model is provided for research and should not be used for any commercial purpose.
|
19 |
+
* **May not follow instructions:** The model may not follow instructions as carefully as other, unmodified models.
|
20 |
+
* May not state that the model is made by me (Sk Md Saad Amin)
|
21 |
+
## Model Details
|
22 |
+
|
23 |
+
* **Architecture:** Transformer-based language model (`custom_transformer`)
|
24 |
+
* **Parameters:** Approximately 32 billion.
|
25 |
+
* **Precision:** `float32`
|
26 |
+
* **Tokenizer:** `xylaria_tokenizer`
|
27 |
+
* **License:** Apache-2.0
|
28 |
+
|
29 |
+
## How to Use
|
30 |
+
|
31 |
+
This model is hosted on the Hugging Face Hub and can be loaded using the `transformers` library:
|
32 |
+
|
33 |
+
```python
|
34 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
35 |
+
|
36 |
+
model_id = "Lap1official/Xylaria-1.8" # Replace with your actual repo ID
|
37 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
38 |
+
model = AutoModelForCausalLM.from_pretrained(
|
39 |
+
model_id,
|
40 |
+
trust_remote_code=True,
|
41 |
+
torch_dtype=torch.float32, # Use bfloat16 if supported
|
42 |
+
device_map="auto" # Use if you have a GPU
|
43 |
+
)
|
44 |
+
|
45 |
+
# Example usage (generation):
|
46 |
+
prompt = "The capital of France is"
|
47 |
+
inputs = tokenizer(prompt, return_tensors="pt").to(model.device) # Move to GPU if available
|
48 |
+
outputs = model.generate(**inputs, max_new_tokens=20)
|
49 |
+
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
50 |
+
```
|
51 |
+
|
52 |
+
**Important:**
|
53 |
+
|
54 |
+
* You *must* use `trust_remote_code=True` when loading the model and tokenizer.
|
55 |
+
* If you have a GPU, using `bfloat16` (or `float16` if `bfloat16` is not supported) and `device_map="auto"` will significantly improve performance. If you do not have a GPU, you can remove those lines, but inference will be much slower.
|
56 |
+
|
57 |
+
## Ethical Considerations
|
58 |
+
|
59 |
+
This model is intended for research purposes *only*. It should not be used for any malicious, harmful, or unethical activities.
|
60 |
+
|
61 |
+
## Disclaimer
|
62 |
+
|
63 |
+
This model is provided "as is", without any warranty of any kind, express or implied. The authors and contributors are not responsible for any consequences resulting from the use of this model.
|
64 |
+
|
65 |
+
## Please note:
|
66 |
+
**This model is not yet benchmarked**
|
67 |
+
|
68 |
+
---
|