forbiddensoul90
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -55,7 +55,7 @@ dtype = None # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for
|
|
55 |
load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
|
56 |
|
57 |
model, tokenizer = FastLanguageModel.from_pretrained(
|
58 |
-
model_name = "
|
59 |
max_seq_length = max_seq_length,
|
60 |
dtype = dtype,
|
61 |
load_in_4bit = load_in_4bit,
|
@@ -93,10 +93,10 @@ The Luxembourgish Chain-of-Thought (CoT) dataset used for fine-tuning consists o
|
|
93 |
## Benchmarking
|
94 |
- *Under progress*
|
95 |
|
|
|
|
|
96 |
|
97 |
<img src="https://www.luxprovide.lu/wp-content/themes/luxprovide2023/public/images/logo/logo_notagline_color_blue.4b07cb.svg" alt="LuxProvide Logo" width="50%">
|
98 |
|
99 |
<img src="https://docs.lxp.lu/FAQ/images/MeluXina_Logo.png" alt="Meluxina Logo" width="50%">
|
100 |
|
101 |
-
## Acknowledgments
|
102 |
-
This work leverages computational resources and support from [Meluxina](https://www.luxprovide.lu/meluxina) by LuxProvide.
|
|
|
55 |
load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
|
56 |
|
57 |
model, tokenizer = FastLanguageModel.from_pretrained(
|
58 |
+
model_name = "aiplanet/Lux-Llama",
|
59 |
max_seq_length = max_seq_length,
|
60 |
dtype = dtype,
|
61 |
load_in_4bit = load_in_4bit,
|
|
|
93 |
## Benchmarking
|
94 |
- *Under progress*
|
95 |
|
96 |
+
## Acknowledgments
|
97 |
+
This work leverages computational resources and support from [Meluxina](https://www.luxprovide.lu/meluxina) by LuxProvide.
|
98 |
|
99 |
<img src="https://www.luxprovide.lu/wp-content/themes/luxprovide2023/public/images/logo/logo_notagline_color_blue.4b07cb.svg" alt="LuxProvide Logo" width="50%">
|
100 |
|
101 |
<img src="https://docs.lxp.lu/FAQ/images/MeluXina_Logo.png" alt="Meluxina Logo" width="50%">
|
102 |
|
|
|
|