forbiddensoul90 commited on
Commit
f057d84
·
verified ·
1 Parent(s): 3e5a6b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -55,7 +55,7 @@ dtype = None # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for
55
  load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
56
 
57
  model, tokenizer = FastLanguageModel.from_pretrained(
58
- model_name = "forbiddensoul90/lux_lora_2",
59
  max_seq_length = max_seq_length,
60
  dtype = dtype,
61
  load_in_4bit = load_in_4bit,
@@ -93,10 +93,10 @@ The Luxembourgish Chain-of-Thought (CoT) dataset used for fine-tuning consists o
93
  ## Benchmarking
94
  - *Under progress*
95
 
 
 
96
 
97
  <img src="https://www.luxprovide.lu/wp-content/themes/luxprovide2023/public/images/logo/logo_notagline_color_blue.4b07cb.svg" alt="LuxProvide Logo" width="50%">
98
 
99
  <img src="https://docs.lxp.lu/FAQ/images/MeluXina_Logo.png" alt="Meluxina Logo" width="50%">
100
 
101
- ## Acknowledgments
102
- This work leverages computational resources and support from [Meluxina](https://www.luxprovide.lu/meluxina) by LuxProvide.
 
55
  load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
56
 
57
  model, tokenizer = FastLanguageModel.from_pretrained(
58
+ model_name = "aiplanet/Lux-Llama",
59
  max_seq_length = max_seq_length,
60
  dtype = dtype,
61
  load_in_4bit = load_in_4bit,
 
93
  ## Benchmarking
94
  - *Under progress*
95
 
96
+ ## Acknowledgments
97
+ This work leverages computational resources and support from [Meluxina](https://www.luxprovide.lu/meluxina) by LuxProvide.
98
 
99
  <img src="https://www.luxprovide.lu/wp-content/themes/luxprovide2023/public/images/logo/logo_notagline_color_blue.4b07cb.svg" alt="LuxProvide Logo" width="50%">
100
 
101
  <img src="https://docs.lxp.lu/FAQ/images/MeluXina_Logo.png" alt="Meluxina Logo" width="50%">
102