Safetensors
qwen2
StarscreamDeceptions commited on
Commit
c529b5a
·
verified ·
1 Parent(s): 74eab29

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -50,11 +50,11 @@ For more details, please refer to our [Hugging Face page](https://huggingface.co
50
 
51
  Marco-LLM includes a 7B parameter model based on the Transformer architecture. The key features of Marco-LLM are:
52
 
53
- -Multilingual Training: The model is trained on a large-scale multilingual dataset covering 29 languages, including both high-resource languages (e.g., English, Chinese) and low-resource languages (e.g., Kazakh, Nepali).
54
 
55
- -Enhanced Tokenizer: An improved tokenizer is used to better handle multilingual data, ensuring higher efficiency and accuracy in tokenization.
56
 
57
- -Post-Training: Marco-LLM supports various post-training methods, such as Supervised Fine-tuning (SFT) and Direct Preference Optimization (DPO), to further enhance performance for specific tasks and languages.
58
  ## Usage
59
 
60
  It is not advised to use the base language models for direct text generation tasks. Instead, it is recommended to apply post-training methods such as Supervised Fine-tuning (SFT), Reinforcement Learning with Human Feedback (RLHF), or continued pretraining to adapt the models for specific use cases.
 
50
 
51
  Marco-LLM includes a 7B parameter model based on the Transformer architecture. The key features of Marco-LLM are:
52
 
53
+ - Multilingual Training: The model is trained on a large-scale multilingual dataset covering 29 languages, including both high-resource languages (e.g., English, Chinese) and low-resource languages (e.g., Kazakh, Nepali).
54
 
55
+ - Enhanced Tokenizer: An improved tokenizer is used to better handle multilingual data, ensuring higher efficiency and accuracy in tokenization.
56
 
57
+ - Post-Training: Marco-LLM supports various post-training methods, such as Supervised Fine-tuning (SFT) and Direct Preference Optimization (DPO), to further enhance performance for specific tasks and languages.
58
  ## Usage
59
 
60
  It is not advised to use the base language models for direct text generation tasks. Instead, it is recommended to apply post-training methods such as Supervised Fine-tuning (SFT), Reinforcement Learning with Human Feedback (RLHF), or continued pretraining to adapt the models for specific use cases.