Update README.md
Browse files
README.md
CHANGED
@@ -50,11 +50,11 @@ For more details, please refer to our [Hugging Face page](https://huggingface.co
|
|
50 |
|
51 |
Marco-LLM includes a 7B parameter model based on the Transformer architecture. The key features of Marco-LLM are:
|
52 |
|
53 |
-
-Multilingual Training: The model is trained on a large-scale multilingual dataset covering 29 languages, including both high-resource languages (e.g., English, Chinese) and low-resource languages (e.g., Kazakh, Nepali).
|
54 |
|
55 |
-
-Enhanced Tokenizer: An improved tokenizer is used to better handle multilingual data, ensuring higher efficiency and accuracy in tokenization.
|
56 |
|
57 |
-
-Post-Training: Marco-LLM supports various post-training methods, such as Supervised Fine-tuning (SFT) and Direct Preference Optimization (DPO), to further enhance performance for specific tasks and languages.
|
58 |
## Usage
|
59 |
|
60 |
It is not advised to use the base language models for direct text generation tasks. Instead, it is recommended to apply post-training methods such as Supervised Fine-tuning (SFT), Reinforcement Learning with Human Feedback (RLHF), or continued pretraining to adapt the models for specific use cases.
|
|
|
50 |
|
51 |
Marco-LLM includes a 7B parameter model based on the Transformer architecture. The key features of Marco-LLM are:
|
52 |
|
53 |
+
- Multilingual Training: The model is trained on a large-scale multilingual dataset covering 29 languages, including both high-resource languages (e.g., English, Chinese) and low-resource languages (e.g., Kazakh, Nepali).
|
54 |
|
55 |
+
- Enhanced Tokenizer: An improved tokenizer is used to better handle multilingual data, ensuring higher efficiency and accuracy in tokenization.
|
56 |
|
57 |
+
- Post-Training: Marco-LLM supports various post-training methods, such as Supervised Fine-tuning (SFT) and Direct Preference Optimization (DPO), to further enhance performance for specific tasks and languages.
|
58 |
## Usage
|
59 |
|
60 |
It is not advised to use the base language models for direct text generation tasks. Instead, it is recommended to apply post-training methods such as Supervised Fine-tuning (SFT), Reinforcement Learning with Human Feedback (RLHF), or continued pretraining to adapt the models for specific use cases.
|