Safetensors
qwen2
StarscreamDeceptions commited on
Commit
74eab29
·
verified ·
1 Parent(s): 0f03d01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -1
README.md CHANGED
@@ -32,4 +32,51 @@ language:
32
  - pl
33
  base_model:
34
  - Qwen/Qwen2-7B
35
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  - pl
33
  base_model:
34
  - Qwen/Qwen2-7B
35
+ ---
36
+
37
+
38
+ # Marco-LLM-GLO
39
+
40
+ ## Introduction
41
+
42
+ Marco-LLM is a series of advanced multilingual language models designed to bridge the performance gap between high-resource languages and low-resource languages. This repository contains the Marco-LLM base language model with 7 billion parameters.
43
+
44
+ The model has undergone extensive multilingual continual pretraining on a diverse dataset containing over 5 trillion tokens, with a particular focus on enhancing performance in low-resource languages while maintaining strong capabilities in high-resource languages like English and Chinese.
45
+
46
+ Compared to state-of-the-art open-source language models, Marco-LLM demonstrates significant improvements in multilingual tasks, including machine translation, question answering, and reasoning across multiple languages.
47
+ For more details, please refer to our [Hugging Face page](https://huggingface.co/AIDC-AI/Marco-LLM-GLO).
48
+
49
+ ## Model Details
50
+
51
+ Marco-LLM includes a 7B parameter model based on the Transformer architecture. The key features of Marco-LLM are:
52
+
53
+ -Multilingual Training: The model is trained on a large-scale multilingual dataset covering 29 languages, including both high-resource languages (e.g., English, Chinese) and low-resource languages (e.g., Kazakh, Nepali).
54
+
55
+ -Enhanced Tokenizer: An improved tokenizer is used to better handle multilingual data, ensuring higher efficiency and accuracy in tokenization.
56
+
57
+ -Post-Training: Marco-LLM supports various post-training methods, such as Supervised Fine-tuning (SFT) and Direct Preference Optimization (DPO), to further enhance performance for specific tasks and languages.
58
+ ## Usage
59
+
60
+ It is not advised to use the base language models for direct text generation tasks. Instead, it is recommended to apply post-training methods such as Supervised Fine-tuning (SFT), Reinforcement Learning with Human Feedback (RLHF), or continued pretraining to adapt the models for specific use cases.
61
+
62
+
63
+ ## Citation
64
+
65
+ If you find our work helpful, please give us a citation.
66
+ ```
67
+ @article{unique_identifier,
68
+
69
+ title={Marco-LLM: Bridging Languages via Massive Multilingual Training for Cross-Lingual Enhancement},
70
+
71
+ journal={arXiv},
72
+
73
+ volume={},
74
+
75
+ number={2412.04003},
76
+
77
+ year={2024},
78
+
79
+ url={https://arxiv.org/abs/2412.04003}
80
+
81
+ }
82
+ ```