Commit
·
91b458a
1
Parent(s):
813660a
Update README.md
Browse files
README.md
CHANGED
@@ -10,18 +10,18 @@ tags:
|
|
10 |
# Mistral-7B-v0.1 for Italian Language Text Generation
|
11 |
|
12 |
## Overview
|
13 |
-
`Mistral-7B-v0.1` is a state-of-the-art Large Language Model (LLM) specifically pre-trained for generating text. With its 7 billion parameters, it's built to excel in benchmarks and outperforms even some larger models like the Llama 2 13B
|
14 |
|
15 |
## Model Architecture
|
16 |
-
The Mistral-7B-v0.1 model is a transformer-based model that can handle a variety of tasks including but not limited to translation, summarization, and text completion. It's particularly designed for the Italian language and can be fine-tuned for specific tasks
|
17 |
|
18 |
## Capabilities
|
19 |
- **Vocabulary Size**: 32,000 tokens, allowing for a broad range of inputs and outputs.
|
20 |
- **Hidden Size**: 4,096 dimensions, providing rich internal representations.
|
21 |
-
- **Intermediate Size**: 14,336 dimensions, which contributes to the model's ability to process and generate complex sentences
|
22 |
|
23 |
## Performance
|
24 |
-
Mistral-7B has been demonstrated to perform exceptionally well across a range of benchmarks, making it a reliable choice for developers and researchers working with the Italian language
|
25 |
|
26 |
|
27 |
## How to Use
|
|
|
10 |
# Mistral-7B-v0.1 for Italian Language Text Generation
|
11 |
|
12 |
## Overview
|
13 |
+
`Mistral-7B-v0.1` is a state-of-the-art Large Language Model (LLM) specifically pre-trained for generating text. With its 7 billion parameters, it's built to excel in benchmarks and outperforms even some larger models like the Llama 2 13B.
|
14 |
|
15 |
## Model Architecture
|
16 |
+
The Mistral-7B-v0.1 model is a transformer-based model that can handle a variety of tasks including but not limited to translation, summarization, and text completion. It's particularly designed for the Italian language and can be fine-tuned for specific tasks.
|
17 |
|
18 |
## Capabilities
|
19 |
- **Vocabulary Size**: 32,000 tokens, allowing for a broad range of inputs and outputs.
|
20 |
- **Hidden Size**: 4,096 dimensions, providing rich internal representations.
|
21 |
+
- **Intermediate Size**: 14,336 dimensions, which contributes to the model's ability to process and generate complex sentences.
|
22 |
|
23 |
## Performance
|
24 |
+
Mistral-7B has been demonstrated to perform exceptionally well across a range of benchmarks, making it a reliable choice for developers and researchers working with the Italian language.
|
25 |
|
26 |
|
27 |
## How to Use
|