File size: 1,994 Bytes
96bca39
 
 
 
 
 
f4ad2fa
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55c91cc
f4ad2fa
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
---
license: mit
language:
- it
- en
pipeline_tag: translation
---

# OratioAI
Sequecne to Sequence anguage translation, implimenting the methodes outlined in *'attention is all you need'*

1.	Input Tokenization:
The source and target sentences are tokenized using custom WordPiece tokenizers. Tokens are mapped to embeddings via the InputEmbeddings module, scaled by the model dimension.
2.	Positional Encoding:
Positional information is added to token embeddings using a fixed sinusoidal encoding strategy.
3.	Encoding Phase:
The encoder processes the source sequence, transforming token embeddings into contextual representations using stacked EncoderBlock modules.
4.	Decoding Phase:
The decoder autoregressively generates target tokens by attending to both previous tokens and encoder outputs. Cross-attention layers align source and target sequences effectively.
5.	Projection:
Final decoder outputs are projected into the target vocabulary space for token prediction.
6.	Output Generation:
Decoding is performed using a beam search or greedy approach to produce the final translated sentence.




| Resource                          | Description                                              |
|-----------------------------------|----------------------------------------------------------|
| [Training Space](https://huggingface.co/spaces/torinriley/OratioAI) | Hugging Face Space for training and testing the model.    |
| [GitHub Source Code](https://github.com/torinriley/OratioAI) | Source code repository for the translation project.       |
| [Attention Is All You Need](https://arxiv.org/pdf/1706.03762) | Original paper on the transformer architecture published from google |

| Dataset                          | Description                                              |
|-----------------------------------|----------------------------------------------------------|
| [Dataset](https://opus.nlpl.eu/Europarl/en&it/v8/Europarl) | Dataset Used for main model training.    |