--- license: mit language: - it - en pipeline_tag: translation --- # OratioAI Sequecne to Sequence anguage translation, implimenting the methodes outlined in *'attention is all you need'* 1. Input Tokenization: The source and target sentences are tokenized using custom WordPiece tokenizers. Tokens are mapped to embeddings via the InputEmbeddings module, scaled by the model dimension. 2. Positional Encoding: Positional information is added to token embeddings using a fixed sinusoidal encoding strategy. 3. Encoding Phase: The encoder processes the source sequence, transforming token embeddings into contextual representations using stacked EncoderBlock modules. 4. Decoding Phase: The decoder autoregressively generates target tokens by attending to both previous tokens and encoder outputs. Cross-attention layers align source and target sequences effectively. 5. Projection: Final decoder outputs are projected into the target vocabulary space for token prediction. 6. Output Generation: Decoding is performed using a beam search or greedy approach to produce the final translated sentence. | Resource | Description | |-----------------------------------|----------------------------------------------------------| | [Training Space](https://huggingface.co/spaces/torinriley/OratioAI) | Hugging Face Space for training and testing the model. | | [GitHub Source Code](https://github.com/torinriley/OratioAI) | Source code repository for the translation project. | | [Attention Is All You Need](https://arxiv.org/pdf/1706.03762) | Original paper on the transformer architecture published from google | | Dataset | Description | |-----------------------------------|----------------------------------------------------------| | [Dataset](https://opus.nlpl.eu/Europarl/en&it/v8/Europarl) | Dataset Used for main model training. |