This model is an Instruction-Tuned version of Llama 3.2 400M Amharic.

How to use

Chat Format

Given the nature of the training data, the phi-2 instruct model is best suited for prompts using the chat format as follows. You can provide the prompt as a question with a generic template as follows:

<|im_start|>user
αŒ₯ያቄ?<|im_end|>
<|im_start|>assistant

For example:

<|im_start|>user
αˆΆαˆ΅α‰΅ α‹¨αŠ ααˆͺካ αˆ€αŒˆαˆ«α‰΅ αŒ₯α‰€αˆ΅αˆαŠ<|im_end|>
<|im_start|>assistant

where the model generates the text after <|im_start|>assistant .

Sample inference code

First, you need to install the latest version of transformers

pip install -Uq transformers

You can use this model directly with a pipeline for text generation:

from transformers import pipeline

llama3_am = pipeline(
    "text-generation",
    model="rasyosef/Llama-3.2-400M-Amharic-Instruct",
    device_map="auto"
  )

messages = [{"role": "user", "content": "αˆΆαˆ΅α‰΅ α‹¨αŠ ααˆͺካ αˆ€αŒˆαˆ«α‰΅ αŒ₯α‰€αˆ΅αˆαŠ"}]
llama3_am(messages, max_new_tokens=128, repetition_penalty=1.1, return_full_text=False)

Output:

[{'generated_text': '1. ግα‰₯ፅ 2. αŠ“α‹­αŒ„αˆͺα‹« 3. αŒ‹αŠ“'}]
Downloads last month
42
Safetensors
Model size
413M params
Tensor type
F32
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for rasyosef/Llama-3.2-400M-Amharic-Instruct

Finetuned
(2)
this model

Collection including rasyosef/Llama-3.2-400M-Amharic-Instruct