Hawoly18 commited on
Commit
ab296df
·
verified ·
1 Parent(s): 9784699

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +47 -0
app.py ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ os.environ['HF_HOME'] = 'E:/huggingface_cache'
3
+
4
+ import gradio as gr
5
+ from transformers import AutoTokenizer, AutoModelForCausalLM
6
+ import torch
7
+
8
+ # Load the model and tokenizer
9
+ tokenizer = AutoTokenizer.from_pretrained("Hawoly18/llama3.2-3B-Wolof")
10
+ model = AutoModelForCausalLM.from_pretrained("Hawoly18/llama3.2-3B-Wolof")
11
+
12
+ if tokenizer.pad_token is None:
13
+ tokenizer.pad_token = tokenizer.eos_token
14
+
15
+ # Function to generate responses
16
+ def generate_response(question, max_length=512):
17
+ input_text = f"Question: {question}\nRéponse:"
18
+ input_ids = tokenizer.encode(input_text, return_tensors='pt', padding=True, truncation=True)
19
+ attention_mask = input_ids != tokenizer.pad_token_id
20
+
21
+ with torch.no_grad():
22
+ output_ids = model.generate(
23
+ input_ids,
24
+ max_length=max_length,
25
+ attention_mask=attention_mask,
26
+ pad_token_id=tokenizer.eos_token_id,
27
+ eos_token_id=tokenizer.eos_token_id,
28
+ num_beams=5, # Beam search for better quality
29
+ no_repeat_ngram_size=2, # Prevent n-gram repetition
30
+ early_stopping=True
31
+ )
32
+ response = tokenizer.decode(output_ids[0], skip_special_tokens=True)
33
+ response = response.replace(input_text, "").strip()
34
+ return response
35
+
36
+ # Define the Gradio interface
37
+ interface = gr.Interface(
38
+ fn=generate_response,
39
+ inputs="text",
40
+ outputs="text",
41
+ title="Model Adia Géneration de Réponse en Wolof",
42
+ description="Posez une question relative à l'entrepreneuriat en Afrique",
43
+ examples=[["yan jumtukaay ci xaral yi BSE moom mën a dimbali ndax moom mën woyal sama liggéey ci entrepreneur yi"]]
44
+ )
45
+
46
+ # Launch the interface
47
+ interface.launch(share=True)