|
import os |
|
import gradio as gr |
|
import whisper |
|
from gtts import gTTS |
|
import io |
|
import almlapi |
|
|
|
|
|
os.environ["ALML_API_KEY"] = "701b35863e6d4a7b81bdcad2e6f3c880" |
|
|
|
|
|
model = whisper.load_model("base") |
|
|
|
|
|
def process_audio(file_path): |
|
try: |
|
|
|
audio = whisper.load_audio(file_path) |
|
result = model.transcribe(audio) |
|
text = result["text"] |
|
|
|
|
|
response = almlapi.call_o1( |
|
api_key=os.environ.get("ALML_API_KEY"), |
|
prompt=text, |
|
model="o1" |
|
) |
|
|
|
|
|
response_message = response["generated_text"].strip() |
|
|
|
|
|
tts = gTTS(response_message) |
|
response_audio_io = io.BytesIO() |
|
tts.write_to_fp(response_audio_io) |
|
response_audio_io.seek(0) |
|
|
|
|
|
with open("response.mp3", "wb") as audio_file: |
|
audio_file.write(response_audio_io.getvalue()) |
|
|
|
|
|
return response_message, "response.mp3" |
|
|
|
except Exception as e: |
|
|
|
return f"An error occurred: {e}", None |
|
|
|
|
|
title = "Voice-to-Voice AI Chatbot with AL/ML API" |
|
description = "Developed by [Adnan Tariq](https://www.linkedin.com/in/adnaantariq/) with ❤️" |
|
article = "### Instructions\n1. Upload an audio file.\n2. Wait for the transcription.\n3. Listen to the chatbot's response." |
|
|
|
|
|
iface = gr.Interface( |
|
fn=process_audio, |
|
inputs=gr.Audio(type="filepath"), |
|
outputs=[gr.Textbox(label="Response Text"), gr.Audio(label="Response Audio")], |
|
live=True, |
|
title=title, |
|
description=description, |
|
theme="dark", |
|
article=article |
|
) |
|
|
|
|
|
iface.launch() |
|
|