import torch from transformers import AutoModelForSeq2SeqLM, AutoTokenizer # Load your pre-trained model and tokenizer model = AutoModelForSeq2SeqLM.from_pretrained("Vuks/sanchit_whisper") tokenizer = AutoTokenizer.from_pretrained("Vuks/sanchit_whisper") def handle(request, context): inputs = tokenizer(request["inputs"], return_tensors="pt") outputs = model(**inputs) return outputs