Spaces:
Runtime error
Runtime error
File size: 599 Bytes
b95f6dc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 |
from flask import Flask, request, jsonify
import torch
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
import transformers
app = Flask(__name)
@app.route('/api/generate_response', methods=['POST'])
def generate_response():
data = request.get_json()
prompt = data.get('prompt')
token_limit = data.get('token_limit')
# Your model loading and inference code here (from the code you provided)
# ...
responses = mistral_model(prompt, token_limit)
return jsonify({"responses": responses})
if __name__ == "__main__":
app.run()
|