File size: 599 Bytes
b95f6dc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
from flask import Flask, request, jsonify
import torch
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
import transformers

app = Flask(__name)

@app.route('/api/generate_response', methods=['POST'])
def generate_response():
    data = request.get_json()
    prompt = data.get('prompt')
    token_limit = data.get('token_limit')

    # Your model loading and inference code here (from the code you provided)
    # ...

    responses = mistral_model(prompt, token_limit)

    return jsonify({"responses": responses})

if __name__ == "__main__":
    app.run()