# Q&A Chatbot | |
from langchain.llms import OpenAI | |
from constants import openai_key | |
#from dotenv import load_dotenv | |
#load_dotenv() # take environment variables from .env. | |
import streamlit as st | |
import os | |
## Function to load OpenAI model and get respones | |
def get_openai_response(question): | |
llm=OpenAI(model_name="gpt-3.5-turbo-instruct",temperature=0.5) | |
response=llm(question) | |
return response | |
##initialize our streamlit app | |
st.set_page_config(page_title="Q&A Demo") | |
st.header("Langchain Application") | |
input=st.text_input("Input: ",key="input") | |
response=get_openai_response(input) | |
submit=st.button("Ask the question") | |
## If ask button is clicked | |
if submit: | |
st.subheader("The Response is") | |
st.write(response) | |