Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -1,15 +1,10 @@
|
|
1 |
import os
|
2 |
-
import sys
|
3 |
import subprocess
|
4 |
import streamlit as st
|
5 |
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
|
6 |
import black
|
7 |
from pylint import lint
|
8 |
from io import StringIO
|
9 |
-
import openai
|
10 |
-
|
11 |
-
# Set your OpenAI API key here
|
12 |
-
openai.api_key = "YOUR_OPENAI_API_KEY"
|
13 |
|
14 |
HUGGING_FACE_REPO_URL = "https://huggingface.co/spaces/acecalisto3/DevToolKit"
|
15 |
PROJECT_ROOT = "projects"
|
@@ -159,21 +154,11 @@ def terminal_interface(command, project_name=None):
|
|
159 |
st.session_state.current_state['toolbox']['terminal_output'] = result.stderr
|
160 |
return result.stderr
|
161 |
|
162 |
-
|
163 |
-
|
164 |
-
|
165 |
-
|
166 |
-
|
167 |
-
result = StringIO()
|
168 |
-
sys.stdout = result
|
169 |
-
sys.stderr = result
|
170 |
-
(pylint_stdout, pylint_stderr) = lint.py_run(code, return_std=True)
|
171 |
-
sys.stdout = sys.__stdout__
|
172 |
-
sys.stderr = sys.__stderr__
|
173 |
-
lint_message = pylint_stdout.getvalue() + pylint_stderr.getvalue()
|
174 |
-
st.session_state.current_state['toolbox']['formatted_code'] = formatted_code
|
175 |
-
st.session_state.current_state['toolbox']['lint_message'] = lint_message
|
176 |
-
return formatted_code, lint_message
|
177 |
|
178 |
def summarize_text(text):
|
179 |
summarizer = pipeline("summarization")
|
@@ -187,10 +172,20 @@ def sentiment_analysis(text):
|
|
187 |
st.session_state.current_state['toolbox']['sentiment'] = sentiment[0]
|
188 |
return sentiment[0]
|
189 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
190 |
def translate_code(code, input_language, output_language):
|
191 |
# Define a dictionary to map programming languages to their corresponding file extensions
|
192 |
language_extensions = {
|
193 |
-
|
194 |
}
|
195 |
|
196 |
# Add code to handle edge cases such as invalid input and unsupported programming languages
|
|
|
1 |
import os
|
|
|
2 |
import subprocess
|
3 |
import streamlit as st
|
4 |
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
|
5 |
import black
|
6 |
from pylint import lint
|
7 |
from io import StringIO
|
|
|
|
|
|
|
|
|
8 |
|
9 |
HUGGING_FACE_REPO_URL = "https://huggingface.co/spaces/acecalisto3/DevToolKit"
|
10 |
PROJECT_ROOT = "projects"
|
|
|
154 |
st.session_state.current_state['toolbox']['terminal_output'] = result.stderr
|
155 |
return result.stderr
|
156 |
|
157 |
+
# Chat interface using a selected agent
|
158 |
+
def chat_interface_with_agent(input_text, agent_name):
|
159 |
+
# ... [rest of the chat_interface_with_agent function] ...
|
160 |
+
|
161 |
+
# ... [rest of the workspace_interface, add_code_to_workspace, terminal_interface, code_editor_interface functions] ...
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
162 |
|
163 |
def summarize_text(text):
|
164 |
summarizer = pipeline("summarization")
|
|
|
172 |
st.session_state.current_state['toolbox']['sentiment'] = sentiment[0]
|
173 |
return sentiment[0]
|
174 |
|
175 |
+
# ... [rest of the translate_code function, but remove the OpenAI API call and replace it with your own logic] ...
|
176 |
+
|
177 |
+
def generate_code(code_idea):
|
178 |
+
# Replace this with a call to a Hugging Face model or your own logic
|
179 |
+
# For example, using a text-generation pipeline:
|
180 |
+
generator = pipeline('text-generation', model='gpt2')
|
181 |
+
generated_code = generator(code_idea, max_length=100, num_return_sequences=1)[0]['generated_text']
|
182 |
+
st.session_state.current_state['toolbox']['generated_code'] = generated_code
|
183 |
+
return generated_code
|
184 |
+
|
185 |
def translate_code(code, input_language, output_language):
|
186 |
# Define a dictionary to map programming languages to their corresponding file extensions
|
187 |
language_extensions = {
|
188 |
+
|
189 |
}
|
190 |
|
191 |
# Add code to handle edge cases such as invalid input and unsupported programming languages
|