Spaces:
Running
Running
import streamlit as st | |
import datetime | |
import os | |
# Initialize session state variables | |
if 'transcript_history' not in st.session_state: | |
st.session_state.transcript_history = [] | |
# Create containers at the top level | |
st.title("Speech Recognition with Transcript History") | |
# Create the main layout | |
col1, col2 = st.columns([2, 1]) | |
with col1: | |
# Speech recognition component | |
html = """ | |
<!DOCTYPE html> | |
<html> | |
<head> | |
<title>Continuous Speech Demo</title> | |
<style> | |
body { | |
font-family: sans-serif; | |
padding: 20px; | |
max-width: 800px; | |
margin: 0 auto; | |
} | |
button { | |
padding: 10px 20px; | |
margin: 10px 5px; | |
font-size: 16px; | |
} | |
#status { | |
margin: 10px 0; | |
padding: 10px; | |
background: #e8f5e9; | |
border-radius: 4px; | |
} | |
#output { | |
white-space: pre-wrap; | |
padding: 15px; | |
background: #f5f5f5; | |
border-radius: 4px; | |
margin: 10px 0; | |
min-height: 100px; | |
max-height: 400px; | |
overflow-y: auto; | |
} | |
.controls { | |
margin: 10px 0; | |
} | |
</style> | |
</head> | |
<body> | |
<div class="controls"> | |
<button id="start">Start Listening</button> | |
<button id="stop" disabled>Stop Listening</button> | |
<button id="clear">Clear Text</button> | |
</div> | |
<div id="status">Ready</div> | |
<div id="output"></div> | |
<script> | |
if (!('webkitSpeechRecognition' in window)) { | |
alert('Speech recognition not supported'); | |
} else { | |
const recognition = new webkitSpeechRecognition(); | |
const startButton = document.getElementById('start'); | |
const stopButton = document.getElementById('stop'); | |
const clearButton = document.getElementById('clear'); | |
const status = document.getElementById('status'); | |
const output = document.getElementById('output'); | |
let fullTranscript = ''; | |
let lastUpdateTime = Date.now(); | |
// Configure recognition | |
recognition.continuous = true; | |
recognition.interimResults = true; | |
startButton.onclick = () => { | |
try { | |
recognition.start(); | |
status.textContent = 'Listening...'; | |
startButton.disabled = true; | |
stopButton.disabled = false; | |
} catch (e) { | |
console.error(e); | |
status.textContent = 'Error: ' + e.message; | |
} | |
}; | |
stopButton.onclick = () => { | |
recognition.stop(); | |
status.textContent = 'Stopped'; | |
startButton.disabled = false; | |
stopButton.disabled = true; | |
}; | |
clearButton.onclick = () => { | |
fullTranscript = ''; | |
output.textContent = ''; | |
}; | |
recognition.onresult = (event) => { | |
let interimTranscript = ''; | |
let finalTranscript = ''; | |
for (let i = event.resultIndex; i < event.results.length; i++) { | |
const transcript = event.results[i][0].transcript; | |
if (event.results[i].isFinal) { | |
finalTranscript += transcript + '\\n'; | |
} else { | |
interimTranscript += transcript; | |
} | |
} | |
if (finalTranscript || (Date.now() - lastUpdateTime > 5000)) { | |
if (finalTranscript) { | |
fullTranscript += finalTranscript; | |
// Try to communicate with Streamlit | |
try { | |
parent.postMessage({ | |
type: 'transcriptUpdate', | |
text: finalTranscript | |
}, '*'); | |
} catch (e) { | |
console.error('Failed to send transcript:', e); | |
} | |
} | |
lastUpdateTime = Date.now(); | |
} | |
output.textContent = fullTranscript + (interimTranscript ? '... ' + interimTranscript : ''); | |
output.scrollTop = output.scrollHeight; | |
}; | |
recognition.onend = () => { | |
if (!stopButton.disabled) { | |
try { | |
recognition.start(); | |
console.log('Restarted recognition'); | |
} catch (e) { | |
console.error('Failed to restart recognition:', e); | |
status.textContent = 'Error restarting: ' + e.message; | |
startButton.disabled = false; | |
stopButton.disabled = true; | |
} | |
} | |
}; | |
recognition.onerror = (event) => { | |
console.error('Recognition error:', event.error); | |
status.textContent = 'Error: ' + event.error; | |
if (event.error === 'not-allowed' || event.error === 'service-not-allowed') { | |
startButton.disabled = false; | |
stopButton.disabled = true; | |
} | |
}; | |
} | |
</script> | |
</body> | |
</html> | |
""" | |
st.components.v1.html(html, height=400) | |
with col2: | |
# Display transcript history | |
st.subheader("Transcript History") | |
transcript_text = st.empty() | |
# Save transcript function | |
def save_transcript(text): | |
if not os.path.exists('transcripts'): | |
os.makedirs('transcripts') | |
timestamp = datetime.datetime.now().strftime('%Y%m%d_%H%M%S') | |
filename = f"transcripts/transcript_{timestamp}.md" | |
with open(filename, 'w', encoding='utf-8') as f: | |
f.write(text) | |
# Display full transcript | |
if st.session_state.transcript_history: | |
full_transcript = "\n".join(st.session_state.transcript_history) | |
transcript_text.text_area("Full Transcript", value=full_transcript, height=300) | |
# Save and download buttons | |
col1, col2 = st.columns(2) | |
with col1: | |
if st.button("Save Transcript"): | |
save_transcript(full_transcript) | |
st.success("Transcript saved!") | |
with col2: | |
st.download_button( | |
label="Download Transcript", | |
data=full_transcript, | |
file_name=f"transcript_{datetime.datetime.now().strftime('%Y%m%d_%H%M%S')}.md", | |
mime="text/markdown" | |
) | |
# Listen for transcript updates | |
if st.session_state.get('transcriptUpdate'): | |
new_text = st.session_state.transcriptUpdate | |
st.session_state.transcript_history.append(new_text) | |
st.session_state.transcriptUpdate = None # Clear the update | |
st.experimental_rerun() |