whisper-realtime / index.html
freddyaboulton's picture
Upload folder using huggingface_hub
b0614ed verified
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>Real-time Whisper Transcription</title>
<style>
:root {
--primary-gradient: linear-gradient(135deg, #f9a45c 0%, #e66465 100%);
--background-cream: #faf8f5;
--background-cream-end: #f7f5f2;
/* Slightly warmer end color for body gradient */
--text-dark: #2d2d2d;
--transcript-bg: #ffffff;
/* White background for transcript area */
--transcript-border: #e0e0e0;
/* Light border for transcript items */
}
body {
font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Oxygen, Ubuntu, Cantarell, sans-serif;
margin: 0;
padding: 0;
/* Apply a subtle vertical gradient to the body */
background: linear-gradient(to bottom, var(--background-cream), var(--background-cream-end));
color: var(--text-dark);
min-height: 100vh;
}
.hero {
background: var(--primary-gradient);
color: white;
padding: 2.5rem 2rem;
text-align: center;
}
.hero h1 {
font-size: 2.5rem;
margin: 0;
font-weight: 600;
letter-spacing: -0.5px;
}
.hero p {
font-size: 1rem;
margin-top: 0.5rem;
opacity: 0.9;
}
.container {
max-width: 1000px;
margin: 2.5rem auto;
/* Increased top/bottom margin */
padding: 0 2rem;
}
.transcript-container {
border-radius: 12px;
/* Slightly larger radius */
box-shadow: 0 4px 15px rgba(0, 0, 0, 0.08);
/* Enhanced shadow */
padding: 1.5rem;
height: 350px;
/* Increased height */
overflow-y: auto;
margin-bottom: 2rem;
/* Increased margin */
border: 1px solid rgba(0, 0, 0, 0.05);
/* Softer border */
background-color: var(--transcript-bg);
/* Use the new variable */
}
.controls {
text-align: center;
margin: 1.5rem 0;
}
button {
background: var(--primary-gradient);
color: white;
border: none;
padding: 10px 20px;
font-size: 0.95rem;
border-radius: 6px;
cursor: pointer;
transition: all 0.2s ease;
font-weight: 500;
min-width: 180px;
position: relative;
padding-right: 50px;
}
button:hover {
transform: translateY(-1px);
box-shadow: 0 4px 12px rgba(230, 100, 101, 0.15);
}
button:active {
transform: translateY(0);
}
/* Transcript text styling */
.transcript-container p {
margin: 0.6rem 0;
/* Increased vertical margin */
padding: 0.8rem 1rem;
/* Increased padding */
background: var(--background-cream);
/* Use the lighter cream for contrast */
border-radius: 6px;
/* Slightly larger radius */
line-height: 1.5;
/* Improved line spacing */
font-size: 0.98rem;
/* Slightly larger font */
border-left: 3px solid var(--transcript-border);
/* Add a subtle left border */
transition: background-color 0.2s ease;
/* Smooth hover effect */
}
.transcript-container p:hover {
background-color: #fdfbf9;
/* Slightly change background on hover */
}
/* Custom scrollbar - update track color */
.transcript-container::-webkit-scrollbar {
width: 8px;
/* Slightly wider scrollbar */
}
.transcript-container::-webkit-scrollbar-track {
background: var(--background-cream-end);
/* Match body end gradient */
border-radius: 4px;
}
.transcript-container::-webkit-scrollbar-thumb {
background: #e66465;
border-radius: 3px;
opacity: 0.8;
}
.transcript-container::-webkit-scrollbar-thumb:hover {
background: #f9a45c;
}
/* Add styles for toast notifications */
.toast {
position: fixed;
top: 20px;
left: 50%;
transform: translateX(-50%);
padding: 16px 24px;
border-radius: 4px;
font-size: 14px;
z-index: 1000;
display: none;
box-shadow: 0 2px 5px rgba(0, 0, 0, 0.2);
}
.toast.error {
background-color: #f44336;
color: white;
}
.toast.warning {
background-color: #ffd700;
color: black;
}
/* Add styles for audio visualization */
.icon-with-spinner {
display: flex;
align-items: center;
justify-content: center;
gap: 12px;
min-width: 180px;
}
.spinner {
width: 20px;
height: 20px;
border: 2px solid white;
border-top-color: transparent;
border-radius: 50%;
animation: spin 1s linear infinite;
flex-shrink: 0;
}
.pulse-container {
display: flex;
align-items: center;
justify-content: center;
gap: 12px;
min-width: 180px;
}
.pulse-circle {
width: 20px;
height: 20px;
border-radius: 50%;
background-color: white;
opacity: 0.2;
flex-shrink: 0;
transform: translateX(-0%) scale(var(--audio-level, 1));
transition: transform 0.1s ease;
}
/* Styles for the mute button */
.mute-toggle {
position: absolute;
right: 10px;
top: 50%;
transform: translateY(-50%);
width: 24px;
height: 24px;
cursor: pointer;
display: flex;
align-items: center;
justify-content: center;
}
.mute-toggle svg {
width: 20px;
height: 20px;
stroke: white;
}
/* Adjust layout for button content when mute is present */
.button-content {
display: flex;
align-items: center;
justify-content: center;
width: calc(100% - 40px);
margin-right: 40px;
}
.icon-with-spinner,
.pulse-container {
width: 100%;
}
@keyframes spin {
to {
transform: rotate(360deg);
}
}
</style>
</head>
<body>
<!-- Add toast element after body opening tag -->
<div id="error-toast" class="toast"></div>
<div class="hero">
<h1>Real-time Transcription</h1>
<p>Powered by Groq and FastRTC</p>
</div>
<div class="container">
<div class="transcript-container" id="transcript">
</div>
<div class="controls">
<button id="start-button">Start Recording</button>
</div>
</div>
<script>
let peerConnection;
let webrtc_id;
let audioContext, analyser, audioSource;
let audioLevel = 0;
let animationFrame;
let isMuted = false;
const startButton = document.getElementById('start-button');
const transcriptDiv = document.getElementById('transcript');
// SVG Icons
const micIconSVG = `
<svg xmlns="http://www.w3.org/2000/svg" width="100%" height="100%" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<path d="M12 1a3 3 0 0 0-3 3v8a3 3 0 0 0 6 0V4a3 3 0 0 0-3-3z"></path>
<path d="M19 10v2a7 7 0 0 1-14 0v-2"></path>
<line x1="12" y1="19" x2="12" y2="23"></line>
<line x1="8" y1="23" x2="16" y2="23"></line>
</svg>`;
const micMutedIconSVG = `
<svg xmlns="http://www.w3.org/2000/svg" width="100%" height="100%" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round">
<path d="M12 1a3 3 0 0 0-3 3v8a3 3 0 0 0 6 0V4a3 3 0 0 0-3-3z"></path>
<path d="M19 10v2a7 7 0 0 1-14 0v-2"></path>
<line x1="12" y1="19" x2="12" y2="23"></line>
<line x1="8" y1="23" x2="16" y2="23"></line>
<line x1="1" y1="1" x2="23" y2="23"></line>
</svg>`;
function showError(message) {
const toast = document.getElementById('error-toast');
toast.textContent = message;
toast.style.display = 'block';
// Hide toast after 5 seconds
setTimeout(() => {
toast.style.display = 'none';
}, 5000);
}
async function handleMessage(event) {
// Handle any WebRTC data channel messages if needed
const eventJson = JSON.parse(event.data);
if (eventJson.type === "error") {
showError(eventJson.message);
} else if (eventJson.type === "send_input") {
const response = await fetch('/send_input', {
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({
webrtc_id: webrtc_id,
transcript: ""
})
});
}
console.log('Received message:', event.data);
}
function updateButtonState() {
// Remove existing mute listener if present
const existingMuteButton = startButton.querySelector('.mute-toggle');
if (existingMuteButton) {
existingMuteButton.removeEventListener('click', toggleMute);
existingMuteButton.remove();
}
if (peerConnection && (peerConnection.connectionState === 'connecting' || peerConnection.connectionState === 'new')) {
startButton.innerHTML = `
<div class="button-content">
<div class="icon-with-spinner">
<div class="spinner"></div>
<span>Connecting...</span>
</div>
</div>
`;
startButton.disabled = true;
} else if (peerConnection && peerConnection.connectionState === 'connected') {
startButton.innerHTML = `
<div class="button-content">
<div class="pulse-container">
<div class="pulse-circle"></div>
<span>Stop Recording</span>
</div>
</div>
<div class="mute-toggle" title="${isMuted ? 'Unmute' : 'Mute'}">
${isMuted ? micMutedIconSVG : micIconSVG}
</div>
`;
startButton.disabled = false;
const muteButton = startButton.querySelector('.mute-toggle');
if (muteButton) {
muteButton.addEventListener('click', toggleMute);
}
} else {
startButton.innerHTML = 'Start Recording';
startButton.disabled = false;
}
}
function toggleMute(event) {
event.stopPropagation();
if (!peerConnection || peerConnection.connectionState !== 'connected') return;
isMuted = !isMuted;
console.log("Mute toggled:", isMuted);
peerConnection.getSenders().forEach(sender => {
if (sender.track && sender.track.kind === 'audio') {
sender.track.enabled = !isMuted;
console.log(`Audio track ${sender.track.id} enabled: ${!isMuted}`);
}
});
updateButtonState();
}
function setupAudioVisualization(stream) {
audioContext = new (window.AudioContext || window.webkitAudioContext)();
analyser = audioContext.createAnalyser();
audioSource = audioContext.createMediaStreamSource(stream);
audioSource.connect(analyser);
analyser.fftSize = 64;
const dataArray = new Uint8Array(analyser.frequencyBinCount);
function updateAudioLevel() {
analyser.getByteFrequencyData(dataArray);
const average = Array.from(dataArray).reduce((a, b) => a + b, 0) / dataArray.length;
audioLevel = average / 255;
const pulseCircle = document.querySelector('.pulse-circle');
if (pulseCircle) {
pulseCircle.style.setProperty('--audio-level', 1 + audioLevel);
}
animationFrame = requestAnimationFrame(updateAudioLevel);
}
updateAudioLevel();
}
async function setupWebRTC() {
const config = __RTC_CONFIGURATION__;
peerConnection = new RTCPeerConnection(config);
const timeoutId = setTimeout(() => {
const toast = document.getElementById('error-toast');
toast.textContent = "Connection is taking longer than usual. Are you on a VPN?";
toast.className = 'toast warning';
toast.style.display = 'block';
// Hide warning after 5 seconds
setTimeout(() => {
toast.style.display = 'none';
}, 5000);
}, 5000);
try {
const stream = await navigator.mediaDevices.getUserMedia({
audio: true
});
setupAudioVisualization(stream);
stream.getTracks().forEach(track => {
peerConnection.addTrack(track, stream);
});
// Add connection state change listener
peerConnection.addEventListener('connectionstatechange', () => {
console.log('connectionstatechange', peerConnection.connectionState);
if (peerConnection.connectionState === 'connected') {
clearTimeout(timeoutId);
const toast = document.getElementById('error-toast');
toast.style.display = 'none';
}
updateButtonState();
});
peerConnection.onicecandidate = ({ candidate }) => {
if (candidate) {
console.debug("Sending ICE candidate", candidate);
fetch('/webrtc/offer', {
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({
candidate: candidate.toJSON(),
webrtc_id: webrtc_id,
type: "ice-candidate",
})
})
}
};
// Create data channel for messages
const dataChannel = peerConnection.createDataChannel('text');
dataChannel.onmessage = handleMessage;
// Create and send offer
const offer = await peerConnection.createOffer();
await peerConnection.setLocalDescription(offer);
webrtc_id = Math.random().toString(36).substring(7);
const response = await fetch('/webrtc/offer', {
method: 'POST',
headers: { 'Content-Type': 'application/json' },
body: JSON.stringify({
sdp: peerConnection.localDescription.sdp,
type: peerConnection.localDescription.type,
webrtc_id: webrtc_id
})
});
const serverResponse = await response.json();
if (serverResponse.status === 'failed') {
showError(serverResponse.meta.error === 'concurrency_limit_reached'
? `Too many connections. Maximum limit is ${serverResponse.meta.limit}`
: serverResponse.meta.error);
stop();
startButton.textContent = 'Start Recording';
return;
}
await peerConnection.setRemoteDescription(serverResponse);
// Create event stream to receive transcripts
const eventSource = new EventSource('/transcript?webrtc_id=' + webrtc_id);
eventSource.addEventListener("output", (event) => {
appendTranscript(event.data);
});
} catch (err) {
clearTimeout(timeoutId);
console.error('Error setting up WebRTC:', err);
showError('Failed to establish connection. Please try again.');
stop();
startButton.textContent = 'Start Recording';
}
}
function appendTranscript(text) {
const p = document.createElement('p');
p.textContent = text;
transcriptDiv.appendChild(p);
transcriptDiv.scrollTop = transcriptDiv.scrollHeight;
}
function stop() {
if (animationFrame) {
cancelAnimationFrame(animationFrame);
animationFrame = null;
}
if (audioContext) {
audioContext.close().catch(e => console.error("Error closing AudioContext:", e));
audioContext = null;
analyser = null;
audioSource = null;
}
if (peerConnection) {
if (peerConnection.getSenders) {
peerConnection.getSenders().forEach(sender => {
if (sender.track) {
sender.track.stop();
console.log(`Track ${sender.track.id} stopped.`);
}
});
}
peerConnection.close();
peerConnection = null;
console.log("Peer connection closed.");
}
audioLevel = 0;
isMuted = false;
updateButtonState();
}
startButton.addEventListener('click', (event) => {
if (event.target.closest('.mute-toggle')) {
return;
}
if (peerConnection && peerConnection.connectionState === 'connected') {
console.log("Stop button clicked");
stop();
} else if (!peerConnection || ['new', 'closed', 'failed', 'disconnected'].includes(peerConnection.connectionState)) {
console.log("Start button clicked");
transcriptDiv.innerHTML = '';
setupWebRTC();
updateButtonState();
}
});
</script>
</body>
</html>