Spaces:
Running
Running
import { memo, useCallback, useEffect, useRef, useState } from 'react'; | |
import { toast } from 'react-toastify'; | |
interface ScreenshotSelectorProps { | |
isSelectionMode: boolean; | |
setIsSelectionMode: (mode: boolean) => void; | |
containerRef: React.RefObject<HTMLElement>; | |
} | |
export const ScreenshotSelector = memo( | |
({ isSelectionMode, setIsSelectionMode, containerRef }: ScreenshotSelectorProps) => { | |
const [isCapturing, setIsCapturing] = useState(false); | |
const [selectionStart, setSelectionStart] = useState<{ x: number; y: number } | null>(null); | |
const [selectionEnd, setSelectionEnd] = useState<{ x: number; y: number } | null>(null); | |
const mediaStreamRef = useRef<MediaStream | null>(null); | |
const videoRef = useRef<HTMLVideoElement | null>(null); | |
useEffect(() => { | |
// Cleanup function to stop all tracks when component unmounts | |
return () => { | |
if (videoRef.current) { | |
videoRef.current.pause(); | |
videoRef.current.srcObject = null; | |
videoRef.current.remove(); | |
videoRef.current = null; | |
} | |
if (mediaStreamRef.current) { | |
mediaStreamRef.current.getTracks().forEach((track) => track.stop()); | |
mediaStreamRef.current = null; | |
} | |
}; | |
}, []); | |
const initializeStream = async () => { | |
if (!mediaStreamRef.current) { | |
try { | |
const stream = await navigator.mediaDevices.getDisplayMedia({ | |
audio: false, | |
video: { | |
displaySurface: 'window', | |
preferCurrentTab: true, | |
surfaceSwitching: 'include', | |
systemAudio: 'exclude', | |
}, | |
} as MediaStreamConstraints); | |
// Add handler for when sharing stops | |
stream.addEventListener('inactive', () => { | |
if (videoRef.current) { | |
videoRef.current.pause(); | |
videoRef.current.srcObject = null; | |
videoRef.current.remove(); | |
videoRef.current = null; | |
} | |
if (mediaStreamRef.current) { | |
mediaStreamRef.current.getTracks().forEach((track) => track.stop()); | |
mediaStreamRef.current = null; | |
} | |
setIsSelectionMode(false); | |
setSelectionStart(null); | |
setSelectionEnd(null); | |
setIsCapturing(false); | |
}); | |
mediaStreamRef.current = stream; | |
// Initialize video element if needed | |
if (!videoRef.current) { | |
const video = document.createElement('video'); | |
video.style.opacity = '0'; | |
video.style.position = 'fixed'; | |
video.style.pointerEvents = 'none'; | |
video.style.zIndex = '-1'; | |
document.body.appendChild(video); | |
videoRef.current = video; | |
} | |
// Set up video with the stream | |
videoRef.current.srcObject = stream; | |
await videoRef.current.play(); | |
} catch (error) { | |
console.error('Failed to initialize stream:', error); | |
setIsSelectionMode(false); | |
toast.error('Failed to initialize screen capture'); | |
} | |
} | |
return mediaStreamRef.current; | |
}; | |
const handleCopySelection = useCallback(async () => { | |
if (!isSelectionMode || !selectionStart || !selectionEnd || !containerRef.current) { | |
return; | |
} | |
setIsCapturing(true); | |
try { | |
const stream = await initializeStream(); | |
if (!stream || !videoRef.current) { | |
return; | |
} | |
// Wait for video to be ready | |
await new Promise((resolve) => setTimeout(resolve, 300)); | |
// Create temporary canvas for full screenshot | |
const tempCanvas = document.createElement('canvas'); | |
tempCanvas.width = videoRef.current.videoWidth; | |
tempCanvas.height = videoRef.current.videoHeight; | |
const tempCtx = tempCanvas.getContext('2d'); | |
if (!tempCtx) { | |
throw new Error('Failed to get temporary canvas context'); | |
} | |
// Draw the full video frame | |
tempCtx.drawImage(videoRef.current, 0, 0); | |
// Calculate scale factor between video and screen | |
const scaleX = videoRef.current.videoWidth / window.innerWidth; | |
const scaleY = videoRef.current.videoHeight / window.innerHeight; | |
// Get window scroll position | |
const scrollX = window.scrollX; | |
const scrollY = window.scrollY + 40; | |
// Get the container's position in the page | |
const containerRect = containerRef.current.getBoundingClientRect(); | |
// Offset adjustments for more accurate clipping | |
const leftOffset = -9; // Adjust left position | |
const bottomOffset = -14; // Adjust bottom position | |
// Calculate the scaled coordinates with scroll offset and adjustments | |
const scaledX = Math.round( | |
(containerRect.left + Math.min(selectionStart.x, selectionEnd.x) + scrollX + leftOffset) * scaleX, | |
); | |
const scaledY = Math.round( | |
(containerRect.top + Math.min(selectionStart.y, selectionEnd.y) + scrollY + bottomOffset) * scaleY, | |
); | |
const scaledWidth = Math.round(Math.abs(selectionEnd.x - selectionStart.x) * scaleX); | |
const scaledHeight = Math.round(Math.abs(selectionEnd.y - selectionStart.y) * scaleY); | |
// Create final canvas for the cropped area | |
const canvas = document.createElement('canvas'); | |
canvas.width = Math.round(Math.abs(selectionEnd.x - selectionStart.x)); | |
canvas.height = Math.round(Math.abs(selectionEnd.y - selectionStart.y)); | |
const ctx = canvas.getContext('2d'); | |
if (!ctx) { | |
throw new Error('Failed to get canvas context'); | |
} | |
// Draw the cropped area | |
ctx.drawImage(tempCanvas, scaledX, scaledY, scaledWidth, scaledHeight, 0, 0, canvas.width, canvas.height); | |
// Convert to blob | |
const blob = await new Promise<Blob>((resolve, reject) => { | |
canvas.toBlob((blob) => { | |
if (blob) { | |
resolve(blob); | |
} else { | |
reject(new Error('Failed to create blob')); | |
} | |
}, 'image/png'); | |
}); | |
// Create a FileReader to convert blob to base64 | |
const reader = new FileReader(); | |
reader.onload = (e) => { | |
const base64Image = e.target?.result as string; | |
// Find the textarea element | |
const textarea = document.querySelector('textarea'); | |
if (textarea) { | |
// Get the setters from the BaseChat component | |
const setUploadedFiles = (window as any).__BOLT_SET_UPLOADED_FILES__; | |
const setImageDataList = (window as any).__BOLT_SET_IMAGE_DATA_LIST__; | |
const uploadedFiles = (window as any).__BOLT_UPLOADED_FILES__ || []; | |
const imageDataList = (window as any).__BOLT_IMAGE_DATA_LIST__ || []; | |
if (setUploadedFiles && setImageDataList) { | |
// Update the files and image data | |
const file = new File([blob], 'screenshot.png', { type: 'image/png' }); | |
setUploadedFiles([...uploadedFiles, file]); | |
setImageDataList([...imageDataList, base64Image]); | |
toast.success('Screenshot captured and added to chat'); | |
} else { | |
toast.error('Could not add screenshot to chat'); | |
} | |
} | |
}; | |
reader.readAsDataURL(blob); | |
} catch (error) { | |
console.error('Failed to capture screenshot:', error); | |
toast.error('Failed to capture screenshot'); | |
if (mediaStreamRef.current) { | |
mediaStreamRef.current.getTracks().forEach((track) => track.stop()); | |
mediaStreamRef.current = null; | |
} | |
} finally { | |
setIsCapturing(false); | |
setSelectionStart(null); | |
setSelectionEnd(null); | |
setIsSelectionMode(false); // Turn off selection mode after capture | |
} | |
}, [isSelectionMode, selectionStart, selectionEnd, containerRef, setIsSelectionMode]); | |
const handleSelectionStart = useCallback( | |
(e: React.MouseEvent) => { | |
e.preventDefault(); | |
e.stopPropagation(); | |
if (!isSelectionMode) { | |
return; | |
} | |
const rect = e.currentTarget.getBoundingClientRect(); | |
const x = e.clientX - rect.left; | |
const y = e.clientY - rect.top; | |
setSelectionStart({ x, y }); | |
setSelectionEnd({ x, y }); | |
}, | |
[isSelectionMode], | |
); | |
const handleSelectionMove = useCallback( | |
(e: React.MouseEvent) => { | |
e.preventDefault(); | |
e.stopPropagation(); | |
if (!isSelectionMode || !selectionStart) { | |
return; | |
} | |
const rect = e.currentTarget.getBoundingClientRect(); | |
const x = e.clientX - rect.left; | |
const y = e.clientY - rect.top; | |
setSelectionEnd({ x, y }); | |
}, | |
[isSelectionMode, selectionStart], | |
); | |
if (!isSelectionMode) { | |
return null; | |
} | |
return ( | |
<div | |
className="absolute inset-0 cursor-crosshair" | |
onMouseDown={handleSelectionStart} | |
onMouseMove={handleSelectionMove} | |
onMouseUp={handleCopySelection} | |
onMouseLeave={() => { | |
if (selectionStart) { | |
setSelectionStart(null); | |
} | |
}} | |
style={{ | |
backgroundColor: isCapturing ? 'transparent' : 'rgba(0, 0, 0, 0.1)', | |
userSelect: 'none', | |
WebkitUserSelect: 'none', | |
pointerEvents: 'all', | |
opacity: isCapturing ? 0 : 1, | |
zIndex: 50, | |
transition: 'opacity 0.1s ease-in-out', | |
}} | |
> | |
{selectionStart && selectionEnd && !isCapturing && ( | |
<div | |
className="absolute border-2 border-blue-500 bg-blue-200 bg-opacity-20" | |
style={{ | |
left: Math.min(selectionStart.x, selectionEnd.x), | |
top: Math.min(selectionStart.y, selectionEnd.y), | |
width: Math.abs(selectionEnd.x - selectionStart.x), | |
height: Math.abs(selectionEnd.y - selectionStart.y), | |
}} | |
/> | |
)} | |
</div> | |
); | |
}, | |
); | |