--- title: InternViT Development Test emoji: 🔧 colorFrom: indigo colorTo: purple sdk: docker pinned: false --- # InternViT-6B with CUDA Development Tools This Space uses the PyTorch CUDA development image to properly install flash-attn with NVCC. ## Changes in this Version - Using PyTorch CUDA development image instead of runtime image - Includes NVCC (NVIDIA CUDA Compiler) needed for flash-attn - Specific flash-attn version (1.0.9) compatible with CUDA 11.7 - Enhanced diagnostics to verify flash-attn installation ## Dependencies Added - einops: Required for vision transformer operations - flash-attn: Required for efficient attention computation - CUDA build tools for proper compilation ## Instructions 1. Click the "Test Model Loading" button 2. Wait for the model to load and run the test 3. Check the results for success or errors