Spaces:
Build error
Build error
## Trouble Shooting | |
1. #### Out of memory during sampling. | |
- Possible reason: | |
- Too many high-resolution frames for parallel decoding. The default setting will request ca. 66 GB peak VARM. | |
- Try this: | |
- Reduce the number of jointly decoded frames *en_and_decode_n_samples_a_time* in `inference/vista.yaml`. | |
2. #### Get stuck at loading FrozenCLIPEmbedder or FrozenOpenCLIPImageEmbedder. | |
- Possible reason: | |
- A network failure. | |
- Try this: | |
1. Download [openai/clip-vit-large-patch14](https://huggingface.co/openai/clip-vit-large-patch14/tree/main) and [laion/CLIP-ViT-H-14-laion2B-s32B-b79K](https://huggingface.co/laion/CLIP-ViT-H-14-laion2B-s32B-b79K/tree/main) in advance. | |
2. Set *version* of FrozenCLIPEmbedder and FrozenOpenCLIPImageEmbedder in `vwm/modules/encoders/modules.py` to the new paths of `pytorch_model.bin`. | |
3. #### Datasets not yet available during training. | |
- Possible reason: | |
- The installed [sdata](https://github.com/Stability-AI/datapipelines) is not detected. | |
- Try this: | |
- Reinstall in the current project directory. | |
````shell | |
pip3 install -e git+https://github.com/Stability-AI/datapipelines.git@main#egg=sdata | |
```` | |
--- | |
<= Previous: [[Sampling](https://github.com/OpenDriveLab/Vista/blob/main/docs/SAMPLING.md)] | |