Spaces:
Runtime error
A newer version of the Gradio SDK is available:
5.23.3
:computer: CLI Demo
This cli demo allows you to pass in more options and control the model in a fine-grained way, suitable for power users and academic researchers. An examplar command line looks as simple as
python demo.py --data_path <data_path> [additional arguments]
We discuss here first some key attributes:
Procedural Two-Pass Sampling
: We recommend enabling procedural sampling by setting--use_traj_prior True --chunk_strategy <chunk_strategy>
with<chunk_strategy>
set according to the type of the task.Resolution and Aspect-Ratio
: Default image preprocessing include center cropping. All input and output are square images of size $576\times 576$. To overwrite, the code support to pass in--W <W> --H <H>
directly. We recommend passing in--L_short 576
such that the aspect-ratio of original image is kept while the shortest side will be resized to $576$.
Task
Before diving into the command lines, we introduce Task
(specified by --task <task>
) to bucket different usage cases depending on the data constraints in input and output domains (e.g., if the ordering is available).
Task | Type of NVS | Format of <data_path> |
Target Views Sorted? | Input and Target Views Sorted? | Recommended Usage |
---|---|---|---|---|---|
img2img |
set NVS | folder (parsable by ReconfusionParser ) |
:x: | :x: | evaluation, benchmarking |
img2vid |
trajectory NVS | folder (parsable by ReconfusionParser ) |
:white_check_mark: | :white_check_mark: | evaluation, benchmarking |
img2trajvid_s-prob |
trajectory NVS | single image | :white_check_mark: | :white_check_mark: | general |
img2trajvid |
trajectory NVS | folder (parsable by ReconfusionParser ) |
:white_check_mark: | :x: | general |
Format of <data_path>
For img2trajvid_s-prob
task, we are generating a trajectory video following preset camera motions or effects given only one input image, the data format as simple as
<data_path>/
βββ scene_1.png
βββ scene_2.png
βββ scene_3.png
For all the other tasks, we use a folder for each scene that is parsable by ReconfusionParser
(see seva/data_io.py
). It contains (1) a subdirectory containing all views; (2) transforms.json
defining the intrinsics and extrinsics (OpenGL convention) for each image; and (3) train_test_split_*.json
file splitting the input and target views, with *
indicating the number of the input views.
We provide in this release (assets_demo_cli.zip
) several examplar scenes for you to take reference from. Target views is available if you the data are from academic sources, but in the case where target views is unavailble, we will create dummy black images as placeholders (e.g., the garden_flythrough
scene). The general data structure follows
<data_path>/
βββ scene_1/
βββ train_test_split_1.json # for single-view regime
βββ train_test_split_6.json # for sparse-veiw regime
βββ train_test_split_32.json # for semi-dense-view regime
βββ transforms.json
βββ images/
βββ image_0.png
βββ image_1.png
βββ ...
βββ image_1000.png
βββ scene_2
βββ scene_3
You can specify which scene to run by passing in --data_items scene_1,scene_2
to run, for example, scene_1
and scene_2
.
Recommended Usage
img2img
andimg2vid
are recommended to be used for evaluation and benchmarking. These two tasks are used for the quantitative evalution in our paper. The data is converted from academic datasets so the groundtruth target views are available for metric computation. Check thebenchmark
folder for detailed splits we organize to benchmark different NVS models.img2vid
requries both the input and target views to be sorted, which is usually not guaranteed in general usage.img2trajvid_s-prob
is for general usage but only for single-view regime and fixed preset camera control.img2trajvid
is the task designed for general usage since it does not need the ordering of the input views. This is the task used in the gradio demo.
Next we go over all tasks and provide for each task an examplar command line.
img2img
python demo.py \
--data_path <data_path> \
--num_inputs <P> \
--video_save_fps 10
--num_inputs <P>
is only necessary if there are multipletrain_test_split_*.json
files in the scene folder.- The above command works for the dataset without trajectory prior (e.g., DL3DV-140). When the trajectory prior is available given a benchmarking dataset, for example,
orbit
trajectory prior for the CO3D dataset, we use thenearest-gt
chunking strategy by setting--use_traj_prior True --traj_prior orbit --chunking_strategy nearest-gt
. We find this leads to more 3D consistent results. - For all the single-view conditioning test scenarios: we set
--camera_scale <camera_scale>
with<camera_scale>
sweeping 20 different camera scales0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2.0
. - In single-view regime for the RealEstate10K dataset, we find increasing
cfg
is helpful: we additionally set--cfg 6.0
(cfg
is2.0
by default). - For the evaluation in semi-dense-view regime (i.e., DL3DV-140 and Tanks and Temples dataset) with
32
input views, we zero-shot extendT
to fit all input and target views in one forward. Specifically, we set--T 90
for the DL3DV-140 dataset and--T 80
for the Tanks and Temples dataset. - For the evaluation on ViewCrafter split (including the RealEastate10K, CO3D, and Tanks and Temples dataset), we find zero-shot extending
T
to25
to fit all input and target views in one forward is better. Also, the V split uses the original image resolutions: we therefore set--T 25 --L_short 576
.
For example, you can run the following command on the example dl3d140-165f5af8bfe32f70595a1c9393a6e442acf7af019998275144f605b89a306557
with 3 input views:
python demo.py \
--data_path /path/to/assets_demo_cli/ \
--data_items dl3d140-165f5af8bfe32f70595a1c9393a6e442acf7af019998275144f605b89a306557 \
--num_inputs 3 \
--video_save_fps 10
img2vid
python demo.py \
--data_path <data_path> \
--task img2vid \
--replace_or_include_input True \
--num_inputs <P> \
--use_traj_prior True \
--chunk_strategy interp \
--replace_or_include_input True
is necessary here since input views and target views are mutually exclusive, forming a trajectory together in this task, so we need to append back the input views to the generated target views.--num_inputs <P>
is only necessary if there are multipletrain_test_split_*.json
files in the scene folder.- We use
interp
chunking strategy by default. - For the evaluation on ViewCrafter split (including the RealEastate10K, CO3D, and Tanks and Temples dataset), we find zero-shot extending
T
to25
to fit all input and target views in one forward is better. Also, the V split uses the original image resolutions: we therefore set--T 25 --L_short 576
.
img2trajvid_s-prob
python demo.py \
--data_path <data_path> \
--task img2trajvid_s-prob \
--replace_or_include_input True \
--traj_prior orbit \
--cfg 4.0,2.0 \
--guider 1,2 \
--num_targets 111 \
--L_short 576 \
--use_traj_prior True \
--chunk_strategy interp
--replace_or_include_input True
is necessary here since input views and target views are mutually exclusive, forming a trajectory together in this task, so we need to append back the input views to the generated target views.- Default
cfg
should be adusted according totraj_prior
. - Default chunking strategy is
interp
. - Default guider is
--guider 1,2
(instead of1
,1
still works but1,2
is slightly better). camera_scale
(default is2.0
) can be adjusted according totraj_prior
. The model has scale ambiguity with single-view input, especially for panning motions. We encourage to tune upcamera_scale
to10.0
for all panning motions (--traj_prior pan-*/dolly*
) if you expect a larger camera motion.
img2trajvid
Sparse-view regime ($P\leq 8$)
python demo.py \
--data_path <data_path> \
--task img2trajvid \
--num_inputs <P> \
--cfg 3.0,2.0 \
--use_traj_prior True \
--chunk_strategy interp-gt
--num_inputs <P>
is only necessary if there are multipletrain_test_split_*.json
files in the scene folder.- Default
cfg
should be set to3,2
(3
beingcfg
for the first pass, and2
being thecfg
for the second pass). Try to increase thecfg
for the first pass from3
to higher values if you observe blurry areas (usually happens for harder scenes with a fair amount of unseen regions). - Default chunking strategy should be set to
interp+gt
(instead ofinterp
,interp
can work but usually a bit worse). - The
--chunk_strategy_first_pass
is set asgt-nearest
by default. So it can automatically adapt when $P$ is large (up to a thousand frames).
Semi-dense-view regime ($P>9$)
python demo.py \
--data_path <data_path> \
--task img2trajvid \
--num_inputs <P> \
--cfg 3.0 \
--L_short 576 \
--use_traj_prior True \
--chunk_strategy interp
--num_inputs <P>
is only necessary if there are multipletrain_test_split_*.json
files in the scene folder.- Default
cfg
should be set to3
. - Default chunking strategy should be set to
interp
(instead ofinterp-gt
,interp-gt
is also supported but the results do not look good). T
can be overwritten by--T <N>,21
(X being extendedT
for the first pass, and21
being the defaultT
for the second pass).<N>
is dynamically decided now in the code but can also be manually updated. This is useful when you observe that there exist two very dissimilar adjacent anchors which make the interpolation in the second pass impossible. There exist two ways:--T 96,21
: this overwrites theT
in the first pass to be exactly96
.--num_prior_frames_ratio 1.2
: this enlarges T in the first pass dynamically to be1.2
$\times$ larger.