|
--- |
|
license: mit |
|
--- |
|
|
|
## Bibtex |
|
``` |
|
@article{greff2021kubric, |
|
title = {Kubric: a scalable dataset generator}, |
|
author = {Klaus Greff and Francois Belletti and Lucas Beyer and Carl Doersch and |
|
Yilun Du and Daniel Duckworth and David J Fleet and Dan Gnanapragasam and |
|
Florian Golemo and Charles Herrmann and Thomas Kipf and Abhijit Kundu and |
|
Dmitry Lagun and Issam Laradji and Hsueh-Ti (Derek) Liu and Henning Meyer and |
|
Yishu Miao and Derek Nowrouzezahrai and Cengiz Oztireli and Etienne Pot and |
|
Noha Radwan and Daniel Rebain and Sara Sabour and Mehdi S. M. Sajjadi and Matan Sela and |
|
Vincent Sitzmann and Austin Stone and Deqing Sun and Suhani Vora and Ziyu Wang and |
|
Tianhao Wu and Kwang Moo Yi and Fangcheng Zhong and Andrea Tagliasacchi}, |
|
booktitle = {Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, |
|
year = {2022}, |
|
} |
|
``` |
|
|
|
# Kubric |
|
|
|
A data generation pipeline for creating semi-realistic synthetic multi-object |
|
videos with rich annotations such as instance segmentation masks, depth maps, |
|
and optical flow. |
|
|
|
## Motivation and design |
|
We need better data for training and evaluating machine learning systems, especially in the collntext of unsupervised multi-object video understanding. |
|
Current systems succeed on [toy datasets](https://github.com/deepmind/multi_object_datasets), but fail on real-world data. |
|
Progress could be greatly accelerated if we had the ability to create suitable datasets of varying complexity on demand. |
|
Kubric is mainly built on-top of pybullet (for physics simulation) and Blender (for rendering); however, the code is kept modular to potentially support different rendering backends. |
|
|
|
## Getting started |
|
For instructions, please refer to [https://kubric.readthedocs.io](https://kubric.readthedocs.io) |
|
|
|
Assuming you have docker installed, to generate the data above simply execute: |
|
``` |
|
git clone https://github.com/google-research/kubric.git |
|
cd kubric |
|
docker pull kubricdockerhub/kubruntu |
|
docker run --rm --interactive \ |
|
--user $(id -u):$(id -g) \ |
|
--volume "$(pwd):/kubric" \ |
|
kubricdockerhub/kubruntu \ |
|
/usr/bin/python3 examples/helloworld.py |
|
ls output |
|
``` |
|
|
|
Kubric employs **Blender 2.93** (see [here](https://github.com/google-research/kubric/blob/01a08d274234f32f2adc4f7d5666b39490f953ad/docker/Blender.Dockerfile#L48)), so if you want to inspect the generated `*.blend` scene file for interactive inspection (i.e. without needing to render the scene), please make sure you have installed the correct Blender version. |
|
|
|
## Requirements |
|
- A pipeline for conveniently generating video data. |
|
- Physics simulation for automatically generating physical interactions between multiple objects. |
|
- Good control over the complexity of the generated data, so that we can evaluate individual aspects such as variability of objects and textures. |
|
- Realism: Ideally, the ability to span the entire complexity range from CLEVR all the way to real-world video such as YouTube8. This is clearly not feasible, but we would like to get as close as possible. |
|
- Access to rich ground truth information about the objects in a scene for the purpose of evaluation (eg. object segmentations and properties) |
|
- Control the train/test split to evaluate compositionality and systematic generalization (for example on held-out combinations of features or objects) |
|
|
|
|
|
## Challenges and datasets |
|
Generally, we store datasets for the challenges in this [Google Cloud Bucket](https://console.cloud.google.com/storage/browser/kubric-public). |
|
More specifically, these challenges are *dataset contributions* of the Kubric CVPR'22 paper: |
|
* [MOVi: Multi-Object Video](challenges/movi) |
|
* [Texture-Structure in NeRF](challenges/texture_structure_nerf) |
|
* [Optical Flow](challenges/optical_flow) |
|
* [Pre-training Visual Representations](challenges/pretraining_visual) |
|
* [Robust NeRF](challenges/robust_nerf) |
|
* [Multi-View Object Matting](challenges/multiview_matting) |
|
* [Complex BRDFs](challenges/complex_brdf) |
|
* [Single View Reconstruction](challenges/single_view_reconstruction) |
|
* [Video Based Reconstruction](challenges/video_based_reconstruction) |
|
* [Point Tracking](challenges/point_tracking) |
|
|
|
Pointers to additional datasets/workers: |
|
* [ToyBox (from Neural Semantic Fields)](https://nesf3d.github.io) |
|
* [MultiShapeNet (from Scene Representation Transformer)](https://srt-paper.github.io) |
|
* [SyntheticTrio(from Controllable Neural Radiance Fields)](https://github.com/kacperkan/conerf-kubric-dataset#readme) |
|
|
|
## Disclaimer |
|
This is not an official Google Product |
|
|