Fix README typos
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ license: cc-by-4.0
|
|
8 |
|
9 |
The [Monado SLAM datasets
|
10 |
(MSD)](https://huggingface.co/datasets/collabora/monado-slam-datasets), are
|
11 |
-
egocentric visual-inertial SLAM datasets recorded
|
12 |
[Basalt](https://gitlab.com/VladyslavUsenko/basalt)-based inside-out tracking
|
13 |
component of the [Monado](https://monado.dev) project. These have a permissive
|
14 |
license [CC-BY 4.0](http://creativecommons.org/licenses/by/4.0/), meaning you
|
@@ -27,13 +27,13 @@ open-source tracking solution requires a solid measurement pipeline to
|
|
27 |
understand how changes in the system affect tracking quality. For this reason,
|
28 |
the creation of these datasets was essential.
|
29 |
|
30 |
-
These datasets are very specific to the XR use
|
31 |
-
footage recorded from devices such as VR headsets but other devices like phones
|
32 |
or AR glasses might be added in the future. These were made since current SLAM
|
33 |
datasets like EuRoC or TUM-VI were not specific enough for XR, or they didn't
|
34 |
have permissively enough usage licenses.
|
35 |
|
36 |
-
For questions or comments you can use the Hugging Face
|
37 |
[Community](https://huggingface.co/datasets/collabora/monado-slam-datasets/discussions),
|
38 |
join Monado's discord [server](https://discord.gg/8RkJgRJ) and ask in the
|
39 |
`#slam` channel, or send an email to <[email protected]>.
|
@@ -76,7 +76,7 @@ join Monado's discord [server](https://discord.gg/8RkJgRJ) and ask in the
|
|
76 |
- [MIO15_moving_person_props](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIO_others/MIO15_moving_person_props.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIO15_moving_person_props.webm" type="video/webm"/>Video tag not supported.</video></details>
|
77 |
- [MIO16_moving_screens_person_props](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIO_others/MIO16_moving_screens_person_props.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIO16_moving_screens_person_props.webm" type="video/webm"/>Video tag not supported.</video></details>
|
78 |
- [MIP_playing](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing)
|
79 |
-
- [
|
80 |
- [MIPB01_beatsaber_100bills_360_normal](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB01_beatsaber_100bills_360_normal.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB01_beatsaber_100bills_360_normal.webm" type="video/webm"/>Video tag not supported.</video></details>
|
81 |
- [MIPB02_beatsaber_crabrave_360_hard](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB02_beatsaber_crabrave_360_hard.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB02_beatsaber_crabrave_360_hard.webm" type="video/webm"/>Video tag not supported.</video></details>
|
82 |
- [MIPB03_beatsaber_countryrounds_360_expert](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB03_beatsaber_countryrounds_360_expert.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB03_beatsaber_countryrounds_360_expert.webm" type="video/webm"/>Video tag not supported.</video></details>
|
@@ -100,48 +100,48 @@ join Monado's discord [server](https://discord.gg/8RkJgRJ) and ask in the
|
|
100 |
## Valve Index datasets
|
101 |
|
102 |
These datasets were recorded using a Valve Index with the `vive` driver in
|
103 |
-
Monado and they have
|
104 |
the proprietary OpenVR implementation provided by SteamVR. The exact commit used
|
105 |
in Monado at the time of recording is
|
106 |
[a4e7765d](https://gitlab.freedesktop.org/mateosss/monado/-/commit/a4e7765d7219b06a0c801c7bb33f56d3ea69229d).
|
107 |
The datasets are in the ASL dataset format, the same as the [EuRoC
|
108 |
datasets](https://projects.asl.ethz.ch/datasets/doku.php?id=kmavvisualinertialdatasets).
|
109 |
-
Besides the main EuRoC format files we provide some extra files with raw
|
110 |
-
timestamp data for exploring
|
111 |
|
112 |
-
The dataset is
|
113 |
-
from SLAM systems: camera-IMU and
|
114 |
-
alignment and bias calibration
|
115 |
-
been converted to IMU pose and so on. Most of the post
|
116 |
Basalt
|
117 |
[calibration](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Calibration.md?ref_type=heads#camera-imu-mocap-calibration)
|
118 |
and
|
119 |
[alignment](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Realsense.md?ref_type=heads#generating-time-aligned-ground-truth)
|
120 |
tools, as well as the
|
121 |
[xrtslam-metrics](https://gitlab.freedesktop.org/mateosss/xrtslam-metrics)
|
122 |
-
scripts for Monado tracking. The
|
123 |
-
video][
|
124 |
for use starting from its raw version.
|
125 |
|
126 |
### Data
|
127 |
|
128 |
#### Camera samples
|
129 |
|
130 |
-
In the `vive` driver from Monado we don't have direct access to the camera
|
131 |
device timestamps but only to V4L2 timestamps. These are not exactly hardware
|
132 |
timestamps and have some offset with respect to the device clock in which the
|
133 |
IMU samples are timestamped.
|
134 |
|
135 |
The camera frames can be found in the `camX/data` directory as PNG files with
|
136 |
-
names corresponding to
|
137 |
-
|
138 |
-
|
139 |
-
|
140 |
-
|
141 |
-
|
142 |
-
|
143 |
-
The cameras of the Valve Index
|
144 |
-
streaming at 54fps. They have
|
145 |
Index are RGB you will find only grayscale images in these datasets. The
|
146 |
original images are provided in YUYV422 format but only the luma component is
|
147 |
stored.
|
@@ -152,33 +152,33 @@ the dataset. The resulting trajectory is then aligned with the
|
|
152 |
[`basalt_time_alignment`](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Realsense.md?ref_type=heads#generating-time-aligned-ground-truth)
|
153 |
tool that aligns the rotational velocities of the trajectory with the gyroscope
|
154 |
samples and returns the resulting offset in nanoseconds. That correction is then
|
155 |
-
applied to the dataset. Refer to the
|
156 |
-
[video][
|
157 |
|
158 |
#### IMU samples
|
159 |
|
160 |
The IMU timestamps are device timestamps, they come at about 1000Hz. We provide
|
161 |
an `imu0/data.raw.csv` file that contains the raw measurements without any axis
|
162 |
-
scale
|
163 |
-
scale
|
164 |
ignore those corrections. `imu0/data.extra.csv` contains the arrival time of the
|
165 |
IMU sample to the host computer for algorithms that want to adapt themselves to
|
166 |
-
work
|
167 |
|
168 |
-
####
|
169 |
|
170 |
-
The
|
171 |
SteamVR session providing tracking data through the OpenVR API to Monado. While
|
172 |
-
not as precise as
|
173 |
should still provide pretty good accuracy and precision close to the 1mm range.
|
174 |
There are different attempts at studying the accuracy of SteamVR tracking that
|
175 |
-
you can
|
176 |
[this](https://dl.acm.org/doi/pdf/10.1145/3463914.3463921),
|
177 |
[this](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7956487/pdf/sensors-21-01622.pdf),
|
178 |
or [this](http://doc-ok.org/?p=1478). When a tracking system gets closer to
|
179 |
-
|
180 |
|
181 |
-
The raw
|
182 |
timestamps and as such, the timestamps recorded are from when the host asks
|
183 |
OpenVR for the latest pose with a call to
|
184 |
[`GetDeviceToAbsoluteTrackingPose`](https://github.com/ValveSoftware/openvr/wiki/IVRSystem::GetDeviceToAbsoluteTrackingPose).
|
@@ -193,11 +193,13 @@ instead of this headset origin.
|
|
193 |
There are multiple calibration datasets in the
|
194 |
[`MIC_calibration`](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIC_calibration)
|
195 |
directory. There are camera-focused and IMU-focused calibration datasets. See
|
196 |
-
the
|
197 |
-
|
198 |
-
for more information on what each sequence is.
|
199 |
|
200 |
-
In the
|
|
|
|
|
201 |
|
202 |
- [`calibration.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/calibration.json):
|
203 |
Calibration file produced with the
|
@@ -208,18 +210,18 @@ In the [`MI_valve_index/extras`](https://huggingface.co/datasets/collabora/monad
|
|
208 |
[`MIC04_imucalib1`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIC_calibration/MIC04_imucalib1.zip)
|
209 |
datasets with camera-IMU time offset and IMU bias/misalignment info removed so
|
210 |
that it works with the fully the all the datasets by default which are fully
|
211 |
-
|
212 |
- [`calibration.extra.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/calibration.extra.json):
|
213 |
Same as `calibration.json` but with the cam-IMU time offset and IMU bias and
|
214 |
misalignment information filled in.
|
215 |
- [`factory.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/factory.json):
|
216 |
-
JSON exposed by the headset
|
217 |
-
|
218 |
-
|
219 |
- [`other_calibrations/`](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/extras/other_calibrations):
|
220 |
-
|
221 |
-
|
222 |
-
produced with the
|
223 |
[`basalt_calibrate`](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Calibration.md?ref_type=heads#camera-calibration)
|
224 |
tool, while the corresponding `MICXX_imucalibY` datasets use these datasets as
|
225 |
a starting point and have the `basalt_calibrate_imu` calibration results.
|
@@ -232,26 +234,26 @@ model](https://vladyslavusenko.gitlab.io/basalt-headers/classbasalt_1_1KannalaBr
|
|
232 |
with fish-eye distortion (also known as [OpenCV's
|
233 |
fish-eye](https://docs.opencv.org/3.4/db/d58/group__calib3d__fisheye.html#details)).
|
234 |
|
235 |
-
Calibrations with other camera models might be added later on, otherwise you
|
236 |
-
use the calibration sequences for custom calibrations.
|
237 |
|
238 |
##### IMU model
|
239 |
|
240 |
-
For the default `calibration.json` where all parameters are zero you can ignore
|
241 |
any model and just use the measurements present in `imu0/data.csv` directly. If
|
242 |
-
instead you want to use the raw measurements from `imu0/data.raw.csv` you will
|
243 |
need to apply the Basalt
|
244 |
[accelerometer](https://vladyslavusenko.gitlab.io/basalt-headers/classbasalt_1_1CalibAccelBias.html#details)
|
245 |
and
|
246 |
[gyroscope](https://vladyslavusenko.gitlab.io/basalt-headers/classbasalt_1_1CalibGyroBias.html#details)
|
247 |
-
models that
|
248 |
initial bias. The random walk and white noise parameters were not computed and
|
249 |
default reasonable values are used instead.
|
250 |
|
251 |
#### Post-processing walkthrough
|
252 |
|
253 |
If you are interested in understanding the step-by-step procedure of
|
254 |
-
|
255 |
[MIPB08] dataset.
|
256 |
|
257 |
[](https://www.youtube.com/watch?v=0PX_6PNwrvQ)
|
@@ -259,15 +261,15 @@ postprocessing of the dataset, below is a video detailing the procedure for the
|
|
259 |
### Sequences
|
260 |
|
261 |
- [MIC_calibration](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIC_calibration):
|
262 |
-
Calibration sequences
|
263 |
[this](https://drive.google.com/file/d/1DqKWgePodCpAKJCd_Bz-hfiEQOSnn_k0)
|
264 |
calibration target from Kalibr with the squares of the target having sides of
|
265 |
-
|
266 |
planes of both stereo cameras while others on IMU calibration properly
|
267 |
exciting all six components of the IMU.
|
268 |
-
- [MIP_playing](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIO_others):
|
269 |
-
the user is playing a particular VR game on SteamVR while
|
270 |
-
the datasets.
|
271 |
- [MIPB_beat_saber](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber):
|
272 |
This contains different songs played at different speeds. The fitbeat song
|
273 |
is one that requires a lot of head movement while [MIPB08] is a long 40min
|
@@ -276,17 +278,17 @@ postprocessing of the dataset, below is a video detailing the procedure for the
|
|
276 |
This is a shooting and music game, each dataset is a different level/song.
|
277 |
- [MIPT_thrill_of_the_fight](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPT_thrill_of_the_fight):
|
278 |
This is a boxing game.
|
279 |
-
- [MIO_others](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIO_others):
|
280 |
-
datasets that might be useful, they include
|
281 |
-
the user supposed to be playing some particular game,
|
282 |
-
inspection and scanning/mapping of the room, some very
|
283 |
-
lightweight datasets for quick testing, and some datasets with a lot
|
284 |
-
movement around the environment.
|
285 |
|
286 |
## License
|
287 |
|
288 |
This work is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</a>.
|
289 |
<a rel="license" href="http://creativecommons.org/licenses/by/4.0/"><img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by/4.0/88x31.png" /></a>
|
290 |
|
291 |
-
[
|
292 |
[MIPB08]: https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber
|
|
|
8 |
|
9 |
The [Monado SLAM datasets
|
10 |
(MSD)](https://huggingface.co/datasets/collabora/monado-slam-datasets), are
|
11 |
+
egocentric visual-inertial SLAM datasets recorded to improve the
|
12 |
[Basalt](https://gitlab.com/VladyslavUsenko/basalt)-based inside-out tracking
|
13 |
component of the [Monado](https://monado.dev) project. These have a permissive
|
14 |
license [CC-BY 4.0](http://creativecommons.org/licenses/by/4.0/), meaning you
|
|
|
27 |
understand how changes in the system affect tracking quality. For this reason,
|
28 |
the creation of these datasets was essential.
|
29 |
|
30 |
+
These datasets are very specific to the XR use case as they contain VI-SLAM
|
31 |
+
footage recorded from devices such as VR headsets, but other devices like phones
|
32 |
or AR glasses might be added in the future. These were made since current SLAM
|
33 |
datasets like EuRoC or TUM-VI were not specific enough for XR, or they didn't
|
34 |
have permissively enough usage licenses.
|
35 |
|
36 |
+
For questions or comments, you can use the Hugging Face
|
37 |
[Community](https://huggingface.co/datasets/collabora/monado-slam-datasets/discussions),
|
38 |
join Monado's discord [server](https://discord.gg/8RkJgRJ) and ask in the
|
39 |
`#slam` channel, or send an email to <[email protected]>.
|
|
|
76 |
- [MIO15_moving_person_props](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIO_others/MIO15_moving_person_props.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIO15_moving_person_props.webm" type="video/webm"/>Video tag not supported.</video></details>
|
77 |
- [MIO16_moving_screens_person_props](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIO_others/MIO16_moving_screens_person_props.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIO16_moving_screens_person_props.webm" type="video/webm"/>Video tag not supported.</video></details>
|
78 |
- [MIP_playing](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing)
|
79 |
+
- [MIPB_beat_saber](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber)
|
80 |
- [MIPB01_beatsaber_100bills_360_normal](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB01_beatsaber_100bills_360_normal.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB01_beatsaber_100bills_360_normal.webm" type="video/webm"/>Video tag not supported.</video></details>
|
81 |
- [MIPB02_beatsaber_crabrave_360_hard](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB02_beatsaber_crabrave_360_hard.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB02_beatsaber_crabrave_360_hard.webm" type="video/webm"/>Video tag not supported.</video></details>
|
82 |
- [MIPB03_beatsaber_countryrounds_360_expert](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber/MIPB03_beatsaber_countryrounds_360_expert.zip): <details style="display: inline;cursor: pointer;user-select: none"><summary>Preview 5x</summary><video width="320" height="320" controls preload="none"><source src="https://huggingface.co/datasets/collabora/monado-slam-datasets/resolve/main/M_monado_datasets/MI_valve_index/extras/previews/MIPB03_beatsaber_countryrounds_360_expert.webm" type="video/webm"/>Video tag not supported.</video></details>
|
|
|
100 |
## Valve Index datasets
|
101 |
|
102 |
These datasets were recorded using a Valve Index with the `vive` driver in
|
103 |
+
Monado and they have ground truth from 3 lighthouses tracking the headset through
|
104 |
the proprietary OpenVR implementation provided by SteamVR. The exact commit used
|
105 |
in Monado at the time of recording is
|
106 |
[a4e7765d](https://gitlab.freedesktop.org/mateosss/monado/-/commit/a4e7765d7219b06a0c801c7bb33f56d3ea69229d).
|
107 |
The datasets are in the ASL dataset format, the same as the [EuRoC
|
108 |
datasets](https://projects.asl.ethz.ch/datasets/doku.php?id=kmavvisualinertialdatasets).
|
109 |
+
Besides the main EuRoC format files, we provide some extra files with raw
|
110 |
+
timestamp data for exploring real time timestamp alignment techniques.
|
111 |
|
112 |
+
The dataset is post-processed to reduce as much as possible special treatment
|
113 |
+
from SLAM systems: camera-IMU and ground truth-IMU timestamp alignment, IMU
|
114 |
+
alignment and bias calibration have been applied, lighthouse tracked pose has
|
115 |
+
been converted to IMU pose, and so on. Most of the post-processing was done with
|
116 |
Basalt
|
117 |
[calibration](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Calibration.md?ref_type=heads#camera-imu-mocap-calibration)
|
118 |
and
|
119 |
[alignment](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Realsense.md?ref_type=heads#generating-time-aligned-ground-truth)
|
120 |
tools, as well as the
|
121 |
[xrtslam-metrics](https://gitlab.freedesktop.org/mateosss/xrtslam-metrics)
|
122 |
+
scripts for Monado tracking. The post-processing process is documented in [this
|
123 |
+
video][post-processing-video] which goes through making the [MIPB08] dataset ready
|
124 |
for use starting from its raw version.
|
125 |
|
126 |
### Data
|
127 |
|
128 |
#### Camera samples
|
129 |
|
130 |
+
In the `vive` driver from Monado, we don't have direct access to the camera
|
131 |
device timestamps but only to V4L2 timestamps. These are not exactly hardware
|
132 |
timestamps and have some offset with respect to the device clock in which the
|
133 |
IMU samples are timestamped.
|
134 |
|
135 |
The camera frames can be found in the `camX/data` directory as PNG files with
|
136 |
+
names corresponding to their V4L2 timestamps. The `camX/data.csv` file contains
|
137 |
+
aligned timestamps of each frame. The `camX/data.extra.csv` also contains the
|
138 |
+
original V4L2 timestamp and the "host timestamp" which is the time at which the
|
139 |
+
host computer had the frame ready to use after USB transmission. By separating
|
140 |
+
arrival time and exposure time algorithms can be made to be more robust for
|
141 |
+
real time operation.
|
142 |
+
|
143 |
+
The cameras of the Valve Index have global shutters with a resolution of 960×960
|
144 |
+
streaming at 54fps. They have auto exposure enabled. While the cameras of the
|
145 |
Index are RGB you will find only grayscale images in these datasets. The
|
146 |
original images are provided in YUYV422 format but only the luma component is
|
147 |
stored.
|
|
|
152 |
[`basalt_time_alignment`](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Realsense.md?ref_type=heads#generating-time-aligned-ground-truth)
|
153 |
tool that aligns the rotational velocities of the trajectory with the gyroscope
|
154 |
samples and returns the resulting offset in nanoseconds. That correction is then
|
155 |
+
applied to the dataset. Refer to the post-processing walkthrough
|
156 |
+
[video][post-processing-video] for more details.
|
157 |
|
158 |
#### IMU samples
|
159 |
|
160 |
The IMU timestamps are device timestamps, they come at about 1000Hz. We provide
|
161 |
an `imu0/data.raw.csv` file that contains the raw measurements without any axis
|
162 |
+
scale misalignment o bias correction. `imu0/data.csv` has the
|
163 |
+
scale misalignment and bias corrections applied so that the SLAM system can
|
164 |
ignore those corrections. `imu0/data.extra.csv` contains the arrival time of the
|
165 |
IMU sample to the host computer for algorithms that want to adapt themselves to
|
166 |
+
work in real time.
|
167 |
|
168 |
+
#### Ground truth information
|
169 |
|
170 |
+
The ground truth setup consists of three lighthouses 2.0 base stations and a
|
171 |
SteamVR session providing tracking data through the OpenVR API to Monado. While
|
172 |
+
not as precise as other MoCap tracking systems like OptiTrack or Vicon it
|
173 |
should still provide pretty good accuracy and precision close to the 1mm range.
|
174 |
There are different attempts at studying the accuracy of SteamVR tracking that
|
175 |
+
you can check out like
|
176 |
[this](https://dl.acm.org/doi/pdf/10.1145/3463914.3463921),
|
177 |
[this](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7956487/pdf/sensors-21-01622.pdf),
|
178 |
or [this](http://doc-ok.org/?p=1478). When a tracking system gets closer to
|
179 |
+
millimeter accuracy these datasets will no longer be as useful for improving it.
|
180 |
|
181 |
+
The raw ground truth data is stored in `gt/data.raw.csv`. OpenVR does not provide
|
182 |
timestamps and as such, the timestamps recorded are from when the host asks
|
183 |
OpenVR for the latest pose with a call to
|
184 |
[`GetDeviceToAbsoluteTrackingPose`](https://github.com/ValveSoftware/openvr/wiki/IVRSystem::GetDeviceToAbsoluteTrackingPose).
|
|
|
193 |
There are multiple calibration datasets in the
|
194 |
[`MIC_calibration`](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIC_calibration)
|
195 |
directory. There are camera-focused and IMU-focused calibration datasets. See
|
196 |
+
the
|
197 |
+
[README.md](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIC_calibration/README.md)
|
198 |
+
file in there for more information on what each sequence is.
|
199 |
|
200 |
+
In the
|
201 |
+
[`MI_valve_index/extras`](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/extras)
|
202 |
+
directory you can find the following files:
|
203 |
|
204 |
- [`calibration.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/calibration.json):
|
205 |
Calibration file produced with the
|
|
|
210 |
[`MIC04_imucalib1`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/MIC_calibration/MIC04_imucalib1.zip)
|
211 |
datasets with camera-IMU time offset and IMU bias/misalignment info removed so
|
212 |
that it works with the fully the all the datasets by default which are fully
|
213 |
+
post-processed and don't require those fields.
|
214 |
- [`calibration.extra.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/calibration.extra.json):
|
215 |
Same as `calibration.json` but with the cam-IMU time offset and IMU bias and
|
216 |
misalignment information filled in.
|
217 |
- [`factory.json`](https://huggingface.co/datasets/collabora/monado-slam-datasets/blob/main/M_monado_datasets/MI_valve_index/extras/factory.json):
|
218 |
+
JSON file exposed by the headset's firmware with information of the device. It
|
219 |
+
includes camera and display calibration as well as more data that might be of
|
220 |
+
interest. It is not used but included for completeness' sake.
|
221 |
- [`other_calibrations/`](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/extras/other_calibrations):
|
222 |
+
Calibration results obtained from the other calibration datasets. Shown for
|
223 |
+
comparison and ensuring that all of them have similar values.
|
224 |
+
`MICXX_camcalibY` has camera-only calibration produced with the
|
225 |
[`basalt_calibrate`](https://gitlab.com/VladyslavUsenko/basalt/-/blob/master/doc/Calibration.md?ref_type=heads#camera-calibration)
|
226 |
tool, while the corresponding `MICXX_imucalibY` datasets use these datasets as
|
227 |
a starting point and have the `basalt_calibrate_imu` calibration results.
|
|
|
234 |
with fish-eye distortion (also known as [OpenCV's
|
235 |
fish-eye](https://docs.opencv.org/3.4/db/d58/group__calib3d__fisheye.html#details)).
|
236 |
|
237 |
+
Calibrations with other camera models might be added later on, otherwise, you
|
238 |
+
can use the calibration sequences for custom calibrations.
|
239 |
|
240 |
##### IMU model
|
241 |
|
242 |
+
For the default `calibration.json` where all parameters are zero, you can ignore
|
243 |
any model and just use the measurements present in `imu0/data.csv` directly. If
|
244 |
+
instead, you want to use the raw measurements from `imu0/data.raw.csv` you will
|
245 |
need to apply the Basalt
|
246 |
[accelerometer](https://vladyslavusenko.gitlab.io/basalt-headers/classbasalt_1_1CalibAccelBias.html#details)
|
247 |
and
|
248 |
[gyroscope](https://vladyslavusenko.gitlab.io/basalt-headers/classbasalt_1_1CalibGyroBias.html#details)
|
249 |
+
models that use a misalignment-scale correction matrix together with a constant
|
250 |
initial bias. The random walk and white noise parameters were not computed and
|
251 |
default reasonable values are used instead.
|
252 |
|
253 |
#### Post-processing walkthrough
|
254 |
|
255 |
If you are interested in understanding the step-by-step procedure of
|
256 |
+
post-processing of the dataset, below is a video detailing the procedure for the
|
257 |
[MIPB08] dataset.
|
258 |
|
259 |
[](https://www.youtube.com/watch?v=0PX_6PNwrvQ)
|
|
|
261 |
### Sequences
|
262 |
|
263 |
- [MIC_calibration](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIC_calibration):
|
264 |
+
Calibration sequences that record
|
265 |
[this](https://drive.google.com/file/d/1DqKWgePodCpAKJCd_Bz-hfiEQOSnn_k0)
|
266 |
calibration target from Kalibr with the squares of the target having sides of
|
267 |
+
3 cm. Some sequences are focused on camera calibration covering the image
|
268 |
planes of both stereo cameras while others on IMU calibration properly
|
269 |
exciting all six components of the IMU.
|
270 |
+
- [MIP_playing](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIO_others):
|
271 |
+
Datasets in which the user is playing a particular VR game on SteamVR while
|
272 |
+
Monado records the datasets.
|
273 |
- [MIPB_beat_saber](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber):
|
274 |
This contains different songs played at different speeds. The fitbeat song
|
275 |
is one that requires a lot of head movement while [MIPB08] is a long 40min
|
|
|
278 |
This is a shooting and music game, each dataset is a different level/song.
|
279 |
- [MIPT_thrill_of_the_fight](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPT_thrill_of_the_fight):
|
280 |
This is a boxing game.
|
281 |
+
- [MIO_others](https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIO_others):
|
282 |
+
These are other datasets that might be useful, they include play-pretend
|
283 |
+
scenarios in which the user is supposed to be playing some particular game,
|
284 |
+
then there is some inspection and scanning/mapping of the room, some very
|
285 |
+
short and lightweight datasets for quick testing, and some datasets with a lot
|
286 |
+
of movement around the environment.
|
287 |
|
288 |
## License
|
289 |
|
290 |
This work is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</a>.
|
291 |
<a rel="license" href="http://creativecommons.org/licenses/by/4.0/"><img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by/4.0/88x31.png" /></a>
|
292 |
|
293 |
+
[post-processing-video]: https://youtu.be/0PX_6PNwrvQ
|
294 |
[MIPB08]: https://huggingface.co/datasets/collabora/monado-slam-datasets/tree/main/M_monado_datasets/MI_valve_index/MIP_playing/MIPB_beat_saber
|