ETH Zurich - D-INFK - IVC - CVG - Research - Illumination Change Robust Direct SLAM

Illumination Change Robustness in Direct Visual SLAM


Department of Computer Science, ETH Zurich, Switzerland

Abstract

Direct visual odometry and Simultaneous Localization and Mapping (SLAM) methods determine camera poses by means of direct image alignment. This optimizes a photometric cost term based on the Lucas-Kanade method. Many recent works use the brightness constancy assumption in the alignment cost formulation and therefore cannot cope with significant illumination changes. Such changes are especially likely to occur for loop closures in SLAM. Alternatives exist which attempt to match images more robustly. In our paper, we perform a systematic evaluation of real-time capable methods. We determine their accuracy and robustness in the context of odometry and of loop closures, both on real images as well as synthetic datasets with simulated lighting changes. We find that for real images, a Census-based method outperforms the others.

Publications

  • Illumination Change Robustness in Direct Visual SLAM (ICRA 2017) [pdf] [bib]

Datasets

We base our datasets on the TUM RGB-D benchmark and ICL-NUIM synthetic datasets with depth noise added to synthetic data as done by Choi et al..

Before You Download

We provide our real-world datasets under the Creative Commons CC0 license: To the extent possible under law, the authors have waived all copyright and related or neighboring rights to this work.

For the synthetic datasets based on the ICL-NUIM dataset by A. Handa et al., the original data is used under the CC BY 3.0 license with modifications as described in our paper.

If you use our data, please cite our publication:

@InProceedings{park2017icra,
	author = {Seonwook Park and Thomas Sch\"ops and Marc Pollefeys},
	title = "Illumination Change Robustness in Direct Visual SLAM",
	booktitle = "ICRA",
	year = "2017",
}

Also consider citing the original publication for the ICL-NUIM data.

For using the dataset, please refer to instructions for the TUM RGB-D benchmark. Ground-truth trajectory files can be found at the top-level folder of respective zip archives.

Synthetic Data

These synthetic sequences were generated with tools and trajectories provided by the original ICL-NUIM synthetic datasets with depth noise added via a Python script provided by Redwood.

ETHl/syn1 download zip 29s 880 frames
ETHl/syn1_local download zip
ETHl/syn1_global download zip
ETHl/syn1_loc_glo download zip
ETHl/syn1_flash download zip
ETHl/syn2 download zip 41s 1240 frames
ETHl/syn2_local download zip
ETHl/syn2_global download zip
ETHl/syn2_loc_glo download zip
ETHl/syn2_flash download zip

Real-World Data

These sequences were recorded with a first generation Microsoft Kinect. The format of the dataset is equivalent to that of the TUM RGB-D benchmark (no accelerometer recordings are available). Camera intrinsic parameters used are:

f_x: 538.7 [mm]
f_y: 540.7 [mm]
c_x: 319.2 [pixels]
c_y: 233.6 [pixels]

Depth images are registered via the Kinect firmware.

ETHl/real_local download zip 48s 1505 frames
ETHl/real_global download zip 47s 1416 frames
ETHl/real_flash download zip 46s 1426 frames

Acknowledgments

This work was supported by Google and Qualcomm. We thank Torsten Sattler for helpful comments.


© CVG, ETH Zürich lm@inf.ethz.ch