Navigation auf


Department of Informatics Robotics and Perception Group

Multi-FoV Datasets

The "Multi-FoV" synthetic datasets

We provide two synthetic scenes (vehicle moving in a city, and flying robot hovering in a confined room). For each scene, three different optics were used (perspective, fisheye and catadioptric), but the same sensor is used (keeping the image resolution constant). These datasets were generated using Blender, using a custom omnidirectional camera model, which we release as an open-source patch for Blender. The Blender scene for the urban canyon dataset was generated using the addon Scene City. The Blender file is available here. The Blender scene used for the room environment can be found here.


The camera calibrations, groundtruth trajectories and depth maps are provided in the archives. For any question relative to these datasets, please send an e-mail at rebecq (at) ifi (dot) uzh (dot) ch .


City environment (urban canyon)


Perspective (with depth maps)

Ground truth

Fisheye (with depth maps)
Ground truth
Catadioptric (with depth maps)
Ground truth

Perspective (with depth maps)

Ground truth


Fisheye (with depth maps)

Ground truth


Catadioptric (with depth maps)

Ground truth



This dataset and the Blender patch are released under the Creative Commons license (CC BY-NC-SA 3.0), which is free for non-commercial use (including research). 




Z. Zhang, H. Rebecq, C. Forster, D. Scaramuzza

Benefit of Large Field-of-View Cameras for Visual Odometry

IEEE International Conference on Robotics and Automation (ICRA), Stockholm, 2016.

PDF (PDF, 6 MB) YouTube

Motivation: Study the influence of the field-of-view of cameras on visual odometry

The transition of visual-odometry technology from research demonstrators to commercial applications naturally raises the question: "what is the optimal camera for vision-based motion estimation?" This question is crucial as the choice of camera has a tremendous impact on the robustness and accuracy of the employed visual odometry algorithm.


While many properties of a camera (e.g. resolution, frame-rate, global-shutter/rolling-shutter) could be considered, in this work we focus on evaluating the impact of the camera field-of-view (FoV) and optics (i.e., fisheye or catadioptric) on the quality of the motion estimate. Since the motion-estimation performance depends highly on the geometry of the scene and the motion of the camera, we analyze two common operational environments in mobile robotics: an urban environment and an indoor scene.


To confirm the theoretical observations, we implement a state-of-the-art VO pipeline that works with large FoV fisheye and catadioptric cameras. We evaluate the proposed VO pipeline in both synthetic and real experiments. The experiments point out that it is advantageous to use a large FoV camera (e.g., fisheye or catadioptric) for indoor scenes and a smaller FoV for urban canyon environments.