A multi-sensor traffic scene dataset with omnidirectional video

The development of vehicles that perceive their environment, in particular those using computer vision, indispensably requires large databases of sensor recordings obtained from real cars driven in realistic traffic situations. These datasets should be time shaped for enabling synchronization of sen...

Full description

Bibliographic Details
Main Authors: Koschorrek, Philipp, Piccini, Tommaso, Öberg, Per, Felsberg, Michael, Nielsen, Lars, Mester, Rudolf
Format: Others
Language:English
Published: Linköpings universitet, Datorseende 2013
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-93277
Description
Summary:The development of vehicles that perceive their environment, in particular those using computer vision, indispensably requires large databases of sensor recordings obtained from real cars driven in realistic traffic situations. These datasets should be time shaped for enabling synchronization of sensor data from different sources. Furthermore, full surround environment perception requires high frame rates of synchronized omnidirectional video data to prevent information loss at any speeds. This paper describes an experimental setup and software environment for recording such synchronized multi-sensor data streams and storing them in a new open source format. The dataset consists of sequences recorded in various environments from a car equipped with an omnidirectional multi-camera, height sensors, an IMU, a velocity sensor, and a GPS. The software environment for reading these data sets will be provided to the public, together with a collection of long multi-sensor and multi-camera data streams stored in the developed format.