Abstract
This article presents a dataset recorded with a sensor-equipped research vehicle on public roads in the city of Coventry in the United Kingdom. The sensor suite includes a monocular-, infrared- and smartphone-camera, as well as a LiDAR unit, GPS receiver, smartphone sensors and vehicle CAN bus data logger. Data were collected by day and night in a variety of traffic, weather and road surface conditions with a focus on the correlation between vehicle dynamics and the environmental perception process of automated vehicles.
Keywords: Public road data, Automated vehicles, Computer vision, Vehicle dynamics, SLAM, Target state estimation, Road anomalies
Specifications Table
| Subject | Computer Vision and Pattern Recognition |
| Specific subject area | Environmental Perception of Automated Vehicles |
| Type of data | Matlab Structure (CAN-Bus, Smartphone Sensors) Image (Cameras) Point Cloud (Laser Scanner) |
| How data were acquired |
|
| Data format | Raw |
| Parameters for data collection | Data was collected with a sensor-equipped vehicle while driving naturalistically on public roads. In particular: Ego-vehicle position, speed, and acceleration. Stationary and moving obstacle detection using visual and infrared cameras, and LIDAR sensors. |
| Description of data collection | Measurements were conducted by driving naturalistically during day and night in various traffic, weather and road surface conditions. |
| Data source location | City/Town/Region: Coventry Country: United Kingdom Latitude and longitude (and GPS coordinates) for collected samples/data: GPS driving routes part of dataset |
| Data accessibility | Repository name: Zenodo Data identification number: 3383693 Direct URL to data: https://doi.org/10.5281/zenodo.3383693 |
Value of the Data
|
1. Data
The included files consist of recorded data from four different days of public road driving trials, divided into the sets of alpha, bravo, charlie and delta. Following table presents the specifications of each scenario including the traffic volume of driving (Traff. D.) and parked vehicles (Traff. P.), the weather condition (Sunny, Cloudy, Clear), the time of the day and driving environment key features of each trial. The selection of features also reflects the parametrization of driving scenarios in the simulation software IPG Carmaker. While alpha, bravo and charlie road trials include the monocular and infrared camera, LiDAR scanner, GPS receiver and CAN-bus data logger, delta road trials also consist of additional smartphone's camera and sensors measurement data. All features are summarised in Table 1.
Table 1.
General features of the data set.
| No | Length [min] | Traff. D. | Traff. P. | Weather | Time | Environmental features |
|---|---|---|---|---|---|---|
| a1 | 6.4 | High | High | Cloudy | Afternoon | Inner-city, Parking lot |
| a2 | 8.3 | Low | High | Cloudy | Afternoon | Parking lot, Country road |
| a3 | 8.3 | High | Low | Cloudy | Afternoon | Inner-city, Country road |
| a4 | 8.3 | High | Medium | Cloudy | Afternoon | Inner-city |
| b1 | 6.28 | Medium | Medium | Sunny | Noon | Inner-city |
| b2 | 16.92 | Low | High | Sunny | Noon | Residential area, Road bumps |
| b3 | 25.38 | Medium | High | Sunny | Noon | Residential area, Road bumps, Inner-city |
| c1 | 4.86 | Medium | Medium | Clear | Evening | Country road, Parking lot |
| c2 | 11.54 | Low | Low | Clear | Night | Inner-city, Country road |
| c3 | 22.97 | Medium | High | Clear | Night | Parking lot, Highway, Residential area, Road bumps |
| c4 | 11.44 | Low | High | Clear | Night | Residential area, Road bumps, Highway, Inner-city |
| d1 | 8.3 | Low | Low | Clear | Night | Highway, Residential area |
| d2 | 8.3 | Low | High | Clear | Night | Residential area, Road bumps |
| d3 | 8.3 | Low | High | Clear | Night | Residential area, Road bumps, Highway |
The sensor streams were manually synchronised by matching the first and last frames of each sensor in every scenario. One remark must be made here: As the infrared camera's frame rate is not consistently constant, there might be delays between the infrared and the other sensor streams. A consistent time stamp for the infrared camera will be subject to future updates to the dataset. For more information on the synchronisation procedure, the interested reader should feel free to get in touch with the corresponding author.
To ensure ethically correct data handling, the dataset complies with the General Data Protection Regulation (GDPR). Every video sequence collected by all the three deployed cameras was scanned manually for vehicle number plates and people's faces. The features were then marked with the Matlab Ground Truth Labeller application and tracked with a point tracker [4]. Afterwards, every sequence was analysed again to mitigate human errors. Marked regions were blurred with a Gaussian filter.
Part of the dataset are also the camera calibration parameters and errors. These include in particular the image size, radial distortion, world points, translation vectors, reprojection errors, rotation vectors, intrinsic matrix, focal length, principal point, mean reprojection error, reprojected points and rotation matrices. The error files include the intrinsics and extrinsics errors.
The LiDAR point cloud information is stored in Matlab point cloud objects of the appearance illustrated in following Fig. 1. This information can be used to recreate the real-world trials in a simulation environment like e.g. IPG CarMaker.
Fig. 1.
Structure of the acquired LiDAR point cloud data.
The location branch is providing the 3D-position of each of the 16 laser beams with intensity values in a separate column. Count sums up the number of collected laser beams (in this case 16 times 1808) per point cloud with minimum and maximum X-, Y- and Z-values added to the data object. A separate LiDAR data object was created for every time stamp in steps of 10 Hz.
The acquired CAN-bus data consists of the following information: Number of visible GPS satellites, time, latitude, longitude, velocity, heading, height, vertical velocity, steering angle, all wheel speeds, yaw rate, vehicle speed, longitudinal acceleration, lateral acceleration, handbrake, gear requested, gear, engine speed, coolant temperature, clutch position, brake pressure, brake position, battery voltage, air temperature and accelerator pedal position. The data is stored in Matlab structures whose layout is presented in Fig. 2.
Fig. 2.
Structure of the acquired CAN-bus data.
In addition to previously introduced CAN-bus data saved in different channels, more generic information on the recording session is given in the sections path of the structure. The channels themselves consist of one-dimensional column vectors sorted in 10 Hz frequency steps. In total, we collected 19 GB of infrared, 57 GB of LiDAR, 39 GB of monocular and 2.3 GB of smartphone data.
2. Experimental design, materials, and methods
The LiDAR unit, infrared camera and GPS receiver were mounted on the outside of the research vehicle. The residual sensors are placed inside the car. The exact sensor positions are summarised in following Table 2 and refer to the vehicle coordinate system presented in Fig. 3. The presented tables and figures in this chapter are not part of the data set itself but serve as additional information required for setting up e.g. target state estimation, localisation or mapping algorithms.
Table 2.
Sensor positions in vehicle coordinates.
| Sensor | X [mm] | Y [mm] | Z [mm] |
|---|---|---|---|
| Mono Camera | 1570 | 0 | 1260 |
| Infrared Camera | 1660 | 0 | 1430 |
| LiDAR | 2120 | 0 | 1630 |
| Smartphone Camera | 1420 | 310 | 1150 |
| Smartphone Sensors | 1490 | 0 | 1180 |
Fig. 3.
Vehicle dimensions and sensor positions [5].
The sensor wiring inside the vehicle is shown in Fig. 4 with power supply highlighted as yellow lightnings.
Fig. 4.
Interior sensor wiring.
The smartphone sensor coordinate system is displayed in following Fig. 5.
Fig. 5.

Smartphone accelerometer coordinate system.
To generate the camera calibration and error files, the Matlab Camera Calibrator application was used [6]. With the monocular and smartphone camera, pictures of a calibration pattern (checkerboard) were taken in different angles. With help of these images, aforementioned camera intrinsic parameters were extracted according to Refs. [7,8].
For data analysis we highly recommend Matlab in combination with additional packages like the ‘Statistics and Machine Learning Toolbox’, ‘Deep Learning Toolbox’, ‘Computer Vision Toolbox’ or the ‘Automated Driving Toolbox’. These libraries allow the visualisation of all attached data files and furthermore the fusion of e.g. camera and LiDAR data.
Acknowledgments
This project is part of the Interdisciplinary Training Network for Multi-Actuated Ground Vehicles (ITEAM) European program and has received funding from the European Union's Horizon 2020 research and innovation program under the Marie Sklodowska-Curie grant agreement No 675999.
Conflict of Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
References
- 1.Racelogic . 2019. VBOX VIDEO HD2.https://www.vboxmotorsport.co.uk/index.php/en/products/video-loggers/vbox-video [Google Scholar]
- 2.Lidar Velodyne. 2019. PUCK.https://velodynelidar.com/vlp-16.html [Google Scholar]
- 3.FLIR Systems, Inc., FLIR One Pro. 2019. https://www.flir.co.uk/products/flir-one-pro/?model=435-0007-02 [Google Scholar]
- 4.The Mathworks, Inc., Ground Truth Labeler. 2019. https://uk.mathworks.com/help/driving/ref/groundtruthlabeler-app.html [Google Scholar]
- 5.Ourhouzz.site, Ford Fiesta Interior Dimensions. 2019. http://ourhouzz.site/ford-fiesta-interior-dimensions [Google Scholar]
- 6.The Mathworks, Inc., Single Calibrator App. 2019. https://uk.mathworks.com/help/vision/ug/single-camera-calibrator-app.html [Google Scholar]
- 7.Zhengyou Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000;22(11):1330–1334. [Google Scholar]
- 8.Heikkila J., Silven O. IEEE International Conference on Computer Vision and Pattern Recognition. 1997. A four-step camera calibration procedure with implicit image correction. [Google Scholar]




