🇲🇬Relevant Datasets
Selected relevant datasets towards automated driving and mixed traffic research
Selected Online Open-source Datasets
Dataset
Data Description
(Type and Volume)
Relevant Tasks
and Case Studies
Data Samples Screenshot
Image & video with annotation;
100K video clips & images,1.8TB
Perception:
Semantic segmentation;
Lane detection
Involves 7 sensor stations equipped with more than 60 SOTA and multi-modal sensors, and covered a road network of approximately 3.5 kilometres, R0, R1, R2 three different data sets
Perception;
Digital Twin;
Motion Prediction;
Motion: TFRecord format with object trajectories and corresponding 3D maps for 103,354 segments;
Perception: Lidar and Camera data, labels for 2,030 segments
Motion: Motion Prediction, Interaction, Occupancy, and Flow Prediction, Sim Agents;
Perception: Segmentation, Object Detection & Tracking, Pose Estimation
Motion Prediction
104 hours of videos; GPS/IMU, CAN; etc.
10000 video clips; 12 classes bounding box, tracking ID, class ID;
Currently unreachable
Perception: Object Detection & Tracking
Video, point cloud, GPS, and driver behaviour (speed and wheel); 1000 km
Driving Policy Prediction
Drone-Based Vehicle Trajectory, 1140-minutes of drone videos@30 FPS recorded at 12 different locations
VR Driving Simulation;
Digital Twin;
Sensor Simulation;
Driving Behaviour Analysis;
Safety & Crash Analysis
Drone-based collection; 110500 vehicles; 147 hours; CSV; (Highway, Interaction, Roundabout)
Behaviour Extraction & Analysis; Intention / Behaviour / Motion Prediction;
Imitation Learning;
2 grayscale cameras, 2 color cameras, 4 Edmund optics lenses, 1 3D laser scanner (10 HZ); 6 hours; 50 scenes, 180 GB
Perception: Object Detection & Tracking; Semantic and Instance Segmentation; Road/Lane Detection
1000 driving scenes; 23 object classes annotated with 3D bounding boxes at 2Hz; 1.4M camera images, 390k LIDAR sweeps, 1.4M RADAR sweeps, and 1.4M object bounding boxes in 40k keyframes
Perception: 3D Detection and Tracking; Prediction
Motion Planning, Motion Prediction
1: 3D Tracking Dataset with 113 3D annotated scenes;
2: Sensor Dataset with 1,000 3D annotated scenarios (lidar, ring camera, and stereo sensor data), Lidar Dataset with 20,000 unlabeled scenarios
Perception: 3D Tracking;
Motion Forecasting
Image (video) with annotation;
133K images
Perception:
Semantic segmentation;
Lane detection
Video with semantic annotation;
>140K images (video frames)
Perception:
Semantic segmentation;
Object & Lane detection
Image & video with annotation;
Two image sets:7K and 5K
Perception:
Semantic segmentation;
Lane detection
Video, LiDAR, GPS;
10 videos
Perception:
Semantic segmentation;
Lane detection
LiDAR and stereo images with various position sensors targeting a highly complex urban environment;
tar.gz
SLAM; Odometry
Image (Traffic sign) with annotation
Perception: Object Detection
700+ images; 10+ minutes of high quality 30Hz footage with corresponding semantically labeled images at 1Hz and in part, 15Hz
Perception: Segmentation & Recognition
Video with annotation (bounding box, behavioral label);
347 videos, 170GB
Perception:
Object Detection;
Behaviour Analysis
Video with behavioral label, GPS, vehicle data;
35 videos
Behavior analysis
Video, LiDAR, GPS, vehicle with annotation (bounding box);
300GB
Perception:
Object detection, Object tracking:
End2End learning;
Imitation learning
Video (image) with annotation (vehicle and traffic sign);
3 (vehicle) + several (traffic sign) videos
Perception: Object Detection;
Video (image) with annotation;
5,000 (manual) + 20,100 (semi-auto) frames
Perception:
Object Detection, Semantic Segmentation;
Imitation learning
Image(video), LiDAR, with Semantic and Point cloud Segmentation, 3D bounding;
41,280 (image) +
12,499 (3D) + 390,000 (unlabeled sensor) frames
Perception:
Object Detection, Object Tracking;
End2End Learning;
Imitation Learning
25,000 high-resolution images;
124 semantic object categories;
100 instance-annotated categories;
Global reach, covering 6 continents
Perception:
Street-level Instance Segmentation
56,000 camera images; 7,000 LiDAR sweeps; 75 scenes of 50-100 frames each 10 annotation classes;
Full sensor suite: 1 LiDAR, 8 Cameras, Post-processed GPS/IMU;
Adverse weather conditions (snow)
Perception:
(3D) Object Detection, Object Tracking;
Trajectory Prediction
LiDAR point cloud data; LAS, XML, SHP
Perception:
(3D) Object Detection
1 year, 1000 km; 20 million images along with LIDAR, GPS, and INS ground truth
Perception:
Object Detection, Object Tracking;
Dense Reconstruction;
Localization
Two vehicle cooperation simultaneously in the same location, 410 km of the driving area, 20K LiDAR, 40K RGB, and 240K annotated 3D bounding boxes across 5 vehicle classes
Perception: Vehicle-to-Vehicle Cooperative Perception; (3D) Object Detection, Tracking, Prediction, Localization;
Sim2Real Transfer Learning
Basic safety messages (BSM), vehicle trajectories, and various driver-vehicle interaction data; CSV format
Interactive Behaviour Extraction & Analysis;
Safety Analysis;
Driving Anomaly Detection
Roundabout: 10479 trajectories, 365 mins; Unsignalized Intersection: 14867 trajectories, 433 mins; Lane change: 10933 trajectories,
133 mins; Signalized intersection: 3775 trajectories, 60 mins; High definition maps in lanelet2 format
Intention/Behaviour/Motion Prediction;
Imitation Learning;
Reinforcement Learning;
Interactive Behaviour Extraction & Analysis
Last updated