Skip to content

Latest commit

 

History

History
1157 lines (1115 loc) · 33.3 KB

datasets.md

File metadata and controls

1157 lines (1115 loc) · 33.3 KB

Home  | Behavioral  | Applications  | Datasets  

Click on each entry below to see additional information.

100-Driver | paper | link
    Description: Videos of drivers performing secondary tasks
    Data: driver video
    Annotations: action labels
    @article{2023_T-ITS_Wang,
        author = "Wang, Jing and Li, Wenjing and Li, Fang and Zhang, Jun and Wu, Zhongcheng and Zhong, Zhun and Sebe, Nicu",
        journal = "IEEE Transactions on Intelligent Transportation Systems",
        publisher = "IEEE",
        title = "100-Driver: A Large-Scale, Diverse Dataset for Distracted Driver Classification",
        year = "2023"
    }
    

SynDD1 | paper | link
    Full name: Synthetic Distracted Driving Dataset
    Description: Synthetic dataset for machine learning models to detect and analyze drivers' various distracted behavior and different gaze zones.
    Data: driver video
    Annotations: gaze area labels, action labels, appearance labels
    @article{2023_DiB_Rahman,
        author = "Rahman, Mohammed Shaiqur and Venkatachalapathy, Archana and Sharma, Anuj and Wang, Jiyang and Gursoy, Senem Velipasalar and Anastasiu, David and Wang, Shuo",
        journal = "Data in brief",
        pages = "108793",
        publisher = "Elsevier",
        title = "Synthetic distracted driving (syndd1) dataset for analyzing distracted behaviors and various gaze zones of a driver",
        volume = "46",
        year = "2023"
    }
    

Fatigueview | paper | link
    Description: Multi-camera video dataset for vision-based drowsiness detection.
    Data: driver video
    Annotations: facial landmarks, face/hand bounding boxes, head pose, eye status, pose, drowsiness labels
    @article{2022_T-ITS_Yang,
        author = "Yang, Cong and Yang, Zhenyu and Li, Weiyu and See, John",
        journal = "IEEE Transactions on Intelligent Transportation Systems",
        publisher = "IEEE",
        title = "FatigueView: A Multi-Camera Video Dataset for Vision-Based Drowsiness Detection",
        year = "2022"
    }
    

CoCAtt | paper | link
    Full name: A Cognitive-Conditioned Driver Attention Dataset
    Description: Videos of drivers and driver scenes in automated and manual driving conditions with per-frame gaze and distraction annotations
    Data: driver video, scene video, eye-tracking
    Annotations: distraction state, car telemetry, intention labels
    @inproceedings{2022_ITSC_Shen,
        author = "Shen, Yuan and Wijayaratne, Niviru and Sriram, Pranav and Hasan, Aamir and Du, Peter and Driggs-Campbell, Katherine",
        booktitle = "2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC)",
        organization = "IEEE",
        pages = "32--39",
        title = "CoCAtt: A Cognitive-Conditioned Driver Attention Dataset",
        year = "2022"
    }
    

LBW | paper | link
    Full name: Look Both Ways
    Description: Synchronized videos from scene and driver-facing cameras of drivers performing various maneuvers in traffic
    Data: driver video, scene video, eye-tracking
    @inproceedings{2022_ECCV_Kasahara,
        author = "Kasahara, Isaac and Stent, Simon and Park, Hyun Soo",
        booktitle = "Computer Vision--ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23--27, 2022, Proceedings, Part XIII",
        organization = "Springer",
        pages = "126--142",
        title = "Look Both Ways: Self-supervising Driver Gaze Estimation and Road Scene Saliency",
        year = "2022"
    }
    

55 Rides | paper | link
    Description: Naturalistic dataset recorded by four drivers and annotated by three raters to determine distraction states
    Data: driver video, eye-tracking
    Annotations: distraction state, head pose
    @inproceedings{2021_ETRA_Kubler,
        author = {K{\"u}bler, Thomas C and Fuhl, Wolfgang and Wagner, Elena and Kasneci, Enkelejda},
        booktitle = "ACM Symposium on Eye Tracking Research and Applications",
        pages = "1--8",
        title = "55 Rides: attention annotated head and gaze data during naturalistic driving",
        year = "2021"
    }
    

DAD | paper | link
    Full name: Driver Anomaly Detection
    Description: Videos of normal and anomalous behaviors (manual/visual distractions) of drivers.
    Data: driver video
    Annotations: action labels
    @inproceedings{2021_WACV_Kopuklu,
        author = "Kopuklu, Okan and Zheng, Jiapeng and Xu, Hang and Rigoll, Gerhard",
        booktitle = "Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision",
        pages = "91--100",
        title = "Driver anomaly detection: A dataset and contrastive learning approach",
        year = "2021"
    }
    

MAAD | paper | link
    Full name: Attended Awareness in Driving
    Description: A subset of videos from DR(eye)VE annotated with gaze collected in lab conditions.
    Data: eye-tracking, scene video
    Annotations: task labels
    @inproceedings{2021_ICCVW_Gopinath,
        author = "Gopinath, Deepak and Rosman, Guy and Stent, Simon and Terahata, Katsuya and Fletcher, Luke and Argall, Brenna and Leonard, John",
        booktitle = "Proceedings of the IEEE/CVF International Conference on Computer Vision",
        pages = "3426--3436",
        title = {MAAD: A Model and Dataset for" Attended Awareness" in Driving},
        year = "2021"
    }
    

DGW | paper | link
    Full name: Driver Gaze in the Wild
    Description: Videos of drivers fixating on different areas in the vehicle without constraining their head and eye movements
    Data: driver video
    Annotations: gaze area labels
    @inproceedings{2021_ICCVW_Ghosh,
        author = "Ghosh, Shreya and Dhall, Abhinav and Sharma, Garima and Gupta, Sarthak and Sebe, Nicu",
        booktitle = "ICCVW",
        title = "Speak2label: Using domain knowledge for creating a large scale driver gaze zone estimation dataset",
        year = "2021"
    }
    

TrafficSaliency | paper | link
    Description: 16 videos of driving scenes with gaze data of 28 subjects recorded in the lab with eye-tracker
    Data: eye-tracking, scene video
    @article{2020_T-ITS_Deng,
        author = "Deng, Tao and Yan, Hongmei and Qin, Long and Ngo, Thuyen and Manjunath, BS",
        journal = "IEEE Transactions on Intelligent Transportation Systems",
        number = "5",
        pages = "2146--2154",
        publisher = "IEEE",
        title = "{How do drivers allocate their potential attention? Driving fixation prediction via convolutional neural networks}",
        volume = "21",
        year = "2019"
    }
    

NeuroIV | paper | link
    Full name: Neuromorphic Vision Meets Intelligent Vehicle
    Description: Videos of drivers performing secondary tasks, making hand gestures and observing different regions inside the vehicle recorded with DAVIS and depth sensor
    Data: driver video
    @article{2020_T-ITS_Chen,
        author = {Chen, Guang and Wang, Fa and Li, Weijun and Hong, Lin and Conradt, J{\"o}rg and Chen, Jieneng and Zhang, Zhenyan and Lu, Yiwen and Knoll, Alois},
        journal = "IEEE Transactions on Intelligent Transportation Systems",
        number = "2",
        pages = "1171--1183",
        publisher = "IEEE",
        title = "NeuroIV: Neuromorphic vision meets intelligent vehicle towards safe driving with a new database and baseline evaluations",
        volume = "23",
        year = "2020"
    }
    

LISA v2 | paper | link
    Full name: Laboratory for Intelligent and Safe Automobiles
    Description: Videos of drivers with and without eyeglasses recorded under different lighting conditions
    Data: driver video
    @inproceedings{2020_IV_Rangesh,
        author = "Rangesh, Akshay and Zhang, Bowen and Trivedi, Mohan M",
        booktitle = "IV",
        title = "Driver gaze estimation in the real world: Overcoming the eyeglass challenge",
        year = "2020"
    }
    

DGAZE | paper | link
    Description: A dataset mapping drivers’ gaze to different areas in a static traffic scene in lab conditions
    Data: driver video, scene video
    Annotations: bounding boxes
    @inproceedings{2020_IROS_Dua,
        author = "Dua, Isha and John, Thrupthi Ann and Gupta, Riya and Jawahar, CV",
        booktitle = "IROS",
        title = "DGAZE: Driver Gaze Mapping on Road",
        year = "2020"
    }
    

DMD | paper | link
    Full name: Driving Monitoring Dataset
    Description: A diverse multi-modal dataset of drivers performing various secondary tasks, observing different regions inside the car, and showing signs of drowsiness recorded on-road and in simulation environment
    Data: driver video, scene video, vehicle data
    Annotations: bounding boxes, action labels
    @inproceedings{2020_ECCVW_Ortega,
        author = "Ortega, Juan Diego and Kose, Neslihan and Ca{\\textasciitilde n}as, Paola and Chao, Min-An and Unnervik, Alexander and Nieto, Marcos and Otaegui, Oihana and Salgado, Luis",
        booktitle = "ECCV",
        title = "Dmd: A large-scale multi-modal driver monitoring dataset for attention and alertness analysis",
        year = "2020"
    }
    

PRORETA 4 | paper | link
    Description: Videos of traffic scenes recorded in instrumented vehicle with driver’s gaze data for evaluating accuracy of detecting driver’s current object of fixation
    Data: eye-tracking, driver video, scene video
    @inproceedings{2019_IV_Schwehr,
        author = "Schwehr, Julian and Knaust, Moritz and Willert, Volker",
        booktitle = "IV",
        title = "How to evaluate object-of-fixation detection",
        year = "2019"
    }
    

DADA-2000 | paper | link
    Full name: Driver Attention in Driving Accident Scenarios
    Description: 2000 videos of accident videos collected from video hosting websites with eye-tracking data from 20 subjects collected in the lab.
    Data: eye-tracking, scene video
    Annotations: bounding boxes, accident category labels
    @inproceedings{2019_ITSC_Fang,
        author = "Fang, Jianwu and Yan, Dingxin and Qiao, Jiahuan and Xue, Jianru and Wang, He and Li, Sen",
        booktitle = "ITSC",
        title = "{DADA-2000: Can Driving Accident be Predicted by Driver Attentionƒ Analyzed by A Benchmark}",
        year = "2019"
    }
    

Drive&Act | paper | link
    Description: Videos of drivers performing various driving- and non-driving-related tasks
    Data: driver video
    Annotations: semantic maps, action labels
    @inproceedings{2019_ICCV_Martin,
        author = "Martin, Manuel and Roitberg, Alina and Haurilet, Monica and Horne, Matthias and Rei{\ss}, Simon and Voit, Michael and Stiefelhagen, Rainer",
        booktitle = "ICCV",
        title = "Drive\\&act: A multi-modal dataset for fine-grained driver behavior recognition in autonomous vehicles",
        year = "2019"
    }
    

RLDD | paper | link
    Full name: Real-Life Drowsiness Datase
    Description: Crowdsourced videos of people in various states of drowsiness recorded in indoor environments
    Data: driver video
    Annotations: drowsiness labels
    @inproceedings{2019_CVPRW_Ghoddoosian,
        author = "Ghoddoosian, Reza and Galib, Marnim and Athitsos, Vassilis",
        booktitle = "CVPRW",
        title = "A realistic dataset and baseline temporal model for early drowsiness detection",
        year = "2019"
    }
    

HAD | paper | link
    Full name: HAD HRI Advice Dataset
    Description: A subset of videos from HDD naturalistic dataset annotated with textual advice containing 1) goals – where the vehicle should move and 2) attention – where the vehicle should look
    Data: scene video, vehicle data
    Annotations: goal and attention labels
    @inproceedings{2019_CVPR_Kim,
        author = "Kim, Jinkyu and Misu, Teruhisa and Chen, Yi-Ting and Tawari, Ashish and Canny, John",
        booktitle = "CVPR",
        title = "Grounding human-to-vehicle advice for self-driving vehicles",
        year = "2019"
    }
    

3MDAD | paper | link
    Full name: Multimodal Multiview and Multispectral Driver Action Dataset
    Description: Videos of drivers performing secondary tasks
    Data: driver video
    Annotations: action labels, bounding boxes
    @inproceedings{2019_CAIP_Jegham,
        author = "Jegham, Imen and Ben Khalifa, Anouar and Alouani, Ihsen and Mahjoub, Mohamed Ali",
        booktitle = "Computer Analysis of Images and Patterns: 18th International Conference, CAIP 2019, Salerno, Italy, September 3--5, 2019, Proceedings, Part I 18",
        organization = "Springer",
        pages = "518--529",
        title = "Mdad: A multimodal and multiview in-vehicle driver action dataset",
        year = "2019"
    }
    

EBDD | paper | link
    Full name: EEE BUET Distracted Driving Dataset
    Description: Videos of drivers performing secondary tasks
    Data: driver video
    Annotations: action labels, bounding boxes
    @article{2019_TCSVT_Billah,
        author = "Billah, Tashrif and Rahman, SM Mahbubur and Ahmad, M Omair and Swamy, MNS",
        journal = "IEEE Transactions on Circuits and Systems for Video Technology",
        number = "4",
        pages = "1048--1062",
        publisher = "IEEE",
        title = "Recognizing distractions for assistive driving by tracking body parts",
        volume = "29",
        year = "2018"
    }
    

H3D | paper | link
    Full name: H3D Honda 3D Dataset
    Description: A subset of videos from HDD dataset with 3D bounding boxes and object ids for tracking
    Data: driver video, vehicle data
    Annotations: bounding boxes
    @inproceedings{2019_ICRA_Patil,
        author = "Patil, Abhishek and Malla, Srikanth and Gang, Haiming and Chen, Yi-Ting",
        booktitle = "2019 International Conference on Robotics and Automation (ICRA)",
        organization = "IEEE",
        pages = "9552--9557",
        title = "The h3d dataset for full-surround 3d multi-object detection and tracking in crowded urban scenes",
        year = "2019"
    }
    

DR(eye)VE | paper | link
    Description: Driving videos recorded on-road with corresponding gaze data of the driver
    Data: eye-tracking, scene video, vehicle data
    Annotations: weather and road type labels
    @article{2018_PAMI_Palazzi,
        author = "Palazzi, Andrea and Abati, Davide and Solera, Francesco and Cucchiara, Rita and others",
        journal = "IEEE TPAMI",
        number = "7",
        pages = "1720--1733",
        title = "{Predicting the Driver's Focus of Attention: the DR (eye) VE Project}",
        volume = "41",
        year = "2018"
    }
    

BDD-X | paper | link
    Full name: Berkeley Deep Drive-X (eXplanation) Dataset
    Description: A subset of videos from BDD dataset annotated with textual descriptions of actions performed by the vehicle and explanations justifying those actions
    Data: scene video, vehicle data
    Annotations: action explanations
    @inproceedings{2018_ECCV_Kim,
        author = "Kim, Jinkyu and Rohrbach, Anna and Darrell, Trevor and Canny, John and Akata, Zeynep",
        booktitle = "ECCV",
        title = "Textual explanations for self-driving vehicles",
        year = "2018"
    }
    

HDD | paper | link
    Full name: HDD HRI Driving Dataset
    Description: A large naturalistic driving dataset with driving footage, vehicle telemetry and annotations for vehicle actions and their justifications
    Data: scene video, vehicle data
    Annotations: bounding boxes, action labels
    @inproceedings{2018_CVPR_Ramanishka,
        author = "Ramanishka, Vasili and Chen, Yi-Ting and Misu, Teruhisa and Saenko, Kate",
        booktitle = "CVPR",
        title = "Toward driving scene understanding: A dataset for learning driver behavior and causal reasoning",
        year = "2018"
    }
    

BDD-A | paper | link
    Full name: Berkeley Deep Drive-A (Attention) Dataset
    Description: A set of short video clips extracted from the Berkeley Deep Drive (BDD) dataset with additional eye-tracking data collected in the lab from 45 subjects
    Data: eye-tracking, scene video, vehicle data
    @inproceedings{2018_ACCV_Xia,
        author = "Xia, Ye and Zhang, Danqing and Kim, Jinkyu and Nakayama, Ken and Zipser, Karl and Whitney, David",
        booktitle = "ACCV",
        title = "Predicting driver attention in critical situations",
        year = "2018"
    }
    

C42CN | paper | link
    Description: A multi-modal dataset acquired in a controlled experiment on a driving simulator under 4 conditions: no distraction, cognitive, emotional and sensorimotor distraction.
    Data: eye-tracking, scene video, physiological signal
    @article{2017_NatSciData_Taamneh,
        author = "Taamneh, Salah and Tsiamyrtzis, Panagiotis and Dcosta, Malcolm and Buddharaju, Pradeep and Khatri, Ashik and Manser, Michael and Ferris, Thomas and Wunderlich, Robert and Pavlidis, Ioannis",
        journal = "Scientific Data",
        pages = "170110",
        title = "A multimodal dataset for various forms of distracted driving",
        volume = "4",
        year = "2017"
    }
    

DriveAHead | paper | link
    Description: Videos of drivers with frame-level head pose annotations obtained from a motion-capture system
    Data: driver video
    Annotations: occlusion, head pose, depth
    @inproceedings{2017_CVPRW_Schwarz,
        author = "Schwarz, Anke and Haurilet, Monica and Martinez, Manuel and Stiefelhagen, Rainer",
        booktitle = "Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops",
        pages = "1--10",
        title = "Driveahead-a large-scale driver head pose dataset",
        year = "2017"
    }
    

DDD | paper | link
    Full name: Driver Drowsiness Detection Dataset
    Description: Videos of human subjects simulating different levels of drowsiness while driving in a simulator
    Data: driver video
    Annotations: drowsiness labels
    @inproceedings{2017_ACCV_Weng,
        author = "Weng, Ching-Hua and Lai, Ying-Hsiu and Lai, Shang-Hong",
        booktitle = "ACCV",
        title = "Driver drowsiness detection via a hierarchical temporal deep belief network",
        year = "2016"
    }
    

Dashcam dataset | link
    Description: Driving videos with steering information recorded on road
    Data: scene video
    
    

AUCD2 | paper | link
    Full name: American University in Cairo (AUC) Distracted Driver’s Dataset
    Description: Videos of drivers performing secondary tasks
    Data: driver video
    Annotations: action labels
    @inproceedings{2017_NeurIPS_Abouelnaga,
        author = "Abouelnaga, Yehya and Eraqi, Hesham M. and Moustafa, Mohamed N.",
        booktitle = "NeurIPS Workshop on Machine Learning for Intelligent Transportation Systems",
        title = "eal-time Distracted Driver Posture Classification",
        year = "2017"
    }
    

DROZY | paper | link
    Description: Videos and physiological data from subjects in different drowsiness states after prolonged waking
    Data: driver video, physiological signal
    Annotations: drowsiness labels
    @inproceedings{2016_WACV_Massoz,
        author = "Massoz, Quentin and Langohr, Thomas and Fran{\c{c}}ois, Cl{\'e}mentine and Verly, Jacques G",
        booktitle = "WACV",
        title = "The ULg multimodality drowsiness database (called DROZY) and examples of use",
        year = "2016"
    }
    

TETD | paper | link
    Full name: Traffic Eye Tracking Dataset
    Description: A set of 100 images of traffic scenes with corresponding eye-tracking data from 20 subjects
    Data: eye-tracking, scene images
    @article{2016_T-ITS_Deng,
        author = "Deng, Tao and Yang, Kaifu and Li, Yongjie and Yan, Hongmei",
        journal = "IEEE Transactions on Intelligent Transportation Systems",
        number = "7",
        pages = "2051--2062",
        publisher = "IEEE",
        title = "Where does the driver look? Top-down-based saliency detection in a traffic driving environment",
        volume = "17",
        year = "2016"
    }
    

DAD | paper | link
    Description: Videos of accidents recorded with dashboard cameras sourced from video hosting sites with annotations for accidents and road users involved in them
    Data: scene video
    Annotations: bounding boxes, accident category labels
    @inproceedings{2016_ACCV_Chan,
        author = "Chan, Fu-Hsiang and Chen, Yu-Ting and Xiang, Yu and Sun, Min",
        booktitle = "ACCV",
        title = "Anticipating accidents in dashcam videos",
        year = "2016"
    }
    

Brain4Cars | paper | link
    Description: Synchronized videos from scene and driver-facing cameras of drivers performing various maneuvers in traffic
    Data: driver video, scene video, vehicle data
    Annotations: action labels
    @inproceedings{2015_ICCV_Jain,
        author = "Jain, Ashesh and Koppula, Hema S and Raghavan, Bharad and Soh, Shane and Saxena, Ashutosh",
        booktitle = "ICCV",
        title = "Car that knows before you do: Anticipating maneuvers via learning temporal driving models",
        year = "2015"
    }
    

SFD | link
    Full name: State Farm Distracted Driver Detection
    Description: Videos of drivers performing secondary tasks
    Data: driver video
    Annotations: action labels
    
    

DIPLECS Surrey | paper | link
    Description: Driving videos with steering information recorded in different cars and environments
    Data: scene video, vehicle data
    @article{2015_TranVehTech_Pugeault,
        author = "Pugeault, Nicolas and Bowden, Richard",
        journal = "IEEE Transactions on Vehicular Technology",
        number = "12",
        pages = "5424--5438",
        publisher = "IEEE",
        title = "How much of driving is preattentive?",
        volume = "64",
        year = "2015"
    }
    

YawDD | paper | link
    Full name: Yawning Detection Dataset
    Description: Recordings of human subjects in parked vehicles simulating normal driving, singing and taslking, and yawning
    Data: driver video
    Annotations: bounding boxes, action labels
    @inproceedings{2014_ACM_Abtahi,
        author = "Abtahi, Shabnam and Omidyeganeh, Mona and Shirmohammadi, Shervin and Hariri, Behnoosh",
        booktitle = "Proceedings of the ACM Multimedia Systems Conference",
        title = "{YawDD: A yawning detection dataset}",
        year = "2014"
    }
    

3DDS | paper | link
    Full name: 3D Driving School Dataset
    Description: Videos and eye-tracking data of people playing 3D driving simulator game
    Data: eye-tracking, scene video
    @inproceedings{2011_BMVC_Borji,
        author = "Borji, Ali and Sihite, Dicky N and Itti, Laurent",
        booktitle = "BMVC",
        title = "Computational Modeling of Top-down Visual Attention in Interactive Environments.",
        year = "2011"
    }
    

DIPLECS Sweden | paper | link
    Description: Driving videos with steering information recorded in different cars and environments
    Data: scene video, vehicle data
    @inproceedings{2010_ACCV_Pugeault,
        author = "Pugeault, Nicolas and Bowden, Richard",
        booktitle = "ECCV",
        title = "Learning pre-attentive driving behaviour from holistic visual features",
        year = "2010"
    }
    

BU HeadTracking | paper | link
    Full name: Boston University Head Tracking Dataset
    Description: Videos and head tracking information for multiple human subjects recorded in diverse conditions
    Data: driver video
    Annotations: head pose
    @article{2000_PAMI_LaCascia,
        author = "La Cascia, Marco and Sclaroff, Stan and Athitsos, Vassilis",
        journal = "IEEE Transactions on pattern analysis and machine intelligence",
        number = "4",
        pages = "322--336",
        publisher = "IEEE",
        title = "Fast, reliable head tracking under varying illumination: An approach based on registration of texture-mapped 3D models",
        volume = "22",
        year = "2000"
    }