Computer Vision and Machine Learning Datasets

Computer Vision and Machine Learning requires standardised datasets for evaluation, comparison and learning. Datasets associated with our projects are linked below.

  • Eye Tracking Datasets

    Download page PascalVOC2012 train and Cellpose

    Download page PascalVOC2012 validation

    For our eye tracking-based experiments, we used the PascalVOC2012 dataset, a widely adopted benchmark in computer vision and deep learning research. It is a publicly available dataset designed for image segmentation, detection, and classification, featuring exclusively realistic scenes. The annotated objects fall into 20 distinct classes. We used the latest version of the dataset, which was continuously extended between 2005 and 2012. It comprises 11,630 images with 27,450 annotated objects and 6,929 segmentation masks.

    For our experiments in segmentation and classification tasks, we focused on the segmentation subset of PascalVOC2012, which is divided into training and validation splits. As an additional comparison, we also included a part of the Cellpose dataset, which contains segmentation masks of cells from various microscopy images.

    Gaze data was recorded using a screen-based eye tracker (Tobii Pro Fusion) with a sampling frequency of 120 Hz. Before starting the annotation process, each participant underwent a calibration procedure to ensure accurate eye tracking data. In total, 11 different individuals contributed to the dataset. The setup was designed to guide users through the annotation process in a consistent way: to inform participants about the object of interest, the bounding box and ground truth polygon mask were displayed for 0.5 seconds prior to observation. For very small objects, an initial zoom-in was applied, and there was no time limit for viewing.To ensure clean segmentation-related gaze data, participants were instructed to indicate the start and end of each object inspection by pressing a button. Our annotation tool also allowed users to repeat an observation if needed and to freely navigate within the image using standard mouse interactions such as dragging, scrolling, and zooming.

    The corresponding raw gaze data used in our studies are available through the links above.

  • Larvae Collision Dataset 2 Animals (LCD2A)

    Download

    The collision database (called Larvae Collision Dataset 2 Animals; short: LCD2A) contains 1352 image sequences with approximately 159300 individual images resulting from an interaction analysis (collisions) experiment of Drosophila larvae. The images were acquired by utilizing the FIM2c setup. For further information please refer to

    • The "readme.txt" file included in the archive
    • Risse B., Otto N., Berh D., Jiang X., Kiel M., Klambt C. 2017. "FIM2c: Multicolor, Multipurpose Imaging System to Manipulate and Analyze Animal Behavior." IEEE Transactions on Biomedical Engineering 64, No. 3:610-620
    • Otto N, Risse B, Berh D, Bittern J, Jiang X, Klämbt C. 2016. "Interactions among Drosophila larvae before and during collision." Scientific Reports 11, No. 6: 31564
  • Larvae Collision Dataset 2 to 3 (LCD2t3)

    Download

    This dataset (called "Larvae Collision Dataset 2 to 3" or LCD2t3) represents a refined Version of the original collision database LCD2A resulting from an interaction analysis (collisions) experiment of Drosophila larvae. The images were acquired by utilizing the FIM2c setup. For further information please refer to

    • The "readme.txt" file included in the archive
    • Michels T, Berh D, Jiang X. 2018. "An RJMCMC-based method for tracking and resolving collisions of Drosophila Larvae." IEEE/ACM Transactions on Computational Biology and Bioinformatics 2018.
  • Larvae Heartbeat Dataset

    Download

    The heartbeat database contains 39 image sequences with approximately 52700 individual images showing an (irregular) heartbeat of Drosophila melanogaster pupae. The images were acquired by utilizing the FIM setup. For further information please refer to

    • The "readme.txt" file included in the archive
    • Berh D, Scherzinger A, Otto N, Jiang X, Klämbt C, Risse B. 2018. "Automatic non-invasive heartbeat quantification of Drosophila pupae." Computers in Biology and Medicine 93: 189-199