A composite dataset with eight videos (totaling the pronunciation of seventeen words, with intervals, sagittal plane, and gray scale), for experiments in computer vision, video processing, and articulation investigation of the vocal tract.


In this dataset:- There is no audio.- Sagittal image- Grey Scale


Conveyor belts are the most widespread means of transportation for large quantities of materials in the mining sector. This dataset contains 388 images of structures with and without dirt buildup.

One can use this dataset for experimentation on classifying the dirt buildup.


The data are separated into folders that specify each class of the dataset: Clean and Dirty.


This archive contains images and labels for the Idly-Dosa-Vada (IDV) dataset, for use with Yolo (and Tensorflow) object detection frameworks.


This archive contains images and labels for the Idly-Dosa-Vada (IDV) dataset, for use with Yolo (and Tensorflow) object detection frameworks.

The dataset contains 1009 images, and corresponding labels.

The dataset was created by using euclidaug, using only 6 images per class.


Folder structure after extracting idv-dataset-files.zip:

out_images - contains all training images

out_labels - contains labels for each image, in Yolo format


For usage, refer to the IEEE-DL-TAP instructions, which are derived from https://github.com/prabindh/yolo-bins/tree/master/capacito


Step 1: Generate full list of image files, for use in the training process. In Windows, this is done using the below command:


dir /s/b *.jpg > trainingfile.txt


Step 2: Using the above file, and the labelled images and labels, start the training process with Yolo using instructions at https://github.com/prabindh/yolo-bins/tree/master/capacito


Step 3: Perform inference using Yolo


Understanding causes and effects in mechanical systems is an essential component of reasoning in the physical world. This work poses a new problem of counterfactual learning of object mechanics from visual input. We develop the COPHY benchmark to assess the capacity of the state-of-the-art models for causal physical reasoning in a synthetic 3D environment and propose a model for learning the physical dynamics in a counterfactual setting.


Pedestrian detection has never been an easy task for computer vision and automotive industry. Systems like the advanced driver assistance system (ADAS) highly rely on far infrared (FIR) data captured to detect pedestrians at nighttime. The recent development of deep learning-based detectors has proven the excellent results of pedestrian detection in perfect weather conditions. However, it is still unknown what is the performance in adverse weather conditions.


Prefix _b - means benchmark, otherwise used for training/testing


Each recording folder contains:

  16BitFrames - 16bit original capture without processing.

  16BitTransformed - 16bit capture with low pass filter applied and scaled to 640x480.

  annotations - annotations and 8bit images made from 16BitTransformed.

  carParams.csv - a CAN details with coresponding frame ID.

  weather.txt - weather information in which the recording was made.


Annotations are made in YOLO (You only look once) Darknet format.


To have images without low pass filter applied you should make the following steps:

- Take 16bit images from 16BitFrames folder and open with OpenCV function like: Mat input = imread(<image_full_path>, -1);

- Then use convertTo function like: input.convertTo(output, input.depth(), sc, sh), where output is transformed Mat, sc is scale and sh is shift from carParams.csv file.

- Finally, scale image to 640x480 


ADAM is organized as a half day Challenge, a Satellite Event of the ISBI 2020 conference in Iowa City, Iowa, USA.


Endoscopy is a widely used clinical procedure for the early detection of cancers in hollow-organs such as oesophagus, stomach, and colon. Computer-assisted methods for accurate and temporally consistent localisation and segmentation of diseased region-of-interests enable precise quantification and mapping of lesions from clinical endoscopy videos which is critical for monitoring and surgical planning. Innovations have the potential to improve current medical practices and refine healthcare systems worldwide.

Last Updated On: 
Sat, 02/27/2021 - 05:11

The data set includes three sub-data sets, namely the DAGM2007 data set, the ground crack data set, and the Yibao bottle cap defect data set, which are divided into a training set and a test set, in which the positive and negative samples are unbalanced.


Nextmed project is a software platform for the segmentation and visualization of medical images. It consist on a series of different automatic segmentation algorithms for different anatomical structures and  a platform for the visualization of the results as 3D models.

This dataset contains the .obj and .nrrd files that correspond to the results of applying our automatic lung segmentation algorithm to the LIDC-IDRI dataset.

This dataset relates to 718 of the 1012 LIDC-IDRI scans.


The file consists in a folder for each result whith the .obj and .nrrd files generated by the Nextmed algorithms.


These three datasets cover Western, Chinese and Japanese food used for food instance counting and segmentation evaluation.