The dataset consists of 60285 character image files which has been randomly divided into 54239 (90%) images as training set 6046 (10%) images as test set. The collection of data samples was carried out in two phases. The first phase consists of distributing a tabular form and asking people to write the characters five times each. Filled-in forms were collected from around 200 different individuals in the age group 12-23 years. The second phase was the collection of handwritten sheets such as answer sheets and classroom notes from students in the same age group.

Categories:
257 Views

Water meter dataset. Contains 1244 water meter images. Assembled using a crowdsourcing platform Yandex.Toloka.

Instructions: 

The dataset consists of 1244 images.

File name consists of:

1) water meter id

2) water meter readings

Categories:
472 Views

As one of the research directions at OLIVES Lab @ Georgia Tech, we focus on the robustness of data-driven algorithms under diverse challenging conditions where trained models can possibly be depolyed. To achieve this goal, we introduced a large-sacle (1.M images) object recognition dataset (CURE-OR) which is among the most comprehensive datasets with controlled synthetic challenging conditions. In CURE

Instructions: 

 

 

Image name format : 

"backgroundID_deviceID_objectOrientationID_objectID_challengeType_challengeLevel.jpg"

 

backgroundID: 

1: White 2: Texture 1 - living room 3: Texture 2 - kitchen 4: 3D 1 - living room 5: 3D 2 – office

 

 

objectOrientationID: 

1: Front (0 º) 2: Left side (90 º) 3: Back (180 º) 4: Right side (270 º) 5: Top

 

 

objectID:

 1-100

 

 

challengeType: 

No challenge 02: Resize 03: Underexposure 04: Overexposure 05: Gaussian blur 06: Contrast 07: Dirty lens 1 08: Dirty lens 2 09: Salt & pepper noise 10: Grayscale 11: Grayscale resize 12: Grayscale underexposure 13: Grayscale overexposure 14: Grayscale gaussian blur 15: Grayscale contrast 16: Grayscale dirty lens 1 17: Grayscale dirty lens 2 18: Grayscale salt & pepper noise

challengeLevel: 

A number between [0, 5], where 0 indicates no challenge, 1 the least severe and 5 the most severe challenge. Challenge type 1 (no challenge) and 10 (grayscale) has a level of 0 only. Challenge types 2 (resize) and 11 (grayscale resize) has 4 levels (1 through 4). All other challenges have levels 1 to 5.

 

 

 

Categories:
329 Views

As one of the research directions at OLIVES Lab @ Georgia Tech, we focus on the robustness of data-driven algorithms under diverse challenging conditions where trained models can possibly be depolyed. To achieve this goal, we introduced a large-sacle (~1.72M frames) traffic sign detection video dataset (CURE-TSD) which is among the most comprehensive datasets with controlled synthetic challenging conditions. The video sequences in the 

Instructions: 

The name format of the video files are as follows: “sequenceType_sequenceNumber_challengeSourceType_challengeType_challengeLevel.mp4”

·         sequenceType: 01 – Real data 02 – Unreal data

·         sequenceNumber: A number in between [01 – 49]

·         challengeSourceType: 00 – No challenge source (which means no challenge) 01 – After affect

·         challengeType: 00 – No challenge 01 – Decolorization 02 – Lens blur 03 – Codec error 04 – Darkening 05 – Dirty lens 06 – Exposure 07 – Gaussian blur 08 – Noise 09 – Rain 10 – Shadow 11 – Snow 12 – Haze

·         challengeLevel: A number in between [01-05] where 01 is the least severe and 05 is the most severe challenge.

Test Sequences

We split the video sequences into 70% training set and 30% test set. The sequence numbers corresponding to test set are given below:

[01_04_x_x_x, 01_05_x_x_x, 01_06_x_x_x, 01_07_x_x_x, 01_08_x_x_x, 01_18_x_x_x, 01_19_x_x_x, 01_21_x_x_x, 01_24_x_x_x, 01_26_x_x_x, 01_31_x_x_x, 01_38_x_x_x, 01_39_x_x_x, 01_41_x_x_x, 01_47_x_x_x, 02_02_x_x_x, 02_04_x_x_x, 02_06_x_x_x, 02_09_x_x_x, 02_12_x_x_x, 02_13_x_x_x, 02_16_x_x_x, 02_17_x_x_x, 02_18_x_x_x, 02_20_x_x_x, 02_22_x_x_x, 02_28_x_x_x, 02_31_x_x_x, 02_32_x_x_x, 02_36_x_x_x]

The videos with all other sequence numbers are in the training set. Note that “x” above refers to the variations listed earlier.

The name format of the annotation files are as follows: “sequenceType_sequenceNumber.txt“

Challenge source type, challenge type, and challenge level do not affect the annotations. Therefore, the video sequences that start with the same sequence type and the sequence number have the same annotations.

·         sequenceType: 01 – Real data 02 – Unreal data

·         sequenceNumber: A number in between [01 – 49]

The format of each line in the annotation file (txt) should be: “frameNumber_signType_llx_lly_lrx_lry_ulx_uly_urx_ury”. You can see a visual coordinate system example in our GitHub page.

·         frameNumber: A number in between [001-300]

·         signType: 01 – speed_limit 02 – goods_vehicles 03 – no_overtaking 04 – no_stopping 05 – no_parking 06 – stop 07 – bicycle 08 – hump 09 – no_left 10 – no_right 11 – priority_to 12 – no_entry 13 – yield 14 – parking

Categories:
542 Views

The dataset was built by capturing the static gestures of the American Sign Language (ASL) alphabet, from 8 people, except for the letters J and Z, since they are dynamic gestures. To capture the images, we used a Logitech Brio webcam, with a resolution of 1920 × 1080 pixels, in a university laboratory with artificial lighting. By extracting only the hand region, we defined an area of 400 × 400 pixels for the final image of our dataset.

Instructions: 

Original videos are located in directory "Original videos";

Filenames: first character gestures class (a_1-19.png), second character person gesture (a_1-19.png).

 

Using this dataset, also reference the paper:

Raimundo F. Pinto Jr., Carlos D. B. Borges, Antônio M. A. Almeida, and Iális C. Paula, Jr., “Static Hand Gesture Recognition Based on Convolutional Neural Networks,” Journal of Electrical and Computer Engineering, vol. 2019, Article ID 4167890, 12 pages, 2019. https://doi.org/10.1155/2019/4167890.

 

Categories:
Category: 
1611 Views

As one of the research directions at OLIVES Lab @ Georgia Tech, we focus on the robustness of data-driven algorithms under diverse challenging conditions where trained models can possibly be depolyed.

Instructions: 

The name format of the provided images are as follows: "sequenceType_signType_challengeType_challengeLevel_Index.bmp"

  • sequenceType: 01 - Real data 02 - Unreal data

  • signType: 01 - speed_limit 02 - goods_vehicles 03 - no_overtaking 04 - no_stopping 05 - no_parking 06 - stop 07 - bicycle 08 - hump 09 - no_left 10 - no_right 11 - priority_to 12 - no_entry 13 - yield 14 - parking

  • challengeType: 00 - No challenge 01 - Decolorization 02 - Lens blur 03 - Codec error 04 - Darkening 05 - Dirty lens 06 - Exposure 07 - Gaussian blur 08 - Noise 09 - Rain 10 - Shadow 11 - Snow 12 - Haze

  • challengeLevel: A number in between [01-05] where 01 is the least severe and 05 is the most severe challenge.

  • Index: A number shows different instances of traffic signs in the same conditions.

Categories:
366 Views

The Dataset

The Onera Satellite Change Detection dataset addresses the issue of detecting changes between satellite images from different dates.

Instructions: 

Onera Satellite Change Detection dataset

 

##################################################

Authors: Rodrigo Caye Daudt, rodrigo.daudt@onera.fr

Bertrand Le Saux, bls@ieee.org

Alexandre Boulch, alexandre.boulch@valeo.ai

Yann Gousseau, yann.gousseau@telecom-paristech.fr

 

##################################################

About: This dataset contains registered pairs of 13-band multispectral satellite images obtained by the Sentinel-2 satellites of the Copernicus program. Pixel-level urban change groundtruth is provided. In case of discrepancies in image size, the older images with resolution of 10m per pixel is used. Images vary in spatial resolution between 10m, 20m and 60m. For more information, please refer to Sentinel-2 documentation.

 

For each location, folders imgs_1_rect and imgs_2_rect contain the same images as imgs_1 and imgs_2 resampled at 10m resolution and cropped accordingly for ease of use. The proposed split into train and test images is contained in the train.txt and test.txt files.

For downloading and cropping the images, the Medusa toolbox was used: https://github.com/aboulch/medusa_tb

For precise registration of the images, the GeFolki toolbox was used. https://github.com/aplyer/gefolki

 

##################################################

Labels: The train labels are available in two formats, a .png visualization image and a .tif label image. In the png image, 0 means no change and 255 means change. In the tif image, 0 means no change and 1 means change.

<ROOT_DIR>//cm/ contains: - cm.png - -cm.tif

Please note that prediction images should be formated as the -cm.tif rasters for upload and evaluation on DASE (http://dase.grss-ieee.org/).

(Update June 2020) Alternatively, you can use the test labels which are now provided in a separate archive, and compute standard metrics using the python notebook provided in this repo, along with a fulll script to train and classify fully-convolutional networks for change detection: https://github.com/rcdaudt/fully_convolutional_change_detection 

 

##################################################

Citation: If you use this dataset for your work, please use the following citation:

@inproceedings{daudt-igarss18,

author = {{Caye Daudt}, R. and {Le Saux}, B. and Boulch, A. and Gousseau, Y.},

title = {Urban Change Detection for Multispectral Earth Observation Using Convolutional Neural Networks},

booktitle = {IEEE International Geoscience and Remote Sensing Symposium (IGARSS'2018)},

venue = {Valencia, Spain},

month = {July},

year = {2018},

}

 

##################################################

Copyright: Sentinel Images: This dataset contains modified Copernicus data from 2015-2018. Original Copernicus Sentinel Data available from the European Space Agency (https://sentinel.esa.int).

Change labels: Change maps are released under Creative-Commons BY-NC-SA. For commercial purposes, please contact the authors.

 

Categories:
5016 Views

The data set contains 469 fundus images randomly selected from Telemed-R. All the 469 images are annotated by one senior pediatric ophthalmologist in SEH (Shenzhen Eye Hospital). An image annotation tool is developed based on the "VGG Image Annotator (VIA)" tool from the University of Oxford.The tool is included in this repository. The optic disk and macula in each fundus image are annotated by rectangular bounding boxes.This repository also contains related materials for our case study on this data set.

Instructions: 

All data are in standard image formats (.jpg / .png) or plain text format (.txt).

Categories:
Category: 
163 Views

Pedestrian detection and lane guidance 

Instructions: 

dnf

Categories:
130 Views

This dataset contains sheets of handwritten Telugu characters separated in boxes. It contains vowel, consonant, vowel-consonant and consonant-consonant pairs of Telugu characters. The purpose of this dataset is to act as a benchmark for Telugu handwritting related tasks like character recognition. There are 11 sheet layouts that produce 937 unique Telugu characters. Eighty three writers participated in generating the dataset and contributed 913 sheets in all. Each sheet layout contains 90 characters except the last which contains 83 characters where the last 10 are english numerals 0-9.

Categories:
490 Views

Pages