This dataset contains satellite images of areas of interest surrounding 30 different European airports. It also provides ground-truth annotations of flying airplanes in part of those images to support future research involving flying airplane detection. This dataset is part of the work entitled "Measuring economic activity from space: a case study using flying airplanes and COVID-19" published by the IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing. It contains modified Sentinel-2 data processed by Euro Data Cube.

Instructions: 

Details regarding dataset collection and usage are provided at https://github.com/maups/covid19-custom-script-contest

Categories:
275 Views

The region-based segmentation approach has been a major research area for many medical image applications. A vision guided autonomous system has used region-based segmentation information to operate heavy machinery and locomotive machines intended for computer vision applications. The dataset contains raw images in .png format fro brain tumor in various portions of brain.The dataset can be used fro training and testing. Images are calssified into three main regions as frontal lobe(level -1, level-2), optus-lobe(level-1), medula_lobe(level-1,level-2,level-3).

Categories:
975 Views

<p>This is the image dataset for satellite image processing&nbsp; which is a collection therml infrared and multispectral images .</p>

Instructions: 

Dataset images
Thermal infrared images and multispectral images
image size:512x512
format:
image:.tiff
file :.h5

Categories:
424 Views

In the field of 3D reconstruction, although there exist some standard datasets for evaluating the segmentation results of close-up 3D models, these datasets cannot be used to evaluate the segmentation results of 3D models based on satellite images. To address this issue, we provide a standard dataset for evaluating the segmentation results of satellite images and their corresponding DSMs. In this dataset, the satellite images maintain an exact correspondence with the DSMs, thus the segmentation results of both satellite images and DSMs can be evaluated by our proposed dataset.

Categories:
86 Views

DIDA is a new image-based historical handwritten digit dataset and collected from the Swedish historical handwritten document images between the year 1800 and 1940. It is the largest historical handwritten digit dataset which is introduced to the Optical Character Recognition (OCR) community to help the researchers to test their optical handwritten character recognition methods. To generate DIDA, 250,000 single digits and 200,000 multi-digits are cropped from 75,000 different document images. 

Categories:
428 Views

The accompanying dataset for the CVSports 2021 paper: DeepDarts: Modeling Keypoints as Objects for Automatic Scoring in Darts using a Single Camera

Paper Abstract:

Instructions: 

The recommended way to load the labels is to use the pandas Python package:

import pandas as pd

labels = pd.read_pickle("labels.pkl")

See github repository for more information: https://github.com/wmcnally/deep-darts

Categories:
297 Views

We present a 4D Light Field (LF) video dataset collected by the camera matrix, to be used for designing and testing algorithms and systems for LF video coding and processing. For collecting these videos, a 10x10 LF capture matrix composed of 100 cameras is designed and implemented, and the resolution of each camera is 1920x1056. The videos are taken in real and varying illumination conditions. The dataset contains a total of nine groups of LF videos, of which eight groups are collected with a fixed camera matrix position and a fixed orientation.

Instructions: 

The stream data of all LF videos are stored under the directory LF_video_data, and each video contains 100 stream files. The stream files are in 1920 x 1056 24 frames per second MP4 format without audio. There are 100 mat files in the Calibration_mat directory, which store all camera parameters of 100 lenses within the matrix. The depth estimation results are stored under the directory Depth_evaluation_result, which is for reference only.

Categories:
390 Views

The Jackal UGV, from Clearpath Robotics, was used as the data collecting platform. This skid-steer four-wheel-drive vehicle comes with an onboard IMU, two DC motors with encoders that measure wheel angular speeds, and current sensors that measure motor current outputs. On each side of the robot, the front wheel and back wheel are jointed with a gearbox and so spin together at the same rate and direction. The IMU provided vehicle attitude measurements in terms of Euler angles, as well as linear acceleration and angular rate of the vehicle body in three Euclidean axes.

Instructions: 

Each HDF5 file contains four types of data entries: timestamps, signals, images, and labels. A .ipynb code example is included to demonstrate how to retrieve and format data appropriately. 

The .ipynb script requires h5py, numpy, and matplotlib libraries.

** If you plan to load the entire dataset into your memory, make sure your PC has >16 Gb RAM

Categories:
124 Views

The dermoscopic images considered in the paper "Dermoscopic Image Classification with Neural Style Transfer" are available for public download through the ISIC database (https://www.isic-archive.com/#!/topWithHeader/wideContentTop/main). These are 24-bit JPEG images with a typical resolution of 768 × 512 pixels. However, not all the images in the database are in satisfactory condition.

Categories:
211 Views

In a context of rapid urban evolution, there is a need of surveying cities. Nowadays predictive models based on machine learning require large amount of data to be trained, hence the necessity of providing some public dataset allowing to follow up urban evolution. While most of changes occurs onto the vertical axis, there is no public change detection dataset composed of 3D point clouds and directly annotated according to the change at point level yet.

Instructions: 

Urban Point Clouds simulator

We have developed a simulator to generate time series of point clouds (PCs) for urban datasets. Given a 3D model of a city, the simulator allows us to introduce random changes in the model and generates a synthetic aerial LiDAR Survey (ALS) above the city. The 3D model is in practice issued from a real city, e.g. with a Level of Detail 2 (LoD2) precision. From this model, we extract each existing building as well as the ground. By adding or removing buildings in the model, we can simulate construction or demolition of buildings. Notice that depending on area the ground is not necessarily flat. The simulator allows us to obtain as many 3D PCs over urban changed areas as needed. It could be useful especially for deep learning supervised approaches that require lots of training dates. Moreover, the created PCs are all directly annotated by the simulator according to the changes, thus no time-consuming manual annotation needs to be done with this process.

For each obtained model, the ALS simulation is performed thanks to a flight plan and ray tracing with the Visualisation ToolKit (VTK) python library.  Space between flight lines is computed in accordance to predefined parameters such as resolution, covering between swaths and scanning angle. Following this computation, a flight plan is set with a random starting position and direction of flight in order to introduce more variability between two acquisitions. Moreover, Gaussian noise can be added to simulate errors and lack of precision in LiDAR range measuring and scan direction.

Dataset Description

To conduct fair qualitative and quantitative evaluation of PC change detection techniques, we have build some datasets based on LoD2 models of the first and second districts of Lyon  (https://geo.data.gouv.fr/datasets/0731989349742867f8e659b4d70b707612bece89), France. For each simulation, buildings have been added or removed to introduce changes in the model and to generate a large number of pairs of PCs. We also consider various initial states across simulations, and randomly update the set of buildings from the first date through random addition or deletion of  buildings to create the second landscape. In addition, flight starting position and direction are always set randomly. As a consequence, the acquisition patterns will not be the same between generated PCs, thus each acquisition may not have exactly the same visible or hidden parts.

From terrestrial LiDAR surveying to photogrammetric acquisition by satellite images, there exist many different types of sensors and acquisition pipelines to obtain 3D point clouds for urban areas, resulting in PCs with different characteristics. {By providing different acquisition parameters to our simulator}, our goal was to provide a variety of sub-datasets with heterogeneous qualities to reproduce the real variability of LiDAR sensors or to mimic datasets coming from a photogrammetric pipeline with satellite images (by using a tight scan angle with high noise). Thus, we generated the following sub-datasets:

  • ALS with low resolution, low noise for both dates
  • ALS with high resolution, low noise for both dates
  • ALS with low resolution, high noise for both dates
  • ALS with low resolution, high noise, tight scan angle (mimicking photogrammetric acquisition from satellite images) for both dates
  • Multi-sensor data, with low resolution, high noise at date 1, and high resolution, low~noise at date 2

Notice that sub-datasets 3 and 4 are quite similar but the latter provides less visible facades, thanks to the smaller scanning angle and overlapping percentage.

Finally, for the first configuration (ALS low resolution, low noise), we provided the 3~following different training sets:

  • Small training set: 1 simulation
  • Normal training set: 10 simulations
  • Large training set: 50 simulations

More details about the configuration of acquisition are provided in the documentation file and in the publication Change Detection in Urban Point Clouds: An Experimental Comparison with Simulated 3D Datasets, de Gélis et al. (2021)

Technical details

All PCs are available at PLY format. Each train, val, test folder contains sub-folders containing pairs of PCs : pointCloud0.ply and pointCloud1.ply for both first and second dates.

Each ply file contain the coordinates X Y Z of each points and the label:

  • 0 for unchanged points
  • 1 for points on a new building
  • 2 for for points on a destruction.

The label is given in a scalar field named label_ch. Notice that the first PC (pointCloud0.ply) has a label field even if it is set at 0 for every points because change are set in comparison to previous date.

Citation

If you use this dataset for your work, please use the following citation:

@article{degelis2021change,  title={Change Detection in Urban Point Clouds: An Experimental Comparison with Simulated 3D Datasets},  author={{de G\'elis}, I. and Lef\`evre, S. and Corpetti, T. },  journal={Remote Sensing},  volume={13},  pages={2629},  year={2021},  publisher={Multidisciplinary Digital Publishing Institute}}

Categories:
281 Views

Pages