Our complex street scene(CSS) containing strong light and heavy shadow scenes mainly comes from the Kitti dataset.  Our datasets are captured by driving around Karlsruhe's mid-size city, in rural areas, and on highways. We equipped a standard station wagon with two high-resolution color and grayscale video cameras. Up to 15 cars and 30 pedestrians are visible per image. We aim to verify the performance of the algorithm in specific and complex street scenes.

Categories:
44 Views

The color fractal images with correlated RGB color components were generated using the midpoint displacement alogrithm, using vectorial increments in the RGB color space, according to a multivariate Gaussian distribution specified by the variance-covariance matrix. This data set contains two sets of 25 color fractal images with two color components, of varying complexity expressed as the color fractal dimension, as a function of (i) the Hurst coefficient that was varied from 0.1 to 0.9 in steps of 0.2 and (ii) the correlation coefficient between the red and green color channels.

Instructions: 

This data set is for research purposes only. Please consider citing the paper entitled "Fractal Dimension of Color Fractal Images with Correlated Color Components", IEEE Transactions on Image Processing, 2020: https://doi.org/10.1109/TIP.2020.3011283 

Categories:
57 Views

We introduce HUMAN4D, a large and multimodal 4D dataset that contains a variety of human activities simultaneously captured by a professional marker-based MoCap, a volumetric capture and an audio recording system. By capturing 2 female and 2 male professional actors performing various full-body movements and expressions, HUMAN4D provides a diverse set of motions and poses encountered as part of single- and multi-person daily, physical and social activities (jumping, dancing, etc.), along with multi-RGBD (mRGBD), volumetric and audio data. Despite the existence of multi-view color datasets c

Instructions: 

* At this moment, the paper of this dataset is under review. The dataset is going to be fully published along with the publication of the paper, while in the meanwhile, more parts of the dataset will be uploaded.

The dataset includes multi-view RGBD, 3D/2D pose, volumetric (mesh/point-cloud/3D character) and audio data along with metadata for spatiotemporal alignment.

The full dataset is splitted per subject and per activity per modality.

There are also two benchmarking subsets, H4D1 for single-person and H4D2 for two-person sequences, respectively.

The fornats are:

  • mRGBD: *.png
  • 3D/2D poses: *.npy
  • volumetric (mesh/point-cloud/): *.ply
  • 3D character: *.fbx
  • metadata: *.txt, *.json

 

Categories:
255 Views

This dataset is for light field image augmentaion. The dataset contains 100 pairs of light field image, each of them consists of "original" and "modified". "Original" is light field image with only background, "modified" is light field image with exactly same background and an object on it.

Categories:
44 Views

To improve reproductivity of our papar, we would upload experimental data and resources of evaluations.

Categories:
73 Views

Solving the external perception problem for autonomous vehicles and driver-assistance systems requires accurate and robust driving scene perception in both regularly-occurring driving scenarios (termed “common cases”) and rare outlier driving scenarios (termed “edge cases”). In order to develop and evaluate driving scene perception models at scale, and more importantly, covering potential edge cases from the real world, we take advantage of the MIT-AVT Clustered Driving Scene Dataset and build a subset for the semantic scene segmentation task.

Instructions: 

 

MIT DriveSeg (Semi-auto) Dataset is a set of forward facing frame-by-frame pixel level semantic labeled dataset (coarsely annotated through a novel semiautomatic annotation approach) captured from moving vehicles driving in a range of real world scenarios drawn from MIT Advanced Vehicle Technology (AVT) Consortium data.

 

Technical Summary

Video data - Sixty seven 10 second 720P (1280x720) 30 fps videos (20,100 frames)

Class definitions (12) - vehicle, pedestrian, road, sidewalk, bicycle, motorcycle, building, terrain (horizontal vegetation), vegetation (vertical vegetation), pole, traffic light, and traffic sign

 

Technical Specifications, Open Source Licensing and Citation Information

Ding, L., Glazer, M., Terwilliger, J., Reimer, B. & Fridman, L. (2020). MIT DriveSeg (Semi-auto) Dataset: Large-scale Semi-automated Annotation of Semantic Driving Scenes. Massachusetts Institute of Technology AgeLab Technical Report 2020-2, Cambridge, MA. (pdf)

Ding, L., Terwilliger, J., Sherony, R., Reimer, B. & Fridman, L. (2020). MIT DriveSeg (Manual) Dataset. IEEE Dataport. DOI: 10.21227/nb3n-kk46.

 

Attribution and Contact Information

This work was done in collaboration with the Toyota Collaborative Safety Research Center (CSRC). For more information, click here.

For any questions related to this dataset or requests to remove identifying information please contact driveseg@mit.edu.

 

Categories:
922 Views

Semantic scene segmentation has primarily been addressed by forming representations of single images both with supervised and unsupervised methods. The problem of semantic segmentation in dynamic scenes has begun to recently receive attention with video object segmentation approaches. What is not known is how much extra information the temporal dynamics of the visual scene carries that is complimentary to the information available in the individual frames of the video.

Instructions: 

 

MIT DriveSeg (Manual) Dataset is a forward facing frame-by-frame pixel level semantic labeled dataset captured from a moving vehicle during continuous daylight driving through a crowded city street.

The dataset can be downloaded from the IEEE DataPort or demoed as a video.

 

Technical Summary

Video data - 2 minutes 47 seconds (5,000 frame) 1080P (1920x1080) 30 fps

Class definitions (12) - vehicle, pedestrian, road, sidewalk, bicycle, motorcycle, building, terrain (horizontal vegetation), vegetation (vertical vegetation), pole, traffic light, and traffic sign

 

Technical Specifications, Open Source Licensing and Citation Information

Ding, L., Terwilliger, J., Sherony, R., Reimer, B. & Fridman, L. (2020). MIT DriveSeg (Manual) Dataset for Dynamic Driving Scene Segmentation. Massachusetts Institute of Technology AgeLab Technical Report 2020-1, Cambridge, MA. (pdf)

Ding, L., Terwilliger, J., Sherony, R., Reimer, B. & Fridman, L. (2020). MIT DriveSeg (Manual) Dataset. IEEE Dataport. DOI: 10.21227/mmke-dv03.

 

Related Research

Ding, L., Terwilliger, J., Sherony, R., Reimer. B. & Fridman, L. (2019). Value of Temporal Dynamics Information in Driving Scene Segmentation. arXiv preprint arXiv:1904.00758. (link)

 

Attribution and Contact Information

This work was done in collaboration with the Toyota Collaborative Safety Research Center (CSRC). For more information, click here.

For any questions related to this dataset or requests to remove Identifying information please contact driveseg@mit.edu.

 

Categories:
1615 Views

PRIME-FP20 dataset is established for development and evaluation of retinal vessel segmentation algorithms in ultra-widefield fundus photography. PRIME-FP20 provides 15 high-resolution ultra-widefield fundus photography images acquired using the Optos 200Tx camera (Optos plc, Dunfermline, United Kingdom), the corresponding labeled binary vessel maps, and the corresponding binary masks for the FOV of the images.

Instructions: 

Ultra-widefield fundus photography images and the corresponding labeled vessel maps and binary masks are provided where the file names indicate the correspondence between them.

Currently, only a sample low-resolution image is provided. The full set of high-resolution images will be provided upon the publication of the associated paper, which is currently submitted for review.

Categories:
242 Views

This set contains 1450 fundus images with 899 glaucoma data and 551 normal data.

All text about patient information and the date that the associated images were collected are replaced by 0, which is black.

Categories:
229 Views

The graphical user interface (GUI) for memristor is designed under MATLAB GUIDE to study and visualize it's characteristics in interactive way. The GUI is developed so as to study by taking sinusoidal, square and triangular signal waveform only.

Categories:
118 Views

Pages