Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Dataset . 2017
License: CC BY
Data sources: Datacite
addClaim

Cad 120 Affordance Dataset

Authors: Sawatzky, Johann; Srikantha, Abhilash; Gall, Juergen;

Cad 120 Affordance Dataset

Abstract

% ============================================================================== % CAD 120 Affordance Dataset % Version 1.0 % ------------------------------------------------------------------------------ % If you use the dataset please cite: % % Johann Sawatzky, Abhilash Srikantha, Juergen Gall. % Weakly Supervised Affordance Detection. % IEEE Conference on Computer Vision and Pattern Recognition (CVPR'17) % % and % % H. S. Koppula and A. Saxena. % Physically grounded spatio-temporal object affordances. % European Conference on Computer Vision (ECCV'14) % % Any bugs or questions, please email sawatzky AT iai DOT uni-bonn DOT de. % ============================================================================== This is the CAD 120 Affordance Segmentation Dataset based on the Cornell Activity Dataset CAD 120 (see http://pr.cs.cornell.edu/humanactivities/data.php). Content frames/*.png: RGB frames selected from Cornell Activity Dataset. To find out the location of the frame in the original videos, see video_info.txt. object_crop_images/*.png image crops taken from the selected frames and resized to 321*321. Each crop is a padded bounding box of an object the human interacts with in the video. Due to the padding, the crops may contain background and other objects. In each selected frame, each bounding box was processed. The bounding boxes are already given in the Cornell Activity Dataset. The 5-digit number gives the frame number, the second number gives the bounding box number within the frame. segmentation_mat/*.mat 321*321*6 segmentation masks for the image crops. Each channel corresponds to an affordance (openabe, cuttable, pourable, containable, supportable, holdable, in this order). All pixels belonging to a particular affordance are labeled 1 in the respective channel, otherwise 0. segmentation_png/*.png 321*321 png images, each containing the binary mask for one of the affordances. lists/*.txt Lists containing the train and test sets for two splits. The actor split ensures that train and test images stem from different videos with different actors while the object split ensures that train and test data have no (central) object classes in common. The train sets are additionally subdivided into 3 subsets A,B and C. For the actor split, the subsets stem from different videos. For the object split, each subset contains every third crop of the train set. crop_coordinate_info.txt Maps image crops to their coordinates in the frames. hpose_info.txt Maps frames to 2d human pose coordinates. Hand annotated by us. object_info.txt Maps image crops to the (central) object it contains. visible_affordance_info.txt Maps image crops to affordances visible in this crop %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%55 The crops contain the following object classes: 1.table 2.kettle 3.plate 4.bottle 5.thermal cup 6.knife 7.medicine box 8.can 9.microwave 10.paper box 11.bowl 12.mug Affordances in our set: 1.openable 2.cuttable 3.pourable 4.containable 5.supportable 6.holdable Note that our object affordance labeling differs from the Cornell Activity Dataset: E.g. the cap of a pizza box is considered to be supportable.

Acknowledgments. The work has been financially sup- ported by the DFG projects GA 1927/5-1 (DFG Research Unit FOR 2535 Anticipating Human Behavior) and GA 1927/2-2 (DFG Research Unit FOR 1505 Mapping on De- mand).

{"references": ["Sawatzky, J., Srikantha, A., Gall, J.: Weakly supervised affordance detection. CVPR (2017)"]}

Related Organizations
Keywords

robotics, semantic image segmentation, affordances, convolutional neural network, attributes, anticipating human behavior, computer vision, mapping on demand, weakly supervised learning

  • BIP!
    Impact byBIP!
    selected citations
    These citations are derived from selected sources.
    This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    0
    popularity
    This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
    Average
    influence
    This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    Average
    impulse
    This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
    Average
    OpenAIRE UsageCounts
    Usage byUsageCounts
    visibility views 320
    download downloads 118
  • 320
    views
    118
    downloads
    Powered byOpenAIRE UsageCounts
Powered by OpenAIRE graph
Found an issue? Give us feedback
visibility
download
selected citations
These citations are derived from selected sources.
This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Citations provided by BIP!
popularity
This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
BIP!Popularity provided by BIP!
influence
This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Influence provided by BIP!
impulse
This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
BIP!Impulse provided by BIP!
views
OpenAIRE UsageCountsViews provided by UsageCounts
downloads
OpenAIRE UsageCountsDownloads provided by UsageCounts
0
Average
Average
Average
320
118