ObChange Dataset

DOI

This dataset that can be used to evaluate methods, which are able to detect changed objects when comparing two recordings of the same environment at different time instances. Based on the labeled ground truth objects, it is possible to differentiate between static, moved, removed and novel objects. Dataset Description The dataset was recorded with an Asus Xtion PRO Live mounted on the HSR robot. We provide scenes from five different rooms or parts of rooms, namely a big room, a small room, a living area, a kitchen counter and an office desk. Each room is visited by the robot at least five times while between each run a subset of objects from the YCB Object and Model Set (YCB)[1] is re-arranged in the room. In total we generated 26 recordings. For each recording between three and 17 objects are placed (219 in total). Furthermore, furniture and permanent background objects are slightly rearranged. These changes are not labeled because for most service robot tasks, this is not relevant. Assuming most objects are placed on horizontal surfaces, we extracted planes in each room in a pre-processing step (excluding the floor). For each surface, all frames from the recording where it is visible are extracted and used as the input for ElasticFusion[2]. This results in a total of 34 reconstructed surfaces. We provide pointwise annotation of the YCB objects for each surface reconstruction from each recording. Images of exemplary surface reconstructions can be found here: https://www.acin.tuwien.ac.at/vision-for-robotics/software-tools/obchange/   Dataset Structure The file structure of ObChange.zip is the following: Room    - scene2         - planes              - 0                   - merged_plane_clouds_ds002.pcd                   - merged_plane_clouds_ds002.anno                   - merged_plane_clouds_ds002_GT.anno              - 1                   - merged_plane_clouds_ds002.pcd                   - merged_plane_clouds_ds002.anno                   - merged_plane_clouds_ds002_GT.anno              - ...         table.txt    - scene3 The pcd-file contains the reconstruction of the surface. The merged_plane_clouds_ds002.anno lists the YCB objects visible in the reconstruction and merged_plane_clouds_ds002_GT.anno contains the point indices of the reconstruction corresponding to the YCB objects together with the corresponding object name. The last element for each object is a bool value indicating if the object is on the floor (and was reconstructed by chance). The table.txt lists for each detected plane the centroid, height, convex hull points and plane coefficients. We provide the original input data for each room. The zip-files contain the rosbag file for each recording. Each rosbag contains the tf-tree and the RGB and depth stream, as well as the camera intrinsic. Additionally, the semantically annotated Voxblox[3] reconstruction created with SparseConvNet[4] is provided for each recording.    You may also be interested in Object Change Detection Dataset of Indoor Environments. It uses the same input data, but the ground truth annotation is based on a full room reconstruction instead of individual planes.   Acknowledgements The research leading to these results has received funding from the Austrian Science Fund (FWF) under grant agreement Nos. I3969-N30 (InDex), I3967-N30 (BURG) and from the Austrian Research Promotion Agency (FFG) under grant agreement 879878 (K4R). References [1] B. Calli, A. Singh, J. Bruce, A. Walsman, K. Konolige, S. Srinivasa, P. Abbeel, A. M. Dollar, Yale-CMU-Berkeley dataset for robotic manipulation research, The International Journal of Robotics Research, vol. 36, Issue 3, pp. 261 – 268, April 2017. [2] T. Whelan, S. Leutenegger, R. Salas-Moreno, B. Glocker, A. Davison, ElasticFusion: Dense SLAM without a pose graph, Proceedings of Robotics: Science and Systems, July 2015. [3] H. Oleynikova, Z. Taylor, M. Fehr, R. Siegwart, J. Nieto, Juan, Voxblox: Incremental 3D Euclidean Signed Distance Fields for On-Board MAV Planning, in Proceedings of IEEE International Conference on Intelligent Robots and Systems (IROS), pp. 1366-1373, 2017. [4] B. Graham, M. Engelcke, L. van der Maaten, 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks, Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), pp. 9224 – 9232, 2018.

Identifier
DOI https://doi.org/10.48436/y3ggy-hxp10
Related Identifier IsSupplementTo https://doi.org/10.3389/frobt.2022.828732
Related Identifier IsSupplementTo https://github.com/edith-langer/tidyupvisionpipeline/tree/Frontiers_submission
Related Identifier IsVersionOf https://doi.org/10.48436/r2sxr-tsc12
Metadata Access https://researchdata.tuwien.ac.at/oai2d?verb=GetRecord&metadataPrefix=oai_datacite&identifier=oai:researchdata.tuwien.ac.at:y3ggy-hxp10
Provenance
Creator Langer, Edith
Publisher TU Wien
Publication Year 2022
Rights Creative Commons Attribution 4.0 International; https://creativecommons.org/licenses/by/4.0/legalcode
OpenAccess true
Contact tudata(at)tuwien.ac.at
Representation
Resource Type Dataset
Discipline Other