Hay Oussama Abdul, Huang Xiaoqian, Ayyad Abdulla, Sherif Eslam, Almadhoun Randa, Abdulrahman Yusra, Seneviratne Lakmal, Abusafieh Abdulqader, Zweiri Yahya
Advanced Research and Innovation Center (ARIC), Khalifa University of Science and Technology, Abu Dhabi, UAE.
Faculty of Technology, Department of Computer Science, University of Sunderland, Sunderland, UK.
Sci Data. 2025 Feb 12;12(1):245. doi: 10.1038/s41597-025-04536-5.
Robotic automation requires precise object pose estimation for effective grasping and manipulation. With their high dynamic range and temporal resolution, event-based cameras offer a promising alternative to conventional cameras. Despite their success in tracking, segmentation, classification, obstacle avoidance, and navigation, their use for 6D object pose estimation is relatively unexplored due to the lack of datasets. This paper introduces an extensive dataset based on Yale-CMU-Berkeley (YCB) objects, including event packets with associated poses, spike images, masks, 3D bounding box coordinates, segmented events, and a 3-channel event image for validation. Featuring 13 YCB objects, the dataset covers both cluttered and uncluttered scenes across 18 scenarios with varying speeds and illumination. It contains 306 sequences, totaling over an hour and around 1.5 billion events, making it the largest and most diverse event-based dataset for object pose estimation. This resource aims to support researchers in developing and testing object pose estimation algorithms and solutions.
机器人自动化需要精确的物体位姿估计,以便进行有效的抓取和操作。基于其高动态范围和时间分辨率,基于事件的相机为传统相机提供了一个有前景的替代方案。尽管它们在跟踪、分割、分类、避障和导航方面取得了成功,但由于缺乏数据集,它们在6D物体位姿估计中的应用相对较少被探索。本文介绍了一个基于耶鲁-卡内基梅隆大学-伯克利分校(YCB)物体的广泛数据集,包括带有相关位姿的事件包、脉冲图像、掩码、3D边界框坐标、分割事件以及用于验证的3通道事件图像。该数据集包含13个YCB物体,涵盖了18种场景下的杂乱和整洁场景,速度和光照各不相同。它包含306个序列,总计超过一小时,约15亿个事件,使其成为用于物体位姿估计的最大且最多样化的基于事件的数据集。该资源旨在支持研究人员开发和测试物体位姿估计算法及解决方案。