Mesh 3
☆ DOFS: A Real-world 3D Deformable Object Dataset with Full Spatial Information for Dynamics Model Learning
This work proposes DOFS, a pilot dataset of 3D deformable objects (DOs)
(e.g., elasto-plastic objects) with full spatial information (i.e., top, side,
and bottom information) using a novel and low-cost data collection platform
with a transparent operating plane. The dataset consists of active manipulation
action, multi-view RGB-D images, well-registered point clouds, 3D deformed
mesh, and 3D occupancy with semantics, using a pinching strategy with a
two-parallel-finger gripper. In addition, we trained a neural network with the
down-sampled 3D occupancy and action as input to model the dynamics of an
elasto-plastic object. Our dataset and all CADs of the data collection system
will be released soon on our website.
comment: 5 pages, 6 figures, 2024 CoRL Workshop on Learning Robot Fine and
Dexterous Manipulation: Perception and Control
☆ SS3DM: Benchmarking Street-View Surface Reconstruction with a Synthetic 3D Mesh Dataset NeurIPS 2024
Reconstructing accurate 3D surfaces for street-view scenarios is crucial for
applications such as digital entertainment and autonomous driving simulation.
However, existing street-view datasets, including KITTI, Waymo, and nuScenes,
only offer noisy LiDAR points as ground-truth data for geometric evaluation of
reconstructed surfaces. These geometric ground-truths often lack the necessary
precision to evaluate surface positions and do not provide data for assessing
surface normals. To overcome these challenges, we introduce the SS3DM dataset,
comprising precise \textbf{S}ynthetic \textbf{S}treet-view \textbf{3D}
\textbf{M}esh models exported from the CARLA simulator. These mesh models
facilitate accurate position evaluation and include normal vectors for
evaluating surface normal. To simulate the input data in realistic driving
scenarios for 3D reconstruction, we virtually drive a vehicle equipped with six
RGB cameras and five LiDAR sensors in diverse outdoor scenes. Leveraging this
dataset, we establish a benchmark for state-of-the-art surface reconstruction
methods, providing a comprehensive evaluation of the associated challenges.
For more information, visit our homepage at https://ss3dm.top.
comment: NeurIPS 2024, Track on Datasets and Benchmarks
♻ ☆ Multi-hypotheses Conditioned Point Cloud Diffusion for 3D Human Reconstruction from Occluded Images NeurIPS 2024
3D human shape reconstruction under severe occlusion due to human-object or
human-human interaction is a challenging problem. Parametric models i.e.,
SMPL(-X), which are based on the statistics across human shapes, can represent
whole human body shapes but are limited to minimally-clothed human shapes.
Implicit-function-based methods extract features from the parametric models to
employ prior knowledge of human bodies and can capture geometric details such
as clothing and hair. However, they often struggle to handle misaligned
parametric models and inpaint occluded regions given a single RGB image. In
this work, we propose a novel pipeline, MHCDIFF, Multi-hypotheses Conditioned
Point Cloud Diffusion, composed of point cloud diffusion conditioned on
probabilistic distributions for pixel-aligned detailed 3D human reconstruction
under occlusion. Compared to previous implicit-function-based methods, the
point cloud diffusion model can capture the global consistent features to
generate the occluded regions, and the denoising process corrects the
misaligned SMPL meshes. The core of MHCDIFF is extracting local features from
multiple hypothesized SMPL(-X) meshes and aggregating the set of features to
condition the diffusion model. In the experiments on CAPE and MultiHuman
datasets, the proposed method outperforms various SOTA methods based on SMPL,
implicit functions, point cloud diffusion, and their combined, under synthetic
and real occlusions. Our code is publicly available at
https://donghwankim0101.github.io/projects/mhcdiff/ .
comment: 17 pages, 7 figures, accepted NeurIPS 2024