Mesh 1
☆ MITO: Enabling Non-Line-of-Sight Perception using Millimeter-waves through Real-World Datasets and Simulation Tools
We present MITO, the first dataset of multi-spectral millimeter-wave (mmWave)
images of everyday objects. Unlike visible light, mmWave signals can image
through everyday occlusions (e.g., cardboard boxes, fabric, plastic). However,
due to the dearth of publicly-available mmWave images and the interdisciplinary
challenges in collecting and processing mmWave signals, it remains difficult
today for computer vision researchers to develop mmWave-based non-line-of-sight
perception algorithms and models.
To overcome these challenges, we introduce a real-world dataset and
open-source simulation tool for mmWave imaging. The dataset is acquired using a
UR5 robotic arm with two mmWave radars operating at different frequencies and
an RGB-D camera. Through a signal processing pipeline, we capture and create
over 580 real-world 3D mmWave images from over 76 different objects in the YCB
dataset, a standard dataset for robotics manipulation. We provide real-world
mmWave images in line-of-sight and non-line-of-sight, as well as RGB-D images
and ground truth segmentation masks. We also develop an open-source simulation
tool that can be used to generate synthetic mmWave images for any 3D triangle
mesh, which achieves a median F-Score of 94% when compared to real-world mmWave
images.
We show the usefulness of this dataset and simulation tool in multiple CV
tasks in non-line-of-sight. First, we perform object segmentation for mmWave
images using the segment anything model (SAM), and achieve a median precision
and recall of 92.6% and 64%. Second, we train a classifier that can recognize
objects in non-line-of-sight. It is trained on synthetic images and can
classify real-world images with 85% accuracy.
We believe MITO will be a valuable resource for computer vision researchers
in developing non-line-of-sight perception, similar to how early camera-based
datasets shaped the field.