ENIGMA-360:
A Multi-view Dataset for Human Behavior
Understanding in Industrial Scenarios

Francesco Ragusa1,2, Rosario Leonardi1,2, Michele Mazzamuto2, Daniele Di Mauro2,
Camillo Quattrocchi1, Alessandro Passanisi1, Irene D'Ambra1, Antonino Furnari1,2,
Giovanni Maria Farinella1,2
1Department of Mathematics and Computer Science - University of Catania, Catania, IT
2Next Vision s.r.l. - Spin-off of the University of Catania, Catania, IT
ENIGMA-360 on GitHub
180
Egocentric Videos
180
Exocentric Videos
360
Synchronized Pairs
3
Tasks

Abstract

ENIGMA-360 is a new multi-view dataset acquired in a real industrial scenario. The dataset is composed of 180 egocentric and 180 exocentric procedural videos temporally synchronized, offering complementary information of the same scene. The 360 videos have been labeled with temporal and spatial annotations, enabling the study of different aspects of human behavior in the industrial domain. We provide baseline experiments for 3 tasks: Temporal Action Segmentation, Keystep Recognition, Egocentric Human-Object Interaction Detection. The dataset and its annotations are publicly available.

Multi-viewIndustrialEgocentricExocentric

The ENIGMA360 Dataset

Egocentric Videos

180 videos from the worker's point of view.

Exocentric Videos

180 videos from a fixed camera.

Synchronized Pairs

Each egocentric video is aligned with an exocentric one.

Annotations

Temporal & spatial labels, segmentation masks, 3D models.

Dataset labels example

Data Annotation

Temporal annotations
Hand-object interactions
Segmentation masks
3D models
DINOV2 features

Tasks

Temporal Action Segmentation
Keystep Recognition
Egocentric Human-Object Interaction

People

Francesco Ragusa
Francesco Ragusa
Rosario Leonardi
Rosario Leonardi
Michele Mazzamuto
Michele Mazzamuto
Daniele Di Mauro
Daniele Di Mauro
Camillo Quattrocchi
Camillo Quattrocchi
Alessandro Passanisi
Alessandro Passanisi
Irene D'Ambra
Irene D'Ambra
Antonino Furnari
Antonino Furnari
Giovanni Maria Farinella
Giovanni Maria Farinella