MAINTENANCE ACTIONS DATASET

The first large-scale egocentric dataset for maintenance and repair actions

Egocentric Vision Industrial Maintenance Action Recognition
Download Dataset
Scroll

News

Nov 2024

Paper Accepted

Our paper has been accepted at [Conference Name].

Oct 2024

Benchmark Released

Evaluation server and baselines are now available.

What is MAD?

The Maintenance Actions Dataset (MAD) is a large-scale egocentric video dataset capturing real-world maintenance and repair activities. Recorded from a first-person perspective using head-mounted cameras, MAD provides a unique window into procedural tasks performed by technicians in various maintenance scenarios.

Unlike existing egocentric datasets that focus on daily activities or cooking, MAD specifically targets the industrial maintenance domain, featuring complex tool usage, multi-step procedures, and fine-grained hand-object interactions.

Dataset Highlights

0
Hours of Video
0
Million Frames
0
Participants
0
Action Classes

Characteristics

Head-Mounted Camera

First-person perspective capturing the technician's viewpoint during maintenance tasks.

Tool Interactions

Rich annotations of tool usage and hand-object interactions in maintenance scenarios.

Procedural Tasks

Multi-step procedures with temporal ordering and hierarchical action structure.

Industrial Domain

Real maintenance scenarios including assembly, repair, and inspection tasks.

Temporal Annotations

Dense frame-level annotations with action start/end times and verb-noun labels.

High Resolution

Full HD recordings at high frame rate for detailed action analysis.

Dataset Statistics

Action Distribution Chart

Add your visualization here

Verb-Noun Distribution

Add your visualization here

Annotation Pipeline

Add your visualization here

Explore

Browse sample images from the MAD dataset with annotations. Use filters to explore specific actions, objects, or scenarios.

Select filters and click "Random Sample" to explore the dataset

Full explorer coming soon!
0 / 0

Annotations

No image loaded

Metadata

Download

The dataset is available for academic research purposes. Please read and agree to the terms of use before downloading.

Video Data

Full resolution RGB videos

~XXX GB
Download

Annotations

Action segments, labels, splits

~XX MB
Download

Code & Tools

Data loaders, evaluation scripts

GitHub
Repository

Annotations Repository

All annotations (Train/Val/Test splits) are available at our GitHub repository:

MAD-annotations

Citation

If you use this dataset, please cite our paper:

@article{author2024mad,
    title={MAD: Maintenance Actions Dataset for Egocentric Video Understanding},
    author={Author Name and Co-Author Name},
    journal={Conference/Journal Name},
    year={2024}
}

Benchmark

We provide several benchmark tasks for evaluating methods on MAD:

Action Recognition

Task 1

Task: Assign a (verb, noun) label to a trimmed video segment.

Metrics: Top-1/5 accuracy for verb, noun, and action.

Action Detection

Task 2

Task: Detect action start/end times in untrimmed videos.

Metrics: Mean Average Precision (mAP) @ various IoU thresholds.

Action Anticipation

Task 3

Task: Predict the next action before it starts.

Metrics: Top-5 recall at various anticipation times.

The Team

PI Name

PI Name

Principal Investigator Ben-Gurion University
Team Member 1

Team Member 1

PhD Student Ben-Gurion University
Team Member 2

Team Member 2

PhD Student Ben-Gurion University
Team Member 3

Team Member 3

MSc Student Ben-Gurion University

Research Funding

This research was supported by:

  • Funding Source 1
  • Funding Source 2