Simple Science

Cutting edge science explained simply

# Computer Science# Computer Vision and Pattern Recognition

Introducing the ENIGMA-51 Dataset for Industrial Interaction Study

A new dataset to enhance understanding of human-object interactions in industrial settings.

― 7 min read


ENIGMA-51: A New DatasetENIGMA-51: A New DatasetUnveiledanalysis and worker support.Dataset enhances industrial task
Table of Contents

In our daily lives, we constantly interact with various objects to complete tasks. In workplaces, especially in industrial settings, these interactions can be intricate, requiring specific tools and actions. For example, when fixing equipment, a worker might use tools like a screwdriver or an oscilloscope, while also keeping safety in mind.

To support workers in such environments, it’s important to create intelligent systems that can recognize and help manage these interactions. This is where technology, like smart glasses that capture video while a worker’s hands are free, comes into play. Such systems could guide workers through procedures, warn them about safety risks, and suggest the next steps in their tasks.

This article presents a new dataset known as ENIGMA-51. This dataset was created to study how people interact with objects in industrial settings. It consists of a series of videos showing workers repairing electrical boards while using various tools. These videos have been recorded with detailed annotations to capture each interaction between the worker and the objects involved.

The ENIGMA-51 Dataset

The ENIGMA-51 dataset contains videos collected from 19 participants who performed repair tasks in an industrial environment. Each participant used smart glasses to record their actions as they followed audio instructions. The dataset includes 51 videos, each showing a complete process of repairing an electrical board.

These videos provide a wealth of information about how people engage with tools and machines. For every interaction, the dataset is labeled with specific actions, objects involved, and timeframes.

Purpose of ENIGMA-51

The main goal of the ENIGMA-51 dataset is to facilitate the development of systems that can assist workers in an industrial setting. By understanding how humans interact with objects, we can make better tools that can help improve efficiency and safety. The dataset allows for the study of various tasks related to human-object interactions, such as recognizing actions, predicting future actions, and understanding spoken instructions.

Detailed Interaction Study

Every day, workers perform numerous tasks that involve complex interactions with tools and machinery. In the context of industrial work, these tasks need to be efficient to ensure productivity and safety. The ENIGMA-51 dataset aims to address several key aspects of these interactions.

Action Detection

One of the key tasks in studying human-object interactions is recognizing actions. For instance, understanding when a worker is taking a tool or releasing it can provide insights into their behavior. The ENIGMA-51 dataset allows researchers to detect four main actions: “take,” “release,” “first-contact,” and “de-contact.”

  • Take: When a worker picks up a tool.
  • Release: When a worker puts down a tool.
  • First-contact: When a worker touches a tool for the first time.
  • De-contact: When a worker stops touching a tool.

These actions are vital for creating systems that can analyze and predict worker behavior, contributing to workplace safety and efficiency.

Egocentric Human-Object Interaction Detection

Another significant aspect of human-object interaction is egocentric detection, which refers to recognizing how a worker interacts with objects from their point of view. The dataset focuses on identifying which hand is involved, the state of that hand (whether it is in contact with an object), and the object being handled.

Such detection involves not only recognizing the object but also understanding the context of the interaction. For example, knowing if a worker’s left hand is in contact with a screwdriver can provide insights into the task being performed.

Anticipating Future Interactions

The dataset also allows researchers to anticipate future interactions. By analyzing past actions, systems can predict the next tool a worker might need or when they might need to perform a specific action. This predictive capability can enhance training systems and provide real-time assistance to workers, minimizing mistakes and improving safety.

Natural Language Understanding

In addition to visual data, the ENIGMA-51 dataset captures spoken instructions given to participants during the recording. This information is valuable for developing systems that can understand and interpret natural language commands.

For example, if a worker says, “How do I use the oscilloscope?” the system can recognize the intent and provide relevant guidance based on the context. This ability to match spoken language with actions enhances the usability of intelligent systems in industrial settings.

Data Collection Methodology

The creation of the ENIGMA-51 dataset involved several steps to ensure the data's relevance and usability.

Participants and Environment

A total of 19 participants were selected, each with varying levels of experience in repairing electrical boards. The recordings took place in a real industrial laboratory setting, providing a genuine representation of human-object interactions.

Use of Technology

Participants wore Microsoft HoloLens 2 smart glasses, which enabled them to receive audio instructions while keeping their hands free. The audio instructions guided them through the repair process step by step, ensuring consistency across recordings.

Video and Annotations

Each video was recorded at a resolution of 2272x1278 pixels with a framerate of 30 frames per second. The average length of the videos is around 26.32 minutes, resulting in a total of about 22 hours of footage.

The videos were thoroughly annotated, detailing specific actions, objects, and interaction frames to facilitate various studies related to human behavior.

Data Annotation Process

Accurate data annotation is crucial for the effectiveness of the dataset. The ENIGMA-51 dataset employs a detailed annotation strategy to ensure that each interaction is captured comprehensively.

Temporal Annotations

Interaction frames were identified and marked with timestamps and corresponding verbs that describe the action taking place. A taxonomy of four main verbs was used to classify the actions: “first-contact,” “de-contact,” “take,” and “release.”

Object and Hand Annotations

The dataset includes detailed annotations for both fixed and movable objects. There are 25 object classes documented within the dataset, ranging from tools like screwdrivers and pliers to fixed equipment like power supplies and electric panels.

Hands were also annotated, providing bounding boxes around both hands during interactions. This level of detail allows for an accurate study of how hands engage with tools and objects.

Future Interaction Annotations

To predict upcoming actions, the dataset includes annotations that reflect which objects will be involved in future interactions, along with the estimated time until those interactions begin.

Natural Language Annotations

In addition to visual data, the dataset captures the textual instructions provided to participants. These instructions were analyzed to extract intents and entities, further enriching the dataset’s usability for natural language understanding tasks.

Evaluation and Baseline Results

To demonstrate the applicability and challenge of the ENIGMA-51 dataset, baseline experiments were conducted focusing on four key tasks: action detection, egocentric human-object interaction detection, short-term interaction anticipation, and natural language understanding.

Action Detection Results

Baseline results show that detecting the basic actions is a challenging task, with varying levels of accuracy depending on the specific action being recognized. The dataset's complexity ensures that state-of-the-art methods must be refined to achieve satisfactory results.

Egocentric Human-Object Interaction Detection Results

By applying two different baseline models, the performance of egocentric detection was assessed. The outcomes highlight how incorporating domain-specific data significantly improves detection accuracy.

Short-Term Interaction Anticipation Results

For predicting future interactions, the baseline results revealed the capability of recognizing the next tools and actions. The system demonstrated a high level of accuracy in predicting which object would be used next.

Natural Language Understanding Results

Finally, the natural language understanding task was evaluated using various metrics. The best results were attained using only real data, while the inclusion of generated data led to a decline in performance. This underscores the need for quality, contextually relevant data for effective training in natural language tasks.

Conclusion

The ENIGMA-51 dataset provides a comprehensive framework for studying human-object interactions in industrial environments. With its detailed annotations and real-world context, it serves as an essential resource for developing intelligent systems capable of assisting workers in their tasks.

The findings from the baseline evaluations illustrate both the challenges and opportunities present in this research area. As we continue to explore human behavior through datasets like ENIGMA-51, the potential for creating sophisticated support systems in industrial settings becomes increasingly attainable.

Future Directions

Looking ahead, the research community can build on the insights provided by the ENIGMA-51 dataset. Further studies can focus on improving the accuracy of action detection methods, enhancing natural language understanding capabilities, and creating more intuitive tools for workers.

Overall, the ENIGMA-51 dataset stands as a valuable contribution to understanding and improving human-object interactions in industrial scenarios. As technology advances, the collaboration between human workers and intelligent systems will continue to evolve, leading to safer and more efficient workplaces.

Original Source

Title: ENIGMA-51: Towards a Fine-Grained Understanding of Human-Object Interactions in Industrial Scenarios

Abstract: ENIGMA-51 is a new egocentric dataset acquired in an industrial scenario by 19 subjects who followed instructions to complete the repair of electrical boards using industrial tools (e.g., electric screwdriver) and equipments (e.g., oscilloscope). The 51 egocentric video sequences are densely annotated with a rich set of labels that enable the systematic study of human behavior in the industrial domain. We provide benchmarks on four tasks related to human behavior: 1) untrimmed temporal detection of human-object interactions, 2) egocentric human-object interaction detection, 3) short-term object interaction anticipation and 4) natural language understanding of intents and entities. Baseline results show that the ENIGMA-51 dataset poses a challenging benchmark to study human behavior in industrial scenarios. We publicly release the dataset at https://iplab.dmi.unict.it/ENIGMA-51.

Authors: Francesco Ragusa, Rosario Leonardi, Michele Mazzamuto, Claudia Bonanno, Rosario Scavo, Antonino Furnari, Giovanni Maria Farinella

Last Update: 2023-11-27 00:00:00

Language: English

Source URL: https://arxiv.org/abs/2309.14809

Source PDF: https://arxiv.org/pdf/2309.14809

Licence: https://creativecommons.org/licenses/by/4.0/

Changes: This summary was created with assistance from AI and may have inaccuracies. For accurate information, please refer to the original source documents linked here.

Thank you to arxiv for use of its open access interoperability.

More from authors

Similar Articles