Loading…

The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain

Wearable cameras allow to collect images and videos of humans interacting with the world. While human-object interactions have been thoroughly investigated in third person vision, the problem has been understudied in egocentric settings and in industrial scenarios. To fill this gap, we introduce MEC...

Full description

Saved in:
Bibliographic Details
Main Authors: Ragusa, Francesco, Furnari, Antonino, Livatino, Salvatore, Farinella, Giovanni Maria
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 1577
container_issue
container_start_page 1568
container_title
container_volume
creator Ragusa, Francesco
Furnari, Antonino
Livatino, Salvatore
Farinella, Giovanni Maria
description Wearable cameras allow to collect images and videos of humans interacting with the world. While human-object interactions have been thoroughly investigated in third person vision, the problem has been understudied in egocentric settings and in industrial scenarios. To fill this gap, we introduce MECCANO, the first dataset of egocentric videos to study human-object interactions in industrial-like settings. MECCANO has been acquired by 20 participants who were asked to build a motorbike model, for which they had to interact with tiny objects and tools. The dataset has been explicitly labeled for the task of recognizing human-object interactions from an egocentric perspective. Specifically, each interaction has been labeled both temporally (with action segments) and spatially (with active object bounding boxes). With the proposed dataset, we investigate four different tasks including 1) action recognition, 2) active object detection, 3) active object recognition and 4) egocentric human-object interaction detection, which is a revisited version of the standard human-object interaction detection task. Baseline results show that the MECCANO dataset is a challenging benchmark to study egocentric human-object interactions in industrial-like scenarios. We publicy release the dataset at https://iplab.dmi.unict.it/MECCANO/.
doi_str_mv 10.1109/WACV48630.2021.00161
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9423132</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9423132</ieee_id><sourcerecordid>9423132</sourcerecordid><originalsourceid>FETCH-LOGICAL-i249t-9108a5c497ac5f7f350a3eec0dbd1849407787c7aa26e77e110941ee73ef6a0a3</originalsourceid><addsrcrecordid>eNotkM1OAjEUhauJiYg8gS76AoP9m3bqjgwoJCgbwCW5dO5gkemQaVn49o7R1UlOvnzJOYQ8cjbmnNmnj0m5VYWWbCyY4GPGuOZXZGRNwbXOFVPGFNdkILQSmZUFvyV3MR4Zk5ZbOSDn9SfSt1lZTt5XdAoJIqZnugkVdjFBqHw40PmlgZCt9kd0iS5Cwg5c8m2ItO7ahs4OrcOQOu_o1lfYRuoDhdCT1SX2NZyyk_9COm0b8OGe3NRwijj6zyHZvMzW5Txbrl4X5WSZeaFsyixnBeROWQMur00tcwYS0bFqX_FCWcX6XcYZAKHRGPz9QnFEI7HW0LND8vDn9Yi4O3e-ge57Z5WQXAr5A44QWrU</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain</title><source>IEEE Xplore All Conference Series</source><creator>Ragusa, Francesco ; Furnari, Antonino ; Livatino, Salvatore ; Farinella, Giovanni Maria</creator><creatorcontrib>Ragusa, Francesco ; Furnari, Antonino ; Livatino, Salvatore ; Farinella, Giovanni Maria</creatorcontrib><description>Wearable cameras allow to collect images and videos of humans interacting with the world. While human-object interactions have been thoroughly investigated in third person vision, the problem has been understudied in egocentric settings and in industrial scenarios. To fill this gap, we introduce MECCANO, the first dataset of egocentric videos to study human-object interactions in industrial-like settings. MECCANO has been acquired by 20 participants who were asked to build a motorbike model, for which they had to interact with tiny objects and tools. The dataset has been explicitly labeled for the task of recognizing human-object interactions from an egocentric perspective. Specifically, each interaction has been labeled both temporally (with action segments) and spatially (with active object bounding boxes). With the proposed dataset, we investigate four different tasks including 1) action recognition, 2) active object detection, 3) active object recognition and 4) egocentric human-object interaction detection, which is a revisited version of the standard human-object interaction detection task. Baseline results show that the MECCANO dataset is a challenging benchmark to study egocentric human-object interactions in industrial-like scenarios. We publicy release the dataset at https://iplab.dmi.unict.it/MECCANO/.</description><identifier>EISSN: 2642-9381</identifier><identifier>EISBN: 9781665404778</identifier><identifier>EISBN: 1665404779</identifier><identifier>DOI: 10.1109/WACV48630.2021.00161</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Computer vision ; Conferences ; Motorcycles ; Object detection ; Object recognition ; Taxonomy</subject><ispartof>2021 IEEE Winter Conference on Applications of Computer Vision (WACV), 2021, p.1568-1577</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9423132$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9423132$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Ragusa, Francesco</creatorcontrib><creatorcontrib>Furnari, Antonino</creatorcontrib><creatorcontrib>Livatino, Salvatore</creatorcontrib><creatorcontrib>Farinella, Giovanni Maria</creatorcontrib><title>The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain</title><title>2021 IEEE Winter Conference on Applications of Computer Vision (WACV)</title><addtitle>WACV</addtitle><description>Wearable cameras allow to collect images and videos of humans interacting with the world. While human-object interactions have been thoroughly investigated in third person vision, the problem has been understudied in egocentric settings and in industrial scenarios. To fill this gap, we introduce MECCANO, the first dataset of egocentric videos to study human-object interactions in industrial-like settings. MECCANO has been acquired by 20 participants who were asked to build a motorbike model, for which they had to interact with tiny objects and tools. The dataset has been explicitly labeled for the task of recognizing human-object interactions from an egocentric perspective. Specifically, each interaction has been labeled both temporally (with action segments) and spatially (with active object bounding boxes). With the proposed dataset, we investigate four different tasks including 1) action recognition, 2) active object detection, 3) active object recognition and 4) egocentric human-object interaction detection, which is a revisited version of the standard human-object interaction detection task. Baseline results show that the MECCANO dataset is a challenging benchmark to study egocentric human-object interactions in industrial-like scenarios. We publicy release the dataset at https://iplab.dmi.unict.it/MECCANO/.</description><subject>Computer vision</subject><subject>Conferences</subject><subject>Motorcycles</subject><subject>Object detection</subject><subject>Object recognition</subject><subject>Taxonomy</subject><issn>2642-9381</issn><isbn>9781665404778</isbn><isbn>1665404779</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2021</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotkM1OAjEUhauJiYg8gS76AoP9m3bqjgwoJCgbwCW5dO5gkemQaVn49o7R1UlOvnzJOYQ8cjbmnNmnj0m5VYWWbCyY4GPGuOZXZGRNwbXOFVPGFNdkILQSmZUFvyV3MR4Zk5ZbOSDn9SfSt1lZTt5XdAoJIqZnugkVdjFBqHw40PmlgZCt9kd0iS5Cwg5c8m2ItO7ahs4OrcOQOu_o1lfYRuoDhdCT1SX2NZyyk_9COm0b8OGe3NRwijj6zyHZvMzW5Txbrl4X5WSZeaFsyixnBeROWQMur00tcwYS0bFqX_FCWcX6XcYZAKHRGPz9QnFEI7HW0LND8vDn9Yi4O3e-ge57Z5WQXAr5A44QWrU</recordid><startdate>202101</startdate><enddate>202101</enddate><creator>Ragusa, Francesco</creator><creator>Furnari, Antonino</creator><creator>Livatino, Salvatore</creator><creator>Farinella, Giovanni Maria</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>202101</creationdate><title>The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain</title><author>Ragusa, Francesco ; Furnari, Antonino ; Livatino, Salvatore ; Farinella, Giovanni Maria</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i249t-9108a5c497ac5f7f350a3eec0dbd1849407787c7aa26e77e110941ee73ef6a0a3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer vision</topic><topic>Conferences</topic><topic>Motorcycles</topic><topic>Object detection</topic><topic>Object recognition</topic><topic>Taxonomy</topic><toplevel>online_resources</toplevel><creatorcontrib>Ragusa, Francesco</creatorcontrib><creatorcontrib>Furnari, Antonino</creatorcontrib><creatorcontrib>Livatino, Salvatore</creatorcontrib><creatorcontrib>Farinella, Giovanni Maria</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Ragusa, Francesco</au><au>Furnari, Antonino</au><au>Livatino, Salvatore</au><au>Farinella, Giovanni Maria</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain</atitle><btitle>2021 IEEE Winter Conference on Applications of Computer Vision (WACV)</btitle><stitle>WACV</stitle><date>2021-01</date><risdate>2021</risdate><spage>1568</spage><epage>1577</epage><pages>1568-1577</pages><eissn>2642-9381</eissn><eisbn>9781665404778</eisbn><eisbn>1665404779</eisbn><coden>IEEPAD</coden><abstract>Wearable cameras allow to collect images and videos of humans interacting with the world. While human-object interactions have been thoroughly investigated in third person vision, the problem has been understudied in egocentric settings and in industrial scenarios. To fill this gap, we introduce MECCANO, the first dataset of egocentric videos to study human-object interactions in industrial-like settings. MECCANO has been acquired by 20 participants who were asked to build a motorbike model, for which they had to interact with tiny objects and tools. The dataset has been explicitly labeled for the task of recognizing human-object interactions from an egocentric perspective. Specifically, each interaction has been labeled both temporally (with action segments) and spatially (with active object bounding boxes). With the proposed dataset, we investigate four different tasks including 1) action recognition, 2) active object detection, 3) active object recognition and 4) egocentric human-object interaction detection, which is a revisited version of the standard human-object interaction detection task. Baseline results show that the MECCANO dataset is a challenging benchmark to study egocentric human-object interactions in industrial-like scenarios. We publicy release the dataset at https://iplab.dmi.unict.it/MECCANO/.</abstract><pub>IEEE</pub><doi>10.1109/WACV48630.2021.00161</doi><tpages>10</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2642-9381
ispartof 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), 2021, p.1568-1577
issn 2642-9381
language eng
recordid cdi_ieee_primary_9423132
source IEEE Xplore All Conference Series
subjects Computer vision
Conferences
Motorcycles
Object detection
Object recognition
Taxonomy
title The MECCANO Dataset: Understanding Human-Object Interactions from Egocentric Videos in an Industrial-like Domain
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T22%3A44%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=The%20MECCANO%20Dataset:%20Understanding%20Human-Object%20Interactions%20from%20Egocentric%20Videos%20in%20an%20Industrial-like%20Domain&rft.btitle=2021%20IEEE%20Winter%20Conference%20on%20Applications%20of%20Computer%20Vision%20(WACV)&rft.au=Ragusa,%20Francesco&rft.date=2021-01&rft.spage=1568&rft.epage=1577&rft.pages=1568-1577&rft.eissn=2642-9381&rft.coden=IEEPAD&rft_id=info:doi/10.1109/WACV48630.2021.00161&rft.eisbn=9781665404778&rft.eisbn_list=1665404779&rft_dat=%3Cieee_CHZPO%3E9423132%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i249t-9108a5c497ac5f7f350a3eec0dbd1849407787c7aa26e77e110941ee73ef6a0a3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9423132&rfr_iscdi=true