Loading…

Hand-Priming in Object Localization for Assistive Egocentric Vision

Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains diffi...

Full description

Saved in:
Bibliographic Details
Published in:2020 IEEE Winter Conference on Applications of Computer Vision (WACV) 2020-03, p.3411-3421
Main Authors: Lee, Kyungjun, Shrivastava, Abhinav, Kacorri, Hernisa
Format: Article
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 3421
container_issue
container_start_page 3411
container_title 2020 IEEE Winter Conference on Applications of Computer Vision (WACV)
container_volume
creator Lee, Kyungjun
Shrivastava, Abhinav
Kacorri, Hernisa
description Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.
doi_str_mv 10.1109/WACV45572.2020.9093353
format article
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9093353</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9093353</ieee_id><sourcerecordid>9093353</sourcerecordid><originalsourceid>FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03</originalsourceid><addsrcrecordid>eNotj9FKwzAUhqMguE2fQJC8QGtOTk7SXI4ynVCYFzovR5KmI2O20hRBn97BdvVffPDx_Yw9gigBhH36XNZbRWRkKYUUpRUWkfCKzcHICjQR2ms2k1rJwmIFt2ye80EItGBxxuq169vibUxfqd_z1PONP8Qw8WYI7pj-3JSGnnfDyJc5pzyln8hX-yHEfhpT4NuUT_yO3XTumOP9ZRfs43n1Xq-LZvPyWi-bIilZTYUEFbQBNKolIAOxA609tSDIauVca4lkCNqR8t6pqkPlW9Mp8I48eYEL9nD2phjj7vvU7Mbf3eUw_gMLDUqc</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><source>IEEE Xplore All Conference Series</source><creator>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</creator><creatorcontrib>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</creatorcontrib><description>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</description><identifier>EISSN: 2642-9381</identifier><identifier>EISBN: 1728165539</identifier><identifier>EISBN: 9781728165530</identifier><identifier>DOI: 10.1109/WACV45572.2020.9093353</identifier><language>eng</language><publisher>IEEE</publisher><subject>Cameras ; Computer vision ; Context modeling ; Image segmentation ; Object recognition ; Task analysis ; Visualization</subject><ispartof>2020 IEEE Winter Conference on Applications of Computer Vision (WACV), 2020-03, p.3411-3421</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9093353$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,23909,23910,25118,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9093353$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Lee, Kyungjun</creatorcontrib><creatorcontrib>Shrivastava, Abhinav</creatorcontrib><creatorcontrib>Kacorri, Hernisa</creatorcontrib><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><title>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</title><addtitle>WACV</addtitle><description>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</description><subject>Cameras</subject><subject>Computer vision</subject><subject>Context modeling</subject><subject>Image segmentation</subject><subject>Object recognition</subject><subject>Task analysis</subject><subject>Visualization</subject><issn>2642-9381</issn><isbn>1728165539</isbn><isbn>9781728165530</isbn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>6IE</sourceid><recordid>eNotj9FKwzAUhqMguE2fQJC8QGtOTk7SXI4ynVCYFzovR5KmI2O20hRBn97BdvVffPDx_Yw9gigBhH36XNZbRWRkKYUUpRUWkfCKzcHICjQR2ms2k1rJwmIFt2ye80EItGBxxuq169vibUxfqd_z1PONP8Qw8WYI7pj-3JSGnnfDyJc5pzyln8hX-yHEfhpT4NuUT_yO3XTumOP9ZRfs43n1Xq-LZvPyWi-bIilZTYUEFbQBNKolIAOxA609tSDIauVca4lkCNqR8t6pqkPlW9Mp8I48eYEL9nD2phjj7vvU7Mbf3eUw_gMLDUqc</recordid><startdate>20200301</startdate><enddate>20200301</enddate><creator>Lee, Kyungjun</creator><creator>Shrivastava, Abhinav</creator><creator>Kacorri, Hernisa</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20200301</creationdate><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><author>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Cameras</topic><topic>Computer vision</topic><topic>Context modeling</topic><topic>Image segmentation</topic><topic>Object recognition</topic><topic>Task analysis</topic><topic>Visualization</topic><toplevel>online_resources</toplevel><creatorcontrib>Lee, Kyungjun</creatorcontrib><creatorcontrib>Shrivastava, Abhinav</creatorcontrib><creatorcontrib>Kacorri, Hernisa</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library Online</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection><jtitle>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Lee, Kyungjun</au><au>Shrivastava, Abhinav</au><au>Kacorri, Hernisa</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Hand-Priming in Object Localization for Assistive Egocentric Vision</atitle><jtitle>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</jtitle><stitle>WACV</stitle><date>2020-03-01</date><risdate>2020</risdate><spage>3411</spage><epage>3421</epage><pages>3411-3421</pages><eissn>2642-9381</eissn><eisbn>1728165539</eisbn><eisbn>9781728165530</eisbn><abstract>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</abstract><pub>IEEE</pub><doi>10.1109/WACV45572.2020.9093353</doi><tpages>11</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2642-9381
ispartof 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), 2020-03, p.3411-3421
issn 2642-9381
language eng
recordid cdi_ieee_primary_9093353
source IEEE Xplore All Conference Series
subjects Cameras
Computer vision
Context modeling
Image segmentation
Object recognition
Task analysis
Visualization
title Hand-Priming in Object Localization for Assistive Egocentric Vision
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-31T19%3A43%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Hand-Priming%20in%20Object%20Localization%20for%20Assistive%20Egocentric%20Vision&rft.jtitle=2020%20IEEE%20Winter%20Conference%20on%20Applications%20of%20Computer%20Vision%20(WACV)&rft.au=Lee,%20Kyungjun&rft.date=2020-03-01&rft.spage=3411&rft.epage=3421&rft.pages=3411-3421&rft.eissn=2642-9381&rft_id=info:doi/10.1109/WACV45572.2020.9093353&rft.eisbn=1728165539&rft.eisbn_list=9781728165530&rft_dat=%3Cieee_CHZPO%3E9093353%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9093353&rfr_iscdi=true