Loading…
Hand-Priming in Object Localization for Assistive Egocentric Vision
Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains diffi...
Saved in:
Published in: | 2020 IEEE Winter Conference on Applications of Computer Vision (WACV) 2020-03, p.3411-3421 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 3421 |
container_issue | |
container_start_page | 3411 |
container_title | 2020 IEEE Winter Conference on Applications of Computer Vision (WACV) |
container_volume | |
creator | Lee, Kyungjun Shrivastava, Abhinav Kacorri, Hernisa |
description | Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization. |
doi_str_mv | 10.1109/WACV45572.2020.9093353 |
format | article |
fullrecord | <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9093353</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9093353</ieee_id><sourcerecordid>9093353</sourcerecordid><originalsourceid>FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03</originalsourceid><addsrcrecordid>eNotj9FKwzAUhqMguE2fQJC8QGtOTk7SXI4ynVCYFzovR5KmI2O20hRBn97BdvVffPDx_Yw9gigBhH36XNZbRWRkKYUUpRUWkfCKzcHICjQR2ms2k1rJwmIFt2ye80EItGBxxuq169vibUxfqd_z1PONP8Qw8WYI7pj-3JSGnnfDyJc5pzyln8hX-yHEfhpT4NuUT_yO3XTumOP9ZRfs43n1Xq-LZvPyWi-bIilZTYUEFbQBNKolIAOxA609tSDIauVca4lkCNqR8t6pqkPlW9Mp8I48eYEL9nD2phjj7vvU7Mbf3eUw_gMLDUqc</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><source>IEEE Xplore All Conference Series</source><creator>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</creator><creatorcontrib>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</creatorcontrib><description>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</description><identifier>EISSN: 2642-9381</identifier><identifier>EISBN: 1728165539</identifier><identifier>EISBN: 9781728165530</identifier><identifier>DOI: 10.1109/WACV45572.2020.9093353</identifier><language>eng</language><publisher>IEEE</publisher><subject>Cameras ; Computer vision ; Context modeling ; Image segmentation ; Object recognition ; Task analysis ; Visualization</subject><ispartof>2020 IEEE Winter Conference on Applications of Computer Vision (WACV), 2020-03, p.3411-3421</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9093353$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,23909,23910,25118,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9093353$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Lee, Kyungjun</creatorcontrib><creatorcontrib>Shrivastava, Abhinav</creatorcontrib><creatorcontrib>Kacorri, Hernisa</creatorcontrib><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><title>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</title><addtitle>WACV</addtitle><description>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</description><subject>Cameras</subject><subject>Computer vision</subject><subject>Context modeling</subject><subject>Image segmentation</subject><subject>Object recognition</subject><subject>Task analysis</subject><subject>Visualization</subject><issn>2642-9381</issn><isbn>1728165539</isbn><isbn>9781728165530</isbn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>6IE</sourceid><recordid>eNotj9FKwzAUhqMguE2fQJC8QGtOTk7SXI4ynVCYFzovR5KmI2O20hRBn97BdvVffPDx_Yw9gigBhH36XNZbRWRkKYUUpRUWkfCKzcHICjQR2ms2k1rJwmIFt2ye80EItGBxxuq169vibUxfqd_z1PONP8Qw8WYI7pj-3JSGnnfDyJc5pzyln8hX-yHEfhpT4NuUT_yO3XTumOP9ZRfs43n1Xq-LZvPyWi-bIilZTYUEFbQBNKolIAOxA609tSDIauVca4lkCNqR8t6pqkPlW9Mp8I48eYEL9nD2phjj7vvU7Mbf3eUw_gMLDUqc</recordid><startdate>20200301</startdate><enddate>20200301</enddate><creator>Lee, Kyungjun</creator><creator>Shrivastava, Abhinav</creator><creator>Kacorri, Hernisa</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>20200301</creationdate><title>Hand-Priming in Object Localization for Assistive Egocentric Vision</title><author>Lee, Kyungjun ; Shrivastava, Abhinav ; Kacorri, Hernisa</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Cameras</topic><topic>Computer vision</topic><topic>Context modeling</topic><topic>Image segmentation</topic><topic>Object recognition</topic><topic>Task analysis</topic><topic>Visualization</topic><toplevel>online_resources</toplevel><creatorcontrib>Lee, Kyungjun</creatorcontrib><creatorcontrib>Shrivastava, Abhinav</creatorcontrib><creatorcontrib>Kacorri, Hernisa</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library Online</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection><jtitle>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Lee, Kyungjun</au><au>Shrivastava, Abhinav</au><au>Kacorri, Hernisa</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Hand-Priming in Object Localization for Assistive Egocentric Vision</atitle><jtitle>2020 IEEE Winter Conference on Applications of Computer Vision (WACV)</jtitle><stitle>WACV</stitle><date>2020-03-01</date><risdate>2020</risdate><spage>3411</spage><epage>3421</epage><pages>3411-3421</pages><eissn>2642-9381</eissn><eisbn>1728165539</eisbn><eisbn>9781728165530</eisbn><abstract>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the handpriming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</abstract><pub>IEEE</pub><doi>10.1109/WACV45572.2020.9093353</doi><tpages>11</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 2642-9381 |
ispartof | 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), 2020-03, p.3411-3421 |
issn | 2642-9381 |
language | eng |
recordid | cdi_ieee_primary_9093353 |
source | IEEE Xplore All Conference Series |
subjects | Cameras Computer vision Context modeling Image segmentation Object recognition Task analysis Visualization |
title | Hand-Priming in Object Localization for Assistive Egocentric Vision |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-31T19%3A43%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Hand-Priming%20in%20Object%20Localization%20for%20Assistive%20Egocentric%20Vision&rft.jtitle=2020%20IEEE%20Winter%20Conference%20on%20Applications%20of%20Computer%20Vision%20(WACV)&rft.au=Lee,%20Kyungjun&rft.date=2020-03-01&rft.spage=3411&rft.epage=3421&rft.pages=3411-3421&rft.eissn=2642-9381&rft_id=info:doi/10.1109/WACV45572.2020.9093353&rft.eisbn=1728165539&rft.eisbn_list=9781728165530&rft_dat=%3Cieee_CHZPO%3E9093353%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i428t-214c671374d51571ef166b5d105964aad9552cc6a54bba48f34bd7f41ba5b5b03%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9093353&rfr_iscdi=true |