Loading…
Efficient online structured output learning for keypoint-based object tracking
Efficient keypoint-based object detection methods are used in many real-time computer vision applications. These approaches often model an object as a collection of keypoints and associated descriptors, and detection then involves first constructing a set of correspondences between object and image...
Saved in:
Main Authors: | , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Citations: | Items that cite this one |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c314t-7772b9efba2f075c1c696f032a57f1474465ec0836063d9171fdebd6d0ec26c03 |
---|---|
cites | |
container_end_page | 1901 |
container_issue | |
container_start_page | 1894 |
container_title | |
container_volume | |
creator | Hare, S. Saffari, A. Torr, P. H. S. |
description | Efficient keypoint-based object detection methods are used in many real-time computer vision applications. These approaches often model an object as a collection of keypoints and associated descriptors, and detection then involves first constructing a set of correspondences between object and image keypoints via descriptor matching, and subsequently using these correspondences as input to a robust geometric estimation algorithm such as RANSAC to find the transformation of the object in the image. In such approaches, the object model is generally constructed offline, and does not adapt to a given environment at runtime. Furthermore, the feature matching and transformation estimation stages are treated entirely separately. In this paper, we introduce a new approach to address these problems by combining the overall pipeline of correspondence generation and transformation estimation into a single structured output learning framework. Following the recent trend of using efficient binary descriptors for feature matching, we also introduce an approach to approximate the learned object model as a collection of binary basis functions which can be evaluated very efficiently at runtime. Experiments on challenging video sequences show that our algorithm significantly improves over state-of-the-art descriptor matching techniques using a range of descriptors, as well as recent online learning based approaches. |
doi_str_mv | 10.1109/CVPR.2012.6247889 |
format | conference_proceeding |
fullrecord | <record><control><sourceid>ieee_6IE</sourceid><recordid>TN_cdi_ieee_primary_6247889</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>6247889</ieee_id><sourcerecordid>6247889</sourcerecordid><originalsourceid>FETCH-LOGICAL-c314t-7772b9efba2f075c1c696f032a57f1474465ec0836063d9171fdebd6d0ec26c03</originalsourceid><addsrcrecordid>eNo1kM9KxDAYxCMquK59APGSF2jNl6RJc5Rl_QOLiqjXJU2_SHZrW9L0sG9vxXUuw8CPYRhCroEVAMzcrj5f3wrOgBeKS11V5oRcglRaAOcVPyWZ0dV_VvKMLIApkSsD5oJk47hjs2aCGb4gz2vvgwvYJdp3beiQjilOLk0RG9pPaZgSbdHGLnRf1PeR7vEw9KFLeW3HX6TeoUs0Rev2M3JFzr1tR8yOviQf9-v31WO-eXl4Wt1tcidAplxrzWuDvrbcM106cMoozwS3pfYgtZSqRMcqoebljQENvsG6UQ1Dx5VjYklu_noDIm6HGL5tPGyPd4gfhmNSNg</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Efficient online structured output learning for keypoint-based object tracking</title><source>IEEE Electronic Library (IEL) Conference Proceedings</source><creator>Hare, S. ; Saffari, A. ; Torr, P. H. S.</creator><creatorcontrib>Hare, S. ; Saffari, A. ; Torr, P. H. S.</creatorcontrib><description>Efficient keypoint-based object detection methods are used in many real-time computer vision applications. These approaches often model an object as a collection of keypoints and associated descriptors, and detection then involves first constructing a set of correspondences between object and image keypoints via descriptor matching, and subsequently using these correspondences as input to a robust geometric estimation algorithm such as RANSAC to find the transformation of the object in the image. In such approaches, the object model is generally constructed offline, and does not adapt to a given environment at runtime. Furthermore, the feature matching and transformation estimation stages are treated entirely separately. In this paper, we introduce a new approach to address these problems by combining the overall pipeline of correspondence generation and transformation estimation into a single structured output learning framework. Following the recent trend of using efficient binary descriptors for feature matching, we also introduce an approach to approximate the learned object model as a collection of binary basis functions which can be evaluated very efficiently at runtime. Experiments on challenging video sequences show that our algorithm significantly improves over state-of-the-art descriptor matching techniques using a range of descriptors, as well as recent online learning based approaches.</description><identifier>ISSN: 1063-6919</identifier><identifier>ISBN: 9781467312264</identifier><identifier>ISBN: 1467312266</identifier><identifier>EISBN: 1467312282</identifier><identifier>EISBN: 1467312274</identifier><identifier>EISBN: 9781467312271</identifier><identifier>EISBN: 9781467312288</identifier><identifier>DOI: 10.1109/CVPR.2012.6247889</identifier><language>eng</language><publisher>IEEE</publisher><subject>Adaptation models ; Approximation algorithms ; Computational modeling ; Estimation ; Object detection ; Training ; Vectors</subject><ispartof>2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012, p.1894-1901</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c314t-7772b9efba2f075c1c696f032a57f1474465ec0836063d9171fdebd6d0ec26c03</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/6247889$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,778,782,787,788,2054,27908,54903</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/6247889$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Hare, S.</creatorcontrib><creatorcontrib>Saffari, A.</creatorcontrib><creatorcontrib>Torr, P. H. S.</creatorcontrib><title>Efficient online structured output learning for keypoint-based object tracking</title><title>2012 IEEE Conference on Computer Vision and Pattern Recognition</title><addtitle>CVPR</addtitle><description>Efficient keypoint-based object detection methods are used in many real-time computer vision applications. These approaches often model an object as a collection of keypoints and associated descriptors, and detection then involves first constructing a set of correspondences between object and image keypoints via descriptor matching, and subsequently using these correspondences as input to a robust geometric estimation algorithm such as RANSAC to find the transformation of the object in the image. In such approaches, the object model is generally constructed offline, and does not adapt to a given environment at runtime. Furthermore, the feature matching and transformation estimation stages are treated entirely separately. In this paper, we introduce a new approach to address these problems by combining the overall pipeline of correspondence generation and transformation estimation into a single structured output learning framework. Following the recent trend of using efficient binary descriptors for feature matching, we also introduce an approach to approximate the learned object model as a collection of binary basis functions which can be evaluated very efficiently at runtime. Experiments on challenging video sequences show that our algorithm significantly improves over state-of-the-art descriptor matching techniques using a range of descriptors, as well as recent online learning based approaches.</description><subject>Adaptation models</subject><subject>Approximation algorithms</subject><subject>Computational modeling</subject><subject>Estimation</subject><subject>Object detection</subject><subject>Training</subject><subject>Vectors</subject><issn>1063-6919</issn><isbn>9781467312264</isbn><isbn>1467312266</isbn><isbn>1467312282</isbn><isbn>1467312274</isbn><isbn>9781467312271</isbn><isbn>9781467312288</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2012</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNo1kM9KxDAYxCMquK59APGSF2jNl6RJc5Rl_QOLiqjXJU2_SHZrW9L0sG9vxXUuw8CPYRhCroEVAMzcrj5f3wrOgBeKS11V5oRcglRaAOcVPyWZ0dV_VvKMLIApkSsD5oJk47hjs2aCGb4gz2vvgwvYJdp3beiQjilOLk0RG9pPaZgSbdHGLnRf1PeR7vEw9KFLeW3HX6TeoUs0Rev2M3JFzr1tR8yOviQf9-v31WO-eXl4Wt1tcidAplxrzWuDvrbcM106cMoozwS3pfYgtZSqRMcqoebljQENvsG6UQ1Dx5VjYklu_noDIm6HGL5tPGyPd4gfhmNSNg</recordid><startdate>20120101</startdate><enddate>20120101</enddate><creator>Hare, S.</creator><creator>Saffari, A.</creator><creator>Torr, P. H. S.</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>20120101</creationdate><title>Efficient online structured output learning for keypoint-based object tracking</title><author>Hare, S. ; Saffari, A. ; Torr, P. H. S.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c314t-7772b9efba2f075c1c696f032a57f1474465ec0836063d9171fdebd6d0ec26c03</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2012</creationdate><topic>Adaptation models</topic><topic>Approximation algorithms</topic><topic>Computational modeling</topic><topic>Estimation</topic><topic>Object detection</topic><topic>Training</topic><topic>Vectors</topic><toplevel>online_resources</toplevel><creatorcontrib>Hare, S.</creatorcontrib><creatorcontrib>Saffari, A.</creatorcontrib><creatorcontrib>Torr, P. H. S.</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore (Online service)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hare, S.</au><au>Saffari, A.</au><au>Torr, P. H. S.</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Efficient online structured output learning for keypoint-based object tracking</atitle><btitle>2012 IEEE Conference on Computer Vision and Pattern Recognition</btitle><stitle>CVPR</stitle><date>2012-01-01</date><risdate>2012</risdate><spage>1894</spage><epage>1901</epage><pages>1894-1901</pages><issn>1063-6919</issn><isbn>9781467312264</isbn><isbn>1467312266</isbn><eisbn>1467312282</eisbn><eisbn>1467312274</eisbn><eisbn>9781467312271</eisbn><eisbn>9781467312288</eisbn><abstract>Efficient keypoint-based object detection methods are used in many real-time computer vision applications. These approaches often model an object as a collection of keypoints and associated descriptors, and detection then involves first constructing a set of correspondences between object and image keypoints via descriptor matching, and subsequently using these correspondences as input to a robust geometric estimation algorithm such as RANSAC to find the transformation of the object in the image. In such approaches, the object model is generally constructed offline, and does not adapt to a given environment at runtime. Furthermore, the feature matching and transformation estimation stages are treated entirely separately. In this paper, we introduce a new approach to address these problems by combining the overall pipeline of correspondence generation and transformation estimation into a single structured output learning framework. Following the recent trend of using efficient binary descriptors for feature matching, we also introduce an approach to approximate the learned object model as a collection of binary basis functions which can be evaluated very efficiently at runtime. Experiments on challenging video sequences show that our algorithm significantly improves over state-of-the-art descriptor matching techniques using a range of descriptors, as well as recent online learning based approaches.</abstract><pub>IEEE</pub><doi>10.1109/CVPR.2012.6247889</doi><tpages>8</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 1063-6919 |
ispartof | 2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012, p.1894-1901 |
issn | 1063-6919 |
language | eng |
recordid | cdi_ieee_primary_6247889 |
source | IEEE Electronic Library (IEL) Conference Proceedings |
subjects | Adaptation models Approximation algorithms Computational modeling Estimation Object detection Training Vectors |
title | Efficient online structured output learning for keypoint-based object tracking |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-17T03%3A53%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_6IE&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Efficient%20online%20structured%20output%20learning%20for%20keypoint-based%20object%20tracking&rft.btitle=2012%20IEEE%20Conference%20on%20Computer%20Vision%20and%20Pattern%20Recognition&rft.au=Hare,%20S.&rft.date=2012-01-01&rft.spage=1894&rft.epage=1901&rft.pages=1894-1901&rft.issn=1063-6919&rft.isbn=9781467312264&rft.isbn_list=1467312266&rft_id=info:doi/10.1109/CVPR.2012.6247889&rft.eisbn=1467312282&rft.eisbn_list=1467312274&rft.eisbn_list=9781467312271&rft.eisbn_list=9781467312288&rft_dat=%3Cieee_6IE%3E6247889%3C/ieee_6IE%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c314t-7772b9efba2f075c1c696f032a57f1474465ec0836063d9171fdebd6d0ec26c03%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=6247889&rfr_iscdi=true |