Loading…
A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification
Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Conference Proceeding |
Language: | English |
Subjects: | |
Citations: | Items that cite this one |
Online Access: | Request full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323 |
---|---|
cites | |
container_end_page | 3818 |
container_issue | |
container_start_page | 3810 |
container_title | |
container_volume | |
creator | Kan Liu Bingpeng Ma Wei Zhang Rui Huang |
description | Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art. |
doi_str_mv | 10.1109/ICCV.2015.434 |
format | conference_proceeding |
fullrecord | <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_7410791</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>7410791</ieee_id><sourcerecordid>7410791</sourcerecordid><originalsourceid>FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</originalsourceid><addsrcrecordid>eNotzL1OwzAUQGGDhERbGJlY8gIO1_G1E48hghKpEghKxVa59rVk1CaRk4W3J_xMZ_l0GLsRkAsB5q5tml1egFA5SjxjS4G6lJU0As7ZopAV8FIBXrLlOH4CSFNUesE-6uxtsFPs-ZZOQ5_sMauHgWyynaPslYZEI3XTj-iy0KdsFz31_N6O5LMX8jROKdpulrz1M4whul98xS6CPY50_d8Ve3982DZPfPO8bpt6w51AM3Hn5TxBq1ArUhJKJw5aEPoSpQqFwUojYVAerJZOHyCQtQ6ITMDgZSFX7PbvG4loP6R4sulrX6KA0gj5DbkWUW4</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><source>IEEE Xplore All Conference Series</source><creator>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</creator><creatorcontrib>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</creatorcontrib><description>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</description><identifier>EISSN: 2380-7504</identifier><identifier>EISBN: 1467383910</identifier><identifier>EISBN: 9781467383912</identifier><identifier>DOI: 10.1109/ICCV.2015.434</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Adaptation models ; Feature extraction ; Image color analysis ; Legged locomotion ; Measurement ; Training ; Video sequences</subject><ispartof>2015 IEEE International Conference on Computer Vision (ICCV), 2015, p.3810-3818</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/7410791$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,23909,23910,25118,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/7410791$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Kan Liu</creatorcontrib><creatorcontrib>Bingpeng Ma</creatorcontrib><creatorcontrib>Wei Zhang</creatorcontrib><creatorcontrib>Rui Huang</creatorcontrib><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><title>2015 IEEE International Conference on Computer Vision (ICCV)</title><addtitle>ICCV</addtitle><description>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</description><subject>Adaptation models</subject><subject>Feature extraction</subject><subject>Image color analysis</subject><subject>Legged locomotion</subject><subject>Measurement</subject><subject>Training</subject><subject>Video sequences</subject><issn>2380-7504</issn><isbn>1467383910</isbn><isbn>9781467383912</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2015</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotzL1OwzAUQGGDhERbGJlY8gIO1_G1E48hghKpEghKxVa59rVk1CaRk4W3J_xMZ_l0GLsRkAsB5q5tml1egFA5SjxjS4G6lJU0As7ZopAV8FIBXrLlOH4CSFNUesE-6uxtsFPs-ZZOQ5_sMauHgWyynaPslYZEI3XTj-iy0KdsFz31_N6O5LMX8jROKdpulrz1M4whul98xS6CPY50_d8Ve3982DZPfPO8bpt6w51AM3Hn5TxBq1ArUhJKJw5aEPoSpQqFwUojYVAerJZOHyCQtQ6ITMDgZSFX7PbvG4loP6R4sulrX6KA0gj5DbkWUW4</recordid><startdate>201512</startdate><enddate>201512</enddate><creator>Kan Liu</creator><creator>Bingpeng Ma</creator><creator>Wei Zhang</creator><creator>Rui Huang</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>201512</creationdate><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><author>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2015</creationdate><topic>Adaptation models</topic><topic>Feature extraction</topic><topic>Image color analysis</topic><topic>Legged locomotion</topic><topic>Measurement</topic><topic>Training</topic><topic>Video sequences</topic><toplevel>online_resources</toplevel><creatorcontrib>Kan Liu</creatorcontrib><creatorcontrib>Bingpeng Ma</creatorcontrib><creatorcontrib>Wei Zhang</creatorcontrib><creatorcontrib>Rui Huang</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kan Liu</au><au>Bingpeng Ma</au><au>Wei Zhang</au><au>Rui Huang</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</atitle><btitle>2015 IEEE International Conference on Computer Vision (ICCV)</btitle><stitle>ICCV</stitle><date>2015-12</date><risdate>2015</risdate><spage>3810</spage><epage>3818</epage><pages>3810-3818</pages><eissn>2380-7504</eissn><eisbn>1467383910</eisbn><eisbn>9781467383912</eisbn><coden>IEEPAD</coden><abstract>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</abstract><pub>IEEE</pub><doi>10.1109/ICCV.2015.434</doi><tpages>9</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 2380-7504 |
ispartof | 2015 IEEE International Conference on Computer Vision (ICCV), 2015, p.3810-3818 |
issn | 2380-7504 |
language | eng |
recordid | cdi_ieee_primary_7410791 |
source | IEEE Xplore All Conference Series |
subjects | Adaptation models Feature extraction Image color analysis Legged locomotion Measurement Training Video sequences |
title | A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T02%3A54%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=A%20Spatio-Temporal%20Appearance%20Representation%20for%20Video-Based%20Pedestrian%20Re-Identification&rft.btitle=2015%20IEEE%20International%20Conference%20on%20Computer%20Vision%20(ICCV)&rft.au=Kan%20Liu&rft.date=2015-12&rft.spage=3810&rft.epage=3818&rft.pages=3810-3818&rft.eissn=2380-7504&rft.coden=IEEPAD&rft_id=info:doi/10.1109/ICCV.2015.434&rft.eisbn=1467383910&rft.eisbn_list=9781467383912&rft_dat=%3Cieee_CHZPO%3E7410791%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=7410791&rfr_iscdi=true |