Loading…

A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification

Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts...

Full description

Saved in:
Bibliographic Details
Main Authors: Kan Liu, Bingpeng Ma, Wei Zhang, Rui Huang
Format: Conference Proceeding
Language:English
Subjects:
Citations: Items that cite this one
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323
cites
container_end_page 3818
container_issue
container_start_page 3810
container_title
container_volume
creator Kan Liu
Bingpeng Ma
Wei Zhang
Rui Huang
description Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.
doi_str_mv 10.1109/ICCV.2015.434
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_7410791</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>7410791</ieee_id><sourcerecordid>7410791</sourcerecordid><originalsourceid>FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</originalsourceid><addsrcrecordid>eNotzL1OwzAUQGGDhERbGJlY8gIO1_G1E48hghKpEghKxVa59rVk1CaRk4W3J_xMZ_l0GLsRkAsB5q5tml1egFA5SjxjS4G6lJU0As7ZopAV8FIBXrLlOH4CSFNUesE-6uxtsFPs-ZZOQ5_sMauHgWyynaPslYZEI3XTj-iy0KdsFz31_N6O5LMX8jROKdpulrz1M4whul98xS6CPY50_d8Ve3982DZPfPO8bpt6w51AM3Hn5TxBq1ArUhJKJw5aEPoSpQqFwUojYVAerJZOHyCQtQ6ITMDgZSFX7PbvG4loP6R4sulrX6KA0gj5DbkWUW4</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><source>IEEE Xplore All Conference Series</source><creator>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</creator><creatorcontrib>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</creatorcontrib><description>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</description><identifier>EISSN: 2380-7504</identifier><identifier>EISBN: 1467383910</identifier><identifier>EISBN: 9781467383912</identifier><identifier>DOI: 10.1109/ICCV.2015.434</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Adaptation models ; Feature extraction ; Image color analysis ; Legged locomotion ; Measurement ; Training ; Video sequences</subject><ispartof>2015 IEEE International Conference on Computer Vision (ICCV), 2015, p.3810-3818</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/7410791$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,23909,23910,25118,27902,54530,54907</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/7410791$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Kan Liu</creatorcontrib><creatorcontrib>Bingpeng Ma</creatorcontrib><creatorcontrib>Wei Zhang</creatorcontrib><creatorcontrib>Rui Huang</creatorcontrib><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><title>2015 IEEE International Conference on Computer Vision (ICCV)</title><addtitle>ICCV</addtitle><description>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</description><subject>Adaptation models</subject><subject>Feature extraction</subject><subject>Image color analysis</subject><subject>Legged locomotion</subject><subject>Measurement</subject><subject>Training</subject><subject>Video sequences</subject><issn>2380-7504</issn><isbn>1467383910</isbn><isbn>9781467383912</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2015</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotzL1OwzAUQGGDhERbGJlY8gIO1_G1E48hghKpEghKxVa59rVk1CaRk4W3J_xMZ_l0GLsRkAsB5q5tml1egFA5SjxjS4G6lJU0As7ZopAV8FIBXrLlOH4CSFNUesE-6uxtsFPs-ZZOQ5_sMauHgWyynaPslYZEI3XTj-iy0KdsFz31_N6O5LMX8jROKdpulrz1M4whul98xS6CPY50_d8Ve3982DZPfPO8bpt6w51AM3Hn5TxBq1ArUhJKJw5aEPoSpQqFwUojYVAerJZOHyCQtQ6ITMDgZSFX7PbvG4loP6R4sulrX6KA0gj5DbkWUW4</recordid><startdate>201512</startdate><enddate>201512</enddate><creator>Kan Liu</creator><creator>Bingpeng Ma</creator><creator>Wei Zhang</creator><creator>Rui Huang</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>201512</creationdate><title>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</title><author>Kan Liu ; Bingpeng Ma ; Wei Zhang ; Rui Huang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2015</creationdate><topic>Adaptation models</topic><topic>Feature extraction</topic><topic>Image color analysis</topic><topic>Legged locomotion</topic><topic>Measurement</topic><topic>Training</topic><topic>Video sequences</topic><toplevel>online_resources</toplevel><creatorcontrib>Kan Liu</creatorcontrib><creatorcontrib>Bingpeng Ma</creatorcontrib><creatorcontrib>Wei Zhang</creatorcontrib><creatorcontrib>Rui Huang</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kan Liu</au><au>Bingpeng Ma</au><au>Wei Zhang</au><au>Rui Huang</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification</atitle><btitle>2015 IEEE International Conference on Computer Vision (ICCV)</btitle><stitle>ICCV</stitle><date>2015-12</date><risdate>2015</risdate><spage>3810</spage><epage>3818</epage><pages>3810-3818</pages><eissn>2380-7504</eissn><eisbn>1467383910</eisbn><eisbn>9781467383912</eisbn><coden>IEEPAD</coden><abstract>Pedestrian re-identification is a difficult problem due to the large variations in a person's appearance caused by different poses and viewpoints, illumination changes, and occlusions. Spatial alignment is commonly used to address these issues by treating the appearance of different body parts independently. However, a body part can also appear differently during different phases of an action. In this paper we consider the temporal alignment problem, in addition to the spatial one, and propose a new approach that takes the video of a walking person as input and builds a spatio-temporal appearance representation for pedestrian re-identification. Particularly, given a video sequence we exploit the periodicity exhibited by a walking person to generate a spatio-temporal body-action model, which consists of a series of body-action units corresponding to certain action primitives of certain body parts. Fisher vectors are learned and extracted from individual body-action units and concatenated into the final representation of the walking person. Unlike previous spatio-temporal features that only take into account local dynamic appearance information, our representation aligns the spatio-temporal appearance of a pedestrian globally. Extensive experiments on public datasets show the effectiveness of our approach compared with the state of the art.</abstract><pub>IEEE</pub><doi>10.1109/ICCV.2015.434</doi><tpages>9</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2380-7504
ispartof 2015 IEEE International Conference on Computer Vision (ICCV), 2015, p.3810-3818
issn 2380-7504
language eng
recordid cdi_ieee_primary_7410791
source IEEE Xplore All Conference Series
subjects Adaptation models
Feature extraction
Image color analysis
Legged locomotion
Measurement
Training
Video sequences
title A Spatio-Temporal Appearance Representation for Video-Based Pedestrian Re-Identification
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T02%3A54%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=A%20Spatio-Temporal%20Appearance%20Representation%20for%20Video-Based%20Pedestrian%20Re-Identification&rft.btitle=2015%20IEEE%20International%20Conference%20on%20Computer%20Vision%20(ICCV)&rft.au=Kan%20Liu&rft.date=2015-12&rft.spage=3810&rft.epage=3818&rft.pages=3810-3818&rft.eissn=2380-7504&rft.coden=IEEPAD&rft_id=info:doi/10.1109/ICCV.2015.434&rft.eisbn=1467383910&rft.eisbn_list=9781467383912&rft_dat=%3Cieee_CHZPO%3E7410791%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c149t-cd3ede4a5465e5307c1b61e4d7435f294864e4f5d0a63c6b0feaac0ee9f4fd323%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=7410791&rfr_iscdi=true