Loading…

GHT-based associative memory learning and its application to Human action detection and classification

This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed...

Full description

Saved in:
Bibliographic Details
Published in:Pattern recognition 2013-11, Vol.46 (11), p.3117-3128
Main Authors: Cheng, Shyi-Chyi, Cheng, Kwang-Yu, Chen, Yi-Ping Phoebe
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3
cites cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3
container_end_page 3128
container_issue 11
container_start_page 3117
container_title Pattern recognition
container_volume 46
creator Cheng, Shyi-Chyi
Cheng, Kwang-Yu
Chen, Yi-Ping Phoebe
description This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate. •GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.
doi_str_mv 10.1016/j.patcog.2013.03.027
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1671496336</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S003132031300157X</els_id><sourcerecordid>1671496336</sourcerecordid><originalsourceid>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</originalsourceid><addsrcrecordid>eNqFkcFqGzEQhkVoIa6bN8hBl0Av60orreS9FIpp7YKhl_QsZqVRkNldbaR1IG9fuWt6TGAYadA3_6D5CbnnbMMZV19PmwlmG582NeNiw0rU-oas-FaLquGy_kBWjAleiZqJW_Ip5xNjXJeHFfH7w2PVQUZHIedoA8zhBemAQ0yvtEdIYxifKIyOhjlTmKY-2MLEkc6RHs4DjBTsv9rhjMvtQtu-6AV_hT-Tjx76jHfXc03-_PzxuDtUx9_7X7vvx8pKVs9VK7fSocaOO8-dE85xJQCl0lqhZg12yrWSectKapTzftu1jQYmWuyk68SafFl0pxSfz5hnM4Rsse9hxHjOhqvy7VYJod5HGy6kbFrdFFQuqE0x54TeTCkMkF4NZ-bigDmZxQFzccCwErUubQ_XCZAt9D7BaEP-31trKdu6vch_Wzgsm3kJmEy2AUeLLqSyUeNieHvQX0xanzo</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1513445975</pqid></control><display><type>article</type><title>GHT-based associative memory learning and its application to Human action detection and classification</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</creator><creatorcontrib>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</creatorcontrib><description>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate. •GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</description><identifier>ISSN: 0031-3203</identifier><identifier>EISSN: 1873-5142</identifier><identifier>DOI: 10.1016/j.patcog.2013.03.027</identifier><identifier>CODEN: PTNRA8</identifier><language>eng</language><publisher>Kidlington: Elsevier Ltd</publisher><subject>Action object shapes ; Applied sciences ; Architecture ; Associative memory ; Classification ; Construction ; Detection, estimation, filtering, equalization, prediction ; Exact sciences and technology ; Generalized Hough transform ; Human ; Human action detection and recognition ; Hypergraph ; Image processing ; Information, signal and communications theory ; Intelligence ; Learning ; Signal and communications theory ; Signal processing ; Signal, noise ; Telecommunications and information theory ; Voting</subject><ispartof>Pattern recognition, 2013-11, Vol.46 (11), p.3117-3128</ispartof><rights>2013 Elsevier Ltd</rights><rights>2014 INIST-CNRS</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</citedby><cites>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=27449295$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Cheng, Shyi-Chyi</creatorcontrib><creatorcontrib>Cheng, Kwang-Yu</creatorcontrib><creatorcontrib>Chen, Yi-Ping Phoebe</creatorcontrib><title>GHT-based associative memory learning and its application to Human action detection and classification</title><title>Pattern recognition</title><description>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate. •GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</description><subject>Action object shapes</subject><subject>Applied sciences</subject><subject>Architecture</subject><subject>Associative memory</subject><subject>Classification</subject><subject>Construction</subject><subject>Detection, estimation, filtering, equalization, prediction</subject><subject>Exact sciences and technology</subject><subject>Generalized Hough transform</subject><subject>Human</subject><subject>Human action detection and recognition</subject><subject>Hypergraph</subject><subject>Image processing</subject><subject>Information, signal and communications theory</subject><subject>Intelligence</subject><subject>Learning</subject><subject>Signal and communications theory</subject><subject>Signal processing</subject><subject>Signal, noise</subject><subject>Telecommunications and information theory</subject><subject>Voting</subject><issn>0031-3203</issn><issn>1873-5142</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2013</creationdate><recordtype>article</recordtype><recordid>eNqFkcFqGzEQhkVoIa6bN8hBl0Av60orreS9FIpp7YKhl_QsZqVRkNldbaR1IG9fuWt6TGAYadA3_6D5CbnnbMMZV19PmwlmG582NeNiw0rU-oas-FaLquGy_kBWjAleiZqJW_Ip5xNjXJeHFfH7w2PVQUZHIedoA8zhBemAQ0yvtEdIYxifKIyOhjlTmKY-2MLEkc6RHs4DjBTsv9rhjMvtQtu-6AV_hT-Tjx76jHfXc03-_PzxuDtUx9_7X7vvx8pKVs9VK7fSocaOO8-dE85xJQCl0lqhZg12yrWSectKapTzftu1jQYmWuyk68SafFl0pxSfz5hnM4Rsse9hxHjOhqvy7VYJod5HGy6kbFrdFFQuqE0x54TeTCkMkF4NZ-bigDmZxQFzccCwErUubQ_XCZAt9D7BaEP-31trKdu6vch_Wzgsm3kJmEy2AUeLLqSyUeNieHvQX0xanzo</recordid><startdate>20131101</startdate><enddate>20131101</enddate><creator>Cheng, Shyi-Chyi</creator><creator>Cheng, Kwang-Yu</creator><creator>Chen, Yi-Ping Phoebe</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20131101</creationdate><title>GHT-based associative memory learning and its application to Human action detection and classification</title><author>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2013</creationdate><topic>Action object shapes</topic><topic>Applied sciences</topic><topic>Architecture</topic><topic>Associative memory</topic><topic>Classification</topic><topic>Construction</topic><topic>Detection, estimation, filtering, equalization, prediction</topic><topic>Exact sciences and technology</topic><topic>Generalized Hough transform</topic><topic>Human</topic><topic>Human action detection and recognition</topic><topic>Hypergraph</topic><topic>Image processing</topic><topic>Information, signal and communications theory</topic><topic>Intelligence</topic><topic>Learning</topic><topic>Signal and communications theory</topic><topic>Signal processing</topic><topic>Signal, noise</topic><topic>Telecommunications and information theory</topic><topic>Voting</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Cheng, Shyi-Chyi</creatorcontrib><creatorcontrib>Cheng, Kwang-Yu</creatorcontrib><creatorcontrib>Chen, Yi-Ping Phoebe</creatorcontrib><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Pattern recognition</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cheng, Shyi-Chyi</au><au>Cheng, Kwang-Yu</au><au>Chen, Yi-Ping Phoebe</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>GHT-based associative memory learning and its application to Human action detection and classification</atitle><jtitle>Pattern recognition</jtitle><date>2013-11-01</date><risdate>2013</risdate><volume>46</volume><issue>11</issue><spage>3117</spage><epage>3128</epage><pages>3117-3128</pages><issn>0031-3203</issn><eissn>1873-5142</eissn><coden>PTNRA8</coden><abstract>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate. •GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</abstract><cop>Kidlington</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.patcog.2013.03.027</doi><tpages>12</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0031-3203
ispartof Pattern recognition, 2013-11, Vol.46 (11), p.3117-3128
issn 0031-3203
1873-5142
language eng
recordid cdi_proquest_miscellaneous_1671496336
source ScienceDirect Freedom Collection 2022-2024
subjects Action object shapes
Applied sciences
Architecture
Associative memory
Classification
Construction
Detection, estimation, filtering, equalization, prediction
Exact sciences and technology
Generalized Hough transform
Human
Human action detection and recognition
Hypergraph
Image processing
Information, signal and communications theory
Intelligence
Learning
Signal and communications theory
Signal processing
Signal, noise
Telecommunications and information theory
Voting
title GHT-based associative memory learning and its application to Human action detection and classification
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T22%3A26%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=GHT-based%20associative%20memory%20learning%20and%20its%20application%20to%20Human%20action%20detection%20and%20classification&rft.jtitle=Pattern%20recognition&rft.au=Cheng,%20Shyi-Chyi&rft.date=2013-11-01&rft.volume=46&rft.issue=11&rft.spage=3117&rft.epage=3128&rft.pages=3117-3128&rft.issn=0031-3203&rft.eissn=1873-5142&rft.coden=PTNRA8&rft_id=info:doi/10.1016/j.patcog.2013.03.027&rft_dat=%3Cproquest_cross%3E1671496336%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1513445975&rft_id=info:pmid/&rfr_iscdi=true