Loading…
GHT-based associative memory learning and its application to Human action detection and classification
This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed...
Saved in:
Published in: | Pattern recognition 2013-11, Vol.46 (11), p.3117-3128 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3 |
---|---|
cites | cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3 |
container_end_page | 3128 |
container_issue | 11 |
container_start_page | 3117 |
container_title | Pattern recognition |
container_volume | 46 |
creator | Cheng, Shyi-Chyi Cheng, Kwang-Yu Chen, Yi-Ping Phoebe |
description | This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate.
•GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented. |
doi_str_mv | 10.1016/j.patcog.2013.03.027 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1671496336</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S003132031300157X</els_id><sourcerecordid>1671496336</sourcerecordid><originalsourceid>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</originalsourceid><addsrcrecordid>eNqFkcFqGzEQhkVoIa6bN8hBl0Av60orreS9FIpp7YKhl_QsZqVRkNldbaR1IG9fuWt6TGAYadA3_6D5CbnnbMMZV19PmwlmG582NeNiw0rU-oas-FaLquGy_kBWjAleiZqJW_Ip5xNjXJeHFfH7w2PVQUZHIedoA8zhBemAQ0yvtEdIYxifKIyOhjlTmKY-2MLEkc6RHs4DjBTsv9rhjMvtQtu-6AV_hT-Tjx76jHfXc03-_PzxuDtUx9_7X7vvx8pKVs9VK7fSocaOO8-dE85xJQCl0lqhZg12yrWSectKapTzftu1jQYmWuyk68SafFl0pxSfz5hnM4Rsse9hxHjOhqvy7VYJod5HGy6kbFrdFFQuqE0x54TeTCkMkF4NZ-bigDmZxQFzccCwErUubQ_XCZAt9D7BaEP-31trKdu6vch_Wzgsm3kJmEy2AUeLLqSyUeNieHvQX0xanzo</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1513445975</pqid></control><display><type>article</type><title>GHT-based associative memory learning and its application to Human action detection and classification</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</creator><creatorcontrib>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</creatorcontrib><description>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate.
•GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</description><identifier>ISSN: 0031-3203</identifier><identifier>EISSN: 1873-5142</identifier><identifier>DOI: 10.1016/j.patcog.2013.03.027</identifier><identifier>CODEN: PTNRA8</identifier><language>eng</language><publisher>Kidlington: Elsevier Ltd</publisher><subject>Action object shapes ; Applied sciences ; Architecture ; Associative memory ; Classification ; Construction ; Detection, estimation, filtering, equalization, prediction ; Exact sciences and technology ; Generalized Hough transform ; Human ; Human action detection and recognition ; Hypergraph ; Image processing ; Information, signal and communications theory ; Intelligence ; Learning ; Signal and communications theory ; Signal processing ; Signal, noise ; Telecommunications and information theory ; Voting</subject><ispartof>Pattern recognition, 2013-11, Vol.46 (11), p.3117-3128</ispartof><rights>2013 Elsevier Ltd</rights><rights>2014 INIST-CNRS</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</citedby><cites>FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&idt=27449295$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Cheng, Shyi-Chyi</creatorcontrib><creatorcontrib>Cheng, Kwang-Yu</creatorcontrib><creatorcontrib>Chen, Yi-Ping Phoebe</creatorcontrib><title>GHT-based associative memory learning and its application to Human action detection and classification</title><title>Pattern recognition</title><description>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate.
•GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</description><subject>Action object shapes</subject><subject>Applied sciences</subject><subject>Architecture</subject><subject>Associative memory</subject><subject>Classification</subject><subject>Construction</subject><subject>Detection, estimation, filtering, equalization, prediction</subject><subject>Exact sciences and technology</subject><subject>Generalized Hough transform</subject><subject>Human</subject><subject>Human action detection and recognition</subject><subject>Hypergraph</subject><subject>Image processing</subject><subject>Information, signal and communications theory</subject><subject>Intelligence</subject><subject>Learning</subject><subject>Signal and communications theory</subject><subject>Signal processing</subject><subject>Signal, noise</subject><subject>Telecommunications and information theory</subject><subject>Voting</subject><issn>0031-3203</issn><issn>1873-5142</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2013</creationdate><recordtype>article</recordtype><recordid>eNqFkcFqGzEQhkVoIa6bN8hBl0Av60orreS9FIpp7YKhl_QsZqVRkNldbaR1IG9fuWt6TGAYadA3_6D5CbnnbMMZV19PmwlmG582NeNiw0rU-oas-FaLquGy_kBWjAleiZqJW_Ip5xNjXJeHFfH7w2PVQUZHIedoA8zhBemAQ0yvtEdIYxifKIyOhjlTmKY-2MLEkc6RHs4DjBTsv9rhjMvtQtu-6AV_hT-Tjx76jHfXc03-_PzxuDtUx9_7X7vvx8pKVs9VK7fSocaOO8-dE85xJQCl0lqhZg12yrWSectKapTzftu1jQYmWuyk68SafFl0pxSfz5hnM4Rsse9hxHjOhqvy7VYJod5HGy6kbFrdFFQuqE0x54TeTCkMkF4NZ-bigDmZxQFzccCwErUubQ_XCZAt9D7BaEP-31trKdu6vch_Wzgsm3kJmEy2AUeLLqSyUeNieHvQX0xanzo</recordid><startdate>20131101</startdate><enddate>20131101</enddate><creator>Cheng, Shyi-Chyi</creator><creator>Cheng, Kwang-Yu</creator><creator>Chen, Yi-Ping Phoebe</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20131101</creationdate><title>GHT-based associative memory learning and its application to Human action detection and classification</title><author>Cheng, Shyi-Chyi ; Cheng, Kwang-Yu ; Chen, Yi-Ping Phoebe</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2013</creationdate><topic>Action object shapes</topic><topic>Applied sciences</topic><topic>Architecture</topic><topic>Associative memory</topic><topic>Classification</topic><topic>Construction</topic><topic>Detection, estimation, filtering, equalization, prediction</topic><topic>Exact sciences and technology</topic><topic>Generalized Hough transform</topic><topic>Human</topic><topic>Human action detection and recognition</topic><topic>Hypergraph</topic><topic>Image processing</topic><topic>Information, signal and communications theory</topic><topic>Intelligence</topic><topic>Learning</topic><topic>Signal and communications theory</topic><topic>Signal processing</topic><topic>Signal, noise</topic><topic>Telecommunications and information theory</topic><topic>Voting</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Cheng, Shyi-Chyi</creatorcontrib><creatorcontrib>Cheng, Kwang-Yu</creatorcontrib><creatorcontrib>Chen, Yi-Ping Phoebe</creatorcontrib><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Pattern recognition</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cheng, Shyi-Chyi</au><au>Cheng, Kwang-Yu</au><au>Chen, Yi-Ping Phoebe</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>GHT-based associative memory learning and its application to Human action detection and classification</atitle><jtitle>Pattern recognition</jtitle><date>2013-11-01</date><risdate>2013</risdate><volume>46</volume><issue>11</issue><spage>3117</spage><epage>3128</epage><pages>3117-3128</pages><issn>0031-3203</issn><eissn>1873-5142</eissn><coden>PTNRA8</coden><abstract>This paper, investigating the goal of human-level synthetic intelligence, presents a novel approach to learning an associative memory model using Generalized Hough Transform (GHT) [1]. A human action detection and classification system is also constructed to verify the effectiveness of the proposed GHT-based associative memory model. Existing human action classification systems use machine learning architectures and low-level features to characterize a specific human action. However, existing machine learning architectures often lack restructuring capability, which is an important process of forming the conceptual structures in human-level synthetic intelligence. The gap between low-level features and high-level human intelligence also degrades the performance of existing human action recognition algorithms when the spatial–temporal boundaries of action objects are ambiguous. To eliminate the side effect of temporal ambiguity, the proposed system uses a preprocessing procedure to extract key-frames from a video sequence and provide a compact representation for this video. The image and motion features of patches extracted from each key-frame are collected and used to train an appearance–motion codebook. The training procedure, based on the learnt codebook and GHT, constructs a hypergraph for associative memory learning. For each key-frame of a test video clip, the Hough voting framework is also used to detect salient segments, which are further partitioned into multiple patches, by grouping blocks of similar appearance and motions. The features of the detected patches are used to query the associative memory and retrieve missing patches from key-frames to recall the whole action object. These patches are then used to locate the target action object and classify the action type simultaneously using a probabilistic Hough voting scheme. Results show that the proposed method gives good performance on several publicly available datasets in terms of detection accuracy and recognition rate.
•GHT-based associative memory is proposed to recover missing parts of a video object from parts of detectable patches.•The approach is robust even when we apply it to detect video action objects from a video clip with a cluttered background.•We propose an automatic training procedure for generating efficient and discriminative associative memory models.•Techniques to smartly select the patches and synthesize the target action shape are implemented.</abstract><cop>Kidlington</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.patcog.2013.03.027</doi><tpages>12</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0031-3203 |
ispartof | Pattern recognition, 2013-11, Vol.46 (11), p.3117-3128 |
issn | 0031-3203 1873-5142 |
language | eng |
recordid | cdi_proquest_miscellaneous_1671496336 |
source | ScienceDirect Freedom Collection 2022-2024 |
subjects | Action object shapes Applied sciences Architecture Associative memory Classification Construction Detection, estimation, filtering, equalization, prediction Exact sciences and technology Generalized Hough transform Human Human action detection and recognition Hypergraph Image processing Information, signal and communications theory Intelligence Learning Signal and communications theory Signal processing Signal, noise Telecommunications and information theory Voting |
title | GHT-based associative memory learning and its application to Human action detection and classification |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-30T22%3A26%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=GHT-based%20associative%20memory%20learning%20and%20its%20application%20to%20Human%20action%20detection%20and%20classification&rft.jtitle=Pattern%20recognition&rft.au=Cheng,%20Shyi-Chyi&rft.date=2013-11-01&rft.volume=46&rft.issue=11&rft.spage=3117&rft.epage=3128&rft.pages=3117-3128&rft.issn=0031-3203&rft.eissn=1873-5142&rft.coden=PTNRA8&rft_id=info:doi/10.1016/j.patcog.2013.03.027&rft_dat=%3Cproquest_cross%3E1671496336%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c402t-9484de7eb1df1dd3dd163ae46776e705eb6d940fc040f56dff8b957a039eb4db3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1513445975&rft_id=info:pmid/&rfr_iscdi=true |