Loading…

Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network

Sensor-based human activity recognition (HAR) has drawn extensive attention from the research community due to its potential applications in various domains, including interactive gaming, activity monitoring, healthcare, etc. Although plentiful approaches (i.e., handcrafted feature-based and deep le...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2023-10, Vol.227, p.120132, Article 120132
Main Authors: Raja Sekaran, Sarmela, Han, Pang Ying, Yin, Ooi Shih
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3
cites cdi_FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3
container_end_page
container_issue
container_start_page 120132
container_title Expert systems with applications
container_volume 227
creator Raja Sekaran, Sarmela
Han, Pang Ying
Yin, Ooi Shih
description Sensor-based human activity recognition (HAR) has drawn extensive attention from the research community due to its potential applications in various domains, including interactive gaming, activity monitoring, healthcare, etc. Although plentiful approaches (i.e., handcrafted feature-based and deep learning methods) have been proposed throughout the years, there are still several challenges in developing an efficient and effective HAR system. For instance, handcrafted feature-based methods rely on manual feature engineering by experts and require time-consuming feature selection methods. Conversely, deep learning methods can automatically capture salient features without domain experts. However, some deep learning methods, especially Convolutional Neural Networks (CNN), cannot extract temporal features effectively, which are significant to motion analysis. Unlike CNN, recurrent models are exceptional at capturing temporal characteristics, but these models contain gigantic model parameters, requiring tremendous computation. This may limit the deployment of such models, especially to low-spec or embedded devices. Hence, this paper proposes a lightweight deep learning model, Lightweight Multiheaded TCN (Light-MHTCN), for human activity recognition. Light-MHTCN extracts the multiscale features of the inertial sensor signals through the parallelly organised Convolutional Heads to capture richer information. Further, integrating dilated causal convolutions and residual connections preserves longer-term dependency, which can boost the overall model performance. The performance of Light-MHTCN is assessed on three popular smartphone-based HAR databases: UCI HAR, WISDM V1 and UniMiB SHAR. With only ∼0.21 million parameters, our lightweight model is able to achieve state-of-the-art performance with recognition accuracies of 96.47%, 99.98% and 98.63% on these databases, respectively. •Light-MHTCN requires minimal preprocessing and no manual feature engineering.•Light-MHTCN is lightweight in computation, with only 0.21M parameters.•Light-MHTCN allows multiscale feature extraction due to the parallel architecture.•Light-MHTCN retains longer-term dependency using dilations and residual connections.•Light-MHTCN achieves 96.47% on UCI HAR, 99.98% on WISDM V1 and 98.63% on UniMiB SHAR.
doi_str_mv 10.1016/j.eswa.2023.120132
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_eswa_2023_120132</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417423006346</els_id><sourcerecordid>S0957417423006346</sourcerecordid><originalsourceid>FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3</originalsourceid><addsrcrecordid>eNp9kM9KxDAQh4MouK6-gKe-QGvSdJMNeJHFf7DgQT2HNJ1us7bJkqRd9u1NWc9eZhj4fcPMh9A9wQXBhD3sCwhHVZS4pAUpMaHlBVqQNac544JeogUWK55XhFfX6CaEPcaEY8wXqPsclI-HzlnIaxWgybpxUDZTOprJxFPmQbudNdE4m43B2F3Wm10XjzDXbBj7aDpQTQIjDAfnVZ9pZyfXjzOSJgvx6PzPLbpqVR_g7q8v0ffL89fmLd9-vL5vnra5phjHnJG6BBCtYitREdYIUVYgarUGxlKC1ErUlKtWr6DFKYyJXgsODVOctkq1dInK817tXQgeWnnwJv14kgTL2ZXcy9mVnF3Js6sEPZ4hSJdNBrwM2oDV0Jj0f5SNM__hv8gjduE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network</title><source>Elsevier</source><creator>Raja Sekaran, Sarmela ; Han, Pang Ying ; Yin, Ooi Shih</creator><creatorcontrib>Raja Sekaran, Sarmela ; Han, Pang Ying ; Yin, Ooi Shih</creatorcontrib><description>Sensor-based human activity recognition (HAR) has drawn extensive attention from the research community due to its potential applications in various domains, including interactive gaming, activity monitoring, healthcare, etc. Although plentiful approaches (i.e., handcrafted feature-based and deep learning methods) have been proposed throughout the years, there are still several challenges in developing an efficient and effective HAR system. For instance, handcrafted feature-based methods rely on manual feature engineering by experts and require time-consuming feature selection methods. Conversely, deep learning methods can automatically capture salient features without domain experts. However, some deep learning methods, especially Convolutional Neural Networks (CNN), cannot extract temporal features effectively, which are significant to motion analysis. Unlike CNN, recurrent models are exceptional at capturing temporal characteristics, but these models contain gigantic model parameters, requiring tremendous computation. This may limit the deployment of such models, especially to low-spec or embedded devices. Hence, this paper proposes a lightweight deep learning model, Lightweight Multiheaded TCN (Light-MHTCN), for human activity recognition. Light-MHTCN extracts the multiscale features of the inertial sensor signals through the parallelly organised Convolutional Heads to capture richer information. Further, integrating dilated causal convolutions and residual connections preserves longer-term dependency, which can boost the overall model performance. The performance of Light-MHTCN is assessed on three popular smartphone-based HAR databases: UCI HAR, WISDM V1 and UniMiB SHAR. With only ∼0.21 million parameters, our lightweight model is able to achieve state-of-the-art performance with recognition accuracies of 96.47%, 99.98% and 98.63% on these databases, respectively. •Light-MHTCN requires minimal preprocessing and no manual feature engineering.•Light-MHTCN is lightweight in computation, with only 0.21M parameters.•Light-MHTCN allows multiscale feature extraction due to the parallel architecture.•Light-MHTCN retains longer-term dependency using dilations and residual connections.•Light-MHTCN achieves 96.47% on UCI HAR, 99.98% on WISDM V1 and 98.63% on UniMiB SHAR.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2023.120132</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Dilated convolution ; Human activity recognition ; Lightweight deep learning model ; Multiscale feature extraction ; Temporal convolutional network</subject><ispartof>Expert systems with applications, 2023-10, Vol.227, p.120132, Article 120132</ispartof><rights>2023 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3</citedby><cites>FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3</cites><orcidid>0000-0002-3781-6623 ; 0000-0002-3024-1011 ; 0000-0002-6465-5503</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27923,27924</link.rule.ids></links><search><creatorcontrib>Raja Sekaran, Sarmela</creatorcontrib><creatorcontrib>Han, Pang Ying</creatorcontrib><creatorcontrib>Yin, Ooi Shih</creatorcontrib><title>Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network</title><title>Expert systems with applications</title><description>Sensor-based human activity recognition (HAR) has drawn extensive attention from the research community due to its potential applications in various domains, including interactive gaming, activity monitoring, healthcare, etc. Although plentiful approaches (i.e., handcrafted feature-based and deep learning methods) have been proposed throughout the years, there are still several challenges in developing an efficient and effective HAR system. For instance, handcrafted feature-based methods rely on manual feature engineering by experts and require time-consuming feature selection methods. Conversely, deep learning methods can automatically capture salient features without domain experts. However, some deep learning methods, especially Convolutional Neural Networks (CNN), cannot extract temporal features effectively, which are significant to motion analysis. Unlike CNN, recurrent models are exceptional at capturing temporal characteristics, but these models contain gigantic model parameters, requiring tremendous computation. This may limit the deployment of such models, especially to low-spec or embedded devices. Hence, this paper proposes a lightweight deep learning model, Lightweight Multiheaded TCN (Light-MHTCN), for human activity recognition. Light-MHTCN extracts the multiscale features of the inertial sensor signals through the parallelly organised Convolutional Heads to capture richer information. Further, integrating dilated causal convolutions and residual connections preserves longer-term dependency, which can boost the overall model performance. The performance of Light-MHTCN is assessed on three popular smartphone-based HAR databases: UCI HAR, WISDM V1 and UniMiB SHAR. With only ∼0.21 million parameters, our lightweight model is able to achieve state-of-the-art performance with recognition accuracies of 96.47%, 99.98% and 98.63% on these databases, respectively. •Light-MHTCN requires minimal preprocessing and no manual feature engineering.•Light-MHTCN is lightweight in computation, with only 0.21M parameters.•Light-MHTCN allows multiscale feature extraction due to the parallel architecture.•Light-MHTCN retains longer-term dependency using dilations and residual connections.•Light-MHTCN achieves 96.47% on UCI HAR, 99.98% on WISDM V1 and 98.63% on UniMiB SHAR.</description><subject>Dilated convolution</subject><subject>Human activity recognition</subject><subject>Lightweight deep learning model</subject><subject>Multiscale feature extraction</subject><subject>Temporal convolutional network</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kM9KxDAQh4MouK6-gKe-QGvSdJMNeJHFf7DgQT2HNJ1us7bJkqRd9u1NWc9eZhj4fcPMh9A9wQXBhD3sCwhHVZS4pAUpMaHlBVqQNac544JeogUWK55XhFfX6CaEPcaEY8wXqPsclI-HzlnIaxWgybpxUDZTOprJxFPmQbudNdE4m43B2F3Wm10XjzDXbBj7aDpQTQIjDAfnVZ9pZyfXjzOSJgvx6PzPLbpqVR_g7q8v0ffL89fmLd9-vL5vnra5phjHnJG6BBCtYitREdYIUVYgarUGxlKC1ErUlKtWr6DFKYyJXgsODVOctkq1dInK817tXQgeWnnwJv14kgTL2ZXcy9mVnF3Js6sEPZ4hSJdNBrwM2oDV0Jj0f5SNM__hv8gjduE</recordid><startdate>20231001</startdate><enddate>20231001</enddate><creator>Raja Sekaran, Sarmela</creator><creator>Han, Pang Ying</creator><creator>Yin, Ooi Shih</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-3781-6623</orcidid><orcidid>https://orcid.org/0000-0002-3024-1011</orcidid><orcidid>https://orcid.org/0000-0002-6465-5503</orcidid></search><sort><creationdate>20231001</creationdate><title>Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network</title><author>Raja Sekaran, Sarmela ; Han, Pang Ying ; Yin, Ooi Shih</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Dilated convolution</topic><topic>Human activity recognition</topic><topic>Lightweight deep learning model</topic><topic>Multiscale feature extraction</topic><topic>Temporal convolutional network</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Raja Sekaran, Sarmela</creatorcontrib><creatorcontrib>Han, Pang Ying</creatorcontrib><creatorcontrib>Yin, Ooi Shih</creatorcontrib><collection>CrossRef</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Raja Sekaran, Sarmela</au><au>Han, Pang Ying</au><au>Yin, Ooi Shih</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network</atitle><jtitle>Expert systems with applications</jtitle><date>2023-10-01</date><risdate>2023</risdate><volume>227</volume><spage>120132</spage><pages>120132-</pages><artnum>120132</artnum><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>Sensor-based human activity recognition (HAR) has drawn extensive attention from the research community due to its potential applications in various domains, including interactive gaming, activity monitoring, healthcare, etc. Although plentiful approaches (i.e., handcrafted feature-based and deep learning methods) have been proposed throughout the years, there are still several challenges in developing an efficient and effective HAR system. For instance, handcrafted feature-based methods rely on manual feature engineering by experts and require time-consuming feature selection methods. Conversely, deep learning methods can automatically capture salient features without domain experts. However, some deep learning methods, especially Convolutional Neural Networks (CNN), cannot extract temporal features effectively, which are significant to motion analysis. Unlike CNN, recurrent models are exceptional at capturing temporal characteristics, but these models contain gigantic model parameters, requiring tremendous computation. This may limit the deployment of such models, especially to low-spec or embedded devices. Hence, this paper proposes a lightweight deep learning model, Lightweight Multiheaded TCN (Light-MHTCN), for human activity recognition. Light-MHTCN extracts the multiscale features of the inertial sensor signals through the parallelly organised Convolutional Heads to capture richer information. Further, integrating dilated causal convolutions and residual connections preserves longer-term dependency, which can boost the overall model performance. The performance of Light-MHTCN is assessed on three popular smartphone-based HAR databases: UCI HAR, WISDM V1 and UniMiB SHAR. With only ∼0.21 million parameters, our lightweight model is able to achieve state-of-the-art performance with recognition accuracies of 96.47%, 99.98% and 98.63% on these databases, respectively. •Light-MHTCN requires minimal preprocessing and no manual feature engineering.•Light-MHTCN is lightweight in computation, with only 0.21M parameters.•Light-MHTCN allows multiscale feature extraction due to the parallel architecture.•Light-MHTCN retains longer-term dependency using dilations and residual connections.•Light-MHTCN achieves 96.47% on UCI HAR, 99.98% on WISDM V1 and 98.63% on UniMiB SHAR.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2023.120132</doi><orcidid>https://orcid.org/0000-0002-3781-6623</orcidid><orcidid>https://orcid.org/0000-0002-3024-1011</orcidid><orcidid>https://orcid.org/0000-0002-6465-5503</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2023-10, Vol.227, p.120132, Article 120132
issn 0957-4174
1873-6793
language eng
recordid cdi_crossref_primary_10_1016_j_eswa_2023_120132
source Elsevier
subjects Dilated convolution
Human activity recognition
Lightweight deep learning model
Multiscale feature extraction
Temporal convolutional network
title Smartphone-based human activity recognition using lightweight multiheaded temporal convolutional network
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T17%3A59%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Smartphone-based%20human%20activity%20recognition%20using%20lightweight%20multiheaded%20temporal%20convolutional%20network&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Raja%20Sekaran,%20Sarmela&rft.date=2023-10-01&rft.volume=227&rft.spage=120132&rft.pages=120132-&rft.artnum=120132&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2023.120132&rft_dat=%3Celsevier_cross%3ES0957417423006346%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c300t-61b2ee9fa659416d9924e9ba8e66c301ba9b37afc5ef061b01c897ed6a73faaf3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true