Loading…

Human activity recognition using deep electroencephalography learning

•A deep learning-based framework for classifying EEG artifacts is proposed and evaluated, using a stack of an LSTM Network against a CNN.•Evaluating the state-of-the-art works in classifying raw EEG signals and HAR shows the benefits of using the proposed framework.•The human cognition and emotion d...

Full description

Saved in:
Bibliographic Details
Published in:Biomedical signal processing and control 2020-09, Vol.62, p.102094, Article 102094
Main Authors: Salehzadeh, Amirsaleh, Calitz, Andre P., Greyling, Jean
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3
cites cdi_FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3
container_end_page
container_issue
container_start_page 102094
container_title Biomedical signal processing and control
container_volume 62
creator Salehzadeh, Amirsaleh
Calitz, Andre P.
Greyling, Jean
description •A deep learning-based framework for classifying EEG artifacts is proposed and evaluated, using a stack of an LSTM Network against a CNN.•Evaluating the state-of-the-art works in classifying raw EEG signals and HAR shows the benefits of using the proposed framework.•The human cognition and emotion data extracted by classifying EEG signals captured by a consumer-grade EEG wearable device is emphasized.•Using a consumer-grade EEG wearable device to collect data for HAR provides advantages over the common sensory technologies used in HAR. Electroencephalography (EEG) signals can be contaminated by the noise raised from a non-cerebral artifact and vary in magnitude due to physiological differences between individuals. Hence, EEG has not been extensively applied to the design of Human Activity Recognition (HAR) systems. HAR involves classifying the activities of an individual that are captured by sensory technologies. To address this issue, this paper introduces a deep learning-based framework for classifying EEG artifacts (FCEA), based on a person’s physiological activity. The FCEA proposes an approach for designing a processing pipeline involving a Convolutional Neural Network and a Long Short-Term Memory Recurrent Neural Network, in order to classify raw EEG signals based on the EEG artifacts. To demonstrate the performance of the FCEA, a 3-class dataset of EEG activities was created. A consumer-grade EEG wearable device was used to collect the data from two prefrontal EEG channels, from 8 participants whilst speaking loudly, reading printed text and watching a TV program. These activities include jaw-clenching and head and eye movements, that are part of a wide range of daily human activities. Moreover, these activities cannot easily be detected by using the sensory technologies that have commonly been used in HAR. The comparative performance analysis results demonstrate that the 1- and 2-channel models trained by the FCEA outperform competitive state-of-the-art deep-learning models for HAR and raw EEG data. The deep-learning approach proposed by the FCEA improves raw data processing to obtain a better generalization performance.
doi_str_mv 10.1016/j.bspc.2020.102094
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_bspc_2020_102094</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S1746809420302500</els_id><sourcerecordid>S1746809420302500</sourcerecordid><originalsourceid>FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3</originalsourceid><addsrcrecordid>eNp9kM1qwzAQhEVpoWnaF-jJL2BXK8s_gl5KSJtCoJf2LBRp7cg4lpGcgN--Nm6uPe0yzCw7HyHPQBOgkL80ySH0OmGUzQKjgt-QFRQ8j0ug5e11n_R78hBCQykvC-Arst2dT6qLlB7sxQ5j5FG7urODdV10DrarI4PYR9iiHrzDTmN_VK2rveqPY9Si8t1keiR3lWoDPv3NNfl5335vdvH-6-Nz87aPdUrpEFd5xqECyHlqgKnMCE7z9KAzyBQYlkFZUKFRYVkwkaFAnlYMGZpUAApTpWvClrvauxA8VrL39qT8KIHKGYRs5AxCziDkAmIKvS4hnD67WPQyaDs3MXZqO0jj7H_xX5DsZ7g</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Human activity recognition using deep electroencephalography learning</title><source>Elsevier</source><creator>Salehzadeh, Amirsaleh ; Calitz, Andre P. ; Greyling, Jean</creator><creatorcontrib>Salehzadeh, Amirsaleh ; Calitz, Andre P. ; Greyling, Jean</creatorcontrib><description>•A deep learning-based framework for classifying EEG artifacts is proposed and evaluated, using a stack of an LSTM Network against a CNN.•Evaluating the state-of-the-art works in classifying raw EEG signals and HAR shows the benefits of using the proposed framework.•The human cognition and emotion data extracted by classifying EEG signals captured by a consumer-grade EEG wearable device is emphasized.•Using a consumer-grade EEG wearable device to collect data for HAR provides advantages over the common sensory technologies used in HAR. Electroencephalography (EEG) signals can be contaminated by the noise raised from a non-cerebral artifact and vary in magnitude due to physiological differences between individuals. Hence, EEG has not been extensively applied to the design of Human Activity Recognition (HAR) systems. HAR involves classifying the activities of an individual that are captured by sensory technologies. To address this issue, this paper introduces a deep learning-based framework for classifying EEG artifacts (FCEA), based on a person’s physiological activity. The FCEA proposes an approach for designing a processing pipeline involving a Convolutional Neural Network and a Long Short-Term Memory Recurrent Neural Network, in order to classify raw EEG signals based on the EEG artifacts. To demonstrate the performance of the FCEA, a 3-class dataset of EEG activities was created. A consumer-grade EEG wearable device was used to collect the data from two prefrontal EEG channels, from 8 participants whilst speaking loudly, reading printed text and watching a TV program. These activities include jaw-clenching and head and eye movements, that are part of a wide range of daily human activities. Moreover, these activities cannot easily be detected by using the sensory technologies that have commonly been used in HAR. The comparative performance analysis results demonstrate that the 1- and 2-channel models trained by the FCEA outperform competitive state-of-the-art deep-learning models for HAR and raw EEG data. The deep-learning approach proposed by the FCEA improves raw data processing to obtain a better generalization performance.</description><identifier>ISSN: 1746-8094</identifier><identifier>EISSN: 1746-8108</identifier><identifier>DOI: 10.1016/j.bspc.2020.102094</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Deep learning ; EEG artifact classification ; EEG wearable device ; Human activity recognition</subject><ispartof>Biomedical signal processing and control, 2020-09, Vol.62, p.102094, Article 102094</ispartof><rights>2020 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3</citedby><cites>FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3</cites><orcidid>0000-0002-0977-1683</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Salehzadeh, Amirsaleh</creatorcontrib><creatorcontrib>Calitz, Andre P.</creatorcontrib><creatorcontrib>Greyling, Jean</creatorcontrib><title>Human activity recognition using deep electroencephalography learning</title><title>Biomedical signal processing and control</title><description>•A deep learning-based framework for classifying EEG artifacts is proposed and evaluated, using a stack of an LSTM Network against a CNN.•Evaluating the state-of-the-art works in classifying raw EEG signals and HAR shows the benefits of using the proposed framework.•The human cognition and emotion data extracted by classifying EEG signals captured by a consumer-grade EEG wearable device is emphasized.•Using a consumer-grade EEG wearable device to collect data for HAR provides advantages over the common sensory technologies used in HAR. Electroencephalography (EEG) signals can be contaminated by the noise raised from a non-cerebral artifact and vary in magnitude due to physiological differences between individuals. Hence, EEG has not been extensively applied to the design of Human Activity Recognition (HAR) systems. HAR involves classifying the activities of an individual that are captured by sensory technologies. To address this issue, this paper introduces a deep learning-based framework for classifying EEG artifacts (FCEA), based on a person’s physiological activity. The FCEA proposes an approach for designing a processing pipeline involving a Convolutional Neural Network and a Long Short-Term Memory Recurrent Neural Network, in order to classify raw EEG signals based on the EEG artifacts. To demonstrate the performance of the FCEA, a 3-class dataset of EEG activities was created. A consumer-grade EEG wearable device was used to collect the data from two prefrontal EEG channels, from 8 participants whilst speaking loudly, reading printed text and watching a TV program. These activities include jaw-clenching and head and eye movements, that are part of a wide range of daily human activities. Moreover, these activities cannot easily be detected by using the sensory technologies that have commonly been used in HAR. The comparative performance analysis results demonstrate that the 1- and 2-channel models trained by the FCEA outperform competitive state-of-the-art deep-learning models for HAR and raw EEG data. The deep-learning approach proposed by the FCEA improves raw data processing to obtain a better generalization performance.</description><subject>Deep learning</subject><subject>EEG artifact classification</subject><subject>EEG wearable device</subject><subject>Human activity recognition</subject><issn>1746-8094</issn><issn>1746-8108</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNp9kM1qwzAQhEVpoWnaF-jJL2BXK8s_gl5KSJtCoJf2LBRp7cg4lpGcgN--Nm6uPe0yzCw7HyHPQBOgkL80ySH0OmGUzQKjgt-QFRQ8j0ug5e11n_R78hBCQykvC-Arst2dT6qLlB7sxQ5j5FG7urODdV10DrarI4PYR9iiHrzDTmN_VK2rveqPY9Si8t1keiR3lWoDPv3NNfl5335vdvH-6-Nz87aPdUrpEFd5xqECyHlqgKnMCE7z9KAzyBQYlkFZUKFRYVkwkaFAnlYMGZpUAApTpWvClrvauxA8VrL39qT8KIHKGYRs5AxCziDkAmIKvS4hnD67WPQyaDs3MXZqO0jj7H_xX5DsZ7g</recordid><startdate>202009</startdate><enddate>202009</enddate><creator>Salehzadeh, Amirsaleh</creator><creator>Calitz, Andre P.</creator><creator>Greyling, Jean</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-0977-1683</orcidid></search><sort><creationdate>202009</creationdate><title>Human activity recognition using deep electroencephalography learning</title><author>Salehzadeh, Amirsaleh ; Calitz, Andre P. ; Greyling, Jean</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Deep learning</topic><topic>EEG artifact classification</topic><topic>EEG wearable device</topic><topic>Human activity recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Salehzadeh, Amirsaleh</creatorcontrib><creatorcontrib>Calitz, Andre P.</creatorcontrib><creatorcontrib>Greyling, Jean</creatorcontrib><collection>CrossRef</collection><jtitle>Biomedical signal processing and control</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Salehzadeh, Amirsaleh</au><au>Calitz, Andre P.</au><au>Greyling, Jean</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Human activity recognition using deep electroencephalography learning</atitle><jtitle>Biomedical signal processing and control</jtitle><date>2020-09</date><risdate>2020</risdate><volume>62</volume><spage>102094</spage><pages>102094-</pages><artnum>102094</artnum><issn>1746-8094</issn><eissn>1746-8108</eissn><abstract>•A deep learning-based framework for classifying EEG artifacts is proposed and evaluated, using a stack of an LSTM Network against a CNN.•Evaluating the state-of-the-art works in classifying raw EEG signals and HAR shows the benefits of using the proposed framework.•The human cognition and emotion data extracted by classifying EEG signals captured by a consumer-grade EEG wearable device is emphasized.•Using a consumer-grade EEG wearable device to collect data for HAR provides advantages over the common sensory technologies used in HAR. Electroencephalography (EEG) signals can be contaminated by the noise raised from a non-cerebral artifact and vary in magnitude due to physiological differences between individuals. Hence, EEG has not been extensively applied to the design of Human Activity Recognition (HAR) systems. HAR involves classifying the activities of an individual that are captured by sensory technologies. To address this issue, this paper introduces a deep learning-based framework for classifying EEG artifacts (FCEA), based on a person’s physiological activity. The FCEA proposes an approach for designing a processing pipeline involving a Convolutional Neural Network and a Long Short-Term Memory Recurrent Neural Network, in order to classify raw EEG signals based on the EEG artifacts. To demonstrate the performance of the FCEA, a 3-class dataset of EEG activities was created. A consumer-grade EEG wearable device was used to collect the data from two prefrontal EEG channels, from 8 participants whilst speaking loudly, reading printed text and watching a TV program. These activities include jaw-clenching and head and eye movements, that are part of a wide range of daily human activities. Moreover, these activities cannot easily be detected by using the sensory technologies that have commonly been used in HAR. The comparative performance analysis results demonstrate that the 1- and 2-channel models trained by the FCEA outperform competitive state-of-the-art deep-learning models for HAR and raw EEG data. The deep-learning approach proposed by the FCEA improves raw data processing to obtain a better generalization performance.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.bspc.2020.102094</doi><orcidid>https://orcid.org/0000-0002-0977-1683</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1746-8094
ispartof Biomedical signal processing and control, 2020-09, Vol.62, p.102094, Article 102094
issn 1746-8094
1746-8108
language eng
recordid cdi_crossref_primary_10_1016_j_bspc_2020_102094
source Elsevier
subjects Deep learning
EEG artifact classification
EEG wearable device
Human activity recognition
title Human activity recognition using deep electroencephalography learning
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T20%3A21%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Human%20activity%20recognition%20using%20deep%20electroencephalography%20learning&rft.jtitle=Biomedical%20signal%20processing%20and%20control&rft.au=Salehzadeh,%20Amirsaleh&rft.date=2020-09&rft.volume=62&rft.spage=102094&rft.pages=102094-&rft.artnum=102094&rft.issn=1746-8094&rft.eissn=1746-8108&rft_id=info:doi/10.1016/j.bspc.2020.102094&rft_dat=%3Celsevier_cross%3ES1746809420302500%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c300t-f6541f11643d12a5d94063bc515a1d2518709ceae87295e9e43f2e2ed391e9df3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true