Loading…

Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition

Data annotation for human activity recognition is a well-known challenge for researchers. In particular, annotation in daily life settings relies on self-reporting tools with unknown accuracy. Speech is a promising interface for activity labeling. In this work, we compare the accuracy of two commerc...

Full description

Saved in:
Bibliographic Details
Published in:International journal of human-computer studies 2025-01, Vol.195, p.103404, Article 103404
Main Authors: Levasseur, Guillaume, Tang, Kejia, Bersini, Hugues
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c183t-8f3597fdf9e31dd5e9fc44f1ef5fafa3d6a715914918657afb5645873849e5563
container_end_page
container_issue
container_start_page 103404
container_title International journal of human-computer studies
container_volume 195
creator Levasseur, Guillaume
Tang, Kejia
Bersini, Hugues
description Data annotation for human activity recognition is a well-known challenge for researchers. In particular, annotation in daily life settings relies on self-reporting tools with unknown accuracy. Speech is a promising interface for activity labeling. In this work, we compare the accuracy of two commercially available tools for annotation: voice diaries and connected buttons. We retrofit the water meters of thirty homes in the USA for infrastructure-mediated sensing. Participants are split into equal groups and receive one of the self-reporting tools. The balanced accuracy metric is transferred from the field of machine learning to the evaluation of the annotation performance. Our results show that connected buttons perform significantly better than the voice diary, with 92% median accuracy and 65% median reporting rate. Using questionnaire answers, we highlight that annotation performance is impacted by habit formation and sentiments toward the annotation tool. The use case for data annotation is to disaggregate water meter data into human activities beyond the point of use. We show that it is feasible with a machine-learning model and the corrected annotations. Finally, we formulate recommendations for the design of studies and intelligent environments around the key ideas of proportionality and immediacy. •Self-reporting of human activities enables to train machine learning algorithms.•Speech is not the best interface to report chores in daily life settings.•Internet-connected buttons prove better with 92% accuracy and 65% reporting rate.•Human activities can be inferred from smart water meter data.
doi_str_mv 10.1016/j.ijhcs.2024.103404
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_ijhcs_2024_103404</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S1071581924001873</els_id><sourcerecordid>S1071581924001873</sourcerecordid><originalsourceid>FETCH-LOGICAL-c183t-8f3597fdf9e31dd5e9fc44f1ef5fafa3d6a715914918657afb5645873849e5563</originalsourceid><addsrcrecordid>eNp9kE1LAzEQhnNQsFZ_gZf8ga1Jk-yH4EFK_YCCFz2HmJ10Z-kmSxKL_fduW8-eBmZ4XuZ9CLnjbMEZL-_7BfadTYslW8ppIySTF2TGWcULVfPmilyn1DPGKsnYjKQNDpgTDY6mEcB2FD21wXuwGVrahQHSA13_jBBxAJ_NbroOo4mYgj9RsHNFhDHEjH5Lcwi7RF2ItPsejKfGZtxjPtAINmw9Zgz-hlw6s0tw-zfn5PN5_bF6LTbvL2-rp01heS1yUTuhmsq1rgHB21ZB46yUjoNTzjgj2tJUXDVcNrwuVWXclyqlqitRywaUKsWciHOujSGlCE6PUwkTD5ozfXSle31ypY-u9NnVRD2eKZhe2yNEnSyCt9Di1CHrNuC__C-L9HhL</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition</title><source>Elsevier</source><creator>Levasseur, Guillaume ; Tang, Kejia ; Bersini, Hugues</creator><creatorcontrib>Levasseur, Guillaume ; Tang, Kejia ; Bersini, Hugues</creatorcontrib><description>Data annotation for human activity recognition is a well-known challenge for researchers. In particular, annotation in daily life settings relies on self-reporting tools with unknown accuracy. Speech is a promising interface for activity labeling. In this work, we compare the accuracy of two commercially available tools for annotation: voice diaries and connected buttons. We retrofit the water meters of thirty homes in the USA for infrastructure-mediated sensing. Participants are split into equal groups and receive one of the self-reporting tools. The balanced accuracy metric is transferred from the field of machine learning to the evaluation of the annotation performance. Our results show that connected buttons perform significantly better than the voice diary, with 92% median accuracy and 65% median reporting rate. Using questionnaire answers, we highlight that annotation performance is impacted by habit formation and sentiments toward the annotation tool. The use case for data annotation is to disaggregate water meter data into human activities beyond the point of use. We show that it is feasible with a machine-learning model and the corrected annotations. Finally, we formulate recommendations for the design of studies and intelligent environments around the key ideas of proportionality and immediacy. •Self-reporting of human activities enables to train machine learning algorithms.•Speech is not the best interface to report chores in daily life settings.•Internet-connected buttons prove better with 92% accuracy and 65% reporting rate.•Human activities can be inferred from smart water meter data.</description><identifier>ISSN: 1071-5819</identifier><identifier>DOI: 10.1016/j.ijhcs.2024.103404</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Activity recognition ; Machine learning ; Self-reporting ; Smart homes ; Voice assistants ; Water consumption</subject><ispartof>International journal of human-computer studies, 2025-01, Vol.195, p.103404, Article 103404</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c183t-8f3597fdf9e31dd5e9fc44f1ef5fafa3d6a715914918657afb5645873849e5563</cites><orcidid>0000-0001-5744-8533</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27923,27924</link.rule.ids></links><search><creatorcontrib>Levasseur, Guillaume</creatorcontrib><creatorcontrib>Tang, Kejia</creatorcontrib><creatorcontrib>Bersini, Hugues</creatorcontrib><title>Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition</title><title>International journal of human-computer studies</title><description>Data annotation for human activity recognition is a well-known challenge for researchers. In particular, annotation in daily life settings relies on self-reporting tools with unknown accuracy. Speech is a promising interface for activity labeling. In this work, we compare the accuracy of two commercially available tools for annotation: voice diaries and connected buttons. We retrofit the water meters of thirty homes in the USA for infrastructure-mediated sensing. Participants are split into equal groups and receive one of the self-reporting tools. The balanced accuracy metric is transferred from the field of machine learning to the evaluation of the annotation performance. Our results show that connected buttons perform significantly better than the voice diary, with 92% median accuracy and 65% median reporting rate. Using questionnaire answers, we highlight that annotation performance is impacted by habit formation and sentiments toward the annotation tool. The use case for data annotation is to disaggregate water meter data into human activities beyond the point of use. We show that it is feasible with a machine-learning model and the corrected annotations. Finally, we formulate recommendations for the design of studies and intelligent environments around the key ideas of proportionality and immediacy. •Self-reporting of human activities enables to train machine learning algorithms.•Speech is not the best interface to report chores in daily life settings.•Internet-connected buttons prove better with 92% accuracy and 65% reporting rate.•Human activities can be inferred from smart water meter data.</description><subject>Activity recognition</subject><subject>Machine learning</subject><subject>Self-reporting</subject><subject>Smart homes</subject><subject>Voice assistants</subject><subject>Water consumption</subject><issn>1071-5819</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2025</creationdate><recordtype>article</recordtype><recordid>eNp9kE1LAzEQhnNQsFZ_gZf8ga1Jk-yH4EFK_YCCFz2HmJ10Z-kmSxKL_fduW8-eBmZ4XuZ9CLnjbMEZL-_7BfadTYslW8ppIySTF2TGWcULVfPmilyn1DPGKsnYjKQNDpgTDY6mEcB2FD21wXuwGVrahQHSA13_jBBxAJ_NbroOo4mYgj9RsHNFhDHEjH5Lcwi7RF2ItPsejKfGZtxjPtAINmw9Zgz-hlw6s0tw-zfn5PN5_bF6LTbvL2-rp01heS1yUTuhmsq1rgHB21ZB46yUjoNTzjgj2tJUXDVcNrwuVWXclyqlqitRywaUKsWciHOujSGlCE6PUwkTD5ozfXSle31ypY-u9NnVRD2eKZhe2yNEnSyCt9Di1CHrNuC__C-L9HhL</recordid><startdate>202501</startdate><enddate>202501</enddate><creator>Levasseur, Guillaume</creator><creator>Tang, Kejia</creator><creator>Bersini, Hugues</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0001-5744-8533</orcidid></search><sort><creationdate>202501</creationdate><title>Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition</title><author>Levasseur, Guillaume ; Tang, Kejia ; Bersini, Hugues</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c183t-8f3597fdf9e31dd5e9fc44f1ef5fafa3d6a715914918657afb5645873849e5563</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2025</creationdate><topic>Activity recognition</topic><topic>Machine learning</topic><topic>Self-reporting</topic><topic>Smart homes</topic><topic>Voice assistants</topic><topic>Water consumption</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Levasseur, Guillaume</creatorcontrib><creatorcontrib>Tang, Kejia</creatorcontrib><creatorcontrib>Bersini, Hugues</creatorcontrib><collection>CrossRef</collection><jtitle>International journal of human-computer studies</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Levasseur, Guillaume</au><au>Tang, Kejia</au><au>Bersini, Hugues</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition</atitle><jtitle>International journal of human-computer studies</jtitle><date>2025-01</date><risdate>2025</risdate><volume>195</volume><spage>103404</spage><pages>103404-</pages><artnum>103404</artnum><issn>1071-5819</issn><abstract>Data annotation for human activity recognition is a well-known challenge for researchers. In particular, annotation in daily life settings relies on self-reporting tools with unknown accuracy. Speech is a promising interface for activity labeling. In this work, we compare the accuracy of two commercially available tools for annotation: voice diaries and connected buttons. We retrofit the water meters of thirty homes in the USA for infrastructure-mediated sensing. Participants are split into equal groups and receive one of the self-reporting tools. The balanced accuracy metric is transferred from the field of machine learning to the evaluation of the annotation performance. Our results show that connected buttons perform significantly better than the voice diary, with 92% median accuracy and 65% median reporting rate. Using questionnaire answers, we highlight that annotation performance is impacted by habit formation and sentiments toward the annotation tool. The use case for data annotation is to disaggregate water meter data into human activities beyond the point of use. We show that it is feasible with a machine-learning model and the corrected annotations. Finally, we formulate recommendations for the design of studies and intelligent environments around the key ideas of proportionality and immediacy. •Self-reporting of human activities enables to train machine learning algorithms.•Speech is not the best interface to report chores in daily life settings.•Internet-connected buttons prove better with 92% accuracy and 65% reporting rate.•Human activities can be inferred from smart water meter data.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.ijhcs.2024.103404</doi><orcidid>https://orcid.org/0000-0001-5744-8533</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1071-5819
ispartof International journal of human-computer studies, 2025-01, Vol.195, p.103404, Article 103404
issn 1071-5819
language eng
recordid cdi_crossref_primary_10_1016_j_ijhcs_2024_103404
source Elsevier
subjects Activity recognition
Machine learning
Self-reporting
Smart homes
Voice assistants
Water consumption
title Limits of speech in connected homes: Experimental comparison of self-reporting tools for human activity recognition
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-12T00%3A58%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Limits%20of%20speech%20in%20connected%20homes:%20Experimental%20comparison%20of%20self-reporting%20tools%20for%20human%20activity%20recognition&rft.jtitle=International%20journal%20of%20human-computer%20studies&rft.au=Levasseur,%20Guillaume&rft.date=2025-01&rft.volume=195&rft.spage=103404&rft.pages=103404-&rft.artnum=103404&rft.issn=1071-5819&rft_id=info:doi/10.1016/j.ijhcs.2024.103404&rft_dat=%3Celsevier_cross%3ES1071581924001873%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c183t-8f3597fdf9e31dd5e9fc44f1ef5fafa3d6a715914918657afb5645873849e5563%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true