Loading…
0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis
Abstract Introduction Visual sleep scoring (VS) is affected by inter-expert (difference in scoring between several scorers working on the same recording) and intra-expert variability (evolution in the way to score of a given expert when compared with a reference). Our aim was to quantify inter and i...
Saved in:
Published in: | Sleep (New York, N.Y.) N.Y.), 2018-04, Vol.41 (suppl_1), p.A121-A121 |
---|---|
Main Authors: | , , , , , , , , , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c1944-391c7503a6e8aa31df08e4d38847af3257f9ecb604827d6108baef5efac342083 |
---|---|
cites | |
container_end_page | A121 |
container_issue | suppl_1 |
container_start_page | A121 |
container_title | Sleep (New York, N.Y.) |
container_volume | 41 |
creator | Muto, V Berthomier, C Schmidt, C Vandewalle, G Jaspar, M Devillers, J Chellappa, S Meyer, C Phillips, C Berthomier, P Prado, J Benoit, O Brandewinder, M Mattout, J Maquet, P |
description | Abstract
Introduction
Visual sleep scoring (VS) is affected by inter-expert (difference in scoring between several scorers working on the same recording) and intra-expert variability (evolution in the way to score of a given expert when compared with a reference). Our aim was to quantify inter and intra-expert sleep scoring variability in a group of 6 experts -working at the same sleep center and trained to homogenize their sleep scoring- by using the validated automatic scoring (AS) algorithm ASEEGA, which is fully reproducible by design, as a reference.
Methods
Data were collected in 24 healthy young male participants (mean age 21.6 ± 2.5 years). 4 recordings (data set 1, DS1) were scored by the 6 experts (24 visual scorings) according to the AASM criteria, and by AS, which is based on the analysis of the single EEG channel Cz-Pz. Other 88 recordings (DS2) were scored a few weeks later by the same experts (88 visual scorings) and AS. The epoch-by-epoch agreements (concordance and Cohen kappa coefficient) were computed between all VS, and between VS and AS.
Results
Inter-expert agreement on DS1 decreased as the number of experts increased, from 86% for mean pairwise agreement down to 69% for all 6 experts. Adding AS to the pool of experts barely changed the kappa value, from 0.81 to 0.79. A systematic decrease of the agreements was observed between AS and each single expert between DS1 and DS2 (-3.7% on average).
Conclusion
Inter-expert differences are not restricted to a small proportion of specific epochs that are difficult to score, even when the expert team is very homogeneous. Intra-expert variability is highlighted by the systematic agreement decrease across datasets, and can be interpreted as a scoring drift over time. Even if autoscoring neither provides any ground truth, nor can improve the inter-scorer agreement, it can efficiently cope with the intra-scorer variability, when the AS used is perfectly reproducible and largely insensitive to experimental conditions. These properties are mandatory when dealing with large dataset, making autoscoring methods a sensible option.
Support (If Any)
None. |
doi_str_mv | 10.1093/sleep/zsy061.314 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2036668069</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><oup_id>10.1093/sleep/zsy061.314</oup_id><sourcerecordid>2036668069</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1944-391c7503a6e8aa31df08e4d38847af3257f9ecb604827d6108baef5efac342083</originalsourceid><addsrcrecordid>eNqFUD1PwzAQtRBIlMLOaIkRpT3HjuOwlYqPSpUYCl0jN7kglzQOdiIIvx63ZWe6e3ofd3qEXDOYMMj41NeI7fTHDyDZhDNxQkYsSSDKAntKRsAkixSD5JxceL-FgEXGR-QDOEvoounQRXTWlPvV6Qi_W3QdXWtn9MbUphsCQVf7G3RVWGea9zs6t7s2CLxt6D12X4gNXRvf6_oQNOs7u9OdKQLS9eCNvyRnla49Xv3NMXl7fHidP0fLl6fFfLaMCpYJEfGMFWkCXEtUWnNWVqBQlFwpkeqKx0laZVhsJAgVp6VkoDYaqwQrXXARg-JjcnPMbZ397NF3-db2Ljzh8xi4lFKBzIIKjqrCWe8dVnnrzE67IWeQ7yvND5Xmx0rzUGmw3B4ttm__V_8CsAF5tA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2036668069</pqid></control><display><type>article</type><title>0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis</title><source>Oxford Journals Online</source><source>Alma/SFX Local Collection</source><creator>Muto, V ; Berthomier, C ; Schmidt, C ; Vandewalle, G ; Jaspar, M ; Devillers, J ; Chellappa, S ; Meyer, C ; Phillips, C ; Berthomier, P ; Prado, J ; Benoit, O ; Brandewinder, M ; Mattout, J ; Maquet, P</creator><creatorcontrib>Muto, V ; Berthomier, C ; Schmidt, C ; Vandewalle, G ; Jaspar, M ; Devillers, J ; Chellappa, S ; Meyer, C ; Phillips, C ; Berthomier, P ; Prado, J ; Benoit, O ; Brandewinder, M ; Mattout, J ; Maquet, P</creatorcontrib><description>Abstract
Introduction
Visual sleep scoring (VS) is affected by inter-expert (difference in scoring between several scorers working on the same recording) and intra-expert variability (evolution in the way to score of a given expert when compared with a reference). Our aim was to quantify inter and intra-expert sleep scoring variability in a group of 6 experts -working at the same sleep center and trained to homogenize their sleep scoring- by using the validated automatic scoring (AS) algorithm ASEEGA, which is fully reproducible by design, as a reference.
Methods
Data were collected in 24 healthy young male participants (mean age 21.6 ± 2.5 years). 4 recordings (data set 1, DS1) were scored by the 6 experts (24 visual scorings) according to the AASM criteria, and by AS, which is based on the analysis of the single EEG channel Cz-Pz. Other 88 recordings (DS2) were scored a few weeks later by the same experts (88 visual scorings) and AS. The epoch-by-epoch agreements (concordance and Cohen kappa coefficient) were computed between all VS, and between VS and AS.
Results
Inter-expert agreement on DS1 decreased as the number of experts increased, from 86% for mean pairwise agreement down to 69% for all 6 experts. Adding AS to the pool of experts barely changed the kappa value, from 0.81 to 0.79. A systematic decrease of the agreements was observed between AS and each single expert between DS1 and DS2 (-3.7% on average).
Conclusion
Inter-expert differences are not restricted to a small proportion of specific epochs that are difficult to score, even when the expert team is very homogeneous. Intra-expert variability is highlighted by the systematic agreement decrease across datasets, and can be interpreted as a scoring drift over time. Even if autoscoring neither provides any ground truth, nor can improve the inter-scorer agreement, it can efficiently cope with the intra-scorer variability, when the AS used is perfectly reproducible and largely insensitive to experimental conditions. These properties are mandatory when dealing with large dataset, making autoscoring methods a sensible option.
Support (If Any)
None.</description><identifier>ISSN: 0161-8105</identifier><identifier>EISSN: 1550-9109</identifier><identifier>DOI: 10.1093/sleep/zsy061.314</identifier><language>eng</language><publisher>US: Oxford University Press</publisher><subject>Agreements ; Experts ; Sleep</subject><ispartof>Sleep (New York, N.Y.), 2018-04, Vol.41 (suppl_1), p.A121-A121</ispartof><rights>Sleep Research Society 2018. Published by Oxford University Press [on behalf of the Sleep Research Society]. All rights reserved. For permissions, please email: journals.permissions@oup.com 2018</rights><rights>Copyright © 2018 Sleep Research Society</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c1944-391c7503a6e8aa31df08e4d38847af3257f9ecb604827d6108baef5efac342083</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27903,27904</link.rule.ids></links><search><creatorcontrib>Muto, V</creatorcontrib><creatorcontrib>Berthomier, C</creatorcontrib><creatorcontrib>Schmidt, C</creatorcontrib><creatorcontrib>Vandewalle, G</creatorcontrib><creatorcontrib>Jaspar, M</creatorcontrib><creatorcontrib>Devillers, J</creatorcontrib><creatorcontrib>Chellappa, S</creatorcontrib><creatorcontrib>Meyer, C</creatorcontrib><creatorcontrib>Phillips, C</creatorcontrib><creatorcontrib>Berthomier, P</creatorcontrib><creatorcontrib>Prado, J</creatorcontrib><creatorcontrib>Benoit, O</creatorcontrib><creatorcontrib>Brandewinder, M</creatorcontrib><creatorcontrib>Mattout, J</creatorcontrib><creatorcontrib>Maquet, P</creatorcontrib><title>0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis</title><title>Sleep (New York, N.Y.)</title><description>Abstract
Introduction
Visual sleep scoring (VS) is affected by inter-expert (difference in scoring between several scorers working on the same recording) and intra-expert variability (evolution in the way to score of a given expert when compared with a reference). Our aim was to quantify inter and intra-expert sleep scoring variability in a group of 6 experts -working at the same sleep center and trained to homogenize their sleep scoring- by using the validated automatic scoring (AS) algorithm ASEEGA, which is fully reproducible by design, as a reference.
Methods
Data were collected in 24 healthy young male participants (mean age 21.6 ± 2.5 years). 4 recordings (data set 1, DS1) were scored by the 6 experts (24 visual scorings) according to the AASM criteria, and by AS, which is based on the analysis of the single EEG channel Cz-Pz. Other 88 recordings (DS2) were scored a few weeks later by the same experts (88 visual scorings) and AS. The epoch-by-epoch agreements (concordance and Cohen kappa coefficient) were computed between all VS, and between VS and AS.
Results
Inter-expert agreement on DS1 decreased as the number of experts increased, from 86% for mean pairwise agreement down to 69% for all 6 experts. Adding AS to the pool of experts barely changed the kappa value, from 0.81 to 0.79. A systematic decrease of the agreements was observed between AS and each single expert between DS1 and DS2 (-3.7% on average).
Conclusion
Inter-expert differences are not restricted to a small proportion of specific epochs that are difficult to score, even when the expert team is very homogeneous. Intra-expert variability is highlighted by the systematic agreement decrease across datasets, and can be interpreted as a scoring drift over time. Even if autoscoring neither provides any ground truth, nor can improve the inter-scorer agreement, it can efficiently cope with the intra-scorer variability, when the AS used is perfectly reproducible and largely insensitive to experimental conditions. These properties are mandatory when dealing with large dataset, making autoscoring methods a sensible option.
Support (If Any)
None.</description><subject>Agreements</subject><subject>Experts</subject><subject>Sleep</subject><issn>0161-8105</issn><issn>1550-9109</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><recordid>eNqFUD1PwzAQtRBIlMLOaIkRpT3HjuOwlYqPSpUYCl0jN7kglzQOdiIIvx63ZWe6e3ofd3qEXDOYMMj41NeI7fTHDyDZhDNxQkYsSSDKAntKRsAkixSD5JxceL-FgEXGR-QDOEvoounQRXTWlPvV6Qi_W3QdXWtn9MbUphsCQVf7G3RVWGea9zs6t7s2CLxt6D12X4gNXRvf6_oQNOs7u9OdKQLS9eCNvyRnla49Xv3NMXl7fHidP0fLl6fFfLaMCpYJEfGMFWkCXEtUWnNWVqBQlFwpkeqKx0laZVhsJAgVp6VkoDYaqwQrXXARg-JjcnPMbZ397NF3-db2Ljzh8xi4lFKBzIIKjqrCWe8dVnnrzE67IWeQ7yvND5Xmx0rzUGmw3B4ttm__V_8CsAF5tA</recordid><startdate>20180427</startdate><enddate>20180427</enddate><creator>Muto, V</creator><creator>Berthomier, C</creator><creator>Schmidt, C</creator><creator>Vandewalle, G</creator><creator>Jaspar, M</creator><creator>Devillers, J</creator><creator>Chellappa, S</creator><creator>Meyer, C</creator><creator>Phillips, C</creator><creator>Berthomier, P</creator><creator>Prado, J</creator><creator>Benoit, O</creator><creator>Brandewinder, M</creator><creator>Mattout, J</creator><creator>Maquet, P</creator><general>Oxford University Press</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7X7</scope><scope>7XB</scope><scope>88E</scope><scope>88G</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>K9.</scope><scope>M0S</scope><scope>M1P</scope><scope>M2M</scope><scope>M2O</scope><scope>MBDVC</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PSYQQ</scope><scope>Q9U</scope></search><sort><creationdate>20180427</creationdate><title>0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis</title><author>Muto, V ; Berthomier, C ; Schmidt, C ; Vandewalle, G ; Jaspar, M ; Devillers, J ; Chellappa, S ; Meyer, C ; Phillips, C ; Berthomier, P ; Prado, J ; Benoit, O ; Brandewinder, M ; Mattout, J ; Maquet, P</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1944-391c7503a6e8aa31df08e4d38847af3257f9ecb604827d6108baef5efac342083</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Agreements</topic><topic>Experts</topic><topic>Sleep</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Muto, V</creatorcontrib><creatorcontrib>Berthomier, C</creatorcontrib><creatorcontrib>Schmidt, C</creatorcontrib><creatorcontrib>Vandewalle, G</creatorcontrib><creatorcontrib>Jaspar, M</creatorcontrib><creatorcontrib>Devillers, J</creatorcontrib><creatorcontrib>Chellappa, S</creatorcontrib><creatorcontrib>Meyer, C</creatorcontrib><creatorcontrib>Phillips, C</creatorcontrib><creatorcontrib>Berthomier, P</creatorcontrib><creatorcontrib>Prado, J</creatorcontrib><creatorcontrib>Benoit, O</creatorcontrib><creatorcontrib>Brandewinder, M</creatorcontrib><creatorcontrib>Mattout, J</creatorcontrib><creatorcontrib>Maquet, P</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Health & Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Medical Database (Alumni Edition)</collection><collection>Psychology Database (Alumni)</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>ProQuest Health & Medical Complete (Alumni)</collection><collection>Health & Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>Psychology Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest One Psychology</collection><collection>ProQuest Central Basic</collection><jtitle>Sleep (New York, N.Y.)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Muto, V</au><au>Berthomier, C</au><au>Schmidt, C</au><au>Vandewalle, G</au><au>Jaspar, M</au><au>Devillers, J</au><au>Chellappa, S</au><au>Meyer, C</au><au>Phillips, C</au><au>Berthomier, P</au><au>Prado, J</au><au>Benoit, O</au><au>Brandewinder, M</au><au>Mattout, J</au><au>Maquet, P</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis</atitle><jtitle>Sleep (New York, N.Y.)</jtitle><date>2018-04-27</date><risdate>2018</risdate><volume>41</volume><issue>suppl_1</issue><spage>A121</spage><epage>A121</epage><pages>A121-A121</pages><issn>0161-8105</issn><eissn>1550-9109</eissn><abstract>Abstract
Introduction
Visual sleep scoring (VS) is affected by inter-expert (difference in scoring between several scorers working on the same recording) and intra-expert variability (evolution in the way to score of a given expert when compared with a reference). Our aim was to quantify inter and intra-expert sleep scoring variability in a group of 6 experts -working at the same sleep center and trained to homogenize their sleep scoring- by using the validated automatic scoring (AS) algorithm ASEEGA, which is fully reproducible by design, as a reference.
Methods
Data were collected in 24 healthy young male participants (mean age 21.6 ± 2.5 years). 4 recordings (data set 1, DS1) were scored by the 6 experts (24 visual scorings) according to the AASM criteria, and by AS, which is based on the analysis of the single EEG channel Cz-Pz. Other 88 recordings (DS2) were scored a few weeks later by the same experts (88 visual scorings) and AS. The epoch-by-epoch agreements (concordance and Cohen kappa coefficient) were computed between all VS, and between VS and AS.
Results
Inter-expert agreement on DS1 decreased as the number of experts increased, from 86% for mean pairwise agreement down to 69% for all 6 experts. Adding AS to the pool of experts barely changed the kappa value, from 0.81 to 0.79. A systematic decrease of the agreements was observed between AS and each single expert between DS1 and DS2 (-3.7% on average).
Conclusion
Inter-expert differences are not restricted to a small proportion of specific epochs that are difficult to score, even when the expert team is very homogeneous. Intra-expert variability is highlighted by the systematic agreement decrease across datasets, and can be interpreted as a scoring drift over time. Even if autoscoring neither provides any ground truth, nor can improve the inter-scorer agreement, it can efficiently cope with the intra-scorer variability, when the AS used is perfectly reproducible and largely insensitive to experimental conditions. These properties are mandatory when dealing with large dataset, making autoscoring methods a sensible option.
Support (If Any)
None.</abstract><cop>US</cop><pub>Oxford University Press</pub><doi>10.1093/sleep/zsy061.314</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0161-8105 |
ispartof | Sleep (New York, N.Y.), 2018-04, Vol.41 (suppl_1), p.A121-A121 |
issn | 0161-8105 1550-9109 |
language | eng |
recordid | cdi_proquest_journals_2036668069 |
source | Oxford Journals Online; Alma/SFX Local Collection |
subjects | Agreements Experts Sleep |
title | 0315 Inter- And Intra-expert Variability In Sleep Scoring: Comparison Between Visual And Automatic Analysis |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-26T18%3A22%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=0315%20Inter-%20And%20Intra-expert%20Variability%20In%20Sleep%20Scoring:%20Comparison%20Between%20Visual%20And%20Automatic%20Analysis&rft.jtitle=Sleep%20(New%20York,%20N.Y.)&rft.au=Muto,%20V&rft.date=2018-04-27&rft.volume=41&rft.issue=suppl_1&rft.spage=A121&rft.epage=A121&rft.pages=A121-A121&rft.issn=0161-8105&rft.eissn=1550-9109&rft_id=info:doi/10.1093/sleep/zsy061.314&rft_dat=%3Cproquest_cross%3E2036668069%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c1944-391c7503a6e8aa31df08e4d38847af3257f9ecb604827d6108baef5efac342083%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2036668069&rft_id=info:pmid/&rft_oup_id=10.1093/sleep/zsy061.314&rfr_iscdi=true |