Exploring scoring methods for research studies: Accuracy and variability of visual and automated sleep scoring.
Exploring scoring methods for research studies: Accuracy and variability of visual and automated sleep scoring.
Sleep studies face new challenges in terms of data, objectives and metrics. This requires reappraising the adequacy of existing analysis methods, including scoring methods. Visual and automatic sleep scoring of healthy individuals were compared in terms of reliability (i.e., accuracy and stability) to find a scoring method capable of giving access to the actual data variability without adding exogenous variability. A first dataset (DS1, four recordings) scored by six experts plus an autoscoring algorithm was used to characterize inter-scoring variability. A second dataset (DS2, 88 recordings) scored a few weeks later was used to explore intra-expert variability. Percentage agreements and Conger's kappa were derived from epoch-by-epoch comparisons on pairwise and consensus scorings. On DS1 the number of epochs of agreement decreased when the number of experts increased, ranging from 86% (pairwise) to 69% (all experts). Adding autoscoring to visual scorings changed the kappa value from 0.81 to 0.79. Agreement between expert consensus and autoscoring was 93%. On DS2 the hypothesis of intra-expert variability was supported by a systematic decrease in kappa scores between autoscoring used as reference and each single expert between datasets (.75-.70). Although visual scoring induces inter- and intra-expert variability, autoscoring methods can cope with intra-scorer variability, making them a sensible option to reduce exogenous variability and give access to the endogenous variability in the data.
Chellappa, Sarah
516582b5-3cba-4644-86c9-14c91a4510f2
18 February 2020
Chellappa, Sarah
516582b5-3cba-4644-86c9-14c91a4510f2
Chellappa, Sarah
(2020)
Exploring scoring methods for research studies: Accuracy and variability of visual and automated sleep scoring.
Journal of Sleep Research, 29 (5), [e12994].
(doi:10.1111/jsr.12994).
Abstract
Sleep studies face new challenges in terms of data, objectives and metrics. This requires reappraising the adequacy of existing analysis methods, including scoring methods. Visual and automatic sleep scoring of healthy individuals were compared in terms of reliability (i.e., accuracy and stability) to find a scoring method capable of giving access to the actual data variability without adding exogenous variability. A first dataset (DS1, four recordings) scored by six experts plus an autoscoring algorithm was used to characterize inter-scoring variability. A second dataset (DS2, 88 recordings) scored a few weeks later was used to explore intra-expert variability. Percentage agreements and Conger's kappa were derived from epoch-by-epoch comparisons on pairwise and consensus scorings. On DS1 the number of epochs of agreement decreased when the number of experts increased, ranging from 86% (pairwise) to 69% (all experts). Adding autoscoring to visual scorings changed the kappa value from 0.81 to 0.79. Agreement between expert consensus and autoscoring was 93%. On DS2 the hypothesis of intra-expert variability was supported by a systematic decrease in kappa scores between autoscoring used as reference and each single expert between datasets (.75-.70). Although visual scoring induces inter- and intra-expert variability, autoscoring methods can cope with intra-scorer variability, making them a sensible option to reduce exogenous variability and give access to the endogenous variability in the data.
This record has no associated files available for download.
More information
Accepted/In Press date: 20 January 2020
Published date: 18 February 2020
Identifiers
Local EPrints ID: 479565
URI: http://eprints.soton.ac.uk/id/eprint/479565
ISSN: 0962-1105
PURE UUID: 87a77873-c156-4b00-af2d-e47d2945a677
Catalogue record
Date deposited: 26 Jul 2023 16:37
Last modified: 17 Mar 2024 04:20
Export record
Altmetrics
Contributors
Author:
Sarah Chellappa
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics