RESUMO
The ability of remote research tools to collect granular, high-frequency data on symptoms and digital biomarkers is an important strength because it circumvents many limitations of traditional clinical trials and improves the ability to capture clinically relevant data. This approach allows researchers to capture more robust baselines and derive novel phenotypes for improved precision in diagnosis and accuracy in outcomes. The process for developing these tools however is complex because data need to be collected at a frequency that is meaningful but not burdensome for the participant or patient. Furthermore, traditional techniques, which rely on fixed conditions to validate assessments, may be inappropriate for validating tools that are designed to capture data under flexible conditions. This paper discusses the process for determining whether a digital assessment is suitable for remote research and offers suggestions on how to validate these novel tools.
RESUMO
BACKGROUND: Computerized assessments are already used to derive accurate and reliable measures of cognitive function. Web-based cognitive assessment could improve the accessibility and flexibility of research and clinical assessment, widen participation, and promote research recruitment while simultaneously reducing costs. However, differences in context may influence task performance. OBJECTIVE: This study aims to determine the comparability of an unsupervised, web-based administration of the Cambridge Neuropsychological Test Automated Battery (CANTAB) against a typical in-person lab-based assessment, using a within-subjects counterbalanced design. The study aims to test (1) reliability, quantifying the relationship between measurements across settings using correlational approaches; (2) equivalence, the extent to which test results in different settings produce similar overall results; and (3) agreement, by quantifying acceptable limits to bias and differences between measurement environments. METHODS: A total of 51 healthy adults (32 women and 19 men; mean age 36.8, SD 15.6 years) completed 2 testing sessions, which were completed on average 1 week apart (SD 4.5 days). Assessments included equivalent tests of emotion recognition (emotion recognition task [ERT]), visual recognition (pattern recognition memory [PRM]), episodic memory (paired associate learning [PAL]), working memory and spatial planning (spatial working memory [SWM] and one touch stockings of Cambridge), and sustained attention (rapid visual information processing [RVP]). Participants were randomly allocated to one of the two groups, either assessed in-person in the laboratory first (n=33) or with unsupervised web-based assessments on their personal computing systems first (n=18). Performance indices (errors, correct trials, and response sensitivity) and median reaction times were extracted. Intraclass and bivariate correlations examined intersetting reliability, linear mixed models and Bayesian paired sample t tests tested for equivalence, and Bland-Altman plots examined agreement. RESULTS: Intraclass correlation (ICC) coefficients ranged from ρ=0.23-0.67, with high correlations in 3 performance indices (from PAL, SWM, and RVP tasks; ρ≥0.60). High ICC values were also seen for reaction time measures from 2 tasks (PRM and ERT tasks; ρ≥0.60). However, reaction times were slower during web-based assessments, which undermined both equivalence and agreement for reaction time measures. Performance indices did not differ between assessment settings and generally showed satisfactory agreement. CONCLUSIONS: Our findings support the comparability of CANTAB performance indices (errors, correct trials, and response sensitivity) in unsupervised, web-based assessments with in-person and laboratory tests. Reaction times are not as easily translatable from in-person to web-based testing, likely due to variations in computer hardware. The results underline the importance of examining more than one index to ascertain comparability, as high correlations can present in the context of systematic differences, which are a product of differences between measurement environments. Further work is now needed to examine web-based assessments in clinical populations and in larger samples to improve sensitivity for detecting subtler differences between test settings.
Assuntos
Cognição/fisiologia , Internet/normas , Laboratórios/normas , Testes Neuropsicológicos/normas , Adulto , Feminino , Humanos , Masculino , Reprodutibilidade dos TestesRESUMO
INTRODUCTION: Normative cognitive data can help to distinguish pathological decline from normal aging. This study presents normative data from the Cambridge Neuropsychological Test Automated Battery, using linear regression and nonlinear quantile regression approaches. METHODS: Heinz Nixdorf Recall study participants completed Cambridge Neuropsychological Test Automated Battery tests: paired-associate learning, spatial working memory, and reaction time. Data were available for 1349-1529 healthy adults aged 57-84 years. Linear and nonlinear quantile regression analyses examined age-related changes, adjusting for sex and education. Quantile regression differentiated seven performance bands (percentiles: 97.7, 93.3, 84.1, 50, 15.9, 6.7, and 2.3). RESULTS: Normative data show age-related cognitive decline across all tests, but with quantile regression revealing heterogeneous trajectories of cognitive aging, particularly for the test of episodic memory function (paired-associate learning). DISCUSSION: This study presents normative data from Cambridge Neuropsychological Test Automated Battery in mid-to-late life. Quantile regression can model heterogeneity in age-related cognitive trajectories as seen in the paired-associate learning episodic memory measure.