Volume 15, Issue 2 e12420
RESEARCH ARTICLE
Open Access

Feasibility of an online consensus approach for the diagnosis of cognitive impairment and dementia in rural South Africa

Darina T. Bassil

Corresponding Author

Darina T. Bassil

Harvard Center for Population and Development Studies, Harvard University, Cambridge, Massachusetts, USA

Correspondence

Darina T. Bassil, Harvard Center for Population and Development Studies, Harvard University, 9 Bow Street, Cambridge, MA 02138, USA.

E-mail: [email protected]

Search for more papers by this author
Meagan T. Farrell

Meagan T. Farrell

Harvard Center for Population and Development Studies, Harvard University, Cambridge, Massachusetts, USA

Search for more papers by this author
Albert Weerman

Albert Weerman

Center for Economic and Social Research, University of Southern California, Los Angeles, California, USA

Search for more papers by this author
Muqi Guo

Muqi Guo

Harvard Center for Population and Development Studies, Harvard University, Cambridge, Massachusetts, USA

Search for more papers by this author
Ryan G. Wagner

Ryan G. Wagner

MRC/Wits Rural Public Health and Health Transitions Research Unit (Agincourt), School of Public Health, Faculty of Health Sciences, University of the Witwatersrand, Johannesburg, South Africa

Search for more papers by this author
Adam M. Brickman

Adam M. Brickman

Taub Institute for Research on Alzheimer's Disease and the Aging Brain, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Gertrude H. Sergievsky Center, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Department of Neurology, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Search for more papers by this author
M. Maria Glymour

M. Maria Glymour

Department of Epidemiology and Biostatistics, University of California, San Francisco, California, USA

Search for more papers by this author
Kenneth M. Langa

Kenneth M. Langa

Institute for Healthcare Policy and Innovation, University of Michigan, Ann Arbor, Michigan, USA

Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, Michigan, USA

Institute for Social Research, University of Michigan, Ann Arbor, Michigan, USA

Veterans Affairs Center for Clinical Management Research, Ann Arbor, Michigan, USA

Search for more papers by this author
Jennifer J. Manly

Jennifer J. Manly

Taub Institute for Research on Alzheimer's Disease and the Aging Brain, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Gertrude H. Sergievsky Center, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Department of Neurology, Vagelos College of Physicians and Surgeons, Columbia University, New York, New York, USA

Search for more papers by this author
Brent Tipping

Brent Tipping

Division of Geriatric Medicine, School of Clinical Medicine, Faculty of Health Sciences, University of the Witwatersrand, Johannesburg, South Africa

Search for more papers by this author
India Butler

India Butler

Division of Geriatric Medicine, School of Clinical Medicine, Faculty of Health Sciences, University of the Witwatersrand, Johannesburg, South Africa

Search for more papers by this author
Stephen Tollman

Stephen Tollman

MRC/Wits Rural Public Health and Health Transitions Research Unit (Agincourt), School of Public Health, Faculty of Health Sciences, University of the Witwatersrand, Johannesburg, South Africa

Search for more papers by this author
Lisa F. Berkman

Lisa F. Berkman

Harvard Center for Population and Development Studies, Harvard University, Cambridge, Massachusetts, USA

Department of Social and Behavioral Sciences, Harvard T.H. Chan School of Public Health, Boston, Massachusetts, USA

Search for more papers by this author
First published: 04 April 2023
Citations: 1

Abstract

INTRODUCTION

We describe the development and feasibility of using an online consensus approach for diagnosing cognitive impairment and dementia in rural South Africa.

METHODS

Cognitive assessments, clinical evaluations, and informant interviews from Cognition and Dementia in the Health and Aging in Africa Longitudinal Study (HAALSI Dementia) were reviewed by an expert panel using a web-based platform to assign a diagnosis of cognitively normal, mild cognitive impairment (MCI), or dementia.

RESULTS

Six hundred thirty-five participants were assigned a final diagnostic category, with 298 requiring adjudication conference calls. Overall agreement between each rater's independent diagnosis and final diagnosis (via the portal or consensus conference) was 78.3%. A moderate level of agreement between raters’ individual ratings and the final diagnostic outcomes was observed (average κ coefficient = 0.50).

DISCUSSION

Findings show initial feasibility in using an online consensus approach for the diagnosis of cognitive impairment and dementia in remote, rural, and low-resource settings.

1 INTRODUCTION

Valid and reliable assessments of dementia are essential to advance research, yet difficult to implement in many settings in which research is most urgently needed. To improve dementia diagnostic reliability and accuracy, many research studies recommend a consensus panel approach for the diagnosis of dementia.1-4 In this approach, a panel of expert clinicians meet to review and discuss a study participant's cognitive and clinical profile to adjudicate the diagnosis of dementia,3, 5 using standardized criteria such as the National Institute on Aging–Alzheimer's Association (NIA-AA)6 diagnostic guidelines for Alzheimer's disease and the Diagnostic and Statistical Manual of Mental Disorders, Fifth edition (DSM-5).7 Consensus panels could be especially useful in low- and middle-income countries (LMICs) where population-based research studies using detailed neuropsychological tests remain scarce, with limited to no availability of comprehensive normative population data.8-10 However, in-person consensus conferences can be very costly, time consuming, and impractical, especially in remote and rural areas. Few studies have evaluated the feasibility, validity, and reliability of using a web-based consensus approach for the diagnosis of cognitive impairment and dementia.3, 11

In this paper, we use data from our Cognition and Dementia in the Health and Aging in Africa Longitudinal Study of an INDEPTH community in South Africa (HAALSI Dementia),12 a harmonized sister study to the United States Health and Retirement Study (HRS) Harmonized Cognitive Assessment Protocol (HCAP) Project,13 and implement a rigorous and robust web-based online consensus system for remote capture of diagnostic cognitive outcomes. We built our web interface based on similar systems deployed in the Longitudinal Aging Study in India–Diagnostic Assessment of Dementia (LASI-DAD)11 and the Monogahela–Youghiogheny Health Aging Team study (MYHAT).3 We first describe detailed information on the web interface design, diagnostic criteria for dementia ascertainment, and the consensus expert panel. We also report baseline characteristics and distribution of final diagnostic outcomes for our cohort, assess the feasibility of using a multidisciplinary web-based consensus conference approach in rural South Africa, and identify the key factors responsible for diagnostic variability among raters. A successful web-based consensus panel could be a novel, efficient, and less expensive approach to facilitate the diagnosis of cognitive impairment and dementia, especially in LMICs.

2 METHODS

2.1 Study cohort

HAALSI Dementia is an ongoing prospective cohort study on a stratified subsample of 635 HAALSI participants aged 50 and older in Agincourt, a rural South African community. The study design and methods have been previously described.12 In brief, HAALSI Dementia collects detailed neuropsychological and functional assessments, informant interviews, and neurological and clinical evaluations to enable cross-cultural comparison and cross-calibration with international HCAP and HRS studies. Furthermore, HAALSI Dementia is strategically oversampled for respondents categorized as higher risk for dementia to allow future estimation of dementia prevalence in the parent HAALSI cohort.

The first wave of data collection was completed in January 2020. Data were reviewed by an expert consensus panel to assign diagnosis of cognitively normal, mild cognitive impairment (MCI), or dementia, based on the NIA-AA criteria.6

RESEARCH IN CONTEXT

  1. Systematic Review: We reviewed results from a PubMed search and found that little is currently known about the validity of using online consensus approaches for the diagnosis of cognitive impairment and dementia. Building on the online consensus platform developed by the Longitudinal Aging Study in India–Diagnostic Assessment of Dementia (LASI-DAD) team, we developed and implemented an online clinical consensus approach for the Cognition and Dementia in the Health and Aging in Africa Longitudinal Study (HAALSI Dementia) cohort in rural South Africa.
  2. Interpretation: Our findings demonstrate the feasibility of using an online consensus approach for the diagnosis of cognitive impairment and dementia in a rural South African population.
  3. Future Directions: To further assess the reliability and validity of this web-based approach, our next steps include comparative validation analysis using data from ongoing ancillary magnetic resonance imaging and biomarker substudies.

This study was approved by the University of the Witwatersrand Human Research Ethics Committee (ref.M190443), Harvard T.H. Chan School of Public Health Institutional Research Ethics Board (ref. 18-1459,19-1396), and Mpumalanga Provincial Research and Ethics Committee.

2.2 Online consensus interface: design

Our online consensus website was modeled after LASI-DAD11 and MYHAT3 web portals. Both studies showed the online consensus system to be an efficient, feasible, and valid approach for the clinical diagnosis of dementia. Following the same format and structure as LASI-DAD, our web interface entailed two main pages: assessments and ratings.

Raters independently reviewed all the data on the assessments page, including demographic, cognitive, neurological, and informant measures (Figure S1 in supporting information). Next, raters navigated to the ratings page and worked through the NIA-AA criteria to submit their final diagnosis (Figure S2 in supporting information).

Research participants were anonymous and identified only by a unique case ID. Raters were provided with unique rater IDs to access the web portal securely. The web portal randomly selected and assigned cases for each rater to review. Cases were reviewed one at a time, and any selections made by raters on the web page were autosaved. All recorded data from the site were autogenerated in a database and securely exported to STATA for data analysis and management.

2.2.1 Assessments: cognitive battery, informant interview, and clinical examination

All assessments were administered and captured via tablets using Computer-Assisted Personal Interviewing (CAPI) software. The respondents’ demographic information (age, sex, education) and limited health history were made available on the website. Any narrative text or remarks recorded by the nurses or field workers during visits were translated and displayed on the page. Data on administered cognitive measures, informant scales, and neurological findings were cleaned and managed in STATA and imported securely to the online assessments page (Figure 1). Full details on individual cognitive, informant, and neurological measures have been published.12

Details are in the caption following the image
Flow of HAALSI Dementia activities leading to web-based expert panel diagnoses. HAALSI, Health and Aging in Africa Longitudinal Study

HAALSI Dementia cognitive measures and scores were shown under the respondent's assessments column and were categorized by their principal domain as follows: (1) general cognitive status: Mini-Mental State Examination (MMSE),14 days of the week,15 Telephone Interview for Cognitive Status (TICS),16 Community Screening Inventory for Dementia (CSID),17 and Clinical Dementia Rating (CDR) semi-structured interview;18 (2) episodic memory: immediate and delayed word recall,19 word list recognition,19 logical memory (story recall) immediate,20 delayed and recognition;20 (3) executive function and attention: motor sequences,21 go no/go,22 similarities and differences,23 and Raven's standard Progressive Matrices;24 (4) visuospatial/spatial memory: constructional praxis,19 constructional praxis recall,19 spatial working memory,25 symbol cancellation26 and clock draw;27 (5) language: Boston Naming Test,28 semantic fluency,29 Token Test,30 and phoneme (letter) fluency.31 The respondents’ assessments column also included a “mood” subheading, which outlined the respondent's self-reported measure of depression based on the Center for Epidemiologic Studies Depression (CES-D) scale.32

In the absence of normative population data for our cohort and in an effort to aid the raters with their diagnostic decisions, the following items were also shown for each cognitive measure: (1) weighted means to provide a crude estimate of the participants’ performance relative to the average performance of individuals aged 50 and over across the parent HAALSI cohort, (2) percent of the sample missing the cognitive test, and (3) percent of the sample who performed at minimum and maximum score range. To assess cognitive decline over time, raters were also shown cognitive scores from the word recall task administered during parent HAALSI waves 1 (2014–2015) and 2 (2018–2019). For cognitive tasks that required participants to draw or write on the tablets (constructional praxis, clock draw, and MMSE), raters were able to see and evaluate the actual drawings completed by participants (Figure S3 in supporting information).

As part of the cognitive assessment, respondents also completed a reading assessment to assess basic literacy (letter and number recognition) through secondary level reading skills. Literacy scores were displayed alongside the self-reported education level to assist raters with the interpretation of cognitive test results.

The informant section summarized key demographic and background information (relationship to respondent, years known, frequency informant sees respondent) and presented all informant scales including Informant Questionnaire on Cognitive Decline in the Elderly (IQCODE),33 Blessed Dementia Rating Scale,34 CSI-D Cognitive Activities Questionnaire,17 10/66 Dementia Research Group Informant Questionnaire,35 HRS Activities Questionnaire,13 and Clinical Dementia Rating (CDR) semi-structured interview.18 To help raters navigate through the informant data, different elements from the aforementioned scales were organized under the following headings: (1) respondent's mental status for information on changes in respondent's cognitive abilities, judgment, problem solving, and personal care activities; (2) respondent's activities for changes in Activities of Daily Living (ADLs), Instrumental ADLs (IADLs), leisure activities, as well as a general overview of respondent's social engagement and community affairs; and (3) respondent's behavior for any behavioral and mood changes reported by the informant (Figure S1).

All clinical data were reported under the health history column on the assessments page. Clinical findings were summarized as normal/abnormal for each of the following evaluated neurological exam components: face (muscle weakness, swallow, and tongue mobility), speech, upper and lower limbs mobility, balance, gait, vision, hearing, and any other observations noted by the nurse. Raters were also able to click on any of the neurological exam components for a more detailed overview of clinical findings. Other reported information included blood pressure, pulse, grip strength, and timed up and go readings as well as information on relevant medical history and medications (Figure S1).

2.2.2 Ratings: diagnostic criteria for dementia ascertainment

The assessments page was followed by the ratings page, which showed a series of questions to determine the presence and severity of dementia. The primary diagnostic outcome was determined using the NIA-AA criteria,6 whereby raters had to answer yes/no to impairment for each NIA-AA criteria item (Figure S2). Raters could move freely between the assessments and ratings pages during their review, but were not allowed to move to another case before submitting their rating.

Once the rater responded to all criteria items, the web portal applied the responses to a preprogrammed algorithm designed by our team, to automatically generate the final diagnosis. Using the NIA-AA algorithm, diagnostic outcomes were developed based on the following rules:
  • Cognitively normal: No evidence of cognitive or functional impairment based on objective testing or informant report.
  • MCI: Evidence of cognitive impairment according to objective testing or informant report, but not meeting criteria for dementia, as described in the published work from the Aging, Demographics, and Memory Study (ADAMs) research team.36-39
  • Dementia: Evidence of cognitive impairment according to both the objective tests and informant report and additionally, the participant had (1) evidence that the cognitive impairment interfered with social and/or occupational function, (2) had experienced cognitive decline based on previous parent HAALSI waves, (3) had cognitive impairments that could not be explained by delirium or psychiatric disorders, and (4) had cognitive impairments in at least two cognitive domains.

Prior to submitting the generated final diagnosis for each case, raters additionally: (1) provided an overall CDR score (0, 0.5, 1, 2, 3) to rate the severity of cognitive and functional impairment, (2) rated the presence of aphasia, agnosia, and apraxia to enable diagnostic comparisons to other studies which used the DSM-5 criteria for dementia diagnosis, and (3) provided their level of confidence in the rating (not at all certain, only slightly certain, reasonably certain, very little doubt, absolutely certain).

2.3 Consensus panel

The raters panel consisted of US and South African clinicians with expertise in evaluating individuals with possible cognitive impairment or dementia. Experts on the panel included neurologists, geriatricians, neuropsychologists, as well as researchers with specialized training in the field of dementia. Selected raters had expertise in the clinical and research settings of cognition and dementia, in diverse populations, as well as experience with the neuropsychological assessments and diagnostic criteria used in the study.

All raters completed an online training session on how to navigate the web portal and submit ratings. During training, raters were also required to independently submit ratings for a number of pilot cases and provide feedback on their experience with the portal. Revisions to the web portal were made during the training phase only. Additionally, a detailed manual was provided to all raters, including guidance on the consensus portal, detailed descriptions of cognitive tests and other measures, and an overview of the diagnostic criteria used.

For the online consensus review, each case was reviewed by three raters, randomly chosen at the time of review. Cases were completed at raters’ own pace, so some raters completed more cases than others. All three raters needed to agree on the final NIA-AA diagnosis (cognitively normal, MCI, or dementia). Cases for which there was disagreement on the final diagnostic category among the three independent reviewers were discussed during biweekly consensus conference calls.

Consensus conference calls, which were led by a moderator, were open to all raters on the committee, not just raters who originally reviewed the case, and typically had between three to six raters in attendance. For each discrepant case, the moderator presented all relevant information including the neuropsychological testing profile, informant interview, and clinical observations from the neurological exam to the consensus panel. Members of the consensus panel were also able to log in and see the details of the case currently under discussion. The moderator then focused the discussion on the specific item(s) that led to disagreement among the three reviewers and allowed all consensus members to discuss the case and agree on a final diagnosis (Figure 1).

2.4 Statistical analyses

Distributions of diagnostic outcomes (cognitively normal, MCI, dementia) by age and sex were summarized with frequency counts and percentages. Differences in cognitive measures between diagnostic outcomes were assessed using chi-square tests and one-way analyses of variance (ANOVA), where appropriate. Overall concordance rate was measured as overall agreement between each rater's initial diagnoses (the ratings first entered in the online portal) divided by the final diagnoses (based on consensus agreement via the portal or consensus conference).

An unweighted Cohen's kappa coefficient (κ) was calculated to account for agreement expected to occur by chance alone.40, 41 Cohen's kappa coefficient (κ) was also used to assess the level of agreement among raters by key NIA-AA diagnostic criterions such as impairment based on objective testing and/or informant report, and impairment in social and occupational function, as well as by NIA-AA diagnostic cognitive domains (memory, executive, language, visuospatial functions, and behavioral changes).

Kappa statistics were interpreted using the Landis and Koch method in which strength of agreement was defined as poor (< 0.00), slight (0.00–0.20), fair (0.21–0.40), moderate (0.41–0.60), substantial (0.61–0.80), or almost perfect (0.81–1.00).42

Concordance between rater's diagnosis and final diagnostic outcomes were also compared by rater specialty. Specifically, using the final consensus diagnosis as the gold standard outcome, we coded the diagnosis rating as 1 “cognitively normal,” 2 “MCI,” or 3 “dementia” and calculated the average difference score between each rater's assigned rating and the final consensus rating.

All analyses were performed using STATA 16 (StataCorp) software by authors M.G. and D.T.B.

3 RESULTS

3.1 Participants’ diagnostic outcomes, baseline characteristics and cognitive profile

A total of 635 cases were reviewed by a panel of 11 consensus members (six US and five South African) and assigned a diagnosis of cognitively normal, MCI, or dementia. From July 2020 through October 2021, each case was randomly assigned to three consensus members for independent review. Consensus panel members were compromised of neurologists (n = 5), neuropsychologists (n = 2), geriatricians (n = 2), and dementia researchers (n = 2) with a mean experience of 12 years (range, 6–25 years) and an equal number of men (n = 6) and women (n = 5) on the panel. On average, it took raters between 3 and 5 minutes to review each case, with more complex cases taking approximately 10 to 15 minutes to complete. After the independent individual ratings, 298 cases (47%) showed disagreement in the final diagnosis among the three raters. Discordant cases were resolved during biweekly conference calls and were mostly decisions between cognitively normal and early MCI, or late MCI and dementia.

The mean age in our cohort was 69.87 (standard deviation = 11.55) and the majority of study participants had no formal education (55%). Final diagnostic categories included 314 cognitively normal (49%), 184 MCI (29%), and 137 dementia (22%) cases, which reflects our initial sampling framework to oversample for respondents categorized as higher risk for dementia. The distribution of diagnoses by age and sex is presented in Table 1.

TABLE 1. Distribution of diagnoses (cognitively normal, mild cognitive impairment, dementia) by age and sex.
Diagnosis (N%)
Cognitively normal MCI Dementia Observations
Total 314 (49.5) 184 (29.0) 137 (21.6) 635
Age groups
50–54 41 (91.1) 2 (4.4) 2 (4.4) 45
55–59 80 (75.5) 20 (18.9) 6 (5.7) 106
60–64 55 (64.7) 23 (27.1) 7 (8.2) 85
65–69 51 (57.3) 23 (25.8) 15 (16.9) 89
70–74 38 (46.3) 28 (34.2) 16 (19.5) 82
75+ 49 (21.5) 88 (38.6) 91 (39.9) 228
Sex
Female 182 (46.6) 119 (30.4) 90 (23.0) 391
Male 132 (54.1) 65 (26.6) 47 (19.3) 244
  • Abbreviation: MCI, mild cognitive impairment.

We also compared mean cognitive scores by diagnostic group for 631 participants (four participants were missing the cognitive battery). Overall, participants diagnosed with MCI and dementia performed worse on all cognitive measures (Table 2). Differences in mean cognitive scores between groups were significantly different from zero for all cognitive tests constituting the memory, executive, language, and visuospatial domains (Table 2). Visuospatial tasks, especially those requiring drawing, had high rates of missingness; hence, interpretation of these results may be tenuous.

TABLE 2. Mean score standard deviation for cognitive tests and Informant Questionnaire on Cognitive Decline in the Elderly scale by diagnostic outcomes (cognitively normal, mild cognitive impairment, dementia).

Diagnosis

Cognitively Normal

(n=313)

MCI

(n=183)

Dementia

(n=135)

p-value % missing
TICS 3.2 (0.8) 2.6 (0.8) 2 (0.9) <0.001 0.6
Symbol cancellation 20.3 (12.8) 10.1 (8.2) 8.5 (8.6) <0.001 15.9
CSID 3.9 (0.4) 3.7 (0.6) 3.3 (0.9) <0.001 0.6
Similarities and differences 3.4 (0.7) 3.2 (0.8) 2.8 (1.0) <0.001 0.6
Days of the week 13.2 (1.6) 11.0 (2.7) 8.8 (3.3) <0.001 0.6
Go/No Go 16.0 (3.5) 11.8 (4.6) 8.4 (5.0) <0.001 1.1
Motor sequence 2.8 (0.4) 2.4 (0.7) 1.9 (1.0) <0.001 0.8
Spatial working memory 1.4 (0.9) 1.0 (0.9) 1.0 (0.8) <0.001 6.1
Token test 4.2 (1.4) 3.0 (1.6) 2.5 (1.5) <0.001 0.6
Boston Naming Test 14.4 (1.7) 13.3 (2.8) 11.7 (4.1) <0.001 0.6
Raven's progressive matrices 8.6 (2.5) 6.6 (2.4) 5.6 (2.7) <0.001 5.0
Total immediate word recall 14.7 (3.2) 10.8 (3.8) 8.1 (3.8) <0.001 0.6
Total delayed word recall 4.8 (1.7) 3.0 (1.6) 2.0 (1.5) <0.001 0.6
Word recognition 16.6 (2.4) 14.7 (2.8) 12.9 (3.0) 0.0073 0.6
Logical memory immediate—Story 1 10.3 (3.5) 7.7 (3.8) 5.6 (4.1) <0.001 0.6
Logical memory immediate—Story 2 6.9 (1.7) 5.4 (2.2) 4.1 (2.3) <0.001 0.6
Logical memory delayed—Story 1 8.7 (3.8) 6.0 (3.9) 4.1 (3.6) <0.001 2.4
Logical memory delayed—Story 2 5.9 (2.1) 4.3 (2.4) 3.1 (2.4) <0.001 2.4
Logical memory—Recognition 9.5 (1.7) 8.7 (1.9) 8.1 (1.9) <0.001 0.6
Constructional praxis 6.3 (2.3) 4.3 (2.5) 3.5 (2.1) <0.001 24.4
Constructional praxis recall 3.9 (2.8) 1.9 (2.1) 1.4 (1.8) <0.001 23.9
MMSE 22.2 (2.9) 16.5 (3.5) 12.8 (4.3) <0.001 0.6
Semantic fluency 12.2 (6.8) 10.0 (3.7) 8.7 (3.4) <0.001 0.6
Phoneme fluency 2.9 (3.0) 1.3 (1.9) 0.8 (1.3) <0.001 2.2
CESD total score 10.2 (7.9) 15.3 (9.3) 17.7 (12.2) <0.001 0.6
Clock draw 1.7 (0.8) 1.0 (0.6) 0.9 (0.4) <0.001 43.1
IQCODE 3.1 (0.2) 3.3 (0.3) 3.9 (0.6) <0.001 0.9
  • Abbreviations: CESD, Centers for the Epidemiological Study Depression; CSID, Community Screening Inventory for Dementia; IQCODE, Informant Questionnaire on Cognitive Decline in the Elderly; MCI, mild cognitive impairment; MMSE, Mini-Mental State Examination; SD, standard deviation; TICS, Telephone Interview for Cognitive Status.

3.2 Rater reliability: agreement between individual ratings and final diagnoses

As each case was reviewed by three raters, a total of 1905 individual diagnostic ratings were made. Overall agreement between each rater's initial diagnosis and the final diagnosis (based on the consensus agreement via the portal or consensus conference) was 78.3% (1494/1905). The average κ statistic was 0.50, suggesting a moderate level of agreement between raters’ individual ratings and the final diagnostic outcomes (Table 3). Upon investigating agreement level by diagnostic outcome (at final determination), we observed a higher level of overall rater agreement with cognitively normal (k = 0.61) and dementia (k = 0.57) diagnoses and conversely lower agreement with MCI (k = 0.34). Kappas for individual raters ranged between 0.34 and 0.96, but measurement of inter-rater reliability was not possible due to differences in number of cases reviewed by each rater (Table 3).

TABLE 3. Individual and overall levels of agreement between raters and consensus diagnoses.
Rater Kappa Z Prob>Z Number of cases reviewed
1 0.346 2.88 0.002 37
2 0.553 12.56 <0.001 262
3 0.774 16.38 <0.001 240
4 0.3873 4.34 <0.001 45
5 0.6881 19.81 <0.001 420
6 0.3864 4.37 <0.001 41
7 0.6346 9.17 <0.001 101
8 0.5618 14.06 <0.001 288
9 0.7852 21.07 <0.001 373
10 0.9634 9.03 <0.001 50
11 0.664 6.49 <0.001 46
Cognitively normal 0.6076
MCI 0.3377
Dementia 0.5693
Overall 0.5006
  • Abbreviation: MCI, mild cognitive impairment.

To better understand the reasons for disagreement between raters, we investigated the percent agreement between raters for main NIA-AA diagnostic criterions (cognitive impairment and social and occupational functional impairment) among cases that triggered consensus conference calls. In general, of those cases that were adjudicated between MCI and dementia, we found that discordance among raters was mostly due to difficulty in assessing social and occupational functioning (Figure 2A). Specifically, the three raters almost always agreed on whether the participant had cognitive impairment (91%), but disagreement on the social and occupational functioning criterion was responsible for > 50% of cases sent to adjudication. Another reason for discordance among raters was the absence of normative data; cases between cognitively normal and MCI were sent to adjudication conference calls mainly because of disagreement on the presence of cognitive impairment (>85% of cases; Figure 2A). Upon further exploration of the cognitive impairment criterion by cognitive domain, we found the highest level of disagreement (74%) among raters particularly for tests within the visuospatial domain, followed by the memory domain (70%; Figure 2B).

Details are in the caption following the image
(A) Percent of cases discussed in adjudication conference calls (cognitively normal and MCI, MCI and dementia) due to disagreement in key NIA-AA diagnostic criterions (n = 272). (B) Disagreement in cognitive impairment between raters by NIA-AA cognitive domain for adjudicated sample (n = 272). MCI, mild cognitive impairment; NIA-AA, National Institute on Aging–Alzheimer's Association.

We additionally compared variations in diagnosis of cognitive impairment by rater's professional specialty. We classified raters into two categories: neurologists/geriatricians (seven raters) and neuropsychologists/dementia research psychologists (four raters). An average score of zero indicated congruency between the rater's individual rating and the final consensus rating, while an average score above zero indicated overdiagnosis of MCI/dementia and an average score below zero indicated underdiagnoses of MCI/dementia. Neurologists/geriatricians were more likely to overdiagnose MCI/dementia, averaging positive difference scores compared to neuropsychologists/dementia research psychologists. One neuropsychologist had an average difference score below zero, indicating leniency toward reporting normal cognitive function for study participants who had a final assigned consensus diagnosis of MCI (Figure 3).

Details are in the caption following the image
Overall concordance between individual raters’ diagnoses and final consensus diagnoses (reference) by rater specialty.

4 DISCUSSION

This study is, to the best of our knowledge, the first in South Africa and sub-Saharan Africa to evaluate the feasibility of using a web-based consensus approach for the diagnosis of cognitive impairment and dementia. We assembled a panel of 11 expert US and South African neurologists, neuropsychologists, geriatricians, and researchers with clinical and research experience in dementia, to ascertain dementia diagnoses for the entire HAALSI Dementia cohort (n = 635). All members of the consensus panel found the web-based portal user friendly and well structured in terms of presentation of relevant demographic, cognitive, clinical, and functional information for clinical decision making. The web-based consensus approach proved to be a flexible and efficient methodology as raters were able to securely access the web portal and review cases from anywhere and at any time. The consensus panel offered an improvement in diagnostic confidence and reliability, especially in the absence of normative population data and the observed variability in diagnoses trends by rater specialty.

In this web-based consensus study, in which each case was assigned a diagnostic category by three independent raters, we found an overall diagnostic concordance rate of 78.3% and a moderate level of agreement (k = 0.50) between the overall independent ratings and final diagnosis obtained via consensus. Almost half of the cases required adjudication, with the majority being borderline between normal cognitive function and early MCI or between late MCI and early dementia. Consensus panel members found adjudication discussions on late MCI and early dementia particularly challenging given the heterogenous nature of MCI, which is traditionally viewed as a transitional period between normal aging and dementia.43 We observed the lowest level of overall rater agreement for the MCI category (k = 0.34). The MCI/dementia conundrum mainly stemmed from discordance among raters on whether the participant had cognitive impairment that affected social and occupational function. Contextual factors, such as the high unemployment rates in South Africa, low educational attainment, and cultural norms whereby it is common for younger family members to take over household responsibilities and social activities even if the individual is not cognitively impaired, made it difficult for raters to assess some participants’ true level of social and occupational impairment.44, 45 Our findings build on the existing web-based consensus work reported by our sister study in India, LASI-DAD, which observed significant rater differences in the assessment of “social and community activities” and “home and hobbies” CDR-specific domains.11

Another reason for the observed discordance in raters’ diagnoses is the difference in levels of importance that each rater assigned to the available cognitive and informant measures. This is not surprising in the absence of normative data, especially given the low educational exposure and limited knowledge of psychometric properties within our cohort. Disagreements between raters on impairment were especially low for the visuospatial domain (74%), mainly attributable to the high levels of missingness for measures that required participants to draw (constructional praxis and clock draw). Consequently, it was difficult for raters to assess the visuospatial domain, specifically to decide on impairment for participants with missing data compared to participants who performed poorly on these measures. We have previously reported on missingness in the visuospatial domain and have linked it to the unfamiliarity/uneasiness with holding a pen (or tablet stylus) among participants with low literacy, along with a high prevalence of visual impairment.12, 46 For these reasons, the consensus panel often deemphasized these measures during the diagnostic process. Our in-depth investigation into the causes for discordance among raters will help inform and improve measures for the visuospatial and social/occupation domains for future waves of the longitudinal HAALSI Dementia study.

Potential limitations of our study include that we were not able to estimate inter-rater reliability, due to differences in number of cases reviewed by each rater. However, we computed the average kappa and overall rater reliability and showed reasonable level of agreement between raters’ assigned diagnosis and the final diagnostic outcomes. We did not validate our findings against an “external” reference standard such as pathological results or an in-person clinically diagnosed sample. Our future work includes using magnetic resonance imaging and biomarker substudies to further validate and support the validity of this web-based approach.

Overall, this study further complements the consensus work completed by LASI-DAD, and demonstrates a successful novel web-based consensus panel approach for the diagnosis of cognitive impairment and dementia in a rural South African population with low education level.

This web-based consensus methodology is especially promising for LMICs, as we were able to implement this approach in a very remote, low-literacy, low-resource setting with limited access to health-care services and no access to neuropsychologists and/or dementia specialized professionals. Finally, we will use the assigned consensus diagnoses to develop a diagnostic algorithm to assign dementia probability scores in the larger parent HAALSI cohort.

ACKNOWLEDGMENTS

The authors would like to thank the HAALSI study team, participants, and the Agincourt community. This work was supported by the National Institute on Aging (5R01AG054066). HAALSI Dementia is a collaboration between Harvard Center for Population and Development Studies from Harvard T.H. Chan School of Public Health & MRC/Wits Rural Public Health and Health Transitions Research Unit from School of Public Health at the University of the Witwatersrand in South Africa.

    CONFLICT OF INTEREST STATEMENT

    J.J.M. received grants from NIH, NIA, NIND, NHLBI (institutional payment) and support from Alzheimer's Association for conference attendance. A.M.B. received consulting fees from Cognition Therapeutics, holds patents (including US Patent US9867566B2 and US provisional patent application no.: 63/146,958), and serves on the advisory boards for Albert Einstein College of Medicine and Cogstate (SAB). M.M.G. received grants from NIH/NIA and the Robert Wood Johnson Foundation, royalties from the Oxford University Press, and participates in the SWAN (Study of Women Across the Nation) data safety monitoring board. K.M.L. received grants from the NIA and Alzheimer's Association (institutional payment); consulting fees from Harvard, University of Pennsylvania, University of Minnesota, University of Colorado, Dartmouth, University of Southern California; and payment for expert testimony on a DSMB for a clinical trial at Indiana University. B.T. received consulting fees from Aurai Senior Living and the University of the Witwatersrand Donald Gordon Medical Centre, as well as payments for lectures from the Department of Family Medicine – University of Pretoria and Astellas pharmaceutical. B.T. also serves on the executive committees of the Faculty of Consulting Physicians of South Africa (FCPSA), South African Geriatrics Society (SAGS), Wits Donald Gordon, and the Department of Internal Medicine of the University of the Witwatersrand. All reported conflicts of interest are outside the scope of this published work. The remaining authors have no conflicts of interest to report. Author disclosures are available in the supporting information.

    CONSENT STATEMENT

    All participants gave their informed consent to participate in the HAALSI Dementia study, in accordance with the ethical standards set by the relevant institutional review boards.