Introduction

Alzheimer’s disease (AD) is progressive. It is only diagnosed with certainty when the post-mortem neuropathological examination reveals the presence of amyloid plaques and neurofibrillary tangles in the brains of patients who have suffered from the clinical symptoms prior to their death. Antero-mortem diagnosis thus relies on a set of clinical inclusion and exclusion criteria (e.g., APA 2013; McKhann et al. 2011). Based on current criteria, the diagnosis of probable AD is given when patients have dementia which is defined as cognitive impairment that involves at least two cognitive domains and is sufficiently severe that it interferes with activities of daily living (APA 2013; McKhann et al. 2011). However, dementia corresponds to the end phase of AD. Studies on patients suffering from the autosomal dominant version of the disease and population studies that have followed patients or analyzed their performance retrospectively, suggest that the disease probably starts 10–15 years earlier than when patients typically receive their diagnosis (Amieva et al. 2008; Bateman et al. 2012). This late diagnosis considerably impedes research as it limits the ability to identify the early mechanisms that trigger the disease and contribute to its progression. Furthermore, a considerable effort has been devoted to finding a curative treatment and identifying effective lifestyle prevention strategies for AD. When discovered, such treatment or preventative approaches will have to be provided before the disease has produced major damage to the brain to have the most benefit. These issues have motivated a substantial research effort towards detection at an earlier stage of the disease, prior to dementia.

Mild cognitive impairment (MCI) has been of particular interest in this regard. Individuals with MCI complain about their memory and show mild cognitive deficits when tested with objective cognitive measures (Petersen 2000; Gauthier et al. 2011; Albert et al. 2011), but their cognitive impairment is not sufficiently severe to meet the criteria for dementia. Yet, these individuals have a ten-fold greater risk of progression to dementia than the general population, suggesting that many are actually in a pre-dementia phase of the disease (Gauthier et al. 2011; Petersen 2000). The cognitive assessment performed by neuropsychologists is a central component for the diagnosis of AD at the dementia phase, and it is thus critical to have access to sensitive and specific cognitive measures. Relying on appropriate measures is even more critical during the MCI phase, in which cognitive impairment is mild and functional impact minimal. Differentiating the cognitive deficits related to incipient AD from those related to cognitive fluctuations due to normal aging or other non-morbid conditions is particularly challenging, because impairment is mild at this phase.

Many studies have characterized the cognition of people with MCI relative to that of healthy older adults using cross-sectional designs (for reviews, see Belleville et al. 2008; Belleville et al. 2014a). These studies have provided clinicians with cognitive characterization of this early phase and clinical tools that can be used to determine whether older adults meet the criteria for MCI. However, not all individuals who meet the criteria for MCI will progress to dementia and one major challenge is to identify tests that differentiate between individuals with MCI who will remain stable and those who will progress toward dementia. Many studies have attempted to identify early markers of future progression. These studies rely on a longitudinal design in which MCI participants are tested at entry on a set of cognitive measures and followed over time to identify progressors versus non-progressors. Regression analyses can then be used to assess whether performance on cognitive tests completed at entry indeed predict clinical progression.

Many of these prognostic studies relied on comparable tests and it is therefore possible to pool the data to perform quantitative meta-analyses. This allows us to address many important questions related to early cognitive diagnosis. One major question is whether cognitive tests fare well in predicting progression from MCI to dementia, which is crucial in a context where much effort is devoted to developing biomarkers of progression. Cognitive assessments are widely available and relatively low-cost. Thus, it is important to determine their predictive accuracy and provide quantitative values so that they can be compared to more costly biomarkers. Previous reviews have indicated that the predictive accuracy reported in the literature is extremely variable (Belleville et al. 2014a). This could be due to a wide range of factors, including the cognitive domain measured by the task, length of follow-up, or the characteristics of the patient. Tests might also differ in terms of their specificity and sensitivity. In prognostic studies, sensitivity refers to the correct identification of progressors, whereas specificity refers to the correct rejection of non-progressors.

One important question is to what extent different domains determine overall predictive accuracy, sensitivity, and specificity. For example, some of the most well-accepted clinical criteria have proposed that cued recall (Dubois et al. 2007) or delayed memory (Albert et al. 2011) may be the most highly affected in MCI. It is thus important to assess whether this is the case by conducting an appropriate meta-analysis of the data. Another issue is the role of non-memory domains in early diagnosis. Some studies have found that working memory, executive functions, and language can be impaired in MCI (Belleville et al. 2007; Joubert et al. 2008; Saunders and Summers 2010). However, these studies were based on cross-sectional group comparisons and it is unclear whether these tests are sufficiently sensitive to predict dementia at the individual level. In contrast, non-memory-based tests may have value for specificity, because impairment of non-memory aspects may be more problematic when present. A few studies have examined multiple tests or those that cover a wide range of cognitive functions as predictors of dementia. This strategy may optimize accuracy by pooling tests with a different balance between their sensitivity and specificity (Summers and Saunders 2012; Belleville et al. 2014b).

A related question is the impact that age might have on predictive values of different tests. Age is a potentially important factor because aging itself is accompanied by a specific set of cognitive deficits and because mixed dementia is more frequent in older patients. Thus, older age might be associated with a larger contribution from non-memory domains, particularly executive functions. The length of follow-up is a major issue as well. Studies with shorter follow-ups run the risk of identifying an individual as stable even though they would have progressed to dementia had there been a longer follow-up. Thus, shorter follow-ups increase the likelihood of false negatives and might reduce test sensitivity for the task being evaluated, whereas a longer follow-up should increase sensitivity. A sufficiently long follow-up can be used as a proxy for how far the patient was from the diagnosis when the test was given. Thus, assessing the impact of follow-up length is also informative for identifying whether a task is more useful as a very early marker or as a reflection of imminent progression. A task found to be sensitive at a longer follow-up might be particularly well suited as an early indicator of future dementia. Thus, studies with longer average follow-ups might identify earlier predictors. In turn, a task with a relatively low sensitivity at longer follow-ups but with an excellent value at shorter ones would reflect imminent progression.

Our goal was to conduct a systematic review and meta-analysis to determine the extent to which cognitive measures predict progression from MCI to dementia of the AD type. We used data from longitudinal studies reporting sensitivity and specificity values for cognitive tests to identify individuals with MCI who will progress to a diagnosis of AD in the future. The present study does not replicate or update existing reviews. Other authors have systematically reviewed the diagnostic performance of cognitive tests to cross-sectionally identify people with MCI (Ozer et al. 2016) or dementia (Tsoi et al. 2015), and one study has examined which screening tools best predict progression to dementia in primary care patients (Lischka et al. 2012). However, no review has yet investigated the sensitivity and specificity of a variety of cognitive tests to predict progression to AD among individuals with MCI.

The specific questions addressed in this meta-analysis and systematic review are 1) the predictive value of cognitive measures to predict future progression from MCI to dementia and more specifically, their global (or average) sensitivity (correct identification of progressors) and specificity (correct rejection of non-progressors), 2) the impact of age on predictive accuracy, 3) the average length of follow-up from reported studies and its impact on predictive accuracy, and 4) the benefits of combining different cognitive domains.

Materials and Methods

This systematic review was reported in accordance with the PRISMA statement (Liberati et al. 2009). The protocol was not registered, but it was predetermined, and PICOS statements were used to identify the studies to be included in the review and meta-analyses.

Eligibility Criteria

Criteria for including or excluding articles were determined a priori. The PICOS approach was adapted to formulate the research question. Studies were included if they were longitudinal and designed to assess the sensitivity and specificity of neuropsychological tests to determine future progression from mild cognitive deficit to dementia of the Alzheimer type. Operational inclusion criteria were the following i) articles written in English or French, ii) studies using a prospective or retrospective longitudinal design, iii) at baseline, participants were identified as older adults with MCI and no dementia or a related terminology, iv) neuropsychological or cognitive tests were used as predictors and were sufficiently well described to allow replication, v) participants had a follow-up of at least 1 year, vi) the outcome was a diagnosis of dementia of the Alzheimer disease’s type based on the DSM-III-R criteria (APA 1987), DSM-IV criteria (APA 1994) or National Institute of Neurological and Communication Disorders and Stroke/Alzheimer’s Disease and Related Disorders Association (NINCDS-ADRDA) criteria (McKhann et al. 1984), and supported by a medical examination or a medical consensus panel, and vii) absolute values for false positives, false negatives, true positives, and true negatives were reported or 2x2 tables could be constructed based on reported data. No publication date restriction was imposed.

There were also several exclusion criteria. Studies were excluded i) if the diagnosis of AD was based on a single neuropsychological test or if there was no definition of the criteria used for the diagnosis of AD, ii) if the study targeted populations with other neurological diseases or neurodevelopmental disorders (e.g., HIV associated dementia, Parkinson’s disease, Down Syndrome), iii) if it used a cross-sectional design and iv) if there were insufficient methodological details to allow replication. Furthermore, only the more recent data were included if the same population sample was used in more than one study.

Information Sources and Search

A computer-based search was performed from five electronic databases: MEDLINE using PubMed (1946 - Present), Embase using OVID (1947 - Present), Cochrane (1918 - Present), PsycINFO (1967 - Present), and Web of Science. The last search was carried out on January 29th, 2014. In addition, the reference lists from articles of interest were searched for additional references. The full electronic search strategy for MEDLINE is included in Appendix A (Supplementary material). The search strategy, based on the PICOS approach, was applied through a combination of five concepts that delimited our research question 1) Patient, defined as “elderly population” with “MCI”, 2) Intervention, defined as the index tests, that is, the different neuropsychological predictors, 3) Comparison, defined by the type of analysis expected in the selected studies i.e., those including predictive values for the neuropsychological predictors, 4) Outcome, defined as the predicted outcome which was “Alzheimer’s Disease” and 5) Type of study, which specified that “longitudinal studies” were the type of studies sought. The search strategy was reviewed and validated by a librarian who is an expert in database searches.

Study Selection

The articles produced by the search strategy were first screened based on the titles and then selected by one reviewer (CF) based on the abstracts. One additional reviewer (CH or SB) independently revised the list of potential articles based on the abstracts. The full text of the articles considered to be potentially eligible was then evaluated in detail by the same reviewer for quality assessment and any unresolved issues were discussed with SB. All articles meeting eligibility criteria were included in the qualitative synthesis. Articles reporting data that could be appropriately pooled were included in the meta-analyses: this was done when at least three studies reported data for the same cognitive domain.

Data Collection Process and Data Items

Two reviewers independently extracted data from all eligible studies using a predetermined form (standardized grid). Disagreements between the two reviewers were resolved by consensus or a third reviewer (CH or SB). The data collected from each manuscript included information on the general description of the study, participants, methodological features, results, analysis, and analytic procedure. Table S1 (Supplementary material) shows the items for which information was collected for all papers.

Risk of Bias in Individual Studies

The methodological quality and the risk of bias of each study was assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS) tool (Whiting et al. 2003), as suggested by Cochrane guidelines (Reitsma et al. 2009). A tailored quality assessment grid was constructed to apply the quality assessment criteria proposed in QUADAS 2 (Whiting et al. 2011) adapted to prognostic studies (see Table S2 in Supplementary material). Study quality was assessed by a single reviewer with verification by a second reviewer for i) patient selection and population source (e.g., whether the included sample was representative of the population and whether exclusion criteria were too restrictive), ii) study design (e.g., whether it was a prospective or retrospective study and whether the investigator in charge of diagnosis at follow-up was blind to the baseline performance), iii) flow and follow-up (e.g., length of follow-up), iv) reference standard and outcome (i.e., adequate procedure for diagnosis at follow-up and absence of partial verification bias), v) index tests and prognosis (i.e., adequate description of the neuropsychological battery), and vi) analysis (e.g., whether the cut-off used for calculation of sensitivity and specificity measures was determined using independent norms). This quality assessment allowed us to classify studies as having a low (L), high (H), or low-high (LH) risk of bias.

Analyses

Articles reporting appropriate data were included in the meta-analyses. If the prevalence was not provided, it was calculated by dividing the number of participants who had progressed at follow-up by the total number of participants at follow-up. If the number of true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN) were not available, they were calculated based on the values of sensitivity and specificity using the following formula: TP = sensitivity x number of progressors, TN = specificity x number of non-progressors, FN = number of progressors minus TP, FP = number of non-progressors minus TN.

Variables of Interest

The first aim of the meta-analyses was to obtain global estimates of predictive accuracy for various categories of neuropsychological tests. The study setting was very similar to that of diagnostic test accuracy studies, except that instead of comparing the diagnostic index test results to a gold standard, the index tests were compared to the diagnostic gold standard performed after a certain follow-up period. We therefore used the same methods as for diagnostic test accuracy meta-analyses by obtaining the global sensitivity and specificity for each test category. Sensitivity refers here to the ability of the test to identify MCI patients who will later progress to dementia or more precisely, whether progressors were impaired on the test. Specificity refers to the ability of the test to identify those who will remain stable, that is, whether stable MCI were unimpaired on the test. An ideal test should have 100% sensitivity and 100% specificity, which would result in perfect predictive accuracy. Scores higher than 0.7 are typically considered to be very good, whereas scores above 0.5 indicate good classification (Haynes et al. 2006).

Tests were grouped within 22 cognitive domain categories for systematic review (see Table 3), given the large variety of neuropsychological tests used in the various studies. The classification of tests within cognitive domain categories was performed following study selection, but prior to data analysis, by consensus between the authors and validated by the CIMA-Q cognition expert group, which comprises 12 clinicians and researchers with expertise in the neuropsychology of dementia. This led to the identification of five broad domains corresponding to cognitive components that are well-recognized and believed to depend on distinct neuroanatomical substrates: verbal episodic memory, visual episodic memory, language, executive functions and working memory, and visuo-constructive functions. An additional category was used to include brief and global measures of cognition. We then identified test categories that were judged to reveal different brain or cognitive mechanisms within each domain, or to be differentially affected by dementia. For example, we distinguished immediate from delayed recall, because delayed recall is generally believed to reflect the consolidation processes that occurs within the hippocampus to a larger extent than immediate recall. We distinguished studies using oriented encoding, where the examiner provides cues during the encoding phase, as well as studies using cued recall, where the examiner provides cues during the retrieval phase, because some have argued that providing cues at encoding or retrieval reduces the contribution of executive attention to memory performance.

We separately analyzed studies that identified a combination of tests predicting progression in the systematic review, but they were not included in the meta-analysis. Each cognitive domain category was meta-analyzed separately. The minimum number of studies required to perform a diagnostic accuracy meta-analysis is three (HIQA – IRELAND, IQWiG – GERMANY 2015). The 14 domains that met this criterion were the following 1) among verbal memory tests, verbal immediate recall, paragraph delayed recall, word-list free delayed recall with non-oriented encoding, word-list free delayed recall with oriented encoding, word-list cued delayed recall with oriented encoding, and word-list recognition 2) among visual memory tests, delayed recall 3) among language tests, naming tests, tests of semantic knowledge, and semantic fluency 4) among executive tests, switching tests and working memory tests, 5) among visuo-constructive measures, visuo-spatial tests and visuo-constructive tasks and 6) global measures.

Meta-Analysis

The Cochrane collaboration and others (Macaskill et al. 2010; HIQA – IRELAND, IQWiG – GERMANY 2015; Dahabreh et al. 2012) strongly recommend the use of the random effects hierarchical bivariate models called bivariate (Reitsma et al. 2005) and hierarchical summary receiver operating characteristic (HSROC: Rutter and Gatsonis 2001) for diagnostic test accuracy meta-analyses. The bivariate and HSROC models are equivalent under most common circumstances, for example when no covariate is included or when the sensitivity (Se) and specificity (Sp) of the bivariate model are adjusted for the same covariates. We used only the bivariate model in our analyses due to software implementation technicalities. There are three main reasons to favor a random effects hierarchical bivariate model (Reitsma et al. 2005). First, the main parameters to be estimated - global Se and global Sp - are analyzed jointly because they are negatively correlated (due to the threshold effect). Second, between-study variability is generally high in this type of meta-analysis. Thus, the use of hierarchical random effects models is necessary to appropriately take into account both within and between study variability. Third, the method employs a regression model that naturally allows the inclusion of study-level covariates to perform meta-regressions.

The main assumption of this hierarchical model is at the second level, namely that true values of the logit-transformed Ses and Sps of the individual studies should be independent and follow normal distributions (with unknown means and variances). The main parameters of interest that are estimated are the means of the underlying normal distributions of the Se and Sp values. It is important to use these normal distributions, as a priori the studies have different Se and Sp values due to numerous differences in the study characteristics. Logit-transformed Se and Sp values are difficult to interpret directly. We therefore used the inverse logit transformation of the estimated means to obtain the global Se and Sp values.

Meta-regressions additionally assume homoscedasticity of heterogeneity variances, that is, between-study variances of the logit-sensitivity and logit-specificity parameters must remain approximately constant through the covariate levels. Here, the number of studies was too low to verify these assumptions. However, subgroup analyses were used to confirm meta-regression results without having to assume homoscedasticity. No zero-cell adjustment was needed since the model used binomial distributions to correctly model sampling uncertainty in both Se and Sp.

As mentioned above, a category of cognition was meta-analyzed when it contained at least three studies. These meta-analyses all consisted of a small number of studies (from 3 to 6) with occasionally small sample sizes. For such meta-analyses, the maximum likelihood estimation of the bivariate and HSROC model parameters often suffer from convergence problems (Paul et al. 2010). Therefore, we chose the bivariate Bayesian implementation of the R package, meta4diag (Guo and Riebler 2016), which is based on the Integrated Nested Laplace Approximations method (Rue et al. 2009).

The estimations of the variances and correlation of the joint distribution of Se and Sp values were not precise, because of the small number of studies, and were thus not informative. Since the model uses Bayesian inference, the reported credible intervals (CrI) for global sensitivities and specificities are credible intervals that correspond to the posterior distributions of the parameters. By definition, the estimated parameters lie within the CrIs with a posterior probability of 95% (Robert 2001). In other words, the CrIs must be interpreted to have a 95% certainty of containing the true parameter.

Exploration of Sources of Heterogeneity

We considered mean age and the length of follow-up of study participants to be potential sources of heterogeneity, as these variables could influence the pooled estimates. These analyses were performed through meta-regressions in which these variables were considered to be continuous covariates, one at a time. We also carried out subgroup analyses whenever the CrI of the effect of the covariate on Se or Sp was found to exclude the zero value. These analyses were performed by dividing the studies into two more or less equally numbered subgroups based on the covariate values.

Sensitivity to Prior Distributions

Vague prior distributions used in Bayesian analyses are generally expected to yield findings close to those obtained with other frequently used methods (Spiegelhalter et al. 1999). However, given the small number of studies and the sparse data involved in our meta-analyses, it was important to verify the sensitivity to prior distributions. The prior distributions that are susceptible to have an influence on the results are those of the heterogeneity variances and the correlation. However, as noted by Paul et al. (2010), the influence of the correlation prior is expected to be less for a data set with a lower correlation. Since we did not observe strong correlations in our data, we left the correlation prior as the default of the meta4diag function: normal distribution with a mean of 0 and a variance of 5. Hence, our sensitivity analyses were performed by varying only the prior distribution of the heterogeneity variances. We took two of the priors suggested by (Paul et al. 2010), namely 1) mildly informative inverse gamma distribution with a shape of 0.25 and a rate of 0.025 and 2) the less informative inverse gamma distribution with both the shape and rate equal to 0.001. The priors for the mean parameters and beta-coefficients of the covariates were also used as the default for both sensitivity and specificity, that is, normal prior with a mean of zero and a large variance (exact value of variance not specified by the authors of the package). The two priors for the heterogeneity variances usually yielded very close results, in which case only one was reported.

Publication Bias

The number of studies was insufficient to assess publication bias. Fortunately, this type of bias is less expected to occur for reviews of diagnostic test accuracy than for other types of reviews. In fact, Leeflang (2014) suggested that there is little evidence of publication bias for diagnostic test accuracy studies and hence publication bias is less of a concern for diagnostic test accuracy studies than for other types of reviews (Leeflang 2014).

Subsidiary Analyses Excluding High Risk of Bias Studies

Se and Sp values may be inflated by the inclusion of studies with a high risk of bias (Whiting et al. 2011). We performed subsidiary analyses after removal of all such studies when possible and then compared the results to those obtained when including all studies.

Results

Study Selection

The search strategy identified 4979 citations from electronic databases and six citations from hand searching. After adjusting for duplicates, 3355 citations remained in the pool of papers. After screening titles and abstracts, 313 citations were potentially eligible and full reports were retrieved and analyzed. We then excluded 285 studies which did not meet criteria based on a detailed examination of the full text, thus leaving 28 studies for the systematic review and meta-analysis. All were included in the systematic review. Of these, 21 met criteria described above to be included in the meta-analysis (Fig. 1).

Fig. 1
figure 1

PRISMA 2009 Flow Diagram

Study Characteristics

The included studies and their characteristics are shown in Table 1. Among the 28 selected studies (references listed in Table 1), 22 were prospective, four were retrospective, and two used data from the Alzheimer’s Disease Neuroimaging Initiative. The studies on which the review is based involved a total of 2365 participants with MCI. Participants were followed over a period of 12 to 60 months (mean follow-up: 31 +/−14 months). Among the enrolled MCI participants, 916 (38.7%) later fulfilled criteria for a diagnosis of AD. The mean conversion rate varied from 6 to 39% per year based on studies that used a prospective study design.

Table 1 Characteristics of included Studies

Twenty-one of the 28 studies used Petersen’s criteria to identify MCI at entry, consisting of subjective cognitive complaint, evidence for cognitive impairment (relative to age and education), preservation of general functional abilities, and the absence of diagnosed dementia. Among these 21 studies, 12 included both single and multiple domain amnestic MCI (aMCI), four included only single-domain amnestic MCI (sd-MCI), and five included amnestic and non-amnestic MCI. Four studies used the Global Deterioration Scale (GDS) to identify MCI (i.e., those with a score of 3 on the GDS). Three studies used other terminologies to describe their patients. Albert et al. (2001) selected participants with questionable Alzheimer’s disease (QAD), defined as those with a score of 0.5 for the CDR (Albert et al. 2001). Galton et al. selected participants with “questionable dementia,” defined by the presence of subjective complaints of memory impairment substantiated by an informant, (b) normal activities of daily living, and (c) absence of dementia as evident from an MMSE score of ≥23, and a score of 0.5 on the CDR (Galton et al. 2005). Finally, Tierney et al., selected patients with a three-month history of symptomatic memory problems that interfered with daily functioning (GDS 2 or 3), no diagnosis of dementia, and an MMSE ≥24 or a score ≥ 123 on the Mattis Dementia Rating Scale (Mattis 1988; Tierney et al. 1996).

The mean age at entry for those who progressed to AD ranged from 64.9 to 77.6 years (mean age = 71.8 years) versus 59.3 to 75 years (mean age = 69.8 years) for non-progressors. The percentage of females in the group of progressors varied from 33 to 77% (overall percentage = 54%) vs 32 to 72% in the group of non-progressor MCI (overall percentage = 51%). Mean years of education of participants who progressed to AD varied from 7.5 to 15.6 (mean = 12.1 years) vs 7.7 to 15.9 (mean = 12.6 years) for non-progressors. All included studies used the NINCDS-ARDRA criteria as their outcome for AD.

The studies reported predictive values for a total of 61 neuropsychological tests. Sixteen articles reported the predictive values of individual neuropsychological tests, five reported the predictive value of a combination of several neuropsychological tests, and seven included both individual tests and combinations of tests.

Risk of Bias within Studies

Table 2 shows the risk of bias associated with the included studies. Ten studies showed a low risk of bias, eight a moderate risk of bias, and ten a high risk of bias (all references listed in Table 2). The most frequent limitation comes from the selection process of the sample, which may not be representative of the population of interest. For example, Eckerstrom et al. selected, a posteriori, a sub-sample of non-progressors to match the sample of progressors. Other studies only selected participants who previously underwent particular examinations (e.g., PET scan in Arnaiz et al. 2001 or a CSF examination in Richard et al. 2013). Griffith et al., included participants who were taking cholinesterase inhibitors (13 of 49 MCI). Other studies did not include sufficient details about the selection process (Babins et al. 2008). Many studies relied on a small sample size (< 40 participants: Ahmed et al. 2008; Arnaiz et al. 2001; Didic et al. 2013; Dierckx et al. 2009; Flicker et al. 1991; Galton et al. 2005; Irish et al. 2011; Lekeu et al. 2010; Venneri et al. 2011) or showed a higher progression rate than expected (Gallagher et al. 2010; Irish et al. 2011; Lekeu et al. 2010).

Table 2 Risk of bias within studies

Other aspects of the study design raised issues related to quality, albeit less frequently. For example, five studies did not ensure strict independence between the diagnostic battery and the predictive battery (Anchisi et al. 2005; Arnaiz et al. 2001; Marcos et al. 2006; Tabert et al. 2006; Venneri et al. 2011). Thus, either some of the neuropsychological tests included in the predictive battery were also used to make the baseline or follow-up diagnosis of the participants or the staff administering the diagnostic battery were not blind to the participants’ performance on the predictive battery. Six studies did not provide sufficient information to determine whether the diagnostic and predictive assessments were independent with certainty (Defrancesco et al. 2013; Flicker et al. 1991; Gallagher et al. 2010; Griffith et al. 2006; Lekeu et al. 2010; Visser et al. 2001).

Most studies had a duration of follow-up longer than 12 months and most reported the reasons for loss to follow-up. Only two studies had a relatively short follow-up (12 months), which likely increased the risk of false negatives (Ahmed et al. 2008; Anchisi et al. 2005). Most studies used appropriate procedures for diagnosis. Two studies pooled patients who converted to AD and patients who showed significant cognitive decline (Belleville et al. 2014b; Flicker et al. 1991) and thus compared “stable MCI” to “decliners”. Flicker et al. administrated the reference standard battery for AD diagnosis only to patients who showed decline on the GDS. Thus, this study presents a “partial bias verification”. As per our inclusion criteria, all the predictive tests were well described or relied on tests that are well known and for which an extensive description is available from the literature.

Most studies used ROC curves to calculate optimal cut-offs for the tests and determine predictive accuracy. Twenty studies used regression analyses to assess and compare the predictive accuracy of different tests.

Results from the Systematic Review

A summary of the predictive accuracy values for single domains is presented in Table S3 (Supplementary material). The sensitivities observed from individual studies ranged from 0.18 to 1 and the specificities ranged from 0 to 1. A very large number of tests show very good (≥ 0.70) sensitivity, very good specificity, or both. The tests which obtained the sensitivity and specificity values in the excellent range (≥ 0.90) are shown in Table 3. The table indicates predictive values (number of true positives, true negatives, false positives, false negatives, sensitivity, specificity, and accuracy) for each cognitive domain category. The cut-off values of the tests used in the studies are included, if available. Individual tests showed excellent specificity more often than excellent sensitivity. The tests with best overall accuracy (≥ 0.90) are shown in Table 4. Five individual tests and one global measure had excellent overall predictive value, implying that they combine very high specificity and sensitivity. The tests were 1) the Guild Paragraph, delayed recall, 2) the RAVLT, delayed Recall, 3) Face-name association task, free delayed recall of names, 4) Object function recognition, 5) VOSP Silhouettes, and the ACE Addenbrooke’s Cognitive Examination.

Table 3 Individual tests with excellent specificity or sensitivity, grouped by cognitive domain
Table 4 Individual tests with excellent overall accuracy

Results obtained from combining different neuropsychological tests are provided in Table 4 (Supplementary material). Most studies combined episodic memory tests with global measures or non-memory measures. Studies that examined a combination of memory and non-memory tests obtained very high to excellent predictive accuracy, with a few exceptions. Combining episodic memory tests with executive function tests gave very good sensitivities (from 0.74 to 0.77) and excellent specificities (0.83 to 0.94: Albert et al. 2001; Griffith et al. 2006; Tabert et al. 2006; Tierney et al. 1996). The combination of episodic memory with category fluency allowed for excellent sensitivity (0.84) and very good specificity (0.79: Gallagher et al. 2010). One study reported high sensitivity and specificity (0.88 and 0.87, respectively) when combining episodic memory, language, visual perception, and executive function tasks (Belleville et al. 2014b).

Two studies combining an associative memory task (CANTAB Paired associate learning) with a global measure (Addenbrooke’s Cognitive Examination) reported different results. Ahmed et al. reported excellent accuracy (1 and 0.82 for sensitivity and specificity, respectively), whereas Mitchell et al. reported high sensitivity (0.94), but low specificity (0.40). This may be due to the difference in the duration of follow-up (12 versus 24 months) or study sample (18 versus 82 patients). One study reported that combining associative memory tests with language (category fluency) or visuo-constructive tasks provided very high sensitivities (0.73 to 0.91), but low specificities (0.5 to 0.57: Venneri et al. 2011). The same study reported low sensitivity and specificity values (0.64 and 0.5, respectively) for a combination of associative memory, language measures, and visuo-constructive tasks (Venneri et al. 2011). However, this study was carried out with a small sample size (n = 25).

Combining episodic memory with global measures generally provided very good to excellent predictive accuracy. The combination of an episodic memory test with a brief measure (MMSE orientation) provided very good sensitivity and excellent specificity (0.74 and 0.9, respectively: Defrancesco et al. 2013). Combining episodic memory with global measures and visual perception showed excellent sensitivity (0.92) and very good specificity (0.73: Marcos et al. 2006). One study that combined episodic memory tests with a global measure (CDR mean sum of boxes) obtained low sensitivity (0.66), but high specificity (0.86: Perri et al. 2007b). The combination of different episodic memory modalities gave variable results ranging from low sensitivity (0.44) and high specificity (0.9: Gallagher et al. 2010) to excellent sensitivity and specificity (1: Didic et al. 2013).

Given the large variability observed in the predictive accuracy obtained from individual tests, the meta-analysis presented below, summarizes predictive accuracy values (sensitivity and specificity) by domain subgroups. We then present the results from meta-regression, which examined the interaction of sensitivity and specificity with the duration of follow-up, as a proxy for the time to diagnosis, and age.

Results of the Meta-Analyses

Convergence was achieved in all meta-analyses, except one (associative memory), and only two meta-regressions (verbal immediate recall and paragraph delayed recall) had one prior distribution for which convergence failed. In addition, the second prior distribution analysis failed for short follow-up sub-group for semantic fluency.

Verbal Episodic Memory

Most tests measuring verbal episodic memory reported very good predictive accuracy (all above 0.7) and there was little variability related to the test condition (e.g., delayed or immediate, with or without oriented encoding, with or without cues at recall) or the particular material (e.g., words or paragraph). The only exceptions were for recognition and word-list cued delayed recall with oriented encoding, which had lower sensitivity values than other verbal memory tests. Predictive accuracy for tests of verbal immediate recall was reported in five studies (Galton et al. 2005; Perri et al. 2007b; Didic et al. 2013; Sarazin et al. 2007; Richard et al. 2013). The combined value was 0.745 for sensitivity (95% CrI: 0.511–0.905) and 0.771 for specificity (95% CrI: 0.581–0.897). Predictive accuracy for paragraph delayed recall was reported in five studies (Didic et al. 2013; Galton et al. 2005; Belleville et al. 2014b; Kluger et al. 1999; Perri et al. 2007b). The combined analysis gave a 0.776 sensitivity (95% CrI: 0.541–0.928) and a 0.794 specificity (95% CrI: 0.517–0.961). Word-list free delayed recall with non-oriented encoding was reported in four studies (Eckerstrom et al. 2013; Ewers et al. 2012; Visser et al. 2001; Perri et al. 2007b). The combined analysis gave 0.742 sensitivity (95% CrI: 0.625–0.868) and a 0.814 specificity (95% CrI: 0.517–0.961). Word-list free delayed recall with oriented encoding was also reported in five studies (Anchisi et al. 2005; Gallagher et al. 2010; Sarazin et al. 2007; Belleville et al. 2014b; Irish et al. 2011). The combined analysis gave 0.781 sensitivity (95% CrI: 0.704–0.867) and a 0.797 specificity (95% CrI: 0.651–0.893). Word-list cued delayed recall with oriented encoding was reported in three studies (Sarazin et al. 2007; Didic et al. 2013; Dierckx et al. 2009). The combined analysis gave 0.676 sensitivity (95% CrI: 0.526–0.796) and 0.896 specificity (95% CrI: 0.822–0.947). Word-list recognition was reported in three studies (Galton et al. 2005; Ewers et al. 2012; Gallagher et al. 2010). The combined analysis gave 0.547 sensitivity (95% CrI: 0.428–0.671) and 0.789 specificity (95% CrI: 0.602–0.927).

Visual Episodic Memory

Four studies reported predictive accuracy for visual episodic memory tests (Didic et al. 2013; Griffith et al. 2006; Lekeu et al. 2010; Perri et al. 2007b). The combined analysis showed a relatively low sensitivity, 0.676 (95% CrI: 0.413–0.889), but very good specificity, 0.847 (95% CrI: 0.722–0.913).

Language

The predictive accuracy provided by language tests was relatively high with values just below 0.7 for sensitivity and somewhat higher for specificity. Predictive accuracy for naming tests was reported in six studies (Eckerstrom et al. 2013; Gallagher et al. 2010; Ahmed et al. 2008; Galton et al. 2005; Belleville et al. 2014b; Sarazin et al. 2007). The analysis gave a combined sensitivity score of 0.699 (95% CrI: 0.51–0.868) and a combined specificity score of 0.707 (95% CrI: 0.621–0.813). Data for tests of semantic knowledge was reported in three studies (Didic et al. 2013; Sarazin et al. 2007; Flicker et al. 1991). The analysis gave a combined sensitivity score of 0.703 (95% CrI: 0.442–0.906) and a higher combined specificity of 0.814 (95% CrI: 0.631–0.984). Tests of semantic fluency were examined in six studies (Gallagher et al. 2010; Galton et al. 2005; Ahmed et al. 2008; Ewers et al. 2012; Sarazin et al. 2007; Venneri et al. 2011). The combined sensitivity score was 0.708 (95% CrI: 0.468–0.875) and the combined specificity score 0.7 (95% CrI: 0.539–0.836).

Executive Functions

This domain yielded relatively low predictive accuracy values with better specificity than sensitivity. Three studies reported predictive accuracy for switching tests (Trail test; Ahmed et al. 2008; Sarazin et al. 2007; Ewers et al. 2012). The analysis gave a combined sensitivity score of 0.541 (95% CrI: 0.314–0.701) and a somewhat higher combined specificity score of 0.679 (95% CrI: 0.542–0.797). Three studies reported predictive accuracy for working memory tests (Ewers et al. 2012; Sarazin et al. 2007; Belleville et al. 2014b). The combined sensitivity score was 0.599 (95% CrI: 0.463–0.724) and the specificity was again slightly better with a combined score of 0.679 (95% CrI: 0.573–0.747).

Visuo-Constructive Functions

Four studies reported predictive accuracy for visuo-spatial tests (Gallagher et al. 2010; Marcos et al. 2006; Belleville et al. 2014b; Eckerstrom et al. 2013). The combined sensitivity score was 0.68 (95% CrI: 0.366–0 .914) and the combined specificity score was 0.749 (95% CrI: 0.613–0.873). Three studies reported data from visuo-constructive tasks (Babins et al. 2008; Buchhave et al. 2008; Arnaiz et al. 2001). Their combined value was 0.637 (95% CrI: 0.471–0.778) for sensitivity and 0.643 (95% CrI: 0.496–0.775) for specificity.

Brief or Global Measures

Four studies reported predictive accuracy for brief or global cognitive measures and their combined predictive accuracy was quite high, both in terms of sensitivity and specificity (Ahmed et al. 2008; Galton et al. 2005; Gallagher et al. 2010; Marcos et al. 2006). The analysis gave a combined sensitivity score of 0.852 (95% CrI: 0.73–0.946) and a combined specificity score of 0.757 (95% CrI: 0.468–0.966).

Analyses Excluding the High Risk of Bias Studies

The results of the meta-analyses performed after removal of the 10 studies with a high risk of bias are presented on the right-hand side of Table 5. Three test categories did not require re-analysis as they did not contain any high risk of bias studies (verbal immediate recall, switching tests, and working memory tests). Specificity and sensitivity scores obtained when excluding high-risk studies generally remained close to the values obtained from the initial analyses. In three cases, sensitivities that were above 0.70 fell below that score when excluding high-risk studies: paragraph delayed recall, naming, and semantic fluency. However, for naming and semantic fluency, lower sensitivities were partially compensated by slightly higher specificities. In the case of the visuo-constructive tasks category, the specificity increased from below 0.70 to very good (above 0.70) when excluding high-risk studies. Four test categories could not be re-analyzed because they ended up with only two studies: word-list cued delayed recall with oriented encoding, word-list recognition, semantic knowledge and visuo-spatial tests. No sensitivity analyses were performed for the subgroups as they all contained only three studies.

Table 5 Meta-analyses
Table 6 Meta-regressions assessing the effect of follow-up duration or age of participants at baseline on sensitivity and specificity of neuropsychological tests
Table 7 Meta-analyses of subgroups of studies with long or short follow-up

Results of the Meta-Regression

The results of the meta-regressions are presented in Table 6. There was a negative association between the length of follow-up and specificity values for the naming tests (−0.094, 95% CrI: −0.173 – -0.027: Eckerstrom et al. 2013; Gallagher et al. 2010; Ahmed et al. 2008; Galton et al. 2005; Belleville et al. 2014b; Sarazin et al. 2007). Shorter mean follow-up was associated with higher specificity values. When examining specificity in the length of follow-up subgroups (Table 7), the combined specificity value for the naming tests obtained from the three studies with the longest follow-up (31 to 36 months: Gallagher et al. 2010; Belleville et al. 2014b; Sarazin et al. 2007) was good (0.648; 95% CrI: 0.539–0.738), whereas it was very high (0.852; 95% CrI: 0.703–0.943) for the three studies with the shortest follow-up (12 to 24 months: Eckerstrom et al. 2013; Ahmed et al. 2008; Galton et al. 2005). The second prior distribution of this meta-analysis failed and thus this result should be interpreted with caution.

There was a positive association between the length of follow-up and sensitivity values for tests of semantic fluency (0.143, 95% CrI: 0.035–0.261: Gallagher et al. 2010; Galton et al. 2005; Ahmed et al. 2008; Ewers et al. 2012; Sarazin et al. 2007; Venneri et al. 2011). Longer mean follow-up was associated with higher sensitivity values. When examining sensitivity in the length of follow-up subgroups (Table 7), the three studies with the longest follow-up (31 to 36 months: Gallagher et al. 2010; Sarazin et al. 2007; Venneri et al. 2011) showed a very high sensitivity value (0.842; 95% CrI: 0.569–0.966), whereas those with the shortest follow-up (12 to 24 months) showed a low combined sensitivity of 0.540 (95% CrI: 0.324–0 .722: Galton et al. 2005; Ahmed et al. 2008; Ewers et al. 2012). There was no effect of the length of follow-up for any of the other domains. Furthermore, there was no effect of age on sensitivity or specificity values.

Discussion

Neuropsychological assessment is central to the diagnosis of dementia and identifying individuals who may be in a prodromal phase of AD. People with MCI have a ten-fold higher risk of progressing to dementia every year than people of the same age in the general population. Thus, many people with MCI are actually in a prodromal stage of AD. However, not all patients with MCI progress to dementia and it is thus critical for clinicians to identify tools that can accurately separate those who will remain stable from those who will progress to dementia. Many longitudinal studies have published predictive accuracy values for different cognitive tests. There is a critical need for a systematic analysis of this literature because of the large number of cognitive dimensions that can be measured and because each domain can be assessed with still a larger number of neuropsychological tools.

In this systematic review, we found 28 longitudinal studies that assessed the values of neuropsychological tests to predict progression from MCI to dementia. We selected studies based on strict inclusion and exclusion criteria and hence discarded studies that contained fatal methodological flaws for a meta-analysis of predictive diagnostic test accuracy (for example, those that failed to use clinical criteria to identify progression to dementia or those for which the methodological information provided was insufficient to allow replication). Nevertheless, it was important to assess the general quality level of the remaining studies and whether the methodology could bias the data, as the studies varied somewhat. Based on the quality criteria used here (QUADAS tool and Cochrane guidelines), most studies had a relatively low risk of bias. Furthermore, the studies were relatively homogenous in their methodological approach and relied on well-accepted clinical criteria to identify their patients and outcomes. Most included studies relied on a prospective design and used Petersen’s criteria to select their participants, and all relied on the NINCDS-ARDRA to identify dementia. Yet, 10 of the 28 studies showed a high risk of bias. A few features were found to be problematic. The most frequent problem was related to the selection process, which led to the sample not being representative of the population of interest. This limitation might have a negative impact on the generalizibility of the results from these studies. Another problem was related to a failure to keep predictive tests independent from the gold standard used to identify the outcome (here, progression to dementia). This was found in five studies, and five others failed to report information regarding this criterion. This is an important methodological control because predictive accuracy can be artificially inflated when the predictors are not kept independent from the standard.

We decided to focus on measures of sensitivity and specificity and to report both independently rather than focusing mostly on overall predictive accuracy. Ideally, a test with optimal predictive accuracy should combine excellent sensitivity and specificity. However, the optimal ratio of specificity to sensitivity may also depend on the clinical and research context. For example, clinicians might favor sensitivity over specificity to detect a deadly disease that could be cured if treated. In contrast, specificity might be favored over sensitivity when a disease cannot be treated, when a diagnosis has the potential to result in stigmatization, exclusion or depression, or when treatment has important side effects. In a research context, investigators might favor a different balance between sensitivity and specificity as a function of their research question. Remarkably, our systematic review indicated that many domains and tests show an appropriate balance with very good sensitivity and specificity values.

The studies that we reviewed covered data for a total of 2365 participants who met the criteria for MCI at entry and were followed over an average of 31 months to assess whether they met the criteria for AD type dementia. In total, 916 individuals with MCI were later found to progress to dementia. This represents a progression rate of 38.7%, which is fairly consistent with the literature, considering the 31-month average follow-up (Gauthier et al. 2006). However, the progression rate across individual studies was quite variable, ranging from 6 to 39% per year.

The systematic review examined 61 cognitive tests that evaluated 22 cognitive dimensions. It identified many neuropsychological measures with very good sensitivity for predicting dementia, with some reaching more than 90%. Similarly, many neuropsychological tests revealed excellent specificity values. Five neuropsychological measures had an overall accuracy of greater than 90%. Three were episodic memory tests (Guild paragraph delayed recall, RAVLT delayed recall, and free delayed recall of names from a face-name association) and two measured visual semantics (object function recognition and the VOSP silhouette). One global test measuring different cognitive components (ACE Addenbrooke’s cognitive examination) also yielded excellent overall accuracy. Thus, although the sensitivity to specificity ratio varied for individual tests, many had an appropriate balance between the two and some showed both excellent sensitivity and specificity. This systematic review also examined the predictive value of studies that examined a combination of cognitive measures. The use of a combination of neuropsychological measures is likely to be the best approach to identify future progressors, because the sensitivity to specificity ratio varies for individual domains. Studies that have examined combined markers generally reported very high to excellent predictive accuracy with a good balance between sensitivity and specificity, particularly when they combined memory with executive or language tests.

We performed a meta-analysis of the predictive accuracy for 14 cognitive domain categories that included at least three independent studies. The meta-analysis pooled sensitivity and specificity values to obtain quantitative indicators. The meta-analysis showed that most measures of verbal memory were excellent predictors with very good (≥ 0.7) specificity and sensitivity values. In addition, predictive values from verbal memory tests were barely influenced by the testing conditions. For example, delayed recall did not predict progression better than immediate recall. Similarly, there was no major difference between cued recall and free recall and there was no added benefit from providing orientation at retrieval. This goes against the concept that measures of delayed recall or tests that orient processing at encoding are the best indicators of early AD, because they reflect hippocampal dysfunction (Dubois et al. 2007; Albert et al. 2011). The present data indicate rather that a range of verbal memory tests can be used as appropriate indicators of early AD and that the nature of the task may not profoundly influence predictive accuracy. Contrary to the general finding of high predictive value for verbal memory tasks, two testing conditions were associated with relatively low predictive accuracy: word recognition and word recall with orientation at encoding and cues at retrieval. This is consistent with the notion that AD patients suffer from impaired encoding, because being impaired on tasks that increases encoding is not a good predictor of progression.

Interestingly, some language categories were rather good predictors of future decline, particularly naming and semantic fluency. These tests implicate semantic memory and some form of executive functions, which may explain their ability to predict future decline, as both processes were proposed to be impaired early in MCI (Belleville et al. 2008; Joubert et al. 2008). The predictive accuracy of two language tests (fluency and naming) was modified as a function of the length of follow-up. This indicates that the predictive value of language categories varies with disease stage in the prodromal continuum, contrary to verbal memory, for which predictive accuracy was similar, irrespective of the prodromal stage of the patient. This finding and its implications will be discussed further below.

Several cognitive categories showed better specificity than sensitivity values. This was true for memory tasks that provide support at encoding and retrieval, for example, recognition (0.547 and 0.789 for sensitivity and specificity, respectively) or word-list cued delayed recall with oriented encoding (0.676 and 0.896 for sensitivity and specificity, respectively). The same pattern was found for some non-memory categories as well. For example shifting (sensitivity = 0.541; specificity = 0.679), working memory (sensitivity = 0.599; specificity = 0.667), and semantic knowledge (sensitivity = 0.703; specificity = 0.814) showed higher specificity than sensitivity. Thus, these tests may not be very sensitive to identify future progression, but they might be useful for identifying patients with MCI who will remain stable and thus contribute to reduce the number of false positives.

The meta-regressions examined whether age and length of follow-up determined differences in sensitivity and specificity values. Age had no effect on predictive accuracy. However, conclusions concerning age might be limited by the fact that this aspect was examined across studies and not across individuals. It is therefore possible that the search for an effect was limited by the lack of difference in the average age across studies, because disease onset is likely to be equivalent across different samples. The effect of the length of follow-up is perhaps more informative, as studies have control over this variable, which differed between studies. The predictive accuracy of naming and the semantic fluency category varied as a function of the length of follow-up. Studies with very short follow-ups might increase the likelihood of false negatives, as they do not allow sufficient time for individuals to progress to meet the dementia criteria. It is unlikely that a test would be more sensitive for a shorter than longer follow-up, as symptoms increase with progression. Thus, higher sensitivity for studies with longer follow-ups might reflect such a phenomenon. Semantic fluency followed this pattern. Fluency tasks yielded excellent sensitivity (0.842) for long follow-ups (31 to 36 months), but sensitivity markedly declined (0.540) for short follow-ups (12 to 24 months). Hence, this interaction might reflect the contribution of false negative cases to the data from studies relying on short follow-ups.

Assessing the impact of the length of follow-up is also informative for identifying very early markers. Hence, a reasonably long follow-up can be used as a proxy for how far the patient was from the diagnosis when the test was given. Determining what constitutes a reasonable follow-up is complex, but if approximately 15% of individuals with MCI progress to dementia yearly, and approximately 25% of them remain stable, irrespective of follow-up length, a three-year follow-up would allow the detection of approximately 60% of the MCI progressors, which would increase to 80% for a four-year follow-up. A task found to be predictive, irrespective of the length of follow-up, is a good candidate for an early predictor. A task found to be less predictive at a longer than shorter follow-up may be less well-suited as an early indicator of future dementia and more representative of imminent decline. This pattern was found for the naming category. Although the test showed good specificity at longer follow-ups (0.648 for follow-ups of 31 to 36 months), specificity was markedly increased, and excellent, at shorter follow-ups (0.852 for follow-ups of 12 to 24 months). Sensitivity was unaffected by the length of follow-up, suggesting that naming might fare better as a predictor at shorter rather than longer follow-up and might be better used as a marker of imminent progression rather than an early marker of the disease. This result and its interpretation needs to be confirmed by future studies and meta-analyses, because the effect was not found when examining the alternative distribution and it was based on a relatively small number of studies.

One strength of this meta-analysis was the use of the Bayesian approach, which has several advantages over other frequently used approaches. Simulation studies have shown that the Bayesian method provides better coverage probabilities for global sensitivity and specificity, particularly in the case of sparse data (Paul et al. 2010). One reason for this is that Bayesian inferences generally come with wider credibility intervals (often more realistic) than frequently used alternative methods (Warn et al. 2002). The Bayesian approach yields less biased estimations of variance and correlation parameters (Paul et al. 2010). Other frequently used approaches often experience convergence issues, which are less of an issue with the Bayesian approach (Paul et al. 2010). Finally, the Bayesian approach generally produces an approximate joint posterior distribution of all model parameters. This has the advantage of not only making it possible to test hypotheses, but also to obtain the probability that any given parameter is above or below any given threshold (Rutter and Gatsonis 2001). It also allows the easy computation of point estimates of any functions of Se and Sp, such as predictive values or likelihood ratios, along with their credibility intervals.

This study has limitations. One major limitation is that we focused on individuals meeting the criteria for MCI, which might not represent the earliest stage of AD. Thus, the follow-up periods reported in the included studies are relatively short if one considers that the disease develops over two decades prior to diagnosis. Therefore, it is unclear whether the predictive accuracy identified here is representative of earlier stages and whether it can be extended further back during the prodromal period. It is also possible that a slightly different ensemble of tasks would display different sensitivity and specificity values at an earlier period. Future studies could meta-analyze data from individuals reporting a subjective cognitive decline to assess earlier markers, as these individuals might be in a phase that precedes MCI in the disease continuum (Jessen et al. 2014). Similarly, the outcome of studies interested in pre-dementia diagnosis depends on the type of recruitment at entry and the validity of the classification scheme, for example, how subjective cognitive decline or MCI is diagnosed. Advances in the field and refinement of diagnostic criteria will certainly increase the ability to identify early markers. Another limitation is the large variability in the tasks that were tested across studies. As a result, we focused on cognitive domain categories rather than individual tasks for the meta-analyses. Although this quantitative meta-analysis provides information as to the domains that should be measured for early prediction, it does not identify specific tests. However, the systematic review included in this study identifies the predictive accuracy for a range of neuropsychological tests that map the cognitive domain categories identified in the meta-analysis. Our finding that the use of different testing conditions for memory tasks does not substantially modify predictive accuracy lends support to our approach, as it indicates that sensitivity and specificity do not vary much within broad cognitive domains. As already mentioned, the statistical analyses were limited by the small number of included studies. This does not diminish the reliability of the results, but the high between-study variability was reflected by the large CrI widths. Some statistical assumptions, for example that of normality, could not be verified due to the small number of studies. Also, the reported CrIs are not applicable to a hypothetical future study. For example, using different cut-offs would change the sensitivity and specificity values. We experienced instances of convergence failures. The Bayesian approach that we used is known to be less prone to convergence issues than other frequently used methods (Paul et al. 2010). Yet, when there are relatively few studies with sparse data, convergence can be particularly difficult to achieve because it may lead to very wide estimations of the corresponding credibility intervals. This is even more challenging for meta-regression models as they contain additional parameters. We suspect that the convergence failures were due to small amounts of data compared to the number of model parameters. Finally, sensitivity and specificity values were slightly modified in four cases after excluding studies with a high risk of bias (reduced sensitivity for three, increased specificity for one). In these cases, correct inferences for sensitivities and specificities probably fell somewhere between the original results (left-hand side of Table 5) and those excluding high-risk studies (right-hand side of Table 5).

In conclusion, the results from this meta-analysis are encouraging for those interested in the early identification of AD. They show that neuropsychological assessment, which is affordable and widely accessible, can strongly contribute to predicting dementia while individuals are still in the MCI phase. The meta-analysis revealed very good to excellent predictive accuracy for many cognitive domains, particularly those concerned with verbal memory and semantic processing. Based on the meta-analyzed data, performance on cognitive tests can predict whether MCI patients will progress to dementia at least 3 years prior to the time at which the diagnosis is made and should contribute highly to the development of early indices of AD.