Abstract
Background
Cognitive screening tests (CSTs) are crucial to neuropsychological diagnostics, and thus need to be featured by robust psychometric and diagnostic properties. However, CSTs happen not to meet desirable statistical standards, negatively affecting their level of recommendations and applicability. This study aimed at (a) providing an up-to-date compendium of available CSTs in Italy, (b) report their psychometric and diagnostic properties, and (c) address related limitations.
Methods
This review was implemented by consulting Preferred Reporting Items for Systematic Reviews and Meta-Analyses and pre-registered on the International Prospective Register of Systematic Reviews. Standardization and usability studies focusing on norms, validity, reliability, or sensitivity/specificity (and derived metrics) in adults were considered for eligibility. Quality assessment was performed by means of an ad hoc checklist collecting information on sampling, psychometrics/diagnostics, norming, and feasibility.
Results
Sixty studies were included out of an initial N = 683. Identified CSTs (N = 40) were classified into general, domain-, and disease-specific (N = 17, 7, and 16, respectively), the latter being less statistically robust than remaining categories. Validity and reliability evidence was provided for 29 and 26 CSTs, respectively, sensitivity/specificity for 20 and norms for 33. Prevalence- and post-test-based diagnostic metrics were seldomly represented; factorial structures, ceiling/floor effects, and acceptability rarely investigated; content, face, and ecological validity never assessed.
Discussion
Although available Italian CSTs overall met basic psychometric/diagnostic requirements, their statistical profile often proved to be poor on several properties that are desirable for clinical applications, with a few exceptions among general and domain-specific ones.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Cognitive screening in adults and elders is relevant to both neurological/neuropsychiatric diagnostics and prevention in internal patients with possible brain damage [1], as well as, in turn, to prognosis and interventional management [2]. To screen for cognitive deficits is indeed meant to ease clinicians into determining whether II-level neuropsychological assessment (i.e., an in-depth examination of multiple cognitive/behavioral functions) is needed for a given patients [3].
As aimed at providing practitioners with an optimal compromise between informativity and ease of use within the early detection of changes in cognition [1], cognitive screening tests (CSTs) need to come with robust psychometric and diagnostic properties, representative norms, and evidence of clinical feasibility in target conditions (i.e., clinical populations that they are meant to be administered to) [2, 4] (see Table 1).
However, it has been already acknowledged that widespread CSTs happen to fail reaching the aforementioned statistical standards, in turn negatively affecting their level of recommendations [9]. In this respect, cross-cultural adaptations of CSTs have been specifically highlighted as suffering from psychometric/diagnostic weaknesses [10], and this representing a major issue in the light of the relevance of culture-/language-specificity to cognitive assessment [11].
In Italy, much attention has been historically devoted to providing norms within the development and adaptation of CSTs [12]. However, it is debated whether this focus might have led to neglecting other fundamental statistical aspects when standardizing tests, such as validity, reliability, and diagnostic properties [13].
In light of the above premises, this study aimed at systematically reviewing evidence on originally Italian/adapted-to-Italian CSTs in order to (a) provide an up-to-date compendium of available CSTs in Italy; (b) report their psychometric and diagnostic properties; and (c) address current issues with regard to their development, adaptation, and standardization.
Methods
Search strategy
Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines were consulted [14]. This review was pre-registered on the International Prospective Register of Systematic Reviews (PROSPERO; CRD42021254561: https://www.crd.york.ac.uk/prospero/display_record.php?RecordID=254561).
The following search terms were entered into Scopus and PubMed databases on May 1, 2021 (no date limit set): neuropsych* OR cogniti* AND screen* OR “screening test” OR “screening tool” OR "screening instrument" AND Italy OR Italian. Fields of search were title, abstract, and key words for Scopus whereas title and abstract for PubMed. Only peer-reviewed, full-text contributions written in English/Italian were considered. Hence, non-peer-reviewed literature was not searched for. Further contributions of possible interest were identified within reference lists of included articles/through manual search.
Contributions focusing either on the standardization of Italian/adapted-to-Italian CSTs (i.e., investigations on psychometric/diagnostic or normative studies) or their feasibility/usability in healthy participants (HPs) and in patients with neurological or neuropsychiatric diseases were considered for eligibility. For a non-normative study to be included, at least one property among validity, reliability, and sensitivity/specificity (or related metrics) had to be assessed. Case reports/case series, reviews/meta-analyses, abstracts, research protocols, qualitative studies, and opinion papers were excluded. Among feasibility/usability studies, those focusing on selected clinical populations that would have not allowed sufficient generalizability were not considered. Investigations on proxy-report tools, questionnaires, CSTs for pediatric populations, or requiring ≥45′ to be administered were also excluded in order to improve external validity of conclusions.
Data collection and quality assessment
Screening and eligibility stages were performed by one of the authors (EN.A.) via Rayyan (https://rayyan.qcri.org/welcome); a second author (G.A.) supervised this stage.
Data extraction was performed by two independent collaborators (S.R. and F.C.), whereas one independent author (E.N.A.) supervised this stage and checked extracted data.
Outcomes of interest were (1) sample size, (2) sample representativeness (geographic coverage, exclusion criteria), (3) participants’ demographics, (4) test adaptation procedures, (5) modality of assessment (in-person vs. remote), (6) administration time, (7) validity metrics, (8) reliability metrics (including significant change measures), (9) measures of sensitivity and specificity, (l0) metrics derived from sensitivity and specificity, (11) norming methods, and (12) other psychometric/diagnostic properties (e.g., accuracy, acceptability rate, assessment of ceiling/floor effects).
Formal quality assessment was performed for each CST according to the aforementioned categories by developing an ad hoc checklist (Cognitive Screening Standardization Checklist, CSSC) (see Table 2). The CSSC encompasses two sections, “Sampling” (ranging 0–13) and “Psychometrics, diagnostics, and usability” (ranging 0–29). The first section evaluates the sampling adequacy as for representativeness; the second section focuses on psychometric, diagnostic properties, and feasibility. CSSC total scores range from 0 to 42; a given CST was thus judged as “statistically sound” if scoring ≥21 (i.e., 50% of the maximum) on the CSSC. CSSC items were based on [1, 2] and [7].
Scores were “cumulatively” assigned for each CST by evaluating all available studies on it among those included. Items targeting non-cumulative information which were nonetheless retrievable in multiple studies — e.g., the normative sample size — were scored according to that study providing the highest-quality information — e.g., the highest N.
Quality assessment was performed by one of the authors (S.R.) and supervised by a second, independent one (E.N.A.).
Results
Study selection process is shown in Fig. 1.
Sixty-one studies were ultimately included. Extracted outcomes are reported in Table 3. A summarization of most relevant psychometric and diagnostic properties for each included CST, along with CSSC scores, are reported in Table 4.
The vast majority of contributions were studies mostly aimed at providing normative data (N = 32) — of which, 11 did not report any further relevant statistical property. Twenty-seven studies instead focused on psychometric/diagnostic properties with only marginal/absent attention to normative values (either in the context of clinical usability or not).
Included CSTs fell under the following categories: (a) domain-/disease-nonspecific (in-person: N = 14; remote: N = 3); (b) domain-specific (N = 7), targeting executive functioning, language, memory, and praxis; and (c) disease-specific (N = 16), targeting neurodegenerative disorders (Alzheimer’s, Parkinson’s, and motor neuron diseases), cerebrovascular accidents, neuropsychiatric conditions, infective sub-cortical dementias, delirium, migraine-related subjective cognitive dysfunction, and dementia in the context of intellectual disabilities. Among all the investigations, target clinical populations were present in 33 studies.
Validity was investigated in 37 studies and mostly by convergence (N = 31); divergent validity was assessed in 5 studies, whereas criterion validity in 4 (3 of which via concurrent validity, whereas one via predictive validity). Only 5 of the included studies assessed the factorial structure underlying CSTs by means of dimensionality reduction approaches. No overt evidence of content, face, and ecological validity was detected.
Reliability was investigated in 32 studies and mostly as inter-rater (N = 17), internal consistency (N = 14), and test-retest (N = 12). Parallel forms were developed within 4 studies only.
Although sensitivity and specificity measures were often reported (N = 22), derived metrics (e.g., positive and negative predictive values and likelihood ratios) were provided in 10 studies only.
With respect to norming, regression-based and inferential-error-controlling methods — e.g., tolerance limits and/or Equivalent Scores [12], were highly represented (N = 26). Several studies (N = 17) derived point-estimate cut-offs through receiver-operating characteristic (ROC) analyses.
Acceptability of the CST was overtly examined in 9 studies, while ceiling/floor effects in 11. When applicable, administration time ranged from 2 to 45 min.
Discussion
The present work investigates statistical features of CSTs currently available in Italy, shedding a new light on their clinical and experimental utilization. Information here reported have the potential to promote a more aware and critical usage of CSTs among Italian clinicians, as well as to serve as overall guidelines for researchers either involved in CST development/adaptation/standardization or devoted to addressing open issues on CST psychometrics/diagnostics.
Overall, although psychometrics and diagnostics for a given CST happened not to be assessed within the same study, basic properties and norms were provided within different ones, especially for most widespread CSTs (e.g., Mini-Mental State Examination, MMSE; Montreal Cognitive Assessment, MoCA).
Moreover, although results show a general trend towards focusing on providing only normative data and cut-off values, the majority of included CSTs proved to be supported by sufficient evidence as for basic psychometric/diagnostic requirements are concerned. The present review hints at a relatively high quality of selected global and domain-specific CSTs — e.g., MoCA (CSSC = 34) and Addenbrooke’s Cognitive Examination — Revised (ACE-R; CSSC = 31), Screening for Aphasia in Neurodegeneration (SAND; CSSC = 27), and Frontal Assessment Battery (FAB; CSSC = 24). Disease-specific CSTs were shown to be less statistically robust, with a few exceptions — e.g., ALS Cognitive Behavioral Screen (ALS-CBS; CSSC = 26).
Validity
Findings on validity happened to show misinterpretations of the psychometric concepts and incomplete analyses for certain CSTs.
First, it is worth mentioning that convergent and concurrent validity happened to be mistaken for each other — e.g., concurrent validity being tested by means of correlations instead of regressions, or convergent and concurrent validity being addressed as the same construct [54, 55, 61, 69, 73].
In this regard, one should also note that correlational measures happened not to be meant to assess the same construct as that of the target CSTs — e.g., FAB validity being tested against the MMSE [24].
Moreover, predictive validity happened to be almost never assessed [29] — despite the longitudinal dimension being relevant to the monitoring of patients’ cognitive profile. This may be due to the high cost of performing a proper longitudinal study to assess predictive validity.
It is also worth mentioning that the vast majority of included CSTs lacked divergent validity evidence. This might be due to the fact that different CSTs are commonly found to correlate despite being meant to assess different functions; this because target constructs often overlap to some extent. Researchers are thus encouraged to test divergence by addressing measures that are supposed to deviate from a given CST as far as either construct or face validity is concerned. This could be done by comparing a CST with either a II-level, domain-specific cognitive test, or with a psychodiagnostic tool.
Furthermore, although the need for cognitive measures that are predictive of daily functioning has been highlighted [12], it has to be noted that the ecological validity of CSTs has never been found to be directly investigated within original standardization studies. This may be due to the lack of a wide consensus on how to investigate ecological validity, as well as to the scarce availability of ad hoc scales designed for assessing the specific impact of cognitive disorders on real-life functioning, going beyond a general evaluation of functional disability.
Finally, researchers should take into consideration to explore content validity and factorial structures of CSTs; this equally applying both to those tests postulated to be mono-factorial (e.g., MMSE) and to domain-specific ones (e.g., SAND), which might nonetheless cover multiple cognitive functions.
Reliability
Overall, reliability of Italian CSTs was frequently assessed, although often either incompletely or inefficiently.
When testing reliability of CSTs, it is worth bearing in mind that internal consistency might be problematic: indeed, different items within the same CST may be meant to measure different facets of cognition, this possibly being even truer for multi-domain tests such as the MMSE. This is an aspect that needs further developments.
By contrast, assessing reliability via test-retest or inter-rater may be generalizable to different CST categories and more practically relevant (e.g., clinicians are interested in knowing whether the CST yields similar outcomes/scores when administered in different conditions).
Furthermore, parallel-form reliability was seldom examined, and no CST came with information on its ability to detect significant change [6, 76]. Indeed, although parallel forms reduce the possibility to have “practice effect” (i.e., systematic performance improvements across consecutive assessments), the lack of appropriate methods for detecting clinically meaningful changes over time unrelated to practice has a crucial (even detrimental) impact whenever CSTs are meant to be used longitudinally to monitor the progress of cognitive functions or dysfunctions for either diagnostic or prognostic purposes. Indeed, without thresholds for significant change it is not possible to ascertain whether observed score variations over repeated measurements could be merely traced back to intrinsic and expected, physiological oscillations of performances, or whether they more likely reflect a true cognitive change (worsening or improvement).
Diagnostic properties
The study of the diagnostic properties of CSTs was often addressed within a nosographic-descriptive framework, which, however, might not always fit cognitive semiology [7]. Indeed, a one-to-one correspondence between cognitive profiles and neurological/neuropsychiatric conditions is often not straightforward [13]. Thereupon, the notion of “target condition” within ROC analyses may happen to be elusive, hence limiting the disease-specificity of certain CSTs [2, 7]. The present work indeed highlights the need for identifying more rigorous statistical methods for deriving the optimal cut-off values (e.g., the Youden statistics, an index identifying the best cut-off at the optimal compromise between sensitivity and specificity).
Moreover, although basic diagnostic properties happened to be investigated, less attention has been given to those selectively relevant to screening aims, such as taking into account disease prevalence (e.g., positive and negative predictive values) and allowing an estimation of post-test probability of cognitive impairment (e.g., positive and negative likelihood ratios) [3].
With respect to Italian CSTs, it is noteworthy that diagnostic properties were almost investigated only for 5 out of 16 disease-specific CSTs. Although evidence of case-control discrimination was frequently provided by means of between-group comparisons (e.g., ALS-CBS), it is recommended that sensitivity, specificity, and derived measures be tested in order to statistically substantiate CST applications to target conditions.
Norms
As far as normative data are concerned, although regression-based and inferential-error-controlling techniques were highly represented, it should be noted that a relatively high heterogeneity in approaching norming methods was detected. First, the Equivalent Score method happened to be embraced “incompletely,” by only computing tolerance limits only but not Equivalent Score thresholds [12]. Second, norms were occasionally derived via approaches assuming a normal distribution, this possibly undermining their adequacy as cognitive data often present with overdispersion and skewness [43]. With this last respect, checking for ceiling/floor effects in test scores is encouraged; unfortunately, this was rarely carried out in the studies herewith included.
With a few exceptions, sampling revealed to be overall adequate as far as typical/clinical sample sizes are concerned. However, it has to be noted that geographic coverage of normative samples was often circumscribed. Between-regional differences should nonetheless receive attention as being potential confounders in cognitive testing; this issue has been only recently addressed within the concerning Italian literature [66, 70].
Feasibility
Despite a core feature of a CST is a short administration time [2], it should be noted that several CSTs here reported require up to 20′ to be administered and scored (e.g., ACE-R), in turn limiting their usability in time-restricted settings (e.g., bedside evaluations). By contrast, these “in-depth CSTs” may be more adequate in outpatient settings [2].
With regard to cross-cultural/-linguistic adaptations, it has to be stressed that back-translation approaches have been seldom adopted and culture/language-related issues often not addressed. The latter aspect is of major interest especially for items assessing language functioning, which should instead undergo dedicated, country-specific controls for psycholinguistic predictors (e.g., word frequency for naming tasks) [10].
An overall need for more systematic evidence of acceptability/face validity of CSTs also emerged from the present work. This indeed would help practitioners select a test based on the administration setting, as for instance the assessment of acute patients would benefit from a short, tolerable CST that is clearly recognizable as such by the patient.
Limitations and perspectives
First, it has to be noted that the goodness of a CST is not exhausted in psychometric/diagnostic properties. Indeed, in order for a CST to be introduced in clinical practice, thorough evidence of its applicability in atypical populations should be provided. Moreover, it should be born in mind that evidence of both psychometric and diagnostic soundness may be inferred from applied studies as well. Thereupon, future studies should focus on reviewing available contributions on the clinical usability of Italian CTSs in order to provide a more comprehensive picture on their statistical/methodological quality.
Furthermore, Italian practitioners might benefit from a future review focused on psychometric/diagnostic properties of qualitative/proxy measures of cognition that were not addressed within the present study for generalizability reasons.
Although beyond the aim of this work, it should be then noted that more detailed item-level analyses (Item Response Theory) were conducted in only one of the records included [66]. As being able to provide insights into adaptive testing as well as to help ease interpretations issues, Item Response Theory-based analyses should be taken into consideration when assessing psychometrics/diagnostic properties of CSTs [66].
Finally, it is important to underline that, to the best of the authors’ knowledge, there is not an official, worldwide consensus on the relevant properties to be addressed in cognitive screening, this resulting in the choices being possibly incomplete or selectively reflecting the knowledge of researchers. This latter consideration stresses the importance of developing wider agreement within neurological/neuropsychological societies to ensure higher standards and raise the awareness on the impact of statistical properties on the applicability of CSTs in both applied (e.g., clinical and forensic) and research contexts.
Conclusion
The present work shows that, although available Italian CSTs overall met basic psychometric/diagnostic requirements, their statistical profile often proved to be deficient on several properties that are desirable/needed for clinical applications, with a few exceptions among general and domain-specific CSTs yielding high soundness, namely, the MoCA and ACE-R, and the FAB and SAND, respectively. In particular, this work highlights that:
-
- psychometric/diagnostic properties of disease-specific CSTs happened were poorly examined;
-
- construct and criterion validity should be differentiated and assessed separately;
-
- factorial structure underlying CSTs should be tested for both general and domain-specific ones;
-
- ecological validity of CSTs need to be addressed to provide information relevant to patients’ everyday functioning;
-
- significant change thresholds and alternate versions of CSTs need to be developed in order to improve their longitudinal usage;
-
- a general lack of investigations on sensitivity-/specificity-derived diagnostic metrics selectively relevant to screening aims (i.e., positive and negative predictive values and likelihood ratios) was detected;
-
- a clearer definition of target conditions for a given CST is needed, especially for those thought to be disease-specific;
-
- information on CST acceptability, face validity, and administration time are desirable, as helping an ad hoc usage by practitioners select.
Change history
23 July 2022
A Correction to this paper has been published: https://doi.org/10.1007/s10072-022-06294-3
References
Larner AJ. Introduction to cognitive screening instruments: rationale and desiderata. In A. J. Larner (Ed.). Cognitive screening instruments: a practical approach (2nd ed., pp 3–14). Springer; 2017a.
Larner AJ. The usage of cognitive screening instruments: test characteristics and suspected diagnosis. In A. J. Larner (Ed.). Cognitive screening instruments: a practical approach (2nd ed., pp 315–339). Springer; 2017b.
Bolin E, Lam W. A review of sensitivity, specificity, and likelihood ratios: evaluating the utility of the electrocardiogram as a screening tool in hypertrophic cardiomyopathy. Congenit Heart Dis 2013;8:406–10.
Grigoletto F, Zappalà G, Anderson DW, Lebowitz BD. Norms for the Mini-Mental State Examination in a healthy population. Neurol 1999;53:315–20.
Crawford JR, Garthwaite PH. Comparing patients’ predicted test scores from a regression equation with their obtained scores: a significance test and point estimate of abnormality with accompanying confidence limits. Neuropsychology 2006;20:259.
Arcara G, Bambini V. A test for the assessment of pragmatic abilities and cognitive substrates (APACS): normative data and psychometric properties. Front Psychol 2016;7:70.
Quinn TJ, Takwoingi Y. Assessment of the utility of cognitive screening instruments. In A. J. Larner (Ed.). Cognitive Screening instruments: a practical approach (2nd ed., pp 15–36). Springer; 2017.
Rohling ML, Axelrod BN, Langhinrichsen-Rohling J. Fundamental forensic statistics: statistics every forensic neuropsychologist must know. In: Bush SS, Demakis GJ, Rohling ML (eds) APA Handbook of Forensic Neuropsychology. American Psychiatric Association Publishing; 2017. p. 3–22.
Pottie K, Rahal R, Jaramillo A, Birtwhistle R, Thombs BD, Singh H, Gorber SC, Dunfield L, Shane A, Bacchus M, Bell N. Recommendations on screening for cognitive impairment in older adults. CMAJ 2016;188:37–46.
Uysal-Bozkir Ö, Parlevliet JL, de Rooij SE. Insufficient cross-cultural adaptations and psychometric properties for many translated health assessment scales: a systematic review. J Clin Epidemiol 2013;66:608–18.
Pedraza O, Mungas D. Measurement in cross-cultural neuropsychology. Neuropsychol Rev 2008;18:184–93.
Bianchi A. L’esame neuropsicologico dell’adulto: Applicazioni cliniche e forensi. Giunti Psychometrics; 2013.
Mondini S, Mapelli D, Arcara G. Semeiotica e diagnosi neuropsicologica: metodologia per la valutazione. Carrocci; 2016.
Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. Preferred Reporting Items for Systematic Reviews and Meta-Analyses: the PRISMA statement. J Clin Epidemiol 2009;62:1006–12.
Mazzoni M, Ferroni L, Lombardi L, Del Torto E, Vista M, Moretti P. Mini-Mental State Examination (MMSE): sensitivity in an Italian sample of patients with dementia. Ital J Neurol Sci 1992;13:323–29.
Rocca WA, Bonaiuto S, Lippi A, Luciani P, Pistarelli T, Grandinetti A, Cavarzeran F, Amaducci L. Validation of the Hodkinson abbreviated mental test as a screening instrument for dementia in an Italian population. Neuroepidemiology 1992;11:288–95.
Measso G, Cavarzeran F, Zappalà G, Lebowitz BD, Crook TH, Pirozzolo FJ, Amaducci L, Massari D, Grigoletto F. The Mini-Mental State Examination: normative study of an Italian random sample. Dev Neuropsychol 1993;9:77–85.
Fioravanti M, Nacca D, Buckley AE, Ferrario E, Varetto O, Mogni P, Fabris F. The Italian version of the Alzheimer’s Disease Assessment Scale (ADAS): psychometric and normative characteristics from a normal aged population. Arch Gerontol Geriatr 1994;19:21–30.
Magni E, Binetti G, Bianchetti A, Rozzini R, Trabucchi M. Mini-Mental State Examination: a normative study in Italian elderly population. Eur J Neurol 1996;3:198–202.
Ferrucci L, Del Lungo I, Guralnik JM, Bandinelli S, Benvenuti E, Salani B, Lamponi M, Ubezio C, Benvenuti F, Baroni A. Is the telephone interview for cognitive status a valid alternative in persons who cannot be evaluated by the Mini Mental State Examination? Aging Clin Exp Res 1998;10:332–38.
Cossa FM, Sala SD, Musicco M, Spinnler H, Ubezio MC. The Milan Overall Dementia Assessment and the Mini-Mental State Examination compared: an epidemiological investigation of dementia. Eur J Neurol 1999;6:289–94.
Metitieri T, Geroldi C, Pezzini A, Frisoni GB, Bianchetti A, Trabucchi M. The Itel-MMSE: an Italian telephone version of the Mini-Mental State Examination. Int J Geriatr Psychiatry 2001;16:166–67.
Iavarone A, Ronga B, Pellegrino L, Loré E, Vitaliano S, Galeone F, Carlomagno S. The Frontal Assessment Battery (FAB): normative data from an Italian sample and performances of patients with Alzheimer’s disease and frontotemporal dementia. Funct Neurol 2004;19:191–95.
Appollonio I, Leone M, Isella V, Piamarta F, Consoli T, Villa ML, Forapani E, Russo A, Nichelli P. The Frontal Assessment Battery (FAB): normative values in an Italian population sample. Neurol Sci 2005;26:108–16.
Dal Forno G, Chiovenda P, Bressi F, Ferreri F, Grossi E, Brandt J, Rossini PM, Pasqualetti P. Use of an Italian version of the telephone interview for cognitive status in Alzheimer’s disease. Int J Geriatr Psychiatry 2006;21:126–33.
Michieletto F, Binkin N, Saugo M, Boorson S, Scanlan J. Utilizzo del test Mini-Cog come metodo di screening per la demenza nella popolazione italiana: l’esperienza dello studio Argento. Sanità Pubblica 2006;62:159–72.
Vanacore N, De Carolis A, Sepe-Monti M, Bomboi G, Stazi A, Bianchetti A, Giubilei F. Validity of the Italian telephone version of the mini-mental state examination in the elderly healthy population. Acta Neurol Belg 2006;106:132–36.
Anselmetti S, Poletti S, Ermoli E, Bechi M, Cappa S, Venneri A, Smeraldi E, Cavallaro R. The brief assessment of cognition in schizophrenia. Normative data for the Italian population. Neurol Sci 2008;29:85–92.
Lunardelli A, Mengotti P, Pesavento A, Sverzut A, Zadini A. The Brief Neuropsychological Screening (BNS): valuation of its clinical validity. Eur J Phys Rehabil Med 2009;45:85–91.
Caffarra P, Gardini S, Zonato F, Concari L, Dieci F, Copelli S, Freedman M, Stracciari A, Venneri A. Italian norms for the Freedman version of the Clock Drawing Test. J Clin Exp Neuropsychol 2011;33:982–88.
Girtler N, Brugnolo A, Campus C, De Carli F, Famà F, Ghio L, Robert P, Barbieri MP, Delmonte M, Venuti S, Gabrielli F, Rodriguez G, Nobili F. The short cognitive evaluation battery in cognitive disorders of the elderly-Italian version. Dement Geriatr Cogn Disord 2012;33:255–65.
Costa A, Bagoj E, Monaco M, Zabberoni S, De Rosa S, Mundi C, Caltagirone C, Carlesimo GA. Mini mental Parkinson test: standardization and normative data on an Italian sample. Neurol Sci 2013;34:1797–803.
Isella V, Mapelli C, Morielli N, De Gaspari D, Siri C, Pezzoli G, Antonini A, Poletti M, Bonuccelli U, Picchi L, Napolitano A, Vista M, Veglia M, Piamarta F, Grassi F, Appollonio IM. Psychometric properties of the Italian version of the Scales for Outcomes in Parkinson’s disease-Cognition (SCOPA-Cog). Funct Neurol 2013;28:121.
Isella V, Mapelli C, Morielli N, Siri C, De Gaspari D, Pezzoli G, Antonini A, Poletti M, Bonuccelli U, Picchi L, Napolitano A. Diagnosis of possible mild cognitive impairment in Parkinson’s disease: validity of the SCOPA-Cog. Parkinsonism Relat Disord 2013;19:1160–63.
Timpano F, Pirrotta F, Bonanno L, Marino S, Marra A, Bramanti P, Lanzafame P. Videoconference-based mini mental state examination: a validation study. Telemed J E Health 2013;19:931–37.
Bellelli G, Morandi A, Davis DH, Mazzola P, Turco R, Gentile S, Ryan T, Cash H, Guerini F, Torpillesi T, Del Santo F, Trabucchi M, Annoni G, MacLullich AM. Validation of the 4AT, a new instrument for rapid delirium screening: a study in 234 hospitalised older people. Age Ageing 2014;43:496–502.
Pignatti R, Bertella L, Scarpina F, Mauro A, Portolani E, Calabrese P. Italian version of the Parkinson Neuropsychometric Dementia Assessment (PANDA): a useful instrument to detect cognitive impairments in Parkinson’s disease. J Parkinsons Dis 2014;4:151–160.
Santangelo G, Barone P, Abbruzzese G, Ferini-Strambi L, Antonini A, IRIS Study Group. Validation of the Italian version of Parkinson’s disease-cognitive rating scale (PD-CRS). Neurol Sci 2014;35:537–44.
Conti S, Bonazzi S, Laiacona M, Masina M, Coralli MV. Montreal Cognitive Assessment (MoCA)-Italian version: regression based norms and equivalent scores. Neurol Sci 2015;36:209–214.
Pigliautile M, Chiesi F, Rossetti S, Conestabile Della Staffa M, Ricci M, Federici S, Chiloiro D, Primi C, Mecocci P. Normative data for the ACE-R in an Italian population sample. Neurol Sci 2015;36:2185–90.
Pirrotta F, Timpano F, Bonanno L, Nunnari D, Marino S, Bramanti P, Lanzafame P. Italian validation of Montreal cognitive assessment. Eur J Psychol Assess 2015;31:131–37.
Santangelo G, Siciliano M, Pedone R, Vitale C, Falco F, Bisogno R, Siano P, Barone P, Grossi D, Santangelo F, Trojano L. Normative data for the Montreal Cognitive Assessment in an Italian population sample. Neurol Sci 2015;36:585–91.
Tessari A, Toraldo A, Lunardelli A, Zadini A, Rumiati RI. STIMA: a short screening test for ideo-motor apraxia, selective for action meaning and bodily district. Neurol Sci 2015;36:977–84.
Mancuso M, Varalta V, Sardella L, Capitani D, Zoccolotti P, Antonucci G, Italian OCS Group. Italian normative data for a stroke specific cognitive screening tool: the Oxford Cognitive Screen (OCS). Neurol Sci 2016;37:1713–21.
Poletti B, Solca F, Carelli L, Madotto F, Lafronza A, Faini A, Monti A, Zago S, Calini D, Tiloca C, Doretti A, Verde F, Ratti A, Ticozzi N, Abrahams S, Silani V. The validation of the Italian Edinburgh Cognitive and Behavioural ALS Screen (ECAS). Amyotrophic lateral Sclerosis & Frontotemporal Degeneration 2016;17:489–98.
Ricci M, Pigliautile M, D’Ambrosio V, Ercolani S, Bianchini C, Ruggiero C, Vanacore N, Mecocci P. The clock drawing test as a screening tool in mild cognitive impairment and very mild dementia: a new brief method of scoring and normative data in the elderly. Neurol Sci 2016;37:867–73.
Siciliano M, Raimo S, Tufano D, Basile G, Grossi D, Santangelo F, Trojano L, Santangelo G. The Addenbrooke’s Cognitive Examination Revised (ACE-R) and its sub-scores: normative values in an Italian population sample. Neurol Sci 2016;37:385–92.
Siciliano M, Santangelo G, D’Iorio A, Basile G, Piscopo F, Grossi D, Trojano L. Rouleau version of the Clock Drawing Test: age- and education-adjusted normative data from a wide Italian sample. Clin Neuropsychol 2016;30:1501–16.
Bosco A, Spano G, Caffò AO, Lopez A, Grattagliano I, Saracino G, Pinto K, Hoogeveen F, Lancioni GE. Italians do it worse. Montreal Cognitive Assessment (MoCA) optimal cut-off scores for people with probable Alzheimer’s disease and with probable cognitive impairment. Aging Clin Exp Res 2017;29:1113–20.
Catricalà E, Gobbi E, Battista P, Miozzo A, Polito C, Boschi V, Esposito V, Cuoco S, Barone P, Sorbi S, Cappa SF, Garrard P. SAND: a Screening for Aphasia in NeuroDegeneration. Development and normative data Neurological Sciences 2017;38:1469–83.
Santangelo G, Lagravinese G, Battini V, Chiorri C, Siciliano M, Abbruzzese G, Vitale C, Barone P. The Parkinson’s Disease-Cognitive Rating Scale (PD-CRS): normative values from 268 healthy Italian individuals. Neurol Sci 2017;38:845–53.
Siciliano M, Trojano L, Trojsi F, Greco R, Santoro M, Basile G, Piscopo F, D’Iorio A, Patrone M, Femiano C, Monsurrò M, Tedeschi G, Santangelo G. Edinburgh Cognitive and Behavioural ALS Screen (ECAS)-Italian version: regression based norms and equivalent scores. Neurol Sci 2017;38:1059–68.
Battista P, Catricalà E, Piccininni M, Copetti M, Esposito V, Polito C, Miozzo A, Gobbi E, Cuoco S, Boschi V, Picillo M, Sorbi S, Barone P, Iannaccone S, Garrard P, Logroscino G, Cappa SF. Screening for aphasia in neurodegeneration for the diagnosis of patients with primary progressive aphasia: clinical validity and psychometric properties. Dement Geriatr Cogn Disord 2018;46:243–52.
Crivelli D, Angelillo MT, Grippa E, Colucci A, Nardulli R, Balconi M. When is a novel psychometric measure needed? A preliminary analysis regarding the Cognitive Assessment for Stroke Patients (CASP) battery compared with MMSE and MoCA. Appl Neuropsychol Adult 2018;25:410–16.
Iavarone A, Carpinelli Mazzi M, Russo G, D’Anna F, Peluso S, Mazzeo P, De Luca V, De Michele G, Iaccarino G, Abete P, Milan G, Garofalo E, Musella C, O’Caoimh R, Molloy W, De Joanna G, Manzo V, Ambra FI, Postiglione A, Illario M, Working Group. The Italian version of the quick mild cognitive impairment (Qmci-I) screen: normative study on 307 healthy subjects. Aging Clin Exp Res 2019;31:353–60.
Panebianco M, Zavanone C, Dupont S, Perrone P, Pavone A, Napolitano S, Leger A, Samson Y. The inter-rater reliability of the Italian version of Aphasia Rapid Test (ART) for acute ischemic stroke. Neurol Sci 2019;40:2171–74.
Pigliautile M, Chiesi F, Stablum F, Rossetti S, Primi C, Chiloiro D, Federici S, Mecocci P. Italian version and normative data of Addenbrooke’s Cognitive Examination III. Int Psychogeriatr 2019;31:241–49.
Siciliano M, Chiorri C, Passaniti C, Sant’Elia V, Trojano L, Santangelo G. Comparison of alternate and original forms of the Montreal Cognitive Assessment (MoCA): an Italian normative study. Neurol Sci 2019;40:691–702.
Smirni D, Smirni P, Di Martino G, Fontana ML, Cipolotti L, Oliveri M, Turriziani P. Early detection of memory impairments in older adults: standardization of a short version of the verbal and nonverbal recognition memory test. Neurol Sci 2019;40:97–103.
Belvederi Murri M, Folesani F, Costa S, Morelli AC, Scillitani V, Guaiana G, Biancosino B, Caruso R, Nanni MG, Zerbinati L, Purdon SE, Grassi L. Italian validation of the screen for cognitive impairment in psychiatry. Community Ment Health J 2020;56:1411–18.
Muò R, Raimondo S, Martufi F, Cavagna N, Bassi M, Schindler A. Cross-cultural adaptation and validation of the Italian Aachener Aphasie Bedside Test (I-AABT), a tool for Aphasia assessment in the acute phase. Aphasiology. 2020;1–24.
Pasotti F, Magnani FG, Gallucci M, Salvato G, Ovadia D, Scotto M, Merolla S, Beretta S, Micieli GR, Agostoni EC, Beretta G, Bottini G. Neuropsychological assessment in acute stroke patients. Neurol Sci 2020;41:1259–66.
Russo A, Silvestro M, Garramone F, Tessitore A, Cropano M, Scotto di Clemente F, Trojano L, Tedeschi G, Santangelo G. A subjective cognitive impairments scale for migraine attacks: validation of the Italian version of the MIG-SCOG. Neurol Sci 2020;41(5):1139–43.
Tremolizzo L, Lizio A, Santangelo G, Diamanti S, Lunetta C, Gerardi F, Messina S, La Foresta S, Riva N, Falzone Y, Filippi M, Woolley SC, Sansone VA, Siciliano M, Ferrarese C, Appollonio I, ALS-CBS Italian Study Group. ALS Cognitive Behavioral Screen (ALS-CBS): normative values for the Italian population and clinical usability. Neurol Sci 2020;41:835–41.
Aiello EN, Esposito A, Gramegna C, Gazzaniga V, Zago S, Difonzo T, Appollonio I, Bolognini N. The Frontal Assessment Battery (FAB) and its sub-scales: validation and updated normative data in an Italian population sample. Neurol Sci 2021a;1–6.
Aiello EN, Gramegna C, Esposito A, Zago S, Difonzo T, Maddaluno O, Appollonio I, Bolognini N. The Montreal Cognitive Assessment (MoCA): updated norms and psychometric insights into adaptive testing from healthy individuals in Northern Italy. Aging Clin Exp Res. 2021b;1–8.
Barulli MR, Piccininni M, Brugnolo A, Musarò C, Di Dio C, Capozzo R, Tortelli R, Lucca U, Logroscino G. The Italian Version of the Test Your Memory (TYM-I): a tool to detect mild cognitive impairment in the clinical setting. Front Psychol 2021;11:614920.
De Vreese LP, Gomiero T, De Bastiani E, Marangoni A, Weger E, Mantesso U (2021) Short forms of Prudhoe Cognitive Function Test in adults and aging people with intellectual disabilities: Italian validation study. J Intellect Disability Res: JIDR 65:162–72.
Montanucci C, Chipi E, Salvadori N, Rinaldi R, Eusebi P, Parnetti L. HIV-Dementia Scale as a screening tool for the detection of subcortical cognitive deficits: validation of the Italian version. J Neurol. 2021;1–7.
Carpinelli Mazzi M, Iavarone A, Russo G, Musella C, Milan G, D’Anna F, Garofalo E, Chieffi S, Sannino M, Illario M, De Luca V, Postiglione A, Abete P, with the support of the Working group. Mini-Mental State Examination: new normative values on subjects in Southern Italy. Aging Clin Exp Res 2020;32:699–702.
Brazzelli M, Capitani E, Della Sala S, Spinnler H, Zuffi M. A neuropsychological instrument adding to the description of patients with suspected cortical dementia: the Milan overall dementia assessment. J Neurol Neurosurg Psychiatry 1994;57:1510–17.
Petrazzuoli F, Palmqvist S, Thulesius H, Buono N, Pirrotta E, Cuffari A, Cambielli M, D’Urso M, Farinaro C, Chiumeo F, Marsala V, Wiig EH. A Quick Test of Cognitive Speed: norm-referenced criteria for 121 Italian adults aged 45 to 90 years. International Psychogeriatrics. 2014;1–8.
Pigliautile M, Ricci M, Mioshi E, Ercolani S, Mangialasche F, Monastero R, Croce MF, Federici S, Mecocci P. Validation study of the Italian Addenbrooke’s Cognitive Examination Revised in a young-old and old-old population. Dement Geriatr Cogn Disord 2011;32:301–7.
Pirani A, Brodaty H, Martini E, Zaccherini D, Neviani F, Neri M. The validation of the Italian version of the GPCOG (GPCOG-It): a contribution to cross-national implementation of a screening test for dementia in general practice. Int Psychogeriatr 2010;22:82–90.
Taule T, Søvik M, Lein RK, Wehling E, Aßmus J, Rekand T. Psychometric properties of cognitive assessment in amyotrophic lateral sclerosis: a systematic review. Patient Related Outcome Measures 2020;11:181.
Jacobson NS, Roberts LJ, Berns SB, McGlinchey JB. Methods for defining and determining the clinical significance of treatment effects: description, application, and alternatives. J Consult Clin Psychol 1999;67:300.
Acknowledgements
The authors would like to thank Dr. Francesca Crespi for her contributions.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Ethical approval
None.
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Aiello, E.N., Rimoldi, S., Bolognini, N. et al. Psychometrics and diagnostics of Italian cognitive screening tests: a systematic review. Neurol Sci 43, 821–845 (2022). https://doi.org/10.1007/s10072-021-05683-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10072-021-05683-4