Introduction

The sociodemographic characteristics of HIV at-risk populations vary across the world (UNAIDS 2012). This represents a major challenge for reliably estimating the prevalence of HIV-associated neurocognitive disorder (HAND). In Australia, the main risk group for HIV infection remains men who have sex with men (MSM; >80 %) (The Kirby Institute 2012), and among those many have high premorbid functioning (Law et al. 2011). The Australian HIV epidemic is also characterized by a relatively good treatment response for HIV-infected (HIV+) persons and one of the highest nationally reported rates of viral undetectability for those on combination antiretroviral therapy (CART) (>85 % undetectable) (Law et al. 2011; The Kirby Institute 2012). However, the recently released 2013 Australian HIV Surveillance report indicates that out of the total HIV population who needs to be on CART, only 53 % are (The Kirby Institute 2013).

It is in this context that the current study takes place. The original aim was to assess the effect of HIV and aging on brain functions, as the HIV+ persons in Australia are surviving through older age (Cysique et al. 2013; Lane et al. 2012). Therefore, the current study is by design composed of middle-aged men (median age = 55) with chronic HIV infection (median HIV duration = 20 years) and with stable CART.

As for similar cohorts in other parts of the world, for example, the US MACS cohort (Sacktor et al. 2002) and the North West London cohort (Garvey et al. 2011), a majority of the HIV+ MSM in these cohorts are highly educated (both formal and informal education) and high functioning in addition to being successfully treated. In these cohorts, recent reports of long-term stability of neurocognitive functioning (Cole et al. 2007) and low prevalence of HAND (Garvey et al. 2011) may have been over-generalized in that they were interpreted as evidence of low disease morbidity in general and due to successful CART.

In Western English-speaking countries (Australia, UK, and Canada), and in neuropsychology beyond the field of neuroHIV, there is a reliance on US normative standards to interpret neuropsychological performance. The US normative samples typically represent the current American demographic in terms of age, education, gender, and ethnicity (including migration history and acculturation (Manly et al. 2004)). These norms therefore intrinsically embed performance levels that, while representing the general American population, may not be appropriate for specific subpopulations in the USA and in other countries; this issue that has been recently raised in a UK neuroHIV study (Winston et al. 2013).

In all cases, the use of appropriate normative data is central to valid interpretation of neurocognitive performance. There are two schools of thought in neuropsychology on this matter. Some have advocated the use of large samples, which provide uncorrected and corrected normative standards (corrections for the main demographic effects that are relevant to a given country) (Heaton et al. 2004b). Others argue that relatively smaller local samples, which are as close as possible to the demographic and other (cultural, intellectual) characteristics of a clinical sample are more appropriate (Mitrushina et al. 2005). In many countries, there are no population norms and therefore a matched local control group may be the best option. There are three main issues with having local norms as the default choice in Australia and these are as follows: (1) The use of local comparison groups has variable and uncontrolled ascertainment biases by nature (Williams and Cottle 2011), and tends to be based on smaller samples because of relatively restricted research funding (Australian Neuroscience Society 2010). (2) The collection of a relatively small local sample for a study demands a lot of resources for relatively limited practical use. (3) Relatively small local norms may include sample-specific demographic or other background effects that are otherwise unusual in larger normative samples.

Because of the global nature of neuroHIV research, these issues demand making difficult choices regarding the selection of normative standards. These choices may sometimes be suboptimal and this is probably reflected in the wide variability of HAND prevalence data that have been published so far in the CART era (Kamminga et al. 2013). In Australian neuroHIV research, there are two choices: using US norms or collecting local control data keeping in mind particular HIV demographics as mentioned above.

In assessing the effect of HIV and aging on brain functions in the current cohort of HIV+ and HIV-uninfected (HIV−) healthy controls, which is mainly composed of MSM with high premorbid functioning, we noted that the US normative standards performed suboptimally (see Lane et al. 2012). In the current paper, we therefore assessed whether the use of American population norms versus a local normative reference affected the validity, the nature and the profile of HAND.

The aims of our study were as follows:

  1. 1.

    To determine the rate of neuropsychological impairment in Australian HIV− versus HIV+ individuals using US normative standards (US norms);

  2. 2.

    To compare the HIV effects on neuropsychological impairment generated from the US normative standards versus impairment generated from the demographically comparable local HIV− control group (local norms);

  3. 3.

    To determine which clinical disease and laboratory markers, as well as non-HIV markers impairment risks (i.e., cardiovascular diseases as the cohort was middle-aged per design) predict the degree of impairment and the presence of impairment.

Methods

Participants

Between June 2009 and July 2012, 139 males aged 45 years old and above were enrolled into the HIV and Brain Aging Research Program, a prospective study investigating the effects of HIV infection on the brain in middle-aged persons. Of those 139 participants, 90 were HIV+ and 49 HIV−. HIV+ participants were recruited through the HIV and Neurology Clinics at St. Vincent’s Hospital, Sydney. Fifteen HIV+ individuals had been diagnosed with HAND at some point in the past and our analyses were conducted with and without these cases. Eligible HIV+ participants were: at least 45 years old, on stable CART for at least 6 months, had a nadir CD4 equal to or below 350 cells/mm3, had known HIV duration equal to or greater than 5 years, and had no active opportunistic disease. The 49 HIV− controls were recruited in the metropolitan area of Sydney using community newspapers, community websites, as well as posters and brochures placed at St. Vincent’s Hospital, the University of New South Wales campuses, and the Holdsworth House general practice in Darlinghurst, Sydney. To meet eligibility criteria, control participants were required to be HIV negative on an ELISA test within the past 3 months. Moreover, they were screened for any history of acute cardiovascular disease (CVD) events (see Table 2 legend for exact definition). The majority of the control group participants self-identified as MSM (>85 %), which probably reflects that participants “self-selected” to support research involving their HIV+ counterparts.

Both HIV− and HIV+ participants were excluded from the study based on the following criteria: having a history of neurological disorders predating HIV diagnosis in the HIV+ group such as epilepsy, traumatic brain injury, Parkinson’s Disease, Multiple Sclerosis, Alzheimer’s disease, or Vascular Dementia; or psychiatric disorders on the psychotic axis (e.g., schizophrenia); current substance use disorders (within 12 months of study enrolment); history of loss of consciousness (≥30 min); and being non-proficient in English (see Lane et al. 2012 for English proficiency assessment details). All participants with English as a second language (N = 4 in HIV+ group and N = 2 in HIV− group; see Table 1) considered themselves fluent and stated that they preferred to be assessed in English. Participants were not excluded on the basis of current depressive symptoms or past substance use disorders that predated study entry by 12 months. Recreational use of marijuana was not set as a criterion for exclusion because it would exclude a large number of HIV+ individuals. Hepatitis C (HCV) status was recorded from the participants’ medical records (HCV is actively screened both in tertiary and primary care in Australia) and if HCV+, participants were included only if successfully treated and/or inactive: N = 2 in the HIV+ group (inactive HCV was defined as undetectable HCVRNA). All participants were assessed at their preferred time and without current illnesses that would prevent optimal cognitive performance

Table 1 Demographic characteristics in the HIV+ and HIV− groups

Ethics

All individuals signed an informed consent before participating in the study and the St. Vincent’s Hospital and The University of New South Wales Human Research Ethics Committees approved our protocol.

Procedure

Detailed procedures are reported in Lane et al. (2012). Briefly, we used a standard neuropsychological test battery covering seven ability domains, and that is in widespread use for NeuroAIDS research in the USA (Heaton et al. 2010; Table S1 in supplemental file).

Data analysis

To assess our first aim, we used a battery-wide summary score: the Global Deficit Score (GDS) (Blackstone et al. 2012; Carey et al. 2004; Heaton et al. 2004b). The standard GDS cutoff of ≥0.5 is an established definition of impairment for HIV-related brain injury and that virtually guarantees meeting the international criteria for HAND (Antinori et al. 2007; Blackstone et al. 2012). We first developed the impairment rate computations based on American population norms (US norms). Using US population normative data (Brandt and Benedict 2001; Heaton et al. 2004b), raw scores were converted to normally distributed and demographically uncorrected scaled scores and then demographically corrected T scores (see norming procedure details in supplemental files). T scores were then transformed into deficit scores and averaged to create the GDS. Second, we developed the impairment rate computations based on local Australian scaled scores (local norms). To develop a local normative reference, we used the 49 HIV− control participants’ demographically uncorrected scaled scores (see Table S2) and transformed them into z scores. z scores were then transformed into deficit scores and averaged to create the GDS.

Note that scaled scores, T scores, and z scores are standard scores that have the same linear properties. Moreover, the GDS can be computed from both T scores and z scores (see GDS computation details in Supplemental files). Therefore, for both the US and local norms, we used a GDS greater than or equal to 0.5 as it has typically been used as a cutoff for defining “impairment.” This cutoff indicates that, on average, an individual is at least mildly impaired on at least half of the tests in the battery (Heaton et al. 2004b). When possible, and for clarity of the results and illustrations, we refer to US norm-based GDS or T scores versus local norm-based GDS or z scores.

To assess our second aim, we used the following two strategies: First, we examined the effects of demographics on neuropsychological performance in the HIV− group only. We conducted Pearson correlations between age and education separately and the uncorrected global scaled score, and then used the US norm-based global T scores. These global scores represent an average of the neuropsychological performance across the test battery and are normally distributed. Second, to compare the HIV effects on neuropsychological impairment generated from the USA vs. local norms, we generated Cohen’s d effect sizes between HIV− and HIV+ groups on the seven cognitive domains mean T scores and global T score as well as on the seven cognitive domains mean z scores and global z score.

To assess our third aim, we used the GDS within the HIV+ group as the dependent variable and the disease and laboratory characteristics as well as Instrumental Activities of Daily Living (IADL) and depressive symptoms as the independent variables. The analyses were run using the two types of GDS (US normed versus local normed GDS), as dichotomous and continuous outcomes in order to determine the effects of predictors on the presence and severity of impairment (using χ 2, or logistic regression, and t test, or linear regression as appropriate). The current cohort was recruited to assess the effect of aging and HIV on the brain. One aspect of brain aging in HIV+ middle-aged individuals that is being increasingly recognized is the effect of CVD (Wright et al. 2010); hence this was included in the current study (see Table 2 and legend).

Table 2 Disease and laboratory characteristics in the HIV+ group

For this study, statistical significance was set at the conventional p ≤ 0.05 and effect sizes were used to provide transparency regarding magnitude of differences independently of the p values. Effect sizes (Cohen’s d) and coefficient of correlation (r) were computed using the Effect Size Determination Program (Lipsey and Wilson 2001). Statistical analyses were conducted using the statistical package JMP 10, 2013 SAS Institute Inc.

Results

Demographic characteristics of the study samples are presented in Table 1, and HIV disease, HIV treatment, and CVD risk characteristics are presented in Table 2.

Note also that 15/90 (17 %) cases had been diagnosed with past HAND; therefore, the prevalence data was determined with and without those 15 cases.

First aim: rate of neuropsychological impairment in Australian HIV− versus HIV+ individuals

Figure 1 presents the global impairment rates for the two study groups using the two types of impairment definitions. Panel a provides the data for the entire group and panel b provides the data after exclusion of past HAND cases. We found that the local norm-based GDS provided significantly higher rates of impairment in both groups. All impaired cases detected by US norm-based GDS were also impaired by local norm-based GDS (17.8 %), while 40 % were impaired (Panel A) as classified by local norm-based GDS but not by the US norm-based GDS.

Fig. 1
figure 1

Neuropsychological impairment prevalence rates in the study samples. Panel a entire HIV+ group. Panel b HIV+ group with past HAND cases excluded. Panels a and b within each of the scoring procedures, the USA versus local norms yield to significant differences (p < 0.001) in HAND overall impairment rate. Panel a age was significantly associated with impaired status on the local norm-based GDS (mean age impaired = 58.38 ± 8.02; unimpaired = 52.92 ± 5.65; p = 0.0003), but not on the US norm-based GDS (mean age impaired = 54.76 ± 6.55; unimpaired = 56.36 ± 7.80; p = 0.40). Education was not different between impairment definitions (p > 0.45)

Second aim: to compare the HIV effects on neuropsychological impairment generated from the US norm-based GDS vs. local norm-based GDS

Because demographic effects can mask or influence any HIV effect, we first estimated the magnitude of any demographic effects in the HIV− group. Table 3 presents the correlations between age, education, and the uncorrected scaled scores versus demographically corrected US T scores. As expected, age exerted a negative influence on the uncorrected scores, while education exerted a positive influence on neuropsychological performance. When considering the US T scores, we observed that the demographic corrections successfully diminished demographic influences, but actually showed small, non-significant trends towards over-corrections. That is, compared to the US standards, the Australian HIV− group performed slightly better than expected for their age and education.

Table 3 Age and education correlations on the uncorrected global scaled score in the 49 HIV− individuals

Figure 2 presents the standardized differences (Cohen’s d effect sizes) between the HIV− and HIV+ groups for the domain and global T scores (US norms) as well as z scores (local norms). The two sets of norms generated overall a similar profile. But the magnitude of effect sizes was greater when the local norms were used.

Fig. 2
figure 2

Effect sizes for the domain and global mean US norm-based and local norm-based scores between the HIV− and HIV+ group

Third aim: relationships between impairment definitions, disease markers, and clinical significance, Table 4

Considering the HIV disease factors first, we found that historical AIDS was associated with impaired status using the US norm-based GDS (p = 0.05). Second, we found significant effects of CVD risk scores (which have a large age component) when using the local norm-based GDS (p = 0.003). However, age is a major component of the CVD risk score (Table 2 legend) and because age remains an uncorrected effect in the local norm-based GDS, the significant association between CVD and the this impairment definition was found to be essentially driven by this “age-CVD effect” (see Table 4 legend for further details on this result).

Table 4 Disease and laboratory characteristics predictions of the GDS-based impairment (Correlation coefficients r are reported)

Finally, when considering the clinical significance of impairment status, we found that a decrease in IADL independence was associated with worse cognitive functioning on both US and local norm-based impairment definitions (Table 4). In contrast, we did not detect an effect of depressive complaints on impairment status.

Discussion

Our study assessed whether the use of US norms versus local norms led to significantly different overall neuropsychological impairment rates (validity), and/or led to significant differences in the factors affecting neuropsychological impairment (determinants) or significant differences in patterns of neuropsychological performance (profile).

On the question of validity, the answer is that the local norms yielded a substantially larger HIV effect size, and higher estimated prevalence of HAND than the US norms, albeit with lack of demographic corrections. Note that the standard GDS method and 0.5 cutoff is intended to produce a specificity of approximately 85 % rate for impairment classifications (i.e., about15% false positive), which tends to optimize the balance between sensitivity and specificity (Heaton et al. 2004b; Taylor and Heaton 2001). In fact, the false positive rate for the local norm-based GDS is close to optimal (14.3 % versus only 4.1 % for the US norm-based GDS, Fig. 1); and unsurprisingly, this is associated with a much higher HIV+ impairment rate. In contrast, the US norms apparently underestimated the rate of neuropsychological impairment in both groups.

Our HAND prevalence is higher than those reported in slightly younger, virally controlled cohorts (Heaton et al. 2010; 41 %) and this possibly represents the effect of age and age-related diseases. However, it may also reflect some lack of precision of the local norm-based z scores, particularly for classifying individual participants; as the local norm-based z scores were not demographically corrected, hence impairment estimates with this method were biased by demographic effects (i.e., age and education). Nevertheless, while classification biases probably occurred at the individual participant level, our two groups were comparable with respect to all major demographics (i.e., age, education and gender) as well as our best estimate of premorbid cognition (i.e., WAIS-III predicted VIQ). Therefore, while demographic biases undoubtedly occurred (in both directions) at the individual level, the HAND prevalence may not have been significantly biased at the general population level.

The estimated HAND prevalence in the current study is higher than in a comparable English cohort, which also mostly included highly educated MSM, who were virally suppressed and had an average age of 53 years (HAND prevalence 19 %; Garvey et al. 2011). This study used the brief computerized CogState battery with its large US and UK normative standards (N = 800) that allow age and gender corrections, but not education corrections. By contrast, estimated HAND prevalence in our cohort is lower than that reported in (Simioni et al. 2010) in an optimally treated Swiss cohort (HAND prevalence >70 %). Their sample comprised mostly men, aged in their late 40s, of which 73 % had at least a high school education. Some were on current methadone treatment suggesting that the cohort included participants who had acquired HIV infection through intravenous drug use (the education level in this subgroup was not specifically provided), but HIV infection transmission mode in the overall group was not documented. In addition, some had had previous HIV-related CNS opportunistic infections. Lastly, it is unclear which neuropsychological normative standards were used in this study.

These varying prevalence rates are concerning as they are drawn from similar types of Western cohorts where educational and cultural similarities are thought to allow the use of US normative standards. This phenomenon is likely due to the lack of optimal demographic corrections for the neuropsychological tests that were used in each of those studies (Strauss et al. 2006). This extreme variation in estimated prevalence rate calls for the careful examination of limitations to generalizability of US norms to other English-speaking countries in general and especially in subgroups that may be characterized by unusually high or low premorbid cognitive functioning. A recent NeuroAIDS panel (Joseph et al. 2013) has also emphasized the need for this country-specific effort to improve neuroHIV research standards.

The unusually high level of premorbid functioning in our sample constrains generalization of our findings to the rest of the Australian population. Our results do not suggest that US norms would also perform suboptimally in a more average Australian sample until this has been addressed empirically. Still, accurate neuropsychological classifications of high (or low) functioning individuals may need country-specific norms that correct for performance-based (e.g., reading) estimates of premorbid cognition in addition to the traditional demographic factors as also suggested by other Australian research (Green et al. 2008).

Our results are partially representative of HIV populations on CART in other countries including Europe (Munoz-Moreno et al. 2008) or the US (Heaton et al. 2010), as the level of undetectability for plasma HIV RNA in our cohort was very high (>95 %); however, the high rate of successful viral suppression on CART is comparable to treated participants of the Australian HIV observational database (N = 2439), especially if mortality rate is taken into account (Law et al. 2011). This might serve as a reference for international NeuroAIDS research as well as for Australian HIV research if we consider those who should be on CART and are not (up to 47 %), and are therefore likely to initiate CART only when they become symptomatic or have AIDS (The Kirby Institute 2013). Based on our optimal impairment definition, we found that HAND was detected in a substantial proportion of this optimally treated sample. We confirm that HAND occurs despite what is considered optimal HIV care on CART (Cysique et al. 2011). On the basis of our results, there are concerning implications when more representative cohorts are considered (e.g., those who have initiated CART later than recommended, or cases with multiple psychiatric and neurological confounds, which are fairly common in HIV+ persons including in Australia (The Kirby Institute 2013)). In fact, the largest American study including a more diverse range of HIV+ individuals found a substantially increased rate of impairment in those with higher comorbidity burdens (Heaton et al. 2010), and only 54 % of that unselected sample had low comorbidity levels. These findings argue that further research into the long-term effects of HIV and age on the brain is needed to document HAND prevalence and functional impact in HIV+ persons with broader socioeconomic status and geographical background, as well as including more women and HIV+ persons with comorbidities.

With regards to the question of possible determinants of the neuropsychological impairment, we found that the local norm-based GDS yielded an impairment classification that was significantly associated with age and age-CVD risk. On the other hand, the US norm-based GDS was uniquely associated with historical AIDS. This result emphasizes the impact that selection of a particular normative dataset can have on what conclusions are drawn regarding the nature of HAND. It also reinforces the careful attention to possible needs for risk-population norms, which would allow researchers to assess the detailed connections between CVD, age and HAND using optimal classification standards.

One of our most interesting findings was the strong evidence for the clinical significance of the detected neuropsychological impairment, and this was observed using both local and US standards. We interpret this finding as supporting the inclusion of mild forms of HAND in addition to the moderate to severe forms as detected by the US norms (Heaton et al. 2010).

On the question of the profile of neuropsychological impairment, we showed that the selection of different sets of norms had a major impact on the magnitude of effect sizes but not on the general pattern of performance. Deficits were present especially in learning (episodic memory) and speed of information processing; these are well-recognized hallmarks of HAND in the CART era (Woods et al. 2009). These results imply that norms with greater sensitivity to impairment should be selected in order to allow for timely initiation of both pharmalogical and behavioral interventions for HAND.