Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Introduction

Cognition is defined as “the mental processes by which knowledge is acquired”. These include perception, reasoning, acts of creativity, problem solving, and possibly intuition. Cognition is important for quality of life. The appropriate levels of micronutrients and trace elements are recognized as essential for maintaining cognitive functions [1]. Putting aside the problems of toxicity from excessive intake of trace elements, there are two periods where an unbalanced intake of trace elements can cause disorders with profound alterations of cognition throughout life. These are: (1) childhood development and (2) aging.

  1. (1)

    Micronutrient malnutrition impairs cognitive performance and developmental potential in children [2]. In 2001, Benton [3] reviewed 13 studies that investigated the role of multiple micronutrients on cognition in children aged 6–16 years, of which most reported a positive effect of the micronutrient supplementation, mostly with nonverbal measures. The author theorized that performance on nonverbal tests results, at least in part from basic biological functions, could be influenced by diet. In contrast, verbal intelligence comprises the acquired knowledge that was thought not to be affected by nutrition in the shorter term. Furthermore, it remained unclear whether there are other specific cognitive domains beyond nonverbal intelligence that could be influenced by micronutrient supplementation and whether the effects would depend on other factors (i.e., age and nutritional and socioeconomic status). Since Benton’s review, additional trials have been published in the literature, most of which were conducted in developing countries. Children in developing countries have a more monotonous diet in general, and may have a higher risk of micronutrient deficiencies. Hence, these children might benefit from micronutrient supplementation more than their peers in developed countries [4].

    In 2010, in a posterior study, Eilander et al. [4 systematically reviewed the literature that was current to date and performed a meta-analysis to quantify the effect of multi-plemicronutrient interventions on cognitive performance in children from infancy through late adolescence (i.e., 0–18 years old). Moreover, because they expected heterogeneity among the studies, they explored whether factors such as age, country, nutritional status, duration, and type of micronutrient supplementation would predict the effects of micronutrients on cognition. The meta-analysis suggested the possibility of a small positive effect of multiple micronutrient supplementation on fluid intelligence (reasoning ability), which was not statistically significant, and a positive effect on academic performance (based on a limited number of four trials) in children 5–16 years old. There were no effects on crystallized intelligence (acquired knowledge) and other cognitive domains.

    On the other hand, iodine is required for the production of thyroid hormones, which are necessary for normal brain development and cognition. Globally, >1.9 billion people, including 285 million children, have an inadequate iodine intake [5]. This deficiency is a serious problem and therefore this topic is deeply studied in all internal medicine and endocrinology manuals, which is why this element is not analyzed in this chapter.

  2. (2)

    Trace elements are key regulators of metabolic and physiological pathways known to be altered during the aging process and therefore have the capacity to modulate the rate of biological aging. Optimal intake is required to maintain homeostasis and to increase cell protection. Deficiencies are associated with specific illnesses. However, the contribution of commonly observed life-long suboptimal intake of trace elements to the development and severity of age-related chronic diseases is less appreciated. Additionally, reduced intake of several trace elements has been shown to be particularly challenging for elderly people [4].

    Dementia is one of the most pressing public health problems with social and economic implication. The form called cognitive impairment non-dementia (CIND) represents a subclinical phase of dementia. Different studies have shown a possible effect of micro- and macro-nutrients on cognitive function. Because trace elements are involved in metabolic processes and redox reactions in the central nervous system (CNS), they could influence the cognitive functions. Smorgon et al. [6] evaluated the presence of an eventual correlation between serum trace element concentrations and cognitive function in a group of subjects with CIND and manifest dementia (Alzheimer dementia and vascular dementia), and compared them with a control group. In the study they found a positive correlation between cognitive function and selenium, chrome, cobalt, and iron serum levels, while a negative correlation was observed with copper and aluminum serum levels. Futhermore, some statistically significant differences in selenium, chrome, cobalt, copper, and aluminum concentrations were found among the groups. According to these results, the authors could suppose that selenium, chrome, and cobalt protect cognitive function, that copper influences the evolution of cognitive impairment, while aluminum contributes to the pathogenesis of AD (Fig. 12.1).

    Fig. 12.1
    figure 1

    Minerals and trace elements: relationship with cognition, brain function, Parkinson and Alzheimer disease and other diseases such as Friedreich ataxia

Consideration should be taken into account regarding the handling of trace elements. Because micronutrient deficiencies often coexist and synergistic effects of micronutrients on physical functions may indirectly affect cognition, supplementing children and elderly people with multiple micronutrients could have advantages over single micronutrient supplementation. In contrast, micronutrients might also have antagonistic effects, affecting their bioavailability and their functioning in physiologic processes which could lead to impaired cognitive functioning. Because iron and zinc and copper and manganese compete for intestinal uptake, a high dose of one of these minerals may limit the absorption of the others [7].

Iron

Observational studies have found relationships between iron-deficiency anemia in children and poor cognitive development, poor school achievement, and behavior problems. However, it is difficult to separate the effects of ID anemia (IDA) from other types of deprivation in such studies, and confounding factors may contribute to the association between ID and cognitive deficits [8].

Several possible mechanisms link IDA to altered cognition. Anemic children tend to move around and explore their environment less than children without anemia, which can lead to developmental delays [9].

Conduction of auditory and optic nerve impulses to the brain has been found to be slower in children with IDA. This effect could be associated with changes in nerve myelination, which have been observed in iron-deficient animals [10]. Neurotransmitter synthesis may also be sensitive to ID [11]. Impaired intellectual development in children can be prevented through the treatment or prevention of ID.

On the other hand, iron overload can be related to neurodegenerative disease. Several genetic disorders can lead to pathological accumulation of iron in the body; the body’s tight control of intestinal iron absorption protects it from the adverse effects of iron overload [12]. Iron is required for normal brain and nerve function through its involvement in cellular metabolism and in the synthesis of neurotransmitters and myelin.

However, accumulation of iron excess can result in increased oxidative stress, and the brain is particularly susceptible to oxidative damage. Iron accumulation and oxidative injury are currently under consideration as potential contributors to a number of neurodegenerative diseases such as AD and PD [13, 14].

The abnormal accumulation of iron in the brain does not appear to be a result of increased dietary iron, but rather a disruption in the complex process of cellular iron regulation. Although the mechanisms for this disruption in iron regulation are not yet known, it is currently an active area of biomedical research.

Iron in the Brain

The iron needs of the brain vary with the stage of the life cycle and the cell types in the CNS. Iron is the key component of the many enzymes that involve essential oxidation/reduction reactions, synthesis of neurotransmitters, catabolism of neurotransmitters, and synthetic processes such as the production of myelin.

The highest levels of iron in the brain are found in the basal ganglia, but iron is also found throughout the brain, including the white matter [15]. Both biochemical and histochemical studies reveal white matter throughout the brain as a major site of iron concentration. Iron uptake into the brain is at a maximum during the period of rapid brain growth, coinciding with the peak of myelinogenesis.

Lack of iron in the diet for the first 2 years of life has an important effect on development because that is the time when the majority of brain growth occurs. Although the most rapid brain growth is seen in the months leading up to birth, at birth the brain has only reached 27 % of its adult size and it continues to grow for the next 2 years [16]. It has been shown that iron levels in the brain at birth are 10 % of eventual adult levels, with the remainder accumulating through childhood and young adulthood [15].

Research in humans was directed to impaired iron transport across the placenta in several prenatal conditions (e.g., diabetes mellitus, prenatal alcohol exposure, intrauterine growth retardation, maternal stress). There is direct evidence of decreased brain iron or ID in the offspring son and daughter [17].

Infants born of mothers with nutritional ID during pregnancy are rarely anemic, but they may have lower iron stores and ID sooner in the postnatal period. There is now solid evidence that brain ID can occur even with a normal hemoglobin level.

Brain iron accumulates from birth to early adulthood [15]. Perhaps for this reason, brain iron levels are affected more seriously by ID in the very young than in the adult animal. A brief period of severe ID in the young rat, but not in the adult, resulted in a deficit of brain iron which was not corrected by iron therapy although all signs of systemic ID were reversed [18].

This resistance contrasts with the rapid normalization of hepatic iron and hemoglobin (Hb) concentrations following iron repletion. The failure to reverse iron depletion in the brain with iron treatment seems to be the result of a slow rate of replacement of brain iron compounds [19].

The development of iron-deficient animal models has been an invaluable tool for looking at the consequences of ID. Most animal studies have involved rats because the distribution of iron in their brain is comparable with that of the human brain. Animal studies have shown that ID is associated with hypomyelination of neurons in the developing brain [20].

It must be recognized, however, that there are certain limitations in using animal models. Rats are less mature at birth, and humans and rats have different rates of neuronal development, both of which are particularly important when assessing the effect of early iron deprivation on mental function [21].

There has been considerable research on the possible mechanisms through which IDA affects cognitive function. Most emphasis has been placed on a direct neurochemical effect. ID causes low levels of brain iron, which leads to a reduction in neurotransmitter levels, impaired transmitter function, hypomyelination, and delayed neuromaturation. Another possibility is that the systemic effects of anemia lead to low oxygen delivery to the brain, directly affecting cognition [22].

When ID occurs early in the development of the rat, there are lasting deficits in brain iron, electrophysiological changes, a decrease in the number of dopamine D2 receptors, and alterations in neurotransmitter function, hypomyelination, and persisting behavioral changes that suggest an altered threshold of arousal [23].

The second hypothesis for the mechanism is that ID also has an indirect effect on behavior. IDA infants and children have been shown to be less attentive and less responsive. Lower developmental scores may reflect poorer mother/child interaction because of the child’s reduced responsiveness, and this could result in less effective stimulation of the environment. The lower scores can also reflect poorer interaction with the developmental assessor [24].

To specifically understand the role of iron in neural circuits that underlie learning and memory development and function, in more recent models, iron uptake genes have been genetically manipulated in a tissue- and time-specific manner to generate a non-anemic model of hippocampal ID. The models are capable of isolating the role of iron independent of the potential widespread confounding effects of brain and body ID that accompany maternal dietary restriction (e.g., hypoxia, uptake of other divalent metals [25].

Optimal neurodevelopment is shaped by a variety of factors including growth factors, synaptic activity, and environment. Structures are most sensitive to these factors during rapid development [26].

As noted above, humans are most vulnerable to early ID from late gestation through 2–3 years of age, during the most rapid period of hippocampal structural maturation. Functionally, hippocampus-dependent memory appears and matures between 3 and 18 months of age [27].

This increased metabolic activity is coincident with extensive dendrite arborization, spine formation, and synaptogenesis [28] as well as the maturation of electrophysiological plasticity [29]. In conjunction, the timing and energy demands of hippocampal development with long-term deficits support the vulnerability of the structure to the metabolic consequences of early-life ID [30].

The effects of early-life ID on hippocampus-based learning and memory have been largely ascribed to primary abnormalities in iron-containing proteins, although many effects can be attributed to iron-containing proteins (e.g., reduced neuronal energy capacity).

Iron is necessary for energy production and cellular metabolism because it is essential for many mitochondrial enzymes integral for oxidative phosphorylation and adenosine triphosphate (ATP) production, including cytochromes, Nicotinamide adenine dinucleotide phosphate, and flavoproteins [31]. Adequate energy availability is necessary to support neuronal development and synaptic activity.

At birth, the brain comprises 50 % of resting metabolic energy [32, 33]. Approximately one half of the energy consumption is used to maintain Na+, K+, and Ca2+ gradients necessary for generating membrane potentials required for synaptic transmission. In addition, the generation and maintenance of the complex neuronal structure requires large amounts of energy.

Another important cellular process dependent on iron availability is nucleic acid metabolism. Iron-containing enzymes such as ribonucleotide reductase, deoxyribonucleic acid (DNA) helicase elongation protein 3, and BACH1 are integral for deoxynucleotide triphosphates (dNTP) synthesis, DNA transcription, elongation, and repair, and histone modification [34, 35].

The exact mechanism by which ID induces these acute and persistent gene expression changes is not clear because the experiments utilized maternal dietary restriction models of early IDA and the alterations may be due in part to the contribution of hypoxia. However, the evidence suggests that early life ID affects the regulation of gene expression throughout life in the hippocampus.

Another set of important signaling pathways that are likely affected by ID are found in mitochondria. The cellular functions of mitochondria reach beyond ATP synthesis and include maturation of Fe–S proteins that are crucial for cell function [36, 37]. As part of their function, mitochondria are an important factor in the regulation of intracellular Ca2+ levels. This function is crucial for many aspects of neuronal function, including secondary signaling cascades, neurotransmitter release, and apoptosis [38].

Iron Deficiency and Neural Functioning in Humans

Impaired Intellectual Development in Children

Infancy is considered the age range of highest vulnerability for the CNS because it corresponds with the brain growth spurt and the unfolding fundamental mental and motor processes. Altered behavior and development are among the greatest concerns regarding ID in infancy, especially because the nutrient deficiency is most prevalent in the period between 6 and 24 months of age. Because this age range coincides with a period of maximal brain growth and the unfolding of many neuron developmental processes, several investigators have focused on the question of CNS effects on ID [39].

In observational studies, anemia and ID are associated with cognitive deficits, suggesting that iron supplementation may improve cognitive function and the studies show cross-sectional associations between IDA and poor cognitive function, motor development, behaviour, or school achievement levels [9]. However, it is difficult to separate the effects of IDA from other types of deprivation in such studies, and confounding factors may contribute to the association between ID and cognitive deficits [40].

In observational studies, anemia and ID are associated with cognitive deficits, suggesting that iron supplementation may improve cognitive function. The effect of iron supplementation on a range of health outcomes in infants and young children has been well documented. It is estimated that 47 % of preschool children worldwide have anemia, the highest prevalence of any population group [41].

Longitudinal studies show that ID in infancy is related to poorer cognition in childhood [9]. One systematic review that included seven randomized controlled trials on the effects of supplementary iron in young children with anemia or ID found no evidence of an effect of iron supplementation on psychomotor development [42], while another included 17 randomized controlled trials in children of any age and with any initial iron status, found that iron supplementation was not associated with improved mental development scores in children younger than 5 years [43], or with improved physical growth [44].

A more recent systematic review addressed a range of health risks and benefits of iron supplementation in infants and children 5 years old or younger [45], finding that supplementation led to improvements in cognition and motor development in children with anemia and ID, but was associated with increased risk of death in areas with endemic malaria.

As previously explained deficiency of enzymes involved in the development of parts of the brain is important for cognitive functions such as memory (e.g. the hippocampus). Deficiency and supplementation may have different effects on infants and young children than in other population groups.

Older Children, Adolescents, and Adults

Older children and adolescents are less at risk of anemia than preschool children, but global statistics indicate that approximately 25 % of older children have anemia, as do 30 % of non-pregnant women and 42 % of pregnant women, and 17 % of elderly people (rising to 40–50 % of those admitted to hospital or living in nursing homes), demonstrating that it is a very large and important health problem [41, 46].

A meta-analysis has been published to assess whether iron supplementation improved cognitive domains: concentration, intelligence, memory, psychomotor skills, and scholastic achievement in adults. Evidence was found that iron supplementation improved attention and concentration in adolescents and women, regardless of baseline level of iron status. Iron supplementation also improved performance in intelligence quotient (IQ) tests in adults and children who were anemic at baseline, but had no effect in other groups or on other cognitive domains [47].

The prevalence of depleted iron stores is substantially greater in pre- or perimenopausal women than in postmenopausal women or in men. Poor iron status affects premenopausal women more often than men because of the combination of low dietary iron intake, menstruation, and gestational requirements. Physical performance is affected by poor iron status, including decreases in work productivity, voluntary activity, and athletic performance. Cognitive, affective, behavioral, and neurophysiologic decreases have been associated with poor iron status in premenopausal women [48].

It is important to consider that ID can often be present without anemia [41]. A requisite for successful treatment is the correct diagnosis of depletion and to assess the causes of ID [49].

Iron supplementation may be less effective when there are a number of nutritional problems at baseline (all of which may be contributing to cognitive limitations) than when patients are nutritionally replete except for variations in iron status. Iron and zinc deficiencies often occur together, and zinc deficiency can be exacerbated with a high dose of iron supplements [50]. Zinc may also play a role in cognitive function, therefore iron supplementation could exacerbate cognitive deficits [51].

Although it is not surprising that the brain functions poorly when iron is deficient, the long-term deficits, despite iron repletion, remain mechanistically enigmatic and a fruitful area of research. Furthermore, this research may contribute to defining the time point at which iron repletion can no longer reverse the behavioral phenotype. It would be critical in determining the optimal timing of iron treatment regimens.

Because iron is not only a critical nutrient for brain development but also a potentially toxic element, further research is also necessary to determine optimal iron doses. Arguably, an iron-deficient developing brain that has responded to ID by prematurely expressing large amounts of iron transporters [52] can be at risk for iron overload and generation of reactive oxygen species if large amounts of medicinal iron are suddenly delivered to this “activated system”.

Iron Overload and the Central Nervous System

Several proteins implicated in brain iron homeostasis are involved in disorders associated with abnormal iron metabolism. A basic understanding of mechanisms of iron homeostasis has a clinical relevance, as either accumulation or depletion of intracellular iron may impair normal function and promote cell death.

Iron accumulates in selective brain regions during aging, in acquired neurodegenerative disorders such as AD and PD, and in genetic disorders such as neurodegeneration with brain iron accumulation (NBIA). Dysregulation of iron homeostasis is also a critical feature of FA [53].

Iron is known to catalyze the formation of reactive oxygen species (ROS), such as hydroxyl radical, and initiation or enhancement of lipid peroxidation by reacting with hydrogen peroxide (H2O2) via the Fenton reaction [54]. ROS are highly reactive oxygen-containing radicals that can easily react with other molecules such as protein, DNA, lipids, and antioxidants [55, 56].

$$ {\mathrm{Fe}}^{2+}+{\mathrm{O}}_2\to {\mathrm{Fe}}^{3+}+\bullet {O_2}^{-} $$
$$ 2\bullet {{\mathrm{O}}_2}^{-}+2{\mathrm{H}}^{+}\to {\mathrm{H}}_2{\mathrm{O}}_2+{\mathrm{O}}_2 $$
$$ {\mathrm{Fe}}^{2+}+{\mathrm{H}}_2{\mathrm{O}}_2\to \bullet \mathrm{O}\mathrm{H}+{\mathrm{O}\mathrm{H}}^{-}+{\mathrm{Fe}}^{3+} $$

Iron can also react with lipid peroxides in a way similar to its reaction with H2O2 and produce alkoxy (R ‐ O·) and peroxy radicals (R ‐ O2 ·). The resulting peroxy radical leads to propagation of lipid peroxidation [57].

Lipid peroxidation can proceed until the lipid radicals interact with one another and/or a “chain breaker,” such as vitamin E, forming a no-radical species.

$$ {\mathrm{Fe}}^{2+}+\mathrm{R}-\mathrm{O}\mathrm{O}\mathrm{H}\to {\mathrm{Fe}}^{3+}+\mathrm{R}-{\mathrm{O}}^{-\cdotp } $$
$$ {\mathrm{Fe}}^{3+}+\mathrm{R}-\mathrm{O}\mathrm{O}\mathrm{H}\to {\mathrm{Fe}}^{2+}+\mathrm{H}+\mathrm{R}-{{\mathrm{O}}_2}^{-\cdotp } $$
$$ \mathrm{R}-{\mathrm{O}}^{-\cdotp }+\mathrm{L}-\mathrm{H},\left(\mathrm{L}-\mathrm{O}\mathrm{O}\mathrm{H}\right)\to \mathrm{R}\mathrm{O}\mathrm{H}+{\mathrm{L}}^{\cdotp}\left({\mathrm{L}\mathrm{O}\mathrm{O}}^{\cdotp}\right) $$

Iron compounds such as free Hb can also catalyze peroxidation of purified arachidonic acid and other polyunsaturated fatty acids within normal cell membranes in the presence of H2O2 and •O2− [58].

The CNS is separated from the systemic circulation by the blood–brain barrier (BBB), a tight epithelial barrier analogous to that of the mammalian duodenum, which is the site of absorption of dietary iron. After absorption, iron in its oxidized (ferric) form binds to serum transferrin and is distributed throughout the general circulation. The presence of a BBB explains the relative independence of the brain iron from the total body iron content [59].

Incorporation and Transport of Iron in the Nervous System

Iron is transported to the brain via the BBB which is composed of endothelial cells in small vessels throughout the brain that contain tight junctions which regulate the brain iron levels.

Iron incorporation and transport in the brain depends on interactions between the endothelial cells and astrocytes. Brain endothelial cells express the transferrin receptor 1 (Tf R1) in their luminal membrane. This receptor binds iron-loaded transferrin and internalizes this complex in endosomes.

Within the endosomes, the acid environment facilitates the release of ferric iron from transferrin and this is followed by reduction of ferric to ferrous iron by action of endosomal reductases. The mechanism by which iron is released from transferrin is transported from the interior of the brain endothelial cells to the interstitial fluid is a matter of controversy [60, 61].

One possibility is that ferrous iron is transported from the endosome to the cytosol by the divalent metal transporter-1 (DMT1) and then exported into the extracellular fluid by action of ferroportin [60].

However, there is some disagreement as to the degree of expression of DMT1 and ferroportin and their contribution to iron transport across the brain endothelial cells. Alternatively, it has been proposed that the transferrin-Tf R1 receptor complex is transported within the endosomes from the luminal to the abluminal surface where endosomes release iron at the interface between the endothelial cells and astrocyte end-foot processes [61].

The end-foot processes express ceruloplasmin, which acts as a ferroxidase that oxidizes newly released ferrous iron to ferric iron and binds to the transferrin in the brain interstitial fluid [6062]. Transferrin is the main source of iron for neurons, which express high levels of Tf R1. Whereas transferrin is synthesized by oligodendrocytes, the primary source of transferrin in the brain interstitium is its diffusion from the ventricles.

It has been shown that transferrin uptake and the ratio of iron to transferrin uptake by the brain decrease with age, and the transferrin recycling time increases with age [63].

Neuronal Iron Homeostasis

There is a tight regulation of the cytosolic iron pool in brain cells, and is critical for two reasons: (1) iron is an important source for numerous cytosolic, mitochondrial, and nuclear ferroproteins; and (2) excessive accumulation of free cytosolic ferrous iron predisposes to oxidative stress and cytotoxicity.

Iron regulatory proteins sense cytosolic iron levels and interact with iron-responsive elements and regulate translation of mRNA encoding for proteins involved in iron uptake, storage, and mobilization, including Tf R1, ferritin, and ferroportin.

In the cytosol, the storage protein ferritin sequesters and reduces levels of free iron. Ferritin consists of a heavy [ferritin heavy chain (FTH1)] subunit that catalyzes the rapid oxidation of ferrous to ferric iron and a light (FTL) sub-unit that may be involved in the nucleation of the iron core within the protein shell. Thus, ferritin has a dual function of iron detoxification and iron reserve [64].

Ferritin is also present in axons and may transport iron to the synapse. Ferroportin, present in synaptic vesicles, may allow release of ferrous iron at the synapses [60, 61]. Neuromelanin is an insoluble pigment produced from oxidation of excess cytosolic catechols and is present in granules in dopaminergic neurons of the substantia nigra and in noradrenergic neurons of the locus ceruleus. Neuromelanin binds iron avidly forming stable complexes and sequesters large amounts of iron in those cells [65].

Mitochondrial iron is required for heme biosynthesis and for the generation of iron/sulfur (Fe/S) clusters in many essential enzymes. Iron is transported into the mitochondria by mitoferrin, a transporter expressed in the inner mitochondrial membrane.

Frataxin is a mitochondrial iron chaperone that is involved in the biosynthesis of Fe/S clusters by interactions with critical assembly proteins. [62, 6668] The Fe/S clusters form the prosthetic group of many enzymes of the respiratory chain and also constitute the main mechanism for iron exit from the mitochondria.

The major route of iron export out of the brain is via the cerebral spinal fluid (CSF) and its reabsorption into the blood from the subarachnoid space. The concentration of transferrin in CSF is very low and its capacity to export iron is limited. Lactoferrin, ferritin, and non-protein-bound iron are also present in the CSF and may contribute to iron export. In normal conditions, iron concentration is very low but can increase considerably under pathologic circumstances as discussed below. Microglia and other phagocytic cells are additional important mediators of iron export after cell death and intracerebral hemorrhage.

Iron accumulates in the brain as a function of age, primarily in the form of ferritin, particularly in the microglia and astrocytes, but also in the oligodendrocytes. The brain areas richest in iron are the globus pallidus, substantia nigra, putamen, caudate nucleus, dentate nucleus, and frontal cortex [69].

Impaired regulation of iron homeostasis in those cells may lead to either excessive accumulation of free cytosolic iron or decreased iron availability for critical enzymes. Free cytosolic ferrous iron reacts with endogenously generated hydrogen peroxide to yield hydroxyl radicals, which damage cell membranes. A recent study identified a signaling cascade that links the activation of the N-Methyl-D-aspartate (NMDA) receptors, which mediate glutamate triggered excitotoxicity, with iron homeostasis. This cascade involves activation of nitric oxide synthase and adaptor proteins that interact with ferroportin [70].

Iron deposition or dysregulation occurs in several neurodegenerative disorders, including sporadic AD, PD, FA, and NBIA.

Brain Neurodegenerative Disorders and Iron

Dysregulation of iron homeostasis is also a critical feature of AD, PD, NBIA and FA. Although these diseases have their own distinctive features, they all have one thing in common: accumulating iron in the brain. Increased iron in the brain, rich in oxygen and fatty acids, provides an ideal environment for oxidative stress and possible irreparable tissue damage.

The link between iron and neurodegenerative disease provides potential therapeutic targets for these disorders. If, indeed, iron and/or oxidative processes are involved in the pathogenesis of neurodegenerative disorders, approaches such as iron chelation therapy and antioxidant supplements might help to slow the degenerative processes or ameliorate brain tissue injury.

Alzheimer Disease

AD is a progressive degenerative disease with a gradual deterioration in memory, cognition, behavior, and the ability to perform activities of daily living. Evidence of increased brain metal levels such as iron and copper has been associated with AD [71].

The amyloid-beta (Aβ) plaques in the brain are the hallmark pathologic features of AD and are derived from the cleavage of amyloid precursor protein (APP). Deposition of fibrillar aggregates of Aβ in the brain parenchyma, which is caused by Aβ overproduction, impaired clearance, or both, has been hypothesized to explain the cause of AD [72].

APP has binding sites in its amino-terminal domain and in the Aβ domain for copper and iron. Iron is primarily complexed with ferritin and concentrated in the neuritic processes associated with amyloid plaques [69, 73]. Iron might have a direct impact on plaque formation through its effects on APP processing by alfa-secretase, deposition of Aβ, and oxidative stress [63, 73].

High levels of iron may interact with the Aβ peptide, leading to the reduction of molecular oxygen to superoxide and eventually to H2O2 by reducing iron. It has also been demonstrated that overexpression of the carboxyl terminal fragment of APP (Aβ) significantly reduces the level of copper and iron in the transgenic mouse brain. This suggests a role for APP and Aβ in physiologic metal regulation in AD [74].

Additionally, overexpression of melanotransferrin has been reported in AD. Because purified melanotransferrin can bind iron, it has been proposed as another protein that also might be involved in iron transportation [75].

The role of iron in the pathogenesis of AD is thought to be related to enhance oxidative stress mediated by free iron. Transferrin, ferritin, and iron regulatory protein 2 also have been associated with neurodegeneration in AD. The latter might be responsible for the disturbance in brain iron homeostasis and the overall decompartmentalization of iron and the resulting oxidative processes suggestive of AD [74].

Genetic alterations specific to iron-management proteins, including HEF1 mutations (associated with congenital hemochromatosis) or the transferrin subtype C2, may increase the risk of AD [76].

Friedreich Ataxia

FA is an autosomal recessive neurodegenerative disease characterized by degenerative atrophy of the posterior columns of the spinal cord followed by the spinocerebellar tracts and corticospinal motor tracts, leading to progressive ataxia, sensory loss, and muscle weakness. It is also characterized by degeneration of large sensory neurons of the dorsal root ganglion, cerebellum (particularly the iron-rich dentate nucleus), and cardiomyocytes.

FA is an autosomal recessive disorder resulting from a large guanine adenine adenine (GAA) triplet-repeat expansion in the first intron of the Friedreich ataxia (FRDA) gene, resulting in a reduction in expression of the encoded protein, frataxin [6668, 77].

Frataxin is a mitochondrial protein and is suggested to have a role in mitochondrial iron transport or in iron-sulfur assembly and transport. High levels of iron in the mitochondria can react with superoxide (•O2−) and H2O2 to produce the hydroxyl radical (•OH), which can oxidize cellular components, damage respiratory chain complexes, and result in cellular injury and eventually cell death.

Whereas excess mitochondrial iron is detected in neurons and cardiomyocytes from affected patients, indicating a predisposition to oxidative stress, there is evidence that impairment of heme and Fe/S cluster biosynthesis is the most likely proximate cause of neurodegeneration in this disorder [62, 6668, 77].

Brain and heart cells depend highly on aerobic metabolism and are more susceptible to free radical generation in mitochondria [78, 79]. The effects of treatment with the antioxidants coenzyme Q and vitamin E in patients with FA are being studied, and preliminary results seem to be promising [80].

Parkinson Disease

PD is a progressive disorder manifesting as tremor at rest, bradykinesia, gait abnormalities, rigidity, postural dysfunction, and loss of balance. Iron has been suggested to be responsible for nigrostriatal dopamine neuron degeneration in PD owing to its ability to produce toxic ROS and cause lipid peroxidation [81].

PD is characterized by iron accumulation in dopaminergic neurons of the substantia nigra. Free cytosolic iron may trigger oxidative stress and promote alfa-synuclein aggregation with deposition of Lewy bodies [69].

Lewy bodies can have deleterious effects on the extrapyramidal system and on psychomotor function [82, 83]. The presence of the pigment neuromelanin in the substantia nigra in PD also might result in iron accumulation because neuromelanin can function like ferritin and store iron [84].

Over expression of lactoferrin (a protein that reversibly binds iron) receptors on neurons and microvessels in regions of neuronal degeneration in PD-affected brain tissue suggests a possible link to iron overload in affected brain regions [82].

All these mechanisms suggest that disturbances in iron homeostasis and metabolism in PD occur at several levels, such as iron uptake, storage, intracellular metabolism, release, and posttranscriptional control [85].

As indicated in the preceding text, a disturbance in iron homeostasis can provide a favorable condition in which free iron, via generation of ROS, causes permanent tissue damage. Neuromelanin may exert neuroprotective action at early stages of PD because it prevents free iron accumulation and thus hydroxyl radical production and formation of neurotoxic dopamine quinones [65].

However, in advanced stages, extravasation of neuromelanin granules from dying neurons may attract and activate microglia, causing release of neurotoxic molecules leading to cell injury [65]. Both iron chelation and over expression of iron-sequestering ferritin have been shown to be protective in animal models of PD [69].

Type I Neurodegeneration with Brain Iron Accumulation

Type I NBIA (NBIA-1) was formerly known as Hallervorden-Spatz syndrome or pantothenate kinase-associated neurodegeneration.

NBIA-1 is a rare, genetically determined neurodegenerative disorder characterized by extrapyramidal dysfunction and mental deterioration. Iron accumulates mainly in the globus pallidus and the pars reticularis of the substantia nigra, and presents as brown-pigmented iron deposits [86].

NBIA comprises a clinically and genetically heterogeneous group of disorders that include pantothenate kinase–associated neurodegeneration (PKAN), infantile neuroaxonal dystrophy, neuroferritinopathy, and hereditary aceruloplasminemia [87]. These disorders are characterized by iron accumulation and cell loss affecting primarily the globus pallidus and, in many cases, retinal photoreceptors [87].

PKAN is caused by a mutation of the gene encoding pantothenate kinase 2 (PANK2), the key enzyme in the synthesis of mitochondrial coenzyme A. The mutation in a novel pantothenate kinase gene, PANK2, is predicted to cause the accumulation of cysteine, which binds iron and causes oxidative stress in the iron-rich globus pallidus [88].

Neuroferritinopathy is a dominantly inherited, adult-onset disorder caused by mutations in the ferritin light chain (FTL1) gene [8991]. It results in accumulation of ferritin-iron aggregates in neurons and glial cells of the globus pallidus, substantia nigra, striatum, and cerebellum. Vacuolated glial and neuronal nuclei may be characteristic of the disease [90]. Focal onset limb dystonia or chorea, orolingual dyskinesia, dysarthria, aphonia, and dysphagia are prominent features.

Hereditary aceruloplasminemia is a rare autosomal recessive disorder caused by a mutation of the ceruloplasmin gene, leading to iron overload in the brain and reticuloendothelial system. It manifests with anemia, diabetes, retinal degeneration, and progressive neurologic disorder [92, 93]. A characteristic neuropathologic finding is the presence of enlarged deformed astrocytes and accumulation of spheroid-like, grumose foamy deposits in the astrocytic foot processes [94].

Application of high-resolution imaging technologies such as electron energy-loss spectroscopy and electron tomography may allow early identification of the cell type and intracellular location of iron deposits in patients with neurodegenerative disorders.

If, indeed, iron and/or oxidative processes are involved in the pathogenesis of neurodegenerative disorders, approaches such as iron chelation therapy and antioxidant supplements might help to slow the degenerative processes or to ameliorate brain tissue injury. Several chelators are currently under investigation for treatment of these and other disorders associated with abnormal brain iron homeostasis.

The neuroprotection that is provided by iron chelators in animal models indicates that iron-chelation therapy could be a viable neuroprotective approach for treatment of disorders such as PD or AD.

Manganese

Manganese is an essential nutrient that is common in the environment. It is the fifth most abundant metal and the twelfth most abundant trace element in the earth’s crust [95]. It is necessary for the adequate functioning of the human CNS, but it also has the potential to produce neurotoxic effects when, depending on the route and dose of exposure, it accumulates in an organism (particularly in the brain) exceeding the homeostatic range [96]. It is released into the environment as a product of industrial activities, the use of the manganese-containing pesticides such as Maneb® and Mancozed®, and through the use of methyl-cyclopentadienyl manganese tricarbonyl, as a gasoline antiknock agent (Agency for Toxic Substances and Disease Registry, ATSDR, 2000).

The vast majority of studies on neurotoxic effects of manganese were conducted in occupational settings where exposure occurs mainly through inhalation of airborne particulates (ferroalloy smelting, welding, mining, battery assembly, etc.). Several studies have demonstrated the impact of manganese toxicity in adults, resulting in cognitive, neurological, motor, and psychological impairment [97102].

In the past years, many studies have investigated possible overexposure of children to manganese and subsequently the neuropsychological effects produced. It is generally accepted that children are at greater risk than adults exposed to the same contaminants from the environment [103].

In general, with normal dietary consumption, systemic homeostasis of manganese is maintained. Although very low levels of manganese in air, soil, water, and food are normal, nevertheless, excess exposure can occur by inhalation in areas where there is high manganese concentration in air and dust, or by drinking water that has had long contact with bedrock enriched in manganese, or by consuming high amounts of food sources rich in the trace element. The highest concentrations are found in nuts, legumes, and blueberries teas [104]. Other authors have hypothesized the possibility of overexposure to manganese through ingestion of infant milk formula [105], or even by iatrogenic manganese exposure than occurs in individuals receiving total parenteral nutrition resulting in increased concentrations of manganese in the brain [106]. Furthermore, exposure to at-risk populations with compromised or immature BBBs or underdeveloped excretory pathways (i.e., children) can also result in increased brain manganese levels [107].

Several factors could predispose children to manganese overexposure and subsequent toxic effects. Exposure to manganese by ingestion or inhalation can have different consequences in children than in adults and through different mechanisms [108]. First, children are exposed to a larger amount of manganese from inhalation because of their higher breathing rates (the ratio of inhaled air/weight is much higher in children because of the lower body mass), and greater intestinal absorption rate [103, 109]. Absorption can be as high as 80 % in neonates compared with 1–5 % in adults [11]. Second, high demand for iron linked to growth could further enhance the absorption of ingested manganese [110]. Third, a low excretion rate was observed in infants because of their poorly developed biliary excretion mechanism [111]. In fact, exposure during this period may result in increased delivery of manganese to the brain and other tissues.

Neurotoxic effects resulting from excessive manganese exposure were first described in 1837 by Couper in Scottish laborers who were grinding manganese black oxide in the chemical industry [112]. Neurological symptoms of manganism include decreased memory and concentration, fatigue, headache, vertigo, equilibrium loss, insomnia, tinnitus, trembling of fingers, muscle cramps, rigidity, alteration of libido, and sweating [113]. Many reports of neurotoxic effects in manganese-exposed workers were later published [100], and the definition of manganese intoxication has evolved to include subclinical signs of intoxication indicated by alterations of neurobehavioral functions [114].

Manganese plays a role in immune response, blood sugar homeostasis, ATP regulation, reproduction, digestion, and bone growth [115]. It is a necessary component of metalloenzymes such as manganese superoxide dismutase, arginase, phosphoenol–pyruvate decarboxylase, and glutamine synthetase. This glutamine synthetase enzymeconverts glutamate into glutamine. Glutamine synthetase [116].

Manganese shares several characteristics with iron; both are transition metals with valences of 2+ and 3+ in physiological conditions and proximate ionic radius. In addition, as manganese and iron both strongly bind to transferrin and accumulate in the mitochondria, low iron stores are associated with increased manganese uptake and retention in the blood [117], and increment of the accumulation in CNS. More than 2,000 million people on our pleanet, mainly children and pregnant women (and/or in fertile age), manifest ferropenic anemia after inadequate iron absorption. The potential effects associated to the accumulation in CNS of manganese in these populations represent a sanitary challenge of great magnitude. Manganese can accumulate in the CNS, particularly the basal ganglia but also the cortex. Exposure to manganese has been shown to interfere with several neurotransmitter systems, especially in the dopaminergic system in areas of the brain responsible for motor coordination, attention, and cognition [117, 118]. Manganese is a potent dopamine oxidant, which could explain the toxic lesions in certain dopaminergic brain regions [119]. Excessive exposure could result in dopamine receptor loss or inactivation through damage to the membrane mediated by free radicals or cytotoxic quinones generated by the manganese catalyzing effect on autooxidation of this neurotransmitter [120]. The correlation between manganese and hyperactive behavior is probably a result of the dopaminergic and gamma-aminobutyric acidergic systems, which play a role in hyperactivity in children [121, 122].

One hypothesis for the toxic mechanism of manganese is the production of excess free radicals in the nerve cell, potentiating lipid peroxidation, and resulting in tissue destruction [122, 123]. Manganese neurotoxicity has been extensively studied and a lot has been learned about its mechanism of action at the cellular and molecular levels and the detection of subclinical effects at low exposures. In the last few years, several literature reviews have been published on aspects such as neurotoxic effects on exposed laborers [118, 124], the application of magnetic resonance imaging [125], neuropsychological testing for the assessment of manganese neurotoxicity, manganese neurotoxicity focused on neonates [126], neurotoxicology of chronic manganese exposure in nonhuman primates [127], and manganese exposure and neuropsychological effect on children and adolescents [108, 128].

In most studies, authors observed that manganese exposure was associated with poorer cognitive functions and hyperactive behavior. Many have suggested that manganese exposure is related to cognitive, motor, and behavior deficits in children. Some of them found an adverse effect of manganese on cognitive function, and overall an inverse association between manganese exposure and IQ [129131]. Others studies have focused on motor effects of manganese, finding a positive association. Very few studies focused on the effects of manganese on children’s motor skills, although data on motor effects in adults, occupationally or environmentally exposed to manganese, have been reported [132, 133].

The majority of the studies published have several limitations including sample size, research design, the lack of a validated biomarker of exposure or exposure index for manganese, and the lack of attention to mixed exposure. In this sense, most studies focused on a single agent of exposure and did not measure or adjust for potential effects of other chemicals.

Finally, despite these limitations, it is believed that adverse effects of manganese exposure in children is well demonstrated through different studies. Nevertheless, further investigations should promote preventive strategies to reduce manganese exposure.

Cadmium

Cadmium is a heavy metal found in the earth’s crust that is released to the environment both by natural processes and by human activities such as fossil fuel burning, waste incineration, smelting procedures, mining, from factories of industrial production, mires of residual waters, and the use of phosphate fertilizers [134]. It is used in many industrial processes which include silver plating, paint, plastic stabilizers and nickel-cadmium batteries. The highest exposure to cadmium in humans is dietary. High levels are found in shellfish, liver, and kidney [135]. Soil cadmium is absorbed easily by plants. In general, leafy vegetables such as lettuce and spinach, potatoes and grains, peanuts, soybeans, and sunflower seeds contain high levels of cadmium. Cadmium also tends to concentrate in shellfish from polluted coastal waters. Another important route of exposure is through tobacco smoke (tobacco leafs accumulate high levels of cadmium from the soil). Cadmium blood levels in smokers are approximately twice as high as those of nonsmokers. Finally, occupational exposure is another important source to take into account.

Cadmium is toxic to the CNS of fetuses and infants. During pregnancy it interferes with the placental function, alters various enzymes, and modifies the availability of nutrients and essential elements in the CNS [136].

Neonatal exposure alters the levels of neurotransmitters such as norepinephrine, dopamine, serotonin, and acetylcholine. Cadmium exposure is also associated with increased free radical production in tissues causing damage to the cell membrane and changes in a variety of other physiological functions. Its fetal neurotoxic effects are the indirect result of impaired placental function, enzymatic dysfunctions, and metabolic alteration of essential trace elements for the CNS. Additionally, cadmium has been found to stimulate DNA synthesis and cell multiplication at low levels but increase apoptosis and chromosomal aberrations at high levels, suggesting different effects at low and high levels.

In prior risk assessments, kidney damage had been considered the most sensitive endpoint of cadmium toxicity, and reference levels for urinary cadmium have been established to protect against this effect because cadmium can be ingested or inhaled and can enter into the bloodstream and be stored in the liver or kidneys (EFSA 2009: 1 μg Cadmium/g creatinine, WHO/FAO 2011: 5.24 μg Cd/g creatinine [137]).

Chronic exposure to cadmium induces the production of the metallothionein protein that binds the metal and reduces its toxic effects. However, acute intermittent exposures may elude this mechanism and induce severe toxic responses.

Cadmium is a metal that has no essential biological function and may interfere with normal neurological development via different mechanisms [138]. Many studies have examined the neurological consequences of early exposure to cadmium.

In a recent study, the urinary levels of cadmium in 1,305 Bangladeshi women in the early stages of pregnancy and the levels of their children at 5 years of age were assessed. Both the maternal urinary cadmium levels and the concurrent urinary cadmium levels in their children were inversely associated with the intelligence of the children at 5 years of age [139]. A recent analysis of subsets of children from the National Health and Nutrition Examination Survey (NHANES) (1999–2004) suggested that children who have higher urinary cadmium concentrations may be at increased risk for learning disorders and are more likely to require special needs education [140].

Elevated cadmium exposure in adults has been linked to a variety of neuropsychological deficiencies such as reading difficulties, behavioral problems [141], poor visual-motor performance, complaints of decreased concentration [142], reduced attention, psychomotor speed, and memory [143] as well as lower cognitive scores among elderly adults with [144] or without concomitantly elevated zinc exposure [145].

Antother recent study evaluated the associations between neurocognitive exam scores and a biomarker of cumulative cadmium exposure among adults in the NHANES III. The results provide support for the evidence suggesting that cadmium exposure may be associated with diminished neurocognitive performance in adults. The relationships observed in this study were detected at cadmium exposure levels that are typical of US adults and are below the current WHO/FAO reference level [146].

On the other hand, the usual overlapping exposures to lead and cadmium make difficult the relative contribution of each metal on the observed effects. Lead and cadmium both cross the immature BBB and accumulate in the developing brain [147]. A significant correlation between high levels of cadmium and lead in hair and hyperactivity has been shown in children, with decreased verbal development and lower IQ. Lead and cadmium probably affect different aspects of intelligence. Lead levels are associated with a reduction in IQ, whereas increasing cadmium levels correlate with decreased verbal capacity [148]. At cadmium levels below the median, there was a significant interaction between the two metals that was antagonistic during the early pregnancy period. However, this antagonistic interaction occurred at a very low level for both cadmium and lead [149]. These findings suggest that there may be a dose-dependent interaction between prenatal lead and cadmium with respect to the effects of these heavy metals on neurodevelopment. They also demonstrate the biological complexities of examining the neurodevelopmental effects of co-exposure to multiple toxicants. For this reason, further research is necessary in this area.

Copper and Zinc: Cognition Loss

Serum Free Levels and Cognitive Function

Free copper appears to be a player in cognitive decline. Association of lower cognitive function with increased free copper may imply intoxication from this element. Perhaps when its deregulation is induced by deficiency, the free fraction increases even as total copper in the body decreases, and higher values of free copper were associated with lower cognitive function, which is of pathophysiological importance in cognitive decline [150]. Free copper may be a risk factor in the development of impaired cognition. It has been hypothesized in some studies [151] that there is a relationship between the physiological levels of copper in cognitively normal individuals and their cognitive performance. They showed a significant inverse correlation of the serum levels of free copper with both Mini Mental State Examination and attention-related neuropsychological tests scores [151]. Free copper may modulate attention skills via a disturbing action on the neurons of the locus coeruleus. A hint connecting copper metabolism to neuronal viability of the brain structures involved in attention comes from the fact that in the normal brain, copper deposition is higher in the prefrontal cortex, nucleus caudatus, substantia nigra, and locus coeruleus, all structures that have been associated with attention [151].

Copper Toxicity and Cognition Loss in Alzheimer Disease

There is a strong temporal association with the use of copper plumbing in developed countries and the epidemic of AD. But association does not prove causation, many other things are also associated with development. It has been postulated that the environmental factor is beef eating and AD is a prion disease. Beef eating is certainly associated with development, but there is no supporting evidence that AD is a prion disease [152, 153]. Organic copper is absorbed into the bloodstream and taken up by the liver, which then funnels it into safe channels. The inorganic copper that is absorbed directly into the blood is the part of the free pool which is the readily available and potentially toxic copper of the blood. Some authors have described that the blood free copper pool is increased in AD. Evidence sustains that the higher level correlates negatively with cognition in AD [154]. It has been postulated that inorganic copper ingestion from drinking water and copper supplements is a major factor in triggering the epidemic of AD [152, 154]. Toxicity of this element may be causing a decline in cognition in the aging general population [154].

There are other risk factors for AD fitting with the idea of a toxic role for copper: (a) Age, (b) Apolipoprotein E4 (ApoE4) genotype, (c) High fat diet, (d) elevated homocysteine levels, (e) Certain “iron management” genes alleles (certain hemochromatosis and transferring alleles), and (f) Certain Wilson disease (WD) gene (ATP7b) alleles: [152]:

  1. (a)

    Age: Age fits with any risk factor because it simply increases the amount of exposure [155].

  2. (b)

    Apolipoprotein E4 genotype: It is believed that copper binding ApoE alleles help remove copper from the brain. ApoE4 confers risk (because it has no copper binding cysteine), while ApoE2 is protective (ApoE2 has two copper binding cysteines) and ApoE3 is slightly protective (ApoE3 has one copper binding cysteine).

  3. (c)

    High fat diet: A high fat diet causes risk of AD. There is a correlation between fat intake and prevalence of AD across countries. It seems to be a risk factor for AD and cognition loss, and appears to work in conjunction with inorganic copper ingestion.

  4. (d)

    Iron management genes alleles: Elevated homocysteine levels are a risk factor for AD as well as for artheroscherosis. Homocysteine binds copper, and then oxidizes cholesterol to intermediates toxic neurons. Hemochromatosis and transferring alleles convey an increased risk of AD. Iron and copper cause toxicity in the same way, by increasing oxidant damage.

  5. (e)

    WD gene alleles: Single nucleotide polymorphisms associated with the gene ATP7b, the WD gene, affects risk of AD. Certain alleles of ATP7b cause higher copper levels in WD and these higher levels increase risk of AD.

Copper Toxicity and Mild Cognitive Impairment Subjects

In patients affected by AD, serum copper not bound to ceruloplasmin (“free” copper) appears elevated, slightly but significantly enough to distinguish AD subjects from healthy elderly subjects. Free copper can help in discriminating mild cognitive impairment (MCI) subjects from healthy subjects, but not on an individual basis [156]. The clinical condition of MCI is characterized by memory impairments and is verifiable via objective measures. It precedes the clinical definition of dementia in severity. The importance of an accurate diagnosis of MCI lies in the fact that, despite the mildness of the condition, MCI is normally considered a a precursor of AD. AD is an irreversible, progressive neurodegenerative disorder, characterized by a gradual appearance of cognitive deficits, leading to full dementia. It is a genetically heterogeneous syndrome which includes a broad spectrum of phenotypes. Overt pathological alterations in the AD brain are diverse and include neuron loss, synapse loss, amyloid plaques, neurofibrillary tangles, and microgliosis, as well as functional changes, included metal imbalance, oxidative stress, and changes in cell cycle mediators [156]. Free copper also seems to disturb cognitive performances in healthy subjects. Ceruloplasmin and free copper levels increase in inflammatory conditions. However, copper increase in general circulation can also be explained in terms of its release consequent to neuronal death, and this could be the reason why it is not structural to ceruloplasmin. Free copper is also among the WD diagnostic tools which loses balance in the presence of defects in the ATPase 7B, the protein responsible for the correct copper incorporation into nascent ceruloplasmin. Defects in copper assemblage into ceruloplasmin, because of a minor ATP7B genetic defect such as heterozygosis for WD mutations must be taken into consideration also for AD [156]. Free copper is a fraction of copper loosely bound to and exchanged among albumin and micronutrients such as amino acids (histidine) and peptides. Free copper can easily cross the BBB, probably due to its binding to amino acids in processes mediated by some amino acid transport systems. Metal-related abnormalities (mainly copper, iron, zinc) have been shown to be related to Aß and tau toxicity, leading to AD pathology. CSF Aß, total, and hyperphosphorylated tau proteins are core CSF markers of AD dynamics and have been recently tested as markers for MCI, particularly in relation to brain imaging alterations or clinical symptoms.

Free copper could be a predictor for those patients with a more severe decline [156] and altered serum copper homeostasis predicts cognitive decline in MCI [157]. In one study [157] a significant elevation was observed in the ratio of copper to iron in serum in MCI subjects who subsequently progressed to dementia. This elevation appears to be transient as subjects with early AD were nearly identical to controls and stable MCI subjects and longitudinal data show progressive MCI cases trend downward over time. Altered serum copper homeostasis may serve as a biomarker to identify subjects with subject memory complaints who are at risk of developing further cognitive decline [157].

Zinc Deficiency Associated with Cognition Loss and Alzheimer Disease

Patients with AD are zinc deficient. Zinc has many protective roles in neurons, and zinc deficiency may play a causal role in AD. Zinc therapy appears to at least prevent some cognition decline. In addition to restoring normal levels, it reduces serum free copper levels in AD [158]. Serum zinc declines with age in people for unknown reasons. The neurons of many parts of the brain have high zinc levels, and it is clear that this element plays many critical roles in neurons. In some neurons, high concentrations of zinc are secreted along with glutamate into the synapse. Glutamate initiates firing and zinc quenches, or shuts down, the firing. With inadequate zinc, glutamate-induced firing persists and can damage the neuron [158].

Serum zinc levels decline with aging, but patients with AD present lower levels of serum zinc [159], therefore, patients with AD are zinc deficient by serum status. ZnT3 is the pump that loads synaptical vesicles with the metal. The content of these vesicles is secreted into the synapse and the released zinc plays many important neuronal functions [153].

Extracellular amyloid plaques in the AD brain are avid zinc binders, further depleting available zinc for neurons. One of its important neuronal functions is to limit glutamate neuronal firing. Glutamate excitotoxicity damages neurons and may be a problem in many neurodegenerative diseases. Excess glutamatergic excitotoxicity is believed to be a common occurrence in many neurodegenerative disorders, including AD [158].

Another possible mechanism by which low availability of zinc in the brain can have harmful effects is through failure of adequate inhibition of calcineurin. Increased neuronal calcineurin activity as a causative factor in AD is postulated, because it is increased in AD brain. It affects many downstream biochemical functions adversely. Calcineurin activity is increased by exposure to β-amyloid and inhibited by zinc. It seems increasingly likely that neuronal deficiency is playing an important, perhaps a key, role in decreasing neuronal function and increasing damage leading to cognition loss in AD. The zinc depletion of aging, exaggerated in AD, and the loss of ZnT3 function with aging, exaggerated in AD, leads to severe neuronal deficiency and neuronal damage. Increased zinc in the brain may allow it to displace copper from sites where copper is generating oxidant radicals, and thus reduce the damage from copper.

Zinc therapy significantly slows cognition loss in AD, significantly lowering blood-free copper in patients with AD, which may occur in the brain as well thus limiting copper toxicity in this manner. Zinc could be mediating these effects, interacting with copper, while also independently stabilizing neuronal health [153].

To add to the problem created by systemic zinc deficiency there is another mechanism in the AD brain that depletes neurons of much-needed zinc. The beta amyloid plaques, which build up in the AD brain, are avid binders of this metal. Thus, it seems likely that the neurons of the AD brain are seriously lacking in available zinc and many are probably injured and die as a result [158].

Patients with AD are more zinc deficient than age-matched controls, and that deficiency by amyloid plaques further depletes the neurons of zinc. We pointed out how important adequate zinc is for neuronal health [158], and ingestion of inorganic copper in drinking water and zinc deficiency both contribute to cognition loss [153].

Other Trace Elements

Aging is associated with neurobehavioral deficits. Certain brain areas are more vulnerable to neuronal degeneration than others, reflecting an altered resistance to stress of the tissue itself and/or the lack of adequate immunological defense mechanisms in these regions. Calcium and iron are mediators of the aging process in the normal brain. Enhanced calcium levels are related to apoptosis. Excess concentrations of a number of elements in the brain are capable of producing harmful effects by displacing some essential elements, while in turn, numerous toxic and essential elements have been reported to be imbalanced in AD.

AD is the most common cause of dementia among older people. Trace elements may be important in the pathogenesis of AD. Metal ions are concentrated in senile plaques, neurofibrillary tangles, and CSF. These findings support this notion. Several metals have been proposed as pathogenic cofactors in AD, but various toxic heavy metals (i.e., cadmium, lead, and mercury) are especially prevalent in nature because of their high industrial use. These metals serve no biological function and their presence in tissues reflects contact between the organism and its environment. Metals could be connected to the risk factors for dementia or the pathophysiology of dementia. Although toxicity and the resulting threat to human health are a function of the concentration of a contaminant, chronic exposure to arsenic, cadmium, mercury, and lead at relatively low levels can cause adverse effects [160].

Arsenic exposure induces changes that coincide with most of the developmental, biochemical, pathological, and clinical features of AD and associated disorders [160, 161]. Inorganic arsenic at high doses is a known neurotoxin with both neurodevelopmental and neurocognitive consequences. From a neuropathological standpoint, arsenic exposure has been associated with an increase in the production of β amyloid, hyperphosphorylation of tau protein, oxidative stress, inflammation, endothelial cell dysfunction and angiogenesis, all of which have been linked to cognitive dysfunction and are proposed mechanisms underlying AD [160].

Lead is the most historically pervasive and well-established neurotoxic pollutant. Lead can cause white matter damage, cell death, and changes in cellular architecture. It is believed to interfere with functions essential for neuronal homeostais, such as inhibiting glycolytic enzymes in neurotransmitter metabolism [160]. Workers with high blood lead concentrations present an association among mild impairment of attention, verbal memory, and linguistic processing. Lead affects specific areas in the brain such as the hippocampus and frontal cortex. Neural systems subserving language functions are more sensitive than other cognitive functions to perturbations from the effects of lead. Nervous system symptoms such as irritability, poor attention and concentration, forgetfulness, depressed effect, and sleep disturbance are common after lower doses. Even at very low levels, lead is associated with impaired cognitive function in children. Chronic, low-dose exposure to lead may adversely affect cognitive function in older age in several ways. Several possible mechanisms that could result in structural changes in the brain support the hypothesis that there is a relationship between lead and cognitive decline. Lead could increase apoptosis. It causes changes in cellular architecture, increases oxidative stress, or enhances vascular or inflammatory mechanisms. Lead alters the permeability of the BBB and is accumulated within the astroglia that is an essential element for the maintenance of the neuronal environment. Exposure to lead interferes with several calcium-dependent processes and activates protein kinase C, which has been implicated in neurotoxicity. Exposure to lead in a non-occupational setting is associated with accelerated decline in cognition [160].

Oxidative stress appears to play a major role in chronic cadmium-induced hepatic and renal toxicity. The lateral choroids plexus sequesters mercury, cadmium, arsenic, and lead. This is probably an important mechanism to protect the CSF and the brain from fluxes of heavy metals in the blood. However, mercury and cadmium can directly damage the choroid plexus [160]. Blood cadmium concentrations lower than the level known to produce acute toxicity do not affect cognitive functions [160].