Introduction

Iron deficiency (ID) in prenatal and early postnatal life produces long-term alterations in neurobiological processes as well as disruptions in cognitive, social, and behavioral development. ID is particularly detrimental if it occurs during the fetal period or the first few years of life while neural systems are rapidly developing. Even after iron treatment, cognitive abilities in individuals who were iron deficient early in life are rarely fully restored. This review will cover the effects of ID on neural structure and function, cognition, and behavior during infancy and early childhood.

This review is grounded on 5 core principles of nutrients and brain development, which are especially applicable to ID. (1) The brain is not homogeneous; thus, it should not be viewed as a single organ. The brain consists of different types of cells (e.g., neurons and glial cells) and distinct regions (e.g., striatum, hippocampus, prefrontal cortex, cerebellum) that do not operate in the same way nor respond similarly to environmental factors. Any research on the neurobehavioral effects of early ID must specifically relate to the cell types and brain region hypothesized to be affected. (2) The separate regions of the brain have different developmental trajectories. As a result, it is likely that the greatest effects of ID will occur in brain regions that are rapidly developing and whether that development is specifically dependent on iron. It should be noted that regions with strong anatomical connections to these rapidly developing regions may also be at risk, despite having a later developmental trajectory. This occurs because brain regions typically work together in circuits to optimize behavioral functioning. Thus, insults to one brain region may negatively affect development in other later-developing regions. (3) Timing is critical to understand neurobehavioral abnormalities resulting from ID. Outcomes vary depending on the timing of the insult as well as what is developing in specific regions of the developing brain during the time of the insult. For example, ID during the late prenatal period leads to delayed neural processing speed [1] and deficits in recognition memory that can be detected through event-related potentials (ERP) [2]. The recognition memory deficits are consistent with altered development of the hippocampus, which is predominant during the late fetal and early postnatal period. In contrast, ID acquired between 6 and 36 months postnatally is typically not associated with memory deficits, but rather with abnormalities in affect and motor skills. This later phenotype is consistent with the development of systems governing monoamine transmission in the frontal cortex and mid-brain, respectively, which occurs primarily between birth and 3 years of age. Consequently, a “signature effect” in which a micronutrient deficiency generates one specific neurobehavioral phenotype is not likely. (4) Micronutrient deficiencies rarely occur in isolation. In most malnourished populations, it is likely that several co-occurring nutrient deficiencies that affect brain development are also present (e.g., Vitamin A, Vitamin B12, Vitamin D, iodine, selenium, copper, folate, choline, zinc), and thus it is difficult to determine which nutrient is responsible for particular effects. (5) Correct assessments should be used in order to accurately measure the processes and outcomes of interest. The child’s developmental stage must also be taken into account in order to choose assessments that are appropriate for the child and sensitive enough to test the effects of ID. Unfortunately, there are currently no assessments that are both sensitive and specific enough to detect ID through neurodevelopmental measures alone without supplemental serum iron measures. However, it is possible to use tests that assess brain regions that may have been the most affected by ID. For example, if a hippocampal insult is expected following early life ID, a neurocognitive test of hippocampal function will provide more valuable information than a general cognitive test (i.e., the Bayley Scales of Infant Development) which may not be sensitive enough to detect effects on specific neural systems. Likewise, a test specific to prefrontal cortex function will likely be more appropriate for later postnatal ID.

Iron Biology

The developing brain is a highly metabolic organ, consuming 50 % of the energy utilized by the baby [3]. All highly metabolic, rapidly developing organs have high iron requirements because iron supports fundamental metabolic processes (e.g., ATP generation through its role in cytochromes). As will be discussed below, the risk of ID in the developing human is not equal across all pediatric time epochs. The late fetal/early neonatal period and late infancy/toddler period are times when children are more likely to be iron deficient. The neurobehavioral effects of deficiencies in those time periods are a function of which iron-dependent processes are most active. Three fundamental iron-dependent systems are rapidly developing from mid-late gestation through the first 3 postnatal years: monoamine (e.g., dopamine) neurotransmission, myelination, and the hippocampus. Thus, it is not surprising that behavioral studies of children who are iron deficient during that time period show that they have abnormalities in mood/affect, speed of processing, learning, and memory, respectively [2, 46]. Other iron-sensitive systems developing during that time include the striatum and early pre-frontal cortex as evidenced by abnormalities in attention and other high-end processing following iron deficiencies in toddlerhood [6, 7∙, 8]. The iron-dependent biology behind these effects has been well worked out using multidisciplinary approaches from pre-clinical models through human investigation.

The effects on the monoamine system have been known for 40 years [9]. The enzymes that synthesize dopamine and serotonin, tyrosine and tryptophan hydroxylase, respectively, contain iron and their enzymatic activity is dependent on the iron state. Rodent models of prenatal and postnatal dietary iron deficiency show significant changes not only in dopamine synthesis, but also in expression of dopamine D1 and D2 receptor classes and dopamine and serotonin re-uptake mechanisms. These abnormalities are present during the period of iron deficiency but also in the adult animal long after repletion [10], suggesting long-term or permanent reprogramming of the monoamine system. Human studies (see below) corroborate these findings in the neonate [11] through young adulthood [8] in that they demonstrate short and long-term changes in dopamine dependent affect, frontal lobe function and motor activity.

The effects on the fatty acid profile and myelin content of the brain have also been thoroughly investigated. Iron is essential to the enzymatic activity of desaturases that determine the fatty acid profile of myelin [12, 13]. ID causes significant reductions in myelin content and alters the fatty acid profile of that myelin. In addition, oligodendrocytes that synthesize myelin are highly metabolic cells with high iron requirements. Oligodendrocyte expression of myelin basic protein and PLP-1 & -2 is compromised by ID and the effects last beyond the period of ID [14]. Physiologically, these changes to the myelin system likely contribute to the reduced speed of processing observed in neonates [1], infants [15] and toddlers [4] with current or previous ID.

Neuronal differentiation is a highly metabolic process that requires an adequate and steady source of iron. The efficacy of a neuron is dependent in great part on the complexity of its structure. In the developing hippocampus, this translates into the complexity of the apical dendritic arbor, particularly in area CA-1. This area is where genes that regulate synaptic plasticity are particularly active in changing the structure of the region (e.g., elongation, branching, synaptic head maturation) in response to experience. Thus, the hippocampus is particularly important in fundamental recognition learning and memory of objects and orientation in space. Greater arbor complexity and hippocampal size is associated with better learning and memory performance. Hippocampal dendritic arbor complexity, in turn, is largely dependent on two factors: cellular energy metabolism and glutamatergic neurotransmission. Both are compromised by early life ID [1618] with resultant structural abnormalities [19, 20∙], reduced electrophysiologic potential [21] and compromised learning and memory [19, 22]. The effect is due to iron and not anemia as evidenced by these changes being present in non-anemic models of ID where only hippocampal neuronal iron has been reduced through genetic techniques [19, 20∙]. There is a critical time period for iron delivery to the hippocampus, after which provision of iron will not rescue the phenotype (see below) [20∙].

A more recent set of investigations show that ID alters brain gene expression both during the period of ID and into adulthood. These short- and long-term gene changes are numerous (i.e., over 300 genes) and are found throughout the brain [14], specifically in the hippocampus [23] and are particularly salient to the system abnormalities discussed above; i.e., monoaminergic and glutamatergic neurotransmission [10, 24]; myelination [14] and hippocampal synaptic plasticity [19, 20∙, 23].

Based on the biology described above, it should not be surprising that the iron deficient brain does not function properly and results in abnormal behavior patterns during the period of iron deficiency. However, the more important public health issue revolves around the mechanisms of the long-term effects of early life ID that persist in spite of iron treatment. Unlike the hematologic effects (e.g., anemia, microcytosis) that appear to revert to normal with iron therapy, multiple pre-clinical and clinical studies demonstrate that the brain remains abnormal well into the future. The cost to society of this effect in humans in terms of lost intellectual potential (estimated to 10 IQ points), lost job potential and increased risk of adult psychopathologies including depression, anxiety and schizophrenia [8, 25] is the more concerning consequence of early life ID.

Investigations in pre-clinical models provide mechanistic insight into these long-term effects [10, 14, 23, 24, 26, 27∙]. The studies center on two, not mutually exclusive, potential mechanisms. The first revolves around the concept of critical periods of brain development [28]. Fundamentally, this theory states that if the structure is built wrong during a critical period of development, it remains permanently abnormal. Since the 1950’s studies have demonstrated that certain brain areas require proper stimulation at a specific point in development to ensure normal structure and function. The classic work that Hubel and Weisel performed (for which they received the Nobel Prize) demonstrated the importance of appropriately timed visual input to properly form ocular dominance columns in the occipital cortex [29]. Hensch and colleagues [28] have been instrumental in identifying the cellular markers of the opening and closing of critical periods. The opening of the critical period is heralded by the appearance of parvalbumin positive GABA interneurons, while the closing of the period occurs when these interneurons surround the neurons with perineuronal nets made of chondroitin sulfate. Early life iron deficiency delays the opening of the critical period in the hippocampus, but also closes it prematurely, thereby reducing the window of opportunity to have normal development with treatment [27∙]. The timing of the critical period is during the period of rapid dendritogenesis on the hippocampus and the restoration of iron to an iron deficient hippocampus within the critical period rescues it at the gene, structural and behavioral levels [20∙]. Conversely, late treatment results in abnormalities at all of the levels in adulthood [20∙].

The second concept is the developmental origins of adult health and disease (DoHAD) hypothesis [30]. This concept is driven by environmental epigenetic modification of chromatin during early life that results in life-long abnormal gene processing. Brain derived neurotrophic factor (BDNF) is an important synaptic plasticity growth factor that is suppressed by early life ID in the rodent [26, 31∙]. Its expression remains suppressed in adulthood in spite of iron repletion. BDNF is an epigenetically modifiable gene [32], and the possibility exists that early iron deficiency can alter its expression through epigenetic mechanisms. This area deserves further exploration given the large number of brain genes affected by early life ID.

Ultimately, these abnormalities in biochemistry, metabolism, structure, and neurophysiology translate into abnormal behaviors in the preclinical models. Formerly iron deficient adult rodents exhibit abnormalities in learning and memory [22], social-affective behavior [10, 33], paired-pulse inhibition (a model of schizophrenia) [21], and frontal lobe function [34]. These abnormalities map directly onto the types of behaviors that remain abnormal in humans following early life ID (see below).

Human Studies: Linking Pre-clinical to Clinical Investigations

Overall, ID in infancy and early childhood has been linked to lower IQ and school achievement, poorer attention and memory, slower language development, deficits in motor skills, and disruptions in socioemotional development [7∙, 35]. The effects of ID are most detrimental during the prenatal and early postnatal periods (before 3 years) when rapid myelination is occurring. Myelination deficits may slow the speed of processing or decrease synaptic efficiency. In pre-clinical models, ID decreases expression of serotonin [36] and BDNF [26]. Future research should explore interactions between neurotransmitter systems following early life ID because alterations in expression of several neurotransmitters (e.g., serotonin, dopamine, glutamate) have consequences for psychopathology, including anxiety, depression, schizophrenia, and autism. Anxiety, depression and schizophrenia have all been associated with early life ID [8, 25]. The following sections review recent research on the effects of early ID, broken down by the timing of ID (prenatal vs. postnatal) and type of study (intervention vs. observation).

Observational Studies of Fetal/Neonatal ID

The acute effects of ID in the term or preterm newborn include poorer discrimination memory [2], altered temperament and interactions with the caregiver [11], slower speed of neural processing [1] and a greater number of abnormal neurological reflexes [37]. The risk of neonatal ID to long-term neurobehavioral function was initially reported in a study of children in the lowest quartile of cord serum ferritin at birth who demonstrated significantly lower IQ and poorer language ability, tractability, and fine motor skills at 5 years [38]. One-year-old children who were born to diabetic mothers and had lower cord serum ferritin concentrations at birth showed poorer explicit memory as measured by recalling a series of events after a delay [39]. While diabetes mellitus during pregnancy confers multiple risks to fetal neurodevelopment besides iron deficiency (e.g., hypoglycemia and hypoxia), these factors were controlled for in this series of studies [40]. In 3.5-year-olds whose neonatal ID had resolved, those with lower serum ferritin concentrations at birth had poorer immediate, delayed, and working learning and memory [41]. Indeed, the poorness of their performance was directly related to their degree of neonatal ID. Interestingly, providing children with strategies for the memory tasks partially alleviated problems, suggesting that alternative learning strategies may help children with lower iron levels at birth [41]. Finally, offspring of women who had iron deficiency anemia with a hemoglobin concentration <100 g/L during pregnancy have a 3.7 fold greater risk of schizophrenia as adults. The increase in risk was directly related to the degree of maternal ID during pregnancy [25].

Observation Studies of Postnatal ID

The acute and chronic effects of postnatal ID have been well-chronicled in a series of more than 40 observational studies over the past 30 + years [35]. Acute ID resulting in slower neural conduction velocities have been reported in 6-month-olds with ID [15], which has been attributed to disruptions in myelination. Toddlers with ID have a lower developmental quotient than iron sufficient toddlers, and the gap in cognitive functioning between these groups widens with age [6]. Toddlers currently suffering from ID also have socioemotional alterations, including increased hesitancy and wariness in novel situations [5]. These are thought to relate to alterations in monoamine neurotransmission [35]. Infants with ID anemia demonstrate poorer performance for both gross and fine motor skills at 9 months of age [42] and abnormal sleep patterns [43].

Overall, the cognitive, motor, and socioemotional effects of acute ID in infancy carry forward long-term, persisting into adolescence and adulthood long after ID has resolved [35, 44∙∙]. Early postnatal ID negatively impacts visual processing and myelination, as 4-year-old, formerly iron deficient children have longer visual evoked potential latencies than controls who were never anemic [4]. Abnormal sleep patterns [43], an increased risk of anxiety and depression [8], and poorer inhibitory control [45] have been described in formerly ID children, adolescents and adults.

Intervention Studies

The results of iron intervention studies in infants are mixed, mostly as a function of the differences in intervention strategies. In general, two types of iron interventions have been studied: (1) treatment of children with current ID, and (2) prevention supplementation in children who are at risk for ID. Treatment studies are consistently positive in their results, either hematologically, neurodevelopmentally, or both. There are more mixed results in the prevention studies. This is likely due to the uncertainty of ID in the at-risk population (whether iron sufficient or extremely deficient) or the presence of multiple micronutrient deficiencies, which makes it difficult to isolate the impact of iron. Comparisons in treated versus untreated children (or children’s pre-treatment state) demonstrate that treatment is better than no treatment. Treated children versus always iron sufficient controls show residual effects, which suggests that treatment does not resolve all effects. This latter effect may vary by age at ID detection.

Prenatal Intervention Studies

Prenatal iron and folic acid supplementation in pregnant women from rural Nepal who had a high risk of ID resulted in positive effects on neurodevelopmental outcomes, including inhibitory control, working memory, and motor functioning, at ages 7–9 years compared to the control group [46]. In this same study, supplements between ages 12–36 months did not provide additional benefits [47]. As a result, the timing of supplementation is critical, and it appears that earlier (even during the prenatal period) is better for the developing brain [48, 49∙∙]. Another supplementation study during pregnancy found that iron deficient mothers who began iron supplementation before 20 weeks gestation had infants with higher birth weights, a smaller likelihood of low birth weight (LBW) infants, and fewer preterm LBW infants compared to women who started supplementation if iron deficient at 28 or 38 weeks [50]. This study suggests that iron treatment may improve conditions related to poorer cognitive outcomes, including LBW status.

Postnatal Intervention Studies

Several postnatal intervention studies report marked improvements in cognitive and motor skills. The majority of studies report neurological improvements even after the period of iron treatment, suggesting a sensitive or critical period for certain iron-dependent processes in the developing brain [35]. Most studies of iron supplementation in infants with ID show positive effects in the treatment group [51], and many studies of ID prevention in infancy also demonstrate positive neurological and behavior effects in addition to improved iron status [52, 53]. In iron supplementation in children under 2 years of age, those who were anemic in infancy often fail to catch up with non-anemic infants [54]. In one study, 5-year-olds who received iron supplementation in infancy to correct ID anemia still showed lower physical activity levels, less positive affect, and less verbalization [55], again emphasizing the long-term risk when ID occurs early in life. One study showed no effect of iron supplementation, which was attributed to later timing of supplementation (12–36 months) that could have missed an early period of rapid brain development [56].

Preterm and LBW infants are at higher risk for postnatal ID because iron is accreted in the last trimester. In a study of LBW infants born between 2,000 and 2,500 g who received 0, 1, or 2 mg/kg/day of iron supplements between 6 weeks and 6 months of age, treatment efficacy differed by neurodevelopmental outcome [57∙]. At 3.5 years, although there was no effect of the intervention for IQ, 12.7 % of the LBW placebo (i.e., 0 mg/kg/d) group had scores on the Child Behavior Checklist above the clinical cutoff (T > 63; >90th percentile) versus 2.9, 2.7, and 3.2 % for the 1, 2 mg, and normal birth weight groups, respectively. Consequently, iron supplementation in LBW infants during the first 6 months of life may reduce clinical symptoms in early childhood. Another study of VLBW infants (<1,301 g) randomized infants into early (mean = 14 postnatal days) or late iron supplementation (mean = 61 postnatal days), and the results indicated a trend toward better neurocognitive and motor development for the group with early versus late iron supplementation [58].

In general, neurological domains that are affected by acute ID in toddlerhood show persistent negative effects over time even after correction of ID [7∙, 35]. Cognitive abilities such as math and writing are often poorer in formerly iron deficient children, which is consistent with abnormalities in hippocampal and prefrontal cortical development [8].

Conclusions and Potential Confounders in Attributing Outcomes to ID

Is there evidence of a specific neurodevelopmental biomarker for ID? Although there are multiple behavioral sequelae of ID, it is not likely that any are specific to ID. Traditionally recognized biological abnormalities of early life ID are on monoamine neurotransmission, myelination, and neuronal growth/differentiation with their attendant effects on behavior, mood/affect alterations, reduced speed of processing, and poorer learning and memory. However, one or multiple other nutrient deficiencies can cause abnormalities in each of those domains. For example, manganese toxicity and iodine deficiency can cause monoamine abnormalities. Hypoxia, and iodine and LCPUFA deficiencies can cause hypomyelination. Zinc, oxygen and protein deficiency can cause truncated dendritic trees and abnormal synaptogenesis. Thus, while it would be attractive to have a “signature” effect of early life ID on the brain, this is not likely. Nutrient deficiencies often occur simultaneously, which makes it difficult to determine the effects of specific micronutrient deficiencies on neurodevelopment.

Similarly, non-nutritional risk factors for the brain also co-occur with ID. Internationally adopted children who spent part of their lives in institutional care are an example of a group that has a high prevalence of ID (approximately 26 %) recognized at the time of adoption [59]. However, many of the correlates of ID that are present in internationally adopted children are also residual symptoms of early institutional care, which usually involves some degree of stimulus and/or social deprivation. For example, the duration of institutional care and severity of ID are independent predictors of both IQ and executive function skills 12 months post-adoption [60]. Additionally, the phenotype of ID varies as a result of the timing of the deficiency, so a consistent marker across time is elusive. As a result, it is probable that other factors affecting development may produce a similar phenotype to ID, but there are likely no neurobiological or behavioral markers specific to ID.

Another possible explanation for observed cognitive and social effects with ID includes functional isolation [11]. As discussed above, children with ID are often hesitant and wary and engage less with their caregivers [11, 61]. Caregivers may view the children as less problematic and give them less attention. As a result, children receive less stimulation from caregivers and the environment, which slows brain development.

Research on the effects of ID on neurodevelopment, cognition, and behavior in infancy and early childhood suggests long-term disruptions in functioning even after restoring normal iron levels. Many of these effects are consistent with enduring striatal-nigral and striatal-frontal dopaminergic circuitry alterations. Interestingly, frontal lobe-mediated functions including attention, planning, and inhibitory control are also affected following early ID [6, 8], which would not be expected as the frontal lobe is not experiencing rapid development during very early postnatal life. It is likely that these outcomes are developmental side effects of rapidly developing brain regions such as the striatum and hippocampus, including suboptimal connections between the frontal lobe and these primary regions. As a result, considering the effects of early ID in relation to timing (neurodevelopmental stage), severity, and duration is essential to understanding outcomes in the heterogeneous brains of infants and young children.