Abstract
Language comprehension is a complex process involving an extensive brain network. Brain regions responsible for prosodic processing have been studied in adults; however, much less is known about the neural bases of prosodic processing in children. Using magnetoencephalography (MEG), we mapped regions supporting speech envelope tracking (a marker of prosodic processing) in 80 typically developing children, ages 4–18 years, completing a stories listening paradigm. Neuromagnetic signals coherent with the speech envelope were localized using dynamic imaging of coherent sources (DICS). Across the group, we observed coherence in bilateral perisylvian cortex. We observed age-related increases in coherence to the speech envelope in the right superior temporal gyrus (r = 0.31, df = 78, p = 0.0047) and primary auditory cortex (r = 0.27, df = 78, p = 0.016); age-related decreases in coherence to the speech envelope were observed in the left superior temporal gyrus (r = − 0.25, df = 78, p = 0.026). This pattern may indicate a refinement of the networks responsible for prosodic processing during development, where language areas in the right hemisphere become increasingly specialized for prosodic processing. Altogether, these results reveal a distinct neurodevelopmental trajectory for the processing of prosodic cues, highlighting the presence of supportive language functions in the right hemisphere. Findings from this dataset of typically developing children may serve as a potential reference timeline for assessing children with neurodevelopmental hearing and speech disorders.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Introduction
The seminal work of Broca1 and Wernicke2 led to a left hemisphere model for language, but recent studies have demonstrated a role for the right hemisphere, primarily in the processing of prosody3. Prosody refers to the stress, rhythm, and pitch changes in speech, which can convey information that is both emotional and linguistic in nature4. Emotional prosody refers to cues about the speaker’s affect or attitude5. Linguistic prosody refers to cues about the language structure such as lexical stress contrasts (e.g., permit vs permit) and word boundaries (e.g., hotdog vs hot dog)6,7. Lesion studies in adults have suggested that both types of prosody are processed by the right hemisphere, though some have argued that the left hemisphere is still involved, at least in the processing of linguistic prosody8.
In children, the neural bases of prosodic processing have not been well studied. Right hemisphere injuries are known to affect prosodic processing in childhood9,10,11, though neuroimaging studies suggest bilateral representation12,13. In electroencephalography (EEG) and magnetoencephalography (MEG) studies, the speech envelope is often extracted for analyses because of its ability to capture the rhythm and stress patterns of speech14. The speech envelope does not contain pitch information, and therefore reflects primarily linguistic rather than emotional prosody, but has still been widely used to study prosodic processing in children. For example, an MEG study computed coherence between the speech envelope and low frequency brain rhythms, related to prosody, and found high coherence in the bilateral auditory and temporal regions15. Similarly, an EEG study demonstrated impaired processing of low frequency speech envelopes in children with dyslexia16. However, these studies employed laboratory stimuli or paradigms, such as isolated sentences or word pairs.
Recently, studies of prosodic processing in children have begun using natural continuous speech. For example, 9-month-old infants were shown to be capable of tracking prosodic stress in response to continuous infant-directed speech17. Another study demonstrated speech envelope coherence in children aged 7–8 years during a stories listening task was related to reading level18. Unfortunately, many of these studies involve participants of limited age ranges and fail to describe potential changes in neural support for prosodic processing across development.
Here, we used MEG to map brain regions coherent with the speech envelope during a stories listening paradigm in a large cohort of typically developing children. Based on the right hemispheric specialization of prosodic processing in adults, we hypothesize that neural coherence to the speech envelope will become increasingly right lateralized with age.
Methods
Participants
Eighty children (42 female; ages 4.03–18.99 years, mean = 11.58) participated in this study. All were typically developing native English speakers with no prior history of neurological or psychological deficit and no MEG or MRI contraindications. Informed written consent was obtained from all participants and/or their legal guardians. Compensation was provided for participation and travel costs. The study was approved by the Institutional Review Board at Cincinnati Children’s Hospital Medical Center (OH, USA; data collection site), and the Research Ethics Board at the Hospital for Sick Children (Toronto, ON, Canada; analysis site). All methods and analyses were performed in compliance with the relevant guidelines and regulations and in accordance with the Declaration of Helsinki.
Behavioural measures
Participants were drawn from two studies employing different language assessments as part of their cognitive batteries. Thirty children completed the Peabody Picture Vocabulary Test (PPVT-4; Dunn and Dunn19) and the Expressive Vocabulary Test (EVT-2; Williams20), assaying receptive and expressive vocabulary, respectively. We compute the arithmetic mean for PPVT-4 and EVT-2, to estimate overall language ability. The remaining 50 participants completed the Clinical Evaluation of Language Fundamentals (CELF-421); we report on the Core Language Score, here. All participants (n = 80) completed the Weschler Nonverbal Scale of Ability (WNV; Wechsler and Naglieri22), and the Edinburgh Handedness Inventory (EHI; Oldfield23).
MEG acquisition
MEG data were recorded using a 275-channel whole-head CTF system (CTF MEG Neuro Innovations, Inc., Coquitlam, BC, Canada) at 1200 Hz sampling. Participants were awake, studied in supine position with localization coils placed over the nasion and pre-auricular points for continuous monitoring of head position.
Participants completed a passive stories listening paradigm that has been extensively used by our group (e.g., Barnes-Davis et al.24,25,26). Participants listened to five stories narrated with a neutral affect by a female native English speaker. Stimuli were delivered through bilateral ear inserts connected to a distal transducer and tubing system (Etymotic Research, IL, USA). Each story was presented one sentence at a time in trials that lasted 2–3 s. After each story trial, participants listened to a noise trial consisting of a speech-shaped noise with identical duration, spectral content, and amplitude envelope.
MRI acquisition
Prior to MR imaging, radiopaque fiducial markers were placed over the nasion and preauricular points to facilitate offline co-registration with MEG data. Whole-brain 3D T1-weighted images were acquired at 1 mm isotropic on a 3.0 T Philips Achieva or Ingenia Elition scanner (flip angle = 8°, TE = 3.5 ms, TR = 7.9 ms).
Data analysis
Data preprocessing
MEG data were preprocessed using FieldTrip27 routines running in MATLAB R2020b (MathWorks Inc., Natick, MA, USA). The continuous MEG data were bandpass filtered from 0.1 to 100 Hz and power line noise (60 Hz) was attenuated using a 1 Hz-width notch filter. The data were then downsampled to 256 Hz and decomposed into 30 components using InfoMax Independent Component Analysis (ICA)28. The number of components for the ICA decomposition was chosen based on previous research29,30. Eyeblink and cardiac artifacts were visually identified based on their topographic map and time series, then subsequently rejected (mean number of rejected components = 1.94). Following artifact scrubbing, the data were segmented into 48 epochs of interest, each 3700 ms in duration.
Speech envelope extraction
The speech envelope was calculated as the absolute Hilbert transform of the speech signal, low-pass filtered at 10 Hz (forward and reverse) using a third-order Butterworth filter, and downsampled to 256 Hz to match the scrubbed MEG data.
Head and source modelling
For each participant, a realistic single-shell head model was constructed based on individual T1-weighted MRIs31. A whole-brain source-model was constructed based on the centroids of a ‘100-unit per hemisphere’ parcellation scheme32; the approach yielded 184 locations, across the cerebrum and cerebelli. For each participant, individual source models were generated by warping template positions to individual space.
Source reconstruction and coherence analysis
For each participant, trials were zero-padded to 4 s, filtered from 0.5 to 12 Hz to isolate prosodic and phonemic rhythms, and converted to the frequency domain using a multi-taper fast Fourier transformation with ± 2 Hz smoothing. A cross-spectral density matrix was computed, and stories and noise trials were concatenated for covariance estimation and subsequent spatial filter construction; however, only the stories data were assessed for coherence to the speech envelope using Dynamic Imaging of Coherent Sources (DICS) beamforming33. Individual template positions were then relabelled using the initial 3D template positions defined in MNI space.
Results
Behavioural results
Participants scored average to above average for language assessments (mean for combined PPVT-4 EVT-2 scores = 117.5, SD = 14.5; mean CELF-4 Core Language Score = 106, SD = 13.5). Nonverbal performance was within the average range (mean WNV Total = 107.1, SD = 13.1). The EHI score is computed as a conventional laterality index; we divide the range into 3 equal categories for handedness classification. Across the group, 90.0% of participants were right-handed, 6.25% left-handed, and 3.75% ambidextrous.
Regions supporting speech envelope tracking in childhood
To identify regions coherent with the speech envelope across the group, Z-scores were computed for speech envelope coherence, for each node. A total of 10 regions had coherence values 2 standard deviations greater than the mean (Fig. 1, Table 1). Coherence to the speech envelope was localized in the bilateral frontal and temporal lobes, the right supramarginal gyrus, and the left insula. These 10 regions were used in subsequent correlation analyses, to assess developmental changes in speech envelope tracking.
Age-related changes in speech envelope tracking
Pearson correlations for age versus speech envelope coherence was computed for each of the 10 regions showing significant speech tracking (see, Fig. 1). To control the Type I error rate, we applied an FDR (q = 0.01) across analyses. We observed significant age-related increases in coherence to the speech envelope in the right superior temporal gyrus (r = 0.31, df = 78, p = 0.0047) and the right primary auditory cortex (r = 0.27, df = 78, p = 0.016), and significant age-related decreases in the left superior temporal gyrus (r = − 0.25, df = 78, p = 0.026) (Fig. 2).
For each of the 3 significant regions of interests, partial correlations were computed between speech envelope coherence and age, sex, and handedness; the partial correlations permit characterization of bivariate relationships, while controlling for the other variables. In all cases, only age was significantly correlated to speech envelope coherence (Table 2).
Discussion
Here, we characterize maturation of the language network supporting the processing of prosodic cues. We assessed speech envelope tracking in children performing a stories listening task in MEG, and found that neuromagnetic signals coherent to the speech envelope localized to the bilateral perisylvian cortex. With age, coherence increased in the right temporal lobe and decreased in the left superior temporal gyrus. Findings suggest a distinct neurodevelopmental trajectory for the processing of prosodic cues, where areas in the right hemisphere become increasingly specialized.
Across the group, coherence to the speech envelope was localized to bilateral perisylvian cortex, including the right primary auditory cortex, the right supramarginal gyrus, the bilateral superior temporal gyri, left insula, and the left frontal lobe. These regions are known for auditory processing and language comprehension. Our results are consistent with speech envelope tracking studies in adults34,35,36. The shared regions among adults and children suggest that maturation of speech envelope tracking involves functional reorganization of pre-existing brain regions rather than anatomical reorganization of the network over time. Our coherence values are also consistent with previous studies of neural coherence to the speech envelope in adults34,35.
With age, coherence to the speech envelope increased in right temporal lobe and decreased in one region of the left superior temporal gyrus. These age-related differences are consistent with a previous MEG study that compared typically developing children to healthy adults37. The adult group had higher coherence to the speech envelope than children in a temporal region of interest that included the superior temporal gyrus. Thus, results suggest that young children use a mix of left and right language areas to track the speech envelope, with progressive support by right perisylvian regions with age, as the left perisylvian regions become less tuned to prosodic information related to the speech envelope. This pattern may provide insight into the brain differences underlying the improvement in prosodic processing during development, where the reallocation of neural resources creates a more efficient process.
We note that our study is limited by the omission of discrete pitch reference signals. Pitch is particularly important for conveying emotional information (i.e., emotional prosody)38. Others have documented age-related increases in right hemispheric activity in response to pure tones39. Thus, the processing of pitch may develop in a similar manner, as both pitch tracking and pure tones tracking involve attending to the fundamental frequency of a sound wave. Our stories were narrated by a single speaker who was instructed to speak with a neutral affect. As such, the stimuli have reduced emotional prosody, compared to entirely naturalistic stimuli. Future research involving speech envelope tracking in children should also include speech and noise contrasts, variations in speech-in-noise ratios, or speech containing irony, sarcasm, or other dialogues dependent on prosodic cues. Finally, future large-scale studies may have better sensitivity to potential nonlinearities in age-related effects not observed with the current cohort (see Sharma et al. 202140, for example). Altogether, our results reveal a strong neurodevelopmental trajectory for the processing of prosodic cues and highlight the presence of supportive language functions in the right hemisphere. The results may potentially serve as a reference, where the maturational stage of children with language impairments can be compared to that of their typically developing peers. Our study emphasizes the importance of considering suprasegmental features in language development research. The precise timing of the language network’s reorganization, including its supportive functions, may be crucial for predicting language impairments and informing therapeutic practices.
Data availability
The data analyzed in the current study are not publicly available. Code used for analyses were developed from open-source toolboxes; scripts will be made available from the corresponding author on reasonable request.
References
Broca, P. Rémarques sur la siège de la faculté du langage articulé, suivi dʼune observation dʼaphémie. Bull. Soc. Anat. Paris 2, 330–357 (1861).
Wernicke, C. Der Aphasische Symptomencomplex: Eine Psychologische Studie auf Anatomischer BASIS (Cohn & Weigert, 1874).
Lindell, A. K. In your right mind: Right hemisphere contributions to language processing and production. Neuropsychol. Rev. 16, 131–148 (2006).
Ross, E. D. Affective prosody and the aprosodias. Princip. Behav. Cogn. Neurol. 2, 316–326 (2000).
Ross, E. D. The aprosodias: Functional-anatomic organization of the affective components of language in the right hemisphere. Arch. Neurol. 38, 561–569 (1981).
Beach, C. M. The interpretation of prosodic patterns at points of syntactic structure ambiguity: Evidence for cue trading relations. J. Mem. Lang. 30, 644–663 (1991).
Cutler, A., Dahan, D. & van Donselaar, W. Prosody in the comprehension of spoken language: A literature review. Lang. Speech 40, 141–201 (1997).
Witteman, J., van Ijzendoorn, M. H., van de Velde, D., van Heuven, V. J. J. P. & Schiller, N. O. The nature of hemispheric specialization for linguistic and emotional prosodic perception: A meta-analysis of the lesion literature. Neuropsychologia 49, 3722–3738 (2011).
Cohen, M., Prather, A., Town, P. & Hynd, G. Neurodevelopmental differences in emotional prosody in normal children and children with left and right temporal lobe epilepsy. Brain Lang. 38, 122–134 (1990).
Cohen, M. J., Branch, W. B. & Hynd, G. W. Receptive prosody in children with left or right hemisphere dysfunction. Brain Lang. 47, 171–181 (1994).
Trauner, D. A., Ballantyne, A., Friedland, S. & Chase, C. Disorders of affective and linguistic prosody in children after early unilateral brain damage. Ann. Neurol. 39, 361–367 (1996).
Leipold, S., Abrams, D. A., Karraker, S. & Menon, V. Neural decoding of emotional prosody in voice-sensitive auditory cortex predicts social communication abilities in children. Cereb. Cortex 33, 709–728 (2023).
Wang, A. T., Lee, S. S., Sigman, M. & Dapretto, M. Neural basis of irony comprehension in children with autism: The role of prosody and context. Brain 129, 932–943 (2006).
Myers, B., Lense, M. & Gordon, R. Pushing the envelope: Developments in neural entrainment to speech and the biological underpinnings of prosody perception. Brain Sci. 9, 70 (2019).
Molinaro, N., Lizarazu, M., Lallier, M., Bourguignon, M. & Carreiras, M. Out-of-synchrony speech entrainment in developmental dyslexia. Hum. Brain Mapp. 37, 2767–2783 (2016).
Power, A. J., Colling, L. J., Mead, N., Barnes, L. & Goswami, U. Neural encoding of the speech envelope by children with developmental dyslexia. Brain Lang. 160, 1–10 (2016).
Menn, K. H., Michel, C., Meyer, L., Hoehl, S. & Männel, C. Natural infant-directed speech facilitates neural tracking of prosody. Neuroimage 251, 118991 (2022).
Mandke, K. et al. Neural sampling of the speech signal at different timescales by children with dyslexia. Neuroimage 253, 119077 (2022).
Dunn, L. M. & Dunn, L. M. Peabody Picture Vocabulary Test: PPVT 4 (Pearson Assessments, 2015).
Williams, K. T. Expressive Vocabulary Test, Vol. 42 (Pearson Assessments, 2007).
Semel, E., Wiig, E. H. & Secord, W. A. Clinical Evaluation of Language Fundamentals 4th edn (CELF-4) (Pearson Assessments, 2004).
Wechsler, D. & Naglieri, J. A. Wechsler Nonverbal Scale of Ability (Harcourt Assessments, 2006).
Oldfield, R. C. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia 9, 97–113 (1971).
Barnes-Davis, M. E., Merhar, S. L., Holland, S. K. & Kadis, D. S. Extremely preterm children exhibit increased interhemispheric connectivity for language: Findings from fMRI-constrained MEG analysis. Dev. Sci. 21, e12669 (2018).
Barnes-Davis, M. E. et al. Functional hyperconnectivity during a stories listening task in magnetoencephalography is associated with language gains for children born extremely preterm. Brain Sci. 11, 1271 (2021).
Barnes-Davis, M. E. et al. Extracallosal structural connectivity is positively associated with language performance in well-performing children born extremely preterm. Front. Pediatr. 10, 821121 (2022).
Oostenveld, R., Fries, P., Maris, E. & Schoffelen, J.-M. FieldTrip: Open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011, e156869 (2010).
Bell, A. J. & Sejnowski, T. J. An information-maximization approach to blind separation and blind deconvolution. Neural Comput. 7, 1129–1159 (1995).
Escudero, J., Hornero, R., Abasolo, D., Fernandez, A. & Lopez-Coronado, M. Artifact removal in magnetoencephalogram background activity with independent component analysis. IEEE Trans. Biomed. Eng. 54, 1965–1973 (2007).
Treacher, A. H. et al. MEGnet: Automatic ICA-based artifact removal for MEG using spatiotemporal convolutional neural networks. Neuroimage 241, 118402 (2021).
Nolte, G. The magnetic lead field theorem in the quasi-static approximation and its use for magnetoencephalography forward calculation in realistic volume conductors. Phys. Med. Biol. 48, 3637–3652 (2003).
Shen, X., Tokoglu, F., Papademetris, X. & Constable, R. T. Groupwise whole-brain parcellation from resting-state fMRI data for network node identification. Neuroimage 82, 403–415 (2013).
Gross, J. et al. Dynamic imaging of coherent sources: Studying neural interactions in the human brain. Proc. Natl. Acad. Sci. 98, 694–699 (2001).
Brodbeck, C., Hong, L. E. & Simon, J. Z. Rapid transformation from auditory to linguistic representations of continuous speech. Curr. Biol. 28, 3976-3983.e5 (2018).
Brodbeck, C. et al. Parallel processing in speech perception with local and global representations of linguistic context. Elife 11, e72056 (2022).
Koskinen, M., Kurimo, M., Gross, J., Hyvärinen, A. & Hari, R. Brain activity reflects the predictability of word sequences in listened continuous speech. Neuroimage 219, 116936 (2020).
Kolozsvári, O. B. et al. Coherence between brain activation and speech envelope at word and sentence levels showed age-related differences in low frequency bands. Neurobiol. Lang. 2, 226–253 (2021).
Scherer, K. R. Vocal communication of emotion: A review of research paradigms. Speech Commun. 40, 227–256 (2003).
Yamazaki, H. et al. Cortical hemispheric asymmetries are present at young ages and further develop into adolescence. Hum. Brain Mapp. 39, 941–954 (2018).
Sharma, V. V. et al. Beta synchrony for expressive language lateralizes to the right hemisphere in development. Sci. Rep. 11(1), 3949. https://doi.org/10.1038/s41598-021-83373-z (2021).
Acknowledgements
The study was supported, in part, by grants provided by the Research Institute of Cincinnati Children's Hospital Medical Center [DSK], the SickKids Foundation at the Hospital for Sick Children [DSK/KAG/EWP], and the National Institute of Neurological Disorders and Stroke at the National Institutes of Health [R21NS106631, DSK/HMG]. The content is solely the responsibility of the authors and does not necessarily represent the official views of the funders.
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Li, K.E., Dimitrijevic, A., Gordon, K.A. et al. Age-related increases in right hemisphere support for prosodic processing in children. Sci Rep 13, 15849 (2023). https://doi.org/10.1038/s41598-023-43027-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-43027-8
- Springer Nature Limited