Vocal Learning in Adulthood: Investigating the mechanisms of vocal imitation using MRI of the vocal tract and brain 2015-2018

DOI

This collection contains behavioural and brain activation data from 3 laboratory studies of speech imitation. Each of the three studies involved behavioural and imaging (MRI) test sessions in which participants were familiarised with novel auditory speech targets, and were asked to imitate them as closely as possible. Across the three studies, there were variations in the type of sounds imitated, and in the participant populations tested. The behavioural data in this collection contain measures of speech imitation performance, for different sound categories and at during different stages of the test session. These are expressed using selected acoustic properties of speech recordings, and based on the anatomical configurations of the vocal tract as measured with MRI. Neuroimaging data show the locations of brain activity and speech-related representations as statistical maps in normalised 3D brain space, related to the stages of speech imitation and the type of sound imitated. In addition to the three experimental studies, a database of videos depicts the dynamics of the vocal tract and the sounds of speech during the production of English sentences and syllables from 55 speakers. We are genetically programmed to acquire spoken language from our environment, and infants can master native pronunciation in multiple languages without explicit tuition. However, in adolescence and adulthood we have a limited capacity to achieve accurate pronunciation of unfamiliar languages, and even highly competent users of a language learned in adulthood might speak with a strong, non-native accent. The UK currently lies behind other EU nations in foreign language skills at school and in the workplace, therefore research into skill development has important educational and economic implications. Previous research has used functional MRI to measure how the activity of functional systems in the brain changes as new speech sounds are learned. This work has described the integration of novel speech sounds into a talker's existing speech repertoire as it becomes more familiar. Within this, however, some talkers are more successful than others at attaining native-like pronunciation of new sounds, and this variability is correlated with the activation, size and structural composition of specific brain structures. To date, the cognitive neuroscience of speech learning has assessed performance by measuring or judging the sounds of speech. However, there isn't a simple one-to-one relationship between how speech sounds in the air, and the underlying movements in the vocal tract. Therefore, an important missing piece of the puzzle is an understanding of how vocal articulations relate directly to brain activation during learning. Recent developments in MRI have shown that rapid 'real-time' anatomical scans can be used to create videos of the interior of the mouth and vocal tract, such that we can view how the lips, tongue and voice box are moved and configured to perform speech. We propose to combine real-time MRI with measures of brain structure and function to investigate the relationship between brain and behaviour during the learning of new speech sounds. Our project will focus on short-term learning of novel and unfamiliar vocal sounds by native speakers of English, where the participants will aim to imitate the sounds accurately and with native-like pronunciation. In an MRI scanner, listeners will repeatedly produce these novel sounds, as well as native sounds of English, while scans of the brain will measure neural activity. Interleaved with these scans, we will collect real-time images of the vocal articulators during imitation. Acoustic recordings will be made using an in-scanner microphone, and we will additionally collect high-resolution images of brain structure from each participant. With these data, we will investigate vocal learning in terms of i) the functional brain systems supporting learning, ii) the acoustic accuracy of vocal output and iii) the accuracy of the movements generating the sounds, as well as the relationship between these elements. Further, we can explore how individuals differ in their performance of vocal learning in terms of meeting these acoustic and motor targets, and how this relates to their brain structure and function. Across a series of experiments, we will also investigate how novel sounds are sequenced into new words, and assess the effects of expertise on vocal learning by comparing English speakers with vocal performance experts. Our proposed approach is truly novel, with potential to make groundbreaking developments in the cognitive neuroscience of vocal communication. To deliver the project, we have assembled a uniquely qualified research team with shared and individual expertise in phonetics, cognitive neuroscience and MRI. The data will directly inform our understanding of language learning and accent acquisition. Further, our new methodology has potential future application to other questions, such as the assessment of brain and behaviour relationships in patients with speech impairments following brain injury.

The data are based on 3 experimental studies and a database of speech production involving acoustic recordings of speech, as well as MRI images of speech articulation and neural activation during listening to, and imitating, speech stimuli. Questionnaires were constructed to assess prior language and musical experience. Participants were healthy young adults recruited predominately from a local Psychology department participant pool; in Experiment 3 an additional group of expert participants was recruited from the professional singing community via experimenter contacts, music colleges and online singing interest groups.

Identifier
DOI https://doi.org/10.5255/UKDA-SN-853317
Metadata Access https://datacatalogue.cessda.eu/oai-pmh/v0/oai?verb=GetRecord&metadataPrefix=oai_ddi25&identifier=ad5fff7305c53abe52aea5be4e51372f50e29f0f76472dcce145422e4166c24e
Provenance
Creator McGettigan, C, Royal Holloway, University of London; Miquel, M, Queen Mary University of London; Carey, D, Royal Holloway, University of London; Waters, S, Royal Holloway, University of London; Kanber, E, Royal Holloway, University of London
Publisher UK Data Service
Publication Year 2018
Funding Reference Economic and Social Research Council
Rights Carolyn McGettigan, Royal Holloway, University of London; The Data Collection is available for download to users registered with the UK Data Service.
OpenAccess true
Representation
Language English
Resource Type Numeric; Still image; Audio; Video
Discipline Acoustics; Biology; Engineering Sciences; Life Sciences; Mechanical and industrial Engineering; Mechanics and Constructive Mechanical Engineering; Psychology; Social and Behavioural Sciences
Spatial Coverage Egham, Surrey; United Kingdom