Neural Mechanisms of Perceptual Categorization as Precursors to Speech Perception


Book Description

Perceptual categorization is fundamental to the brain’s remarkable ability to process large amounts of sensory information and efficiently recognize objects including speech. Perceptual categorization is the neural bridge between lower-level sensory and higher-level language processing. A long line of research on the physical properties of the speech signal as determined by the anatomy and physiology of the speech production apparatus has led to descriptions of the acoustic information that is used in speech recognition (e.g., stop consonants place and manner of articulation, voice onset time, aspiration). Recent research has also considered what visual cues are relevant to visual speech recognition (i.e., the visual counter-parts used in lipreading or audiovisual speech perception). Much of the theoretical work on speech perception was done in the twentieth century without the benefit of neuroimaging technologies and models of neural representation. Recent progress in understanding the functional organization of sensory and association cortices based on advances in neuroimaging presents the possibility of achieving a comprehensive and far reaching account of perception in the service of language. At the level of cell assemblies, research in animals and humans suggests that neurons in the temporal cortex are important for encoding biological categories. On the cellular level, different classes of neurons (interneurons and pyramidal neurons) have been suggested to play differential roles in the neural computations underlying auditory and visual categorization. The moment is ripe for a research topic focused on neural mechanisms mediating the emergence of speech representations (including auditory, visual and even somatosensory based forms). Important progress can be achieved by juxtaposing within the same research topic the knowledge that currently exists, the identified lacunae, and the theories that can support future investigations. This research topic provides a snapshot and platform for discussion of current understanding of neural mechanisms underlying the formation of perceptual categories and their relationship to language from a multidisciplinary and multisensory perspective. It includes contributions (reviews, original research, methodological developments) pertaining to the neural substrates, dynamics, and mechanisms underlying perceptual categorization and their interaction with neural processes governing speech perception.




Neural Control of Speech


Book Description

A comprehensive and unified account of the neural computations underlying speech production, offering a theoretical framework bridging the behavioral and the neurological literatures. In this book, Frank Guenther offers a comprehensive, unified account of the neural computations underlying speech production, with an emphasis on speech motor control rather than linguistic content. Guenther focuses on the brain mechanisms responsible for commanding the musculature of the vocal tract to produce articulations that result in an acoustic signal conveying a desired string of syllables. Guenther provides neuroanatomical and neurophysiological descriptions of the primary brain structures involved in speech production, looking particularly at the cerebral cortex and its interactions with the cerebellum and basal ganglia, using basic concepts of control theory (accompanied by nontechnical explanations) to explore the computations performed by these brain regions. Guenther offers a detailed theoretical framework to account for a broad range of both behavioral and neurological data on the production of speech. He discusses such topics as the goals of the neural controller of speech; neural mechanisms involved in producing both short and long utterances; and disorders of the speech system, including apraxia of speech and stuttering. Offering a bridge between the neurological and behavioral literatures on speech production, the book will be a valuable resource for researchers in both fields.




Audiovisual Speech Recognition: Correspondence between Brain and Behavior


Book Description

Perceptual processes mediating recognition, including the recognition of objects and spoken words, is inherently multisensory. This is true in spite of the fact that sensory inputs are segregated in early stages of neuro-sensory encoding. In face-to-face communication, for example, auditory information is processed in the cochlea, encoded in auditory sensory nerve, and processed in lower cortical areas. Eventually, these “sounds” are processed in higher cortical pathways such as the auditory cortex where it is perceived as speech. Likewise, visual information obtained from observing a talker’s articulators is encoded in lower visual pathways. Subsequently, this information undergoes processing in the visual cortex prior to the extraction of articulatory gestures in higher cortical areas associated with speech and language. As language perception unfolds, information garnered from visual articulators interacts with language processing in multiple brain regions. This occurs via visual projections to auditory, language, and multisensory brain regions. The association of auditory and visual speech signals makes the speech signal a highly “configural” percept. An important direction for the field is thus to provide ways to measure the extent to which visual speech information influences auditory processing, and likewise, assess how the unisensory components of the signal combine to form a configural/integrated percept. Numerous behavioral measures such as accuracy (e.g., percent correct, susceptibility to the “McGurk Effect”) and reaction time (RT) have been employed to assess multisensory integration ability in speech perception. On the other hand, neural based measures such as fMRI, EEG and MEG have been employed to examine the locus and or time-course of integration. The purpose of this Research Topic is to find converging behavioral and neural based assessments of audiovisual integration in speech perception. A further aim is to investigate speech recognition ability in normal hearing, hearing-impaired, and aging populations. As such, the purpose is to obtain neural measures from EEG as well as fMRI that shed light on the neural bases of multisensory processes, while connecting them to model based measures of reaction time and accuracy in the behavioral domain. In doing so, we endeavor to gain a more thorough description of the neural bases and mechanisms underlying integration in higher order processes such as speech and language recognition.




Speech Perception and Spoken Word Recognition


Book Description

Speech Perception and Spoken Word Recognition features contributions from the field’s leading scientists, and covers recent developments and current issues in the study of cognitive and neural mechanisms that take patterns of air vibrations and turn them ‘magically’ into meaning. The volume makes a unique theoretical contribution in linking behavioural and cognitive neuroscience research, and cutting across traditional strands of study, such as adult and developmental processing. The book: Focusses on the state of the art in the study of speech perception and spoken word recognition Discusses the interplay between behavioural and cognitive neuroscience evidence, and between adult and developmental research Evaluates key theories in the field and relates them to recent empirical advances, including the relationship between speech perception and speech production, meaning representation and real-time activation, and bilingual and monolingual spoken word recognition Examines emerging areas of study such as word learning and time-course of memory consolidation, and how the science of human speech perception can help computer speech recognition Overall this book presents a renewed focus on theoretical and developmental issues, as well as a multifaceted and broad review of the state of research, in speech perception and spoken word recognition. Particularly interested readers will be researchers of psycholinguistics and adjoining fields as well as advanced undergraduate and postgraduate students.




Phonetics in the Brain


Book Description

Spoken language is a rapidly unfolding signal: a complex code that the listener must crack to understand what is being said. From the structures of the inner ear through to higher-order areas of the brain, a hierarchy of interlinked processes transforms the acoustic signal into a linguistic message within fractions of a second. This Element outlines how we perceive speech and explores what the auditory system needs to achieve to make this possible. It traces a path through the system and discusses the mechanisms that enable us to perceive speech as a coherent sequence of words. This is combined with a brief history of research into language and the brain beginning in the nineteenth century, as well as an overview of the state-of-the-art neuroimaging and analysis techniques that are used to investigate phonetics in the brain today. This title is also available as Open Access on Cambridge Core.




Brain Oscillations in Human Communication


Book Description

Brain oscillations, or neural rhythms, reflect widespread functional connections between large-scale neural networks, as well as within cortical networks. As such they have been related to many aspects of human behaviour. An increasing number of studies have demonstrated the role of brain oscillations at distinct frequency bands in cognitive, sensory and motor tasks. Consequentially, those rhythms also affect diverse aspects of human communication. On the one hand, this comprises verbal communication; a field where the understanding of neural mechanisms has seen huge advances in recent years. Speech is inherently organised in a rhythmic manner. For example, time scales of phonemes and syllables, but also formal prosodic aspects such as intonation and stress, fall into distinct frequency bands. Likewise, neural rhythms in the brain play a role in speech segmentation and coding of continuous speech at multiple time scales, as well as in the production of speech. On the other hand, human communication involves widespread and diverse nonverbal aspects where the role of neural rhythms is far less understood. This can be the enhancement of speech processing through visual signals, thought to be guided via brain oscillations, or the conveying of emotion, which results in differential rhythmic modulations in the observer. Additionally, body movements and gestures often have a communicative purpose and are known to modulate sensorimotor rhythms in the observer. This Research Topic of Frontiers in Human Neuroscience highlights the diverse aspects of human communication that are shaped by rhythmic activity in the brain. Relevant contributions are presented from various fields including cognitive and social neuroscience, neuropsychiatry, and methodology. As such they provide important new insights into verbal and non-verbal communication, pathological changes, and methodological innovations.




Behavioral and Neural Correlates of Human Time Perception, Imagination and Production


Book Description

Rhythmicity is a key component that allows humans to attend to, predict, and respond to the environment. In particular, temporal processing is fundamental to the perception and production of complex sounds, such as speech and music. Neural entrainment hypothesizes that internal oscillators synchronize with external stimuli, providing a unified mechanism for supramodal temporal processing. Extensive research demonstrates the entrainment effect on human time perception for non-speech musical sounds; fewer studies have shown entrainment effects for duration perception in spoken language. To date, it remains unclear how humans encode temporal properties and generate rhythm according to them, and whether and how entrainment mechanisms represent timing information in the brain. My aim is to address these important open questions. In Chapter 1 of my thesis, I reviewed the existing literature and gaps therein. Chapter 2 compared whether entrainment or interval models more accurately predict human time perception. Entrainment models more accurately predicted duration discrimination, but the effect diminished after 2-4 cycles, while interval models predicted more accurately thereafter. Chapter 3 tested entrainment effects on more ecologically valid contexts--speech sounds, and found that entrainment can transfer from tones to speech sounds, suggesting a domain-general entrainment effect with a constraint by acoustical similarity. Chapter 4 examined neural evidence of entrainment in hierarchically organized drumming rhythm. The study found that both auditory and motor regions represent the rhythms imagined by the subjects. A motor-to-auditory information flow was found in all listening conditions without overt movements, suggesting that the motor system actively maintains hierarchical information and exerts a top-down influence on auditory processing and metrical imagery of rhythms. Chapter 5 further investigated rhythm production using self-paced tapping and synchronization, finding that synchronization relies on auditory-motor interaction in beta-band, only observed in individuals who tap relatively stably in the self-paced tapping task without external cues. In summary, this thesis work contributes to the theoretical understanding of how humans perceive, imagine and produce temporal events, particularly in a rhythmic context, at the behavioral and neural levels. My hope is that this work can improve real-life applications and inform work with clinical populations who have timing-related deficits.




Neural Mechanisms of Language


Book Description

This important volume brings together significant findings on the neural bases of spoken language –its processing, use, and organization, including its phylogenetic roots. Employing a potent mix of conceptual and neuroimaging-based approaches, contributors delve deeply into specialized structures of the speech system, locating sensory and cognitive mechanisms involved in listening and comprehension, grasping meanings and storing memories. The novel perspectives revise familiar models by tracing linguistic interactions within and between neural systems, homing in on the brain’s semantic network, exploring the neuroscience behind bilingualism and multilingual fluency, and even making a compelling case for a more nuanced participation of the motor system in speech. From these advances, readers have a more three-dimensional picture of the brain—its functional epicenters, its connections, and the whole—as the seat of language in both wellness and disorders. Included in the topics: · The interaction between storage and computation in morphosyntactic processing. · The role of language in structure-dependent cognition. · Multisensory integration in speech processing: neural mechanisms of cross-modal after-effect. · A neurocognitive view of the bilingual brain. · Causal modeling: methods and their application to speech and language. · A word in the hand: the gestural origins of language. Neural Mechanisms of Language presents a sophisticated mix of detail and creative approaches to understanding brain structure and function, giving neuropsychologists, cognitive neuroscientists, developmental psychologists, cognitive psychologists, and speech/language pathologists new windows onto the research shaping their respective fields.




Timing and Time Perception: Procedures, Measures, & Applications


Book Description

Timing and Time Perception: Procedures, Measures, and Applications is a one-of-a-kind, collective effort to present the most utilized and known methods on timing and time perception. Specifically, it covers methods and analysis on circadian timing, synchrony perception, reaction/response time, time estimation, and alternative methods for clinical/developmental research. The book includes experimental protocols, programming code, and sample results and the content ranges from very introductory to more advanced so as to cover the needs of both junior and senior researchers. We hope that this will be the first step in future efforts to document experimental methods and analysis both in a theoretical and in a practical manner. Contributors are: Patricia V. Agostino, Rocío Alcalá-Quintana, Fuat Balcı, Karin Bausenhart, Richard Block, Ivana L. Bussi, Carlos S. Caldart, Mariagrazia Capizzi, Xiaoqin Chen, Ángel Correa, Massimiliano Di Luca, Céline Z. Duval, Mark T. Elliott, Dagmar Fraser, David Freestone, Miguel A. García-Pérez, Anne Giersch, Simon Grondin, Nori Jacoby, Florian Klapproth, Franziska Kopp, Maria Kostaki, Laurence Lalanne, Giovanna Mioni, Trevor B. Penney, Patrick E. Poncelet, Patrick Simen, Ryan Stables, Rolf Ulrich, Argiro Vatakis, Dominic Ward, Alan M. Wing, Kieran Yarrow, and Dan Zakay.