Blog honen helburua euskaraz hizkuntza patologiei buruz egiten diren ikerketen berri ematea da, eta Marie Pourquié-ren ikerketa proiektu eta interes nagusien aurkeztea.
The Healthy and Impaired Multisensory Talking Brain
Humans are experts in perceiving speech, even though the quality of the auditory speech signal we produce or hear is sub-optimal because of background noise and speaker variability. One reason why we nevertheless experience hardly any perceptual problems when engaged in a face-to-face conversation, is that our brain uses two additional streams of sensory information that generate non-auditory predictions about the upcoming sound. That is, we need to first plan and execute a set of fine-grained motor commands to correctly shape our vocal apparatus before we can produce the correct speech sound and, as a consequence, we actually see these articulatory gestures of an external speaker before we hear the sound. It is well-established that both the preceding motor-information and the preceding visual (i.e., lip-read) information modulate the way in which the self- or externally generated speech sound is processed. However, the effects of motor- and lip-read information on auditory speech processing have always been studied in isolation and current proposal is set-up to determine the multisensory interplay between auditory speech, lip-read speech and self-generated motor commands.