Re: speech/music (Neil Todd )


Subject: Re: speech/music
From:    Neil Todd  <TODD(at)FS4.PSY.MAN.AC.UK>
Date:    Wed, 1 Apr 1998 10:22:57 GMT

Jont Allen wrote: >How do you know? I would expect they are connected at least by the >auditory cortex AC (this is known by now, no? Somebody who knows, help us >here.) The AC is BEFORE speech is recognized, because there is still >a tonotopic representation. That would still be in the bottom up realm, >don't you agree? > >> Some >> information *must* be travelling top-down. It's expectation driven. > Jont here is an edited extract from the chapter I advertised earlier. For references on detailed approaches to the neural architecture of language see for example Seldon (1985), Shallice (1988), Eccles (1989) and Gordon (1990) who marshal the evidence from aphasics to support a three level process of speech perception involving: (a) auditory input feature processing; (b) phonemic identification; and (c) word recognition, in which recognised sequences of phonemes/features trigger lexical access, which then allows access to meaning. Neurological evidence for phonemic processing comes from patients, not with Wernike's aphasia, but with word deafness. In this condition patients have intact hearing, and speech is appreciated as sound, but cannot be understood as language, though music and environmental hearing are intact, as are writing and reading. This implies that pure word deafness is an inability to decode speech sounds, particularly consonants. This selective deficit appears to have two causes: a prephonemic inability to detect rapid changes, and a disorder of phonemic processing. The first cause seems to be associated with lesions around the primary cortex in the left hemisphere (though vowel perception may be accomplished by both hemispheres). The second cause may have its basis in a disrupted phonemic lexicon, possibly caused by lesions to the left posterior superior temporal lobe extending to the angular gyri. Several sources of evidence support a separate third (semantic) stage. Transcortical aphasias and anomias imply that the speech input can be decoupled from semantic understanding. A common area of damage for these patients is the posterior temporal-inferior parietal (T-IP) region, implying that single word semantic processing takes place in the left T-IP region. Finer dissociations of this sort may be taken to further imply that in the normal state, purely verbal knowledge is separate from higher-order visual and other perceptual information (though all the systems involved are highly linked). On the issue of the McGurk effect. There are three principal cues for lip reading, lip modulation rate, maximum lip velocity and maximum lip amplitude, all of which can be extracted from visual image flow, which is coded in the temporal lobe (V5). Recent evidence from blood flow studies (Calvert et al, 1997) of silent lip reading provide evidence that "silent lip reading activates auditory cortical sites also engaged during the perception of heard speech. In addition it appears that auditory cortex may be similary activated by visible pseudospeech but not by nonlinguistic closed-mouth movements. This adds physiological support to the psychological evidence that lipreading modulates the perception of auditory speech at a prelexical level and most likely at the stage of phonetic classification". Calvert et al (1997) Activation of auditory cortex during silent lipreading. Science, 276, 593-596. Eccles, J. (1989) Evolution of the brain: Creation of the self. (Routledge: London). Gordon, B. (1990) Human language. In R.Kesner and D. Olton (Eds.) Neurobiology of Comparative Cognition. (Lawrence Erlbaum: New Jersey). pp 21-49. Shallice, T. (1988) From neuropsychology to mental structure. CUP: New York. Seldon, H.L. (1985) The anatomy of speech perception: Human auditory cortex. In E. Jones and A. Peters (Eds.) Cerebral Cortex. Vol. 4. Association and auditory cortices. Plenum: New York. pp. 273-327. Neil


This message came from the mail archive
http://www.auditory.org/postings/1998/
maintained by:
DAn Ellis <dpwe@ee.columbia.edu>
Electrical Engineering Dept., Columbia University