Units of Linguistic Representation in the Brain Colin...

35
Units of Linguistic Representation in the Brain Colin Phillips University of Delaware Main Collaborators Alec Marantz (MIT) David Poeppel (U. Md.) Tim Roberts (UCSF) Tom Pellathy (U. of Delaware) Krishna Govindarajan (MIT)

Transcript of Units of Linguistic Representation in the Brain Colin...

Units of Linguistic Representation in the Brain

Colin Phillips

University of Delaware

Main Collaborators

Alec Marantz (MIT)David Poeppel (U. Md.)

Tim Roberts (UCSF)Tom Pellathy (U. of Delaware)Krishna Govindarajan (MIT)

Linguistic Computation

• Computation over discrete categories• Computations integrate multiple units at multiple levels of

representation over time

Objectives for Cognitive Neuroscience of Language

• How are discrete linguistic categories represented in the brain?• How is sequential linguistic computation carried out in the brain?

Notice that…

• Localization is not an end in itself• We will know when we have found an answer• Strongly influenced by models in linguistics and psycholinguistics

Outline

• Biomagnetic measurements

• ‘Surprise’ measures: many ways of being shocked• Measures of ‘normal’ processing in syntax and semantics

• Surprise measures and phonological categories• Measures of ‘normal’ processing in extracting phonetic features

Magnetic Field Generation

37-channel biomagnetometer

(magnetoencephalography – MEG)

a)

d)c)

b)Averaged waveforms Magnetic field contour map

MRI-overlay of dipolesource localization

Whole-head reconstructionwith dipole source overlay

Stages in Magnetic Source Imaging

Paradigms in Electrophysiological Studies of Language

• Individual trial analysis

Activity of interest masked by irrelevant activitySome recent attempts

• Averaged responses to non-disrupted processing

Data is much cleanerAbility to interpret depends on highly explicit models of non-disrupted processing

• Averaged responses to disrupted processing

Disruption can be isolated by subtraction of data from control conditionBy far the most productive paradigm to date

Disruption of Semantic Processing

• N400

‘John spread the warm bread with butter/socks.’

‘When you go to bed, turn off the living room light/giraffe.’

‘I ordered a ham and cheese sandwich/scissors.’

(Kutas & Hillyard 1980)

• Elicited by semantic anomaly• Possible generator in anterior medial temporal lobe (McCarthy et al. 1995)

Disruption of Syntactic Processing

• P600

‘The broker hoped to sell the stock was acrook.’

‘The broker persuaded to sell the stock was acrook.’

(Osterhout & Holcomb 1992; Hagoort et al.1993)

• Elicited by Syntactic anomaly• Location of generator uncertain at present; ERP scalp distribution is

broad

Additional Syntactic ‘Surprise’ Responses

• Left Anterior Negativity (~300ms)

What do you wonder who they caught __ at __ by accident?The plane took we to paradise and back.The doctor forced the patient was lying.

(Neville et al. 1991; Kluender & Kutas 1993; Coulson et al. 1998; Osterhout et al. 1994)

• ELAN (~150ms)

The scientist criticized Max’s of proof the theorem.(Neville et al. 1991, Friederici 1995)

• Evidence bearing on correct linguistic analysis

The successful woman congratulated himself on the promotion. P600 elicited(Osterhout & Mobley 1995)

• Time course of different anomaly responses constrain models oflanguage processing (Friederici 1995)

Revised Understanding of N400

• All words elicit N400

• N400 strength varies with cloze probability in sentence contexts, isreduced by priming in word-lists

• N400 decreases across sentence, as constraints on continuationsincrease

• N400 may largely reflect normal process of integrating/updatingsemantic representations – averaging picks out synchronizedsemantic processing

• Syntactic anomaly responses may similarly reflect components ofnormal sentence processing – depends on a suitably explicit modelof undisrupted sentence processing.

Objects of Phonological Perception

• Phonemes

e.g. d, t, b, p, a, i, u

• Collections of features

d = [voice] + [alveolar] + [stop]t = [voiceless] + [alveolar] + [stop]

[voiced] b, d, g, j , ¶, z, z , m, n, [voiceless] p, t, k, f, , s, s , c

• Outcome of Perception is categories

A A A A A A B A A A A A A A A A A B A A A A A A A B …

Mismatch Response to Auditory Oddball (~180-250 msec)

Discontinuous acoustic sensitivity

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

0 10 20 30 40 50 60 70 80

Pro

port

ion

of /d

æ/ R

espo

nses

Voice Onset Time (ms)

B

B

B

B

BB

0-10 10-20 20-30 30-40 40-50 50-600

10

20

30

40

50

60

70

80

90

100

% 'D

iffer

ent'

Res

pons

es

Pairs from /dæ/-/tæ/ VOT Continuum

Results from behavioral identification and discrimination tasks with /dæ/-/tæ/ continuum

• Recent ERP/MEG studies showing effects of discontinuous sensitivity to speech continua

Näätänen et al. 1997: MMF amplitude reflects cross-linguistic vowel difference inFinnish and Estonian speakers

Dehaene-Lambertz 1997: MMN amplitude greater for across-category contrasts thanwithin-category contrasts

‘Categorical Perception’ ≠ Category Labeling

• Discontinuous acoustic sensitivity often co-occurs with categorylabeling, and may be an important substrate for categorization, butthey are not the same

• Discontinuous discrimination without phonological category labeling

Infant speech discrimination (Eimas et al. 1971)Discontinuous discrimination of non-native sounds at short ISI (Werker & Logan 1985)Discontinuous discrimination of non-speech sounds (Pisoni 1977)Discontinuous discrimination of human speech sounds in non-human animals (Kuhl &Miller 1978)

• Phonological category labeling without discontinuous discrimination

Vowel perception (Fry et al. 1962, Pisoni 1973)Perception of words and phonemes in sine-wave speech (Remez et al. 1981, 1994)

0 20 40 60 80 100 120 140 160 180 2000

10

20

30

40

50

60

Stimulus Number

Voi

ce O

nset

Tim

e (m

s)

perceptual boundary

Design for Phonological Mismatch Studies

Responses to /dæ/ as Standard and Deviant Stimulus

Difference Wave for /dæ/ Mismatch (200ms marked)

Overlay of Channels in /dæ/ Difference Wave

MRI-Overlay for /dæ/ Mismatch (1 subject)

0 20 40 60 80 100 120 140 160 180 20010

20

30

40

50

60

70

Stimulus Number

Voi

ce O

nset

Tim

e (m

s)

perceptual boundary

Design for phonological mismatch control

Comparison of /dæ, tæ/ Mismatch with Control

Interim Conclusion

• Phonological mismatch paradigm only elicits MMF when many-to-one ratio matches phonological categories

• Auditory cortex has access to representations of phonologicalcategories

However …

• Does not tell us whether categories are computed by any part ofauditory cortex – information could be feedback from higher area

• Does not automatically extend to additional aspects of discretephonological representations

--phonological categories cued by more complex acoustic cues--phonological features which categories are built out of

Towards Isolating Phonetic Features

[+voice] b, d, g[-voice] p, t, k

• Acoustic cue is similar: voice onset time• Perceptual boundary differs:

b-p ~20msd-t ~28msg-k ~35ms

• Phonological mismatch paradigm, with added within-categoryvariation in place-of-articulation

ta, pa, ka, ka, pa, ta, da, pa, ka, pa, pa, ta, pa, ka, ba, ta, ta, ka, pa,ta, pa, ka, ka, ga, pa, ka …

(Phillips, Pellathy & Marantz, in progress)

[voice] mismatch: response to /bæ, dæ, gæ/ as standard

[voice] mismatch: response to /bæ, dæ, gæ/ as deviant

[voice] mismatch: difference wave to /bæ, dæ, gæ/

Non-disrupted Speech Processing

• Relatively well-understood domain: perception of phonemes andphonological features

• Extremely well-controlled stimuli: synthetic speech can be perfectlycontrolled

Extraction of spectral information for vowel perception

Formants

F1

F2

F3

F4

Fundamentalfrequency (F0)

Spectrogram of /u/ in ‘boot’

Effect of tone-frequency on M100 latency (n=5)(Roberts & Poeppel 1996)

F0 F1 F2 F3a(m) 100 710 1100 2540a(f) 200 850 1220 2810

i(m) 100 280 2250 2890i(f) 200 310 2790 3310

u(m) 100 310 870 2250u(f) 200 370 950 2670

Vowel formants–spectral composition (in Hz)

i (male) i (female) a (male) a (female) u (male) u (female)100

105

110

115

120

125

130

M10

0 la

tenc

y (m

s)

Latency variation due to category (n=6)

Significant effect of vowel category on latency:F(2,5)=27.2, p<0.0001

Vowel Category is best predictor of M100 latency(Poeppel, Phillips, Yellin, et al. 1997)

Male /a/F0: 100HzF1: 720Hz

Female /a/F0: 170HzF1: 720Hz

Male /u/F0: 100HzF1: 300Hz

Female /u/F0: 170HzF1: 300Hz

/a/ /u/

Male voice

Female voice

Single-formant vowel stimuli

F1 determines M100 latency to single-formant vowels (n=9)(Phillips, Govindarajan, Poeppel, et al. 1998)

a (male) a (female) u (male) u (female)130

135

140

145M

100

late

ncy

(ms)

/a/ left /u/ left /a/ right /u/ right100

105

110

115

120

125

M10

0 La

tenc

y (m

s)

Category and Hemisphere

Tone-complexes based on single-formant vowels

F1 affects M100 latency in left-hemisphere,but not in right-hemisphere

In contrast…

• Ragot & Lepaul-Ercole (1996) latency of ERP responses to speech-like sounds tracked F0 rather than formant-like energy bands.

• R&L stimuli contain energy bands which are formant-like, but withunusually high formant amplitude relative to amplitude offundamental

• Increased formant amplitudes evoked responses less likely to trackformant values

• Evoked responses tracking formant frequencies may depend onstimuli having overall spectral shape of speech

Conclusions and Prospects

• Objective is to track down components of linguistic computation

• Pre-requisites

Ability to track activity in space and timeHighly explicit models of real-time linguistic computation

• Surprise-based paradigms continue to yield useful findings,measures of normal processing beginning to provide useful findings

• Further progress requires parallel efforts in behavioralexperimentation, computational modeling and data analysis