Neural Correlates of Auditory Attention in an Exogenous Orienting ...
"Theoretical and Experimental Approaches to Auditory and Visual Attention"
description
Transcript of "Theoretical and Experimental Approaches to Auditory and Visual Attention"
![Page 1: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/1.jpg)
"Theoretical and Experimental Approaches to Auditory and Visual Attention"
Cold Spring Harbor April 20, 2008
Ervin Hafter and Anne-Marie Bonnel Department of Psychology
U.C.Berkeley
“A role for memory in Shared Attention”
![Page 2: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/2.jpg)
Simplest definition of attention:
a process inferred when responses on one task are affected by responding simultaneously to another
![Page 3: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/3.jpg)
S(+)
S(0)
S(+)
S( )
Detection Identification
Energy detection model: compare stimuli in the signal epochs
Two paradigms
S(0) is the level of the standard.
![Page 4: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/4.jpg)
Are
a un
der
the
RO
C
1.00
Signal Levels in z
B
.50
.70
.60
.90
.80
1.5 2.01.00.50 2.5
A
% correct in 2A
FC
Ideal detection based on energy in the signal epochs.
Pro
babi
lity
den
sity
A
B
S(0)
S(+)
S(+)
S(0) λ
λz
Predicts d’B > d’A
![Page 5: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/5.jpg)
But, Macmillan (w/tonal pedestal) and Bonnel et al. (w/visual pededestal) found
d’detection > d’identification
![Page 6: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/6.jpg)
Stock Market National Debt University salaries
Audio demo comparing detection to identification.
![Page 7: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/7.jpg)
Both attributed this to use of Transients
![Page 8: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/8.jpg)
Thinking that, unlike the case with energy, responses to transients might be pre-attentive,
Bonnel et al. (1992) tested in a dual task with independent stimuli on side-by-side LEDs.
![Page 9: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/9.jpg)
Thinking that, unlike the case with energy, responses to transients might be pre-attentive,
Bonnel et al. (1992) tested in a dual task with independent stimuli on side-by-side LEDs.
Detection was comparable to performance found with S instructed to attend to only one LED;
identification showed a tradeoff indicative of a shared attentional resource.
![Page 10: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/10.jpg)
Thinking that, unlike the case with energy, responses to transients might be pre-attentive,
Bonnel et al. (1992) tested in a dual task with independent stimuli on side-by-side LEDs.
Detection was comparable to performance found with S instructed to attend to only one LED;
identification showed a tradeoff indicative of a shared attentional resource.
Caveat: “Perceptual grouping” in identification
![Page 11: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/11.jpg)
S(0)
S(+)
S(-)
Visual Auditory
17 msCRT
500 Hz
Some changes in Berkeley
![Page 12: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/12.jpg)
S(+)
S(0)
S(+)
S( )
Detection Identification
S( )
![Page 13: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/13.jpg)
Ideal energy detection
Pro
babi
lity
den
sity
zλ λ
A
B
S(0)
S(+)
S(0)
S(+)
S(+)
S(0)
S(-)
0 +1-1 +2-2
λ
λ
λ
CC
1.00
Are
a un
der
the
RO
CSignal Levels in z
B
.50
.70
.60
.90
.80
1.5 2.01.00.50 2.5
A
% correct in 2A
FC
CC
![Page 14: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/14.jpg)
Use the strictest possible criterion for asserting that there is no cost of shared attention:
compare performance in the dual-task to that found in the separate single tasks.
![Page 15: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/15.jpg)
Support for transient hypotheses:
1) detection > identification
2) No cost in detection
3) Tradeoff in accord with instructions in identification
Attention-Operating Characteristic
d’au
diti
on
d’vision
3
2
1
3210
50%,50%
20%,80%
50%,50%
20%,80%
80%,20%
80%,20%
are single tasksand
![Page 16: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/16.jpg)
Sharing-Index Attentional Operating Characteristic (SIAOC) normalizes data from a dual task in terms of the single-task controls.
d’dual
d’single
SI =
Sharing Index Vision)
Sha
ring
Ind
ex (
Aud
itio
n) 1.0
.8
.6
.4
.2
.2 .4 .6 .8 1.00
SIAOC
![Page 17: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/17.jpg)
These data clearly imply that detection of transience put no demand on shared attention, unlike discrimination of energy.
SIAOC
Sha
ring
Ind
ex (
Aud
ition
)
Sharing Index (Vision)
.2 .4 .6 .8 1.0
1.0
.8
.6
.4
.2
![Page 18: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/18.jpg)
A more direct test of the idea that detecting transience (change per se) doesn’t require shared attention simply removes transients as information.
![Page 19: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/19.jpg)
A more direct test of the idea that detecting transience (change per se) doesn’t require shared attention simply removes transients as information.
![Page 20: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/20.jpg)
Often called a reminder
Another classic ΔI/I
A classic ΔI/I
![Page 21: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/21.jpg)
d’
Same signal levels
1.0
2.0
3.0
4.0
200 400 600 800 1000 1200 1400 1600 1800
Gap duration (ms)
nogap
Energy detection
Auditory single task
no gap
gap
![Page 22: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/22.jpg)
Dashes suggest absolute identification, i.e. comparisons to long-term, context-coded memory, rather than to a sensory trace of the reminder.
1.0
2.0
3.0
4.0
200 400 600 800 1000 1200 1400 1600 1800
d’
Gap duration (ms)
Same signal levels
![Page 23: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/23.jpg)
1.0
.8
.6
.4
.2
.2 .4 .6 .8 1.0Sharing Index (Vision)
SI-AOC
“larger” “smaller”
< 300 >
“different”
“different”“same”
represent 500-msec signals.
![Page 24: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/24.jpg)
Test Duration {D}
Integration Time {IT}
Single taskD1 = {D}
{D} < {IT}
d’1
A or V
Impact of increased duration. e.g., , on apparent cost of attention.
![Page 25: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/25.jpg)
Test Duration {D}
Integration Time {IT}
Single taskD1 = {D}
{D} < {IT}
d’1
Dual taskD2 = ½ {D}A
V
d’2=.707d’1
d’2=.707d’1
{D} < {IT}
A or V
![Page 26: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/26.jpg)
. d’1
Single task D1 = {IT}
{D} > {IT}
Test Duration {D}
Integration Time {IT}
A or V
![Page 27: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/27.jpg)
. d’1
Single task D1 = {IT}
{D} > {IT}
Test Duration {D}
Integration Time {IT}
.707d’1< d’2 ≤ d’1
.707d’1< d’2 ≤ d’1 Dual task ½{D} < {D2} < {IT}
{D} > {IT}
A or V
![Page 28: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/28.jpg)
. d’1
Single task D1 = {IT}
{D} > {2IT}
Test Duration {D}
Integration Time {IT}
A or V
![Page 29: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/29.jpg)
. d’1
Single task D1 = {IT}
{D} > {2IT}
Test Duration {D}
Integration Time {IT}
d’2
Dual task {D2} = {IT}
{D} > {IT}
d’2
*Here, time sharing produces the same result as no sharing.
A or V
![Page 30: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/30.jpg)
A still stronger test forces the subject to use context-coded memory by simply removing the reminder.
“other”
“other”“standard”
“large” “small”
![Page 31: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/31.jpg)
It seems clear that responses based on context-coded memory were limited by sharing of an attentional resource.
.2 .4 .6 .8 1.0
1.0
.8
.6
.4
.2
Sharing Index (Vision)
Sha
ring
Ind
ex (
Aud
ition
)
![Page 32: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/32.jpg)
Perhaps the reason that use of change per se did not provoke a cost of sharing is that it was done in sensory trace (?rehearsal?) memory?
![Page 33: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/33.jpg)
Subjects can be forced to use sensory-trace memory by roving the standard from trial to trial.
Standard Test
‘louder’
‘louder’
‘softer’
‘softer’
Rove
1
3
2
4
TrialCorrectresponse
![Page 34: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/34.jpg)
da
Unlike the case with context-coding, performance fell as the ephemeral sensory-trace faded over time.
Roved levelsFixed levels
Auditory
Identification
0.5
1.0
1.5
2.0
2.5
3.0
GAP Duration (sec)
1 2 3 4 5 6 7 8
50-ms reminders and signals
*Need higher signals in roving
![Page 35: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/35.jpg)
Most intriguing is that despite very poor performance, especially with long delays, there was no cost of sharing.
1.0
.8
.6
.4
.2
.2 .4 .6 .8 1.0Sharing Index (Vision)
Sha
ring
Ind
ex (
Aud
ition
)
8350
Gap (ms)
1022510310
<Gap (ms)>
rove
![Page 36: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/36.jpg)
loud, dim, green, hot, salty, etc.
CompareCompare
Context-coded, long-term memoryExperience
In typical, everyday life, we label sensory stimuli on the basis of comparisons to long-term memory.
TestSensory/ Neural
![Page 37: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/37.jpg)
CompareCompare
Context-coded, long-term memoryExperience
Sensory/ Neural
T1 T2SOA
Reminder
When presented with an adjacent standard, the response may be to
TestSensory/ Neural
loud, dim, green, hot, salty, etc.
![Page 38: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/38.jpg)
CompareCompare
Context-coded, long-term memoryExperience
TestSensory/ Neural
ReminderSensory/ Neural
When presented with an adjacent standard, the response may be to simply ignore it, labeling the test in accord with long term memory.
T1 T2SOA loud,
dim, green, hot, salty, etc.
![Page 39: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/39.jpg)
CompareCompare
Context-coded, long-term memoryExperience
TestSensory/ Neural
Reminder
Compare
Sensory/ Neural
Rehearsal Memory
Without a reliable context-coded memory, S must compare the test to the reminder the in sensory trace memory.
T1 T2SOA loud,
dim, green, hot, salty, etc.
louder, dimmer, greener, hotter, saltier, etc..
![Page 40: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/40.jpg)
CompareCompare
Context-coded, long-term memoryExperience
TestSensory/ Neural
“loud, dim, green, hot, salty, etc.”
T1 T2SOA
Reminder
Compare
Our audio/video dual-task shows these comparisons to be independent, i.e., no cost of sharing.
Conversely, these comparisons were limited by a shared attentional resource.
Sensory/ Neural
Rehearsal Memory
louder, dimmer, greener, hotter, saltier, etc..
![Page 41: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/41.jpg)
Okay, so comparisons to the sensory trace memory produced no cost of shared attention.
What in the world is trace memory?
![Page 42: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/42.jpg)
Okay, so comparisons to the sensory trace memory produced no cost of shared attention.
What in the world is trace memory?
Recently, we’ve approached this in terms of Weber’s Law.
Bringing the lab up to 1834.
![Page 43: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/43.jpg)
I = k I
In signal detection terms, k can be treated as a multiplicative noise
Weber’s Law
![Page 44: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/44.jpg)
3
1
-1
-3
-5
-7Thr
esho
ld 1
0 lo
g (Δ
I/I)
Pedestal (dB)
55 60 65 70 75
IdentificationPed = Sig = 50 ms Gap = 1s
Without Roving: performance based on based on long-term, labeled memory produce a constant Weber fraction
![Page 45: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/45.jpg)
What happens when comparisons are to a roved standard?
Thresholds go up. But in what way?
To answer this, we parse the data in terms of the individual standards, analyzing performance separately for each pedestal level.
![Page 46: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/46.jpg)
3
1
-1
-3
-5
-7Thr
esho
ld 1
0 lo
g (Δ
I/I
)
Pedestal
55 60 65 70 75
No RoveLabeled memoryRoveTrace memory
![Page 47: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/47.jpg)
3
1
-1
-3
-5
-7
Thr
esho
ld (
dB)
Pedestal (dB)
55 60 65 70 75
multiplicative noise
multiplicative + additive noise
I = k I + c
The change in slope implies a second, additive noise, c.
![Page 48: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/48.jpg)
6
0
-4
-6
-2
4
2
55 60 65 70 75
![Page 49: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/49.jpg)
ΔI = kI + c
What is c?
Perhaps it is a decision noise associated with responding to the stimulus in trace memory?
Maybe it is simply the result of decay in the trace that makes that adds noise to the remembered amplitude code.
![Page 50: "Theoretical and Experimental Approaches to Auditory and Visual Attention"](https://reader035.fdocuments.in/reader035/viewer/2022062314/56813d10550346895da6c9ac/html5/thumbnails/50.jpg)
Our next plan is to go into fMRI in search of sensory rehearsal. Wish us luck.