Thesis Example 5 - Informatics at IUPUI
Transcript of Thesis Example 5 - Informatics at IUPUI
THE EFFECT OF THINK-ALOUD ON TIMED PERFORMANCE
IN USABILITY TESTING
Timothy H. Altom
Submitted to the faculty of the University Graduate School in partial fulfillment of the requirements
for the degree Master of Science
in the School of Informatics Indiana University
August 2006
i
Accepted by the Faculty of Indiana University;
in partial fulfillment of the requirements for the degree of Master of Sciencein Human-Computer Interaction
Master's Thesis Committee
Mark Larew, PhD
Anthony Faiola, PhD
Lois Ann Scheidt, MIS, MPA
ii
TABLE OF CONTENTS
Page LIST OF TABLES vii
LIST OF FIGURES ix
ACKNOWLEDGEMENTS xi
ABSTRACT xiii
CHAPTER ONE: INTRODUCTION AND BACKGROUND 1 Introduction to Subject 1 Importance of Subject 4
CHAPTER TWO: LITERATURE REVIEW 5 Think-Aloud Protocols 5 Time-on-Task as Performance Measurement 8 Use of Dyads 10 Hypothesis 12
CHAPTER THREE: METHODOLOGY 13 Participants 13 Treatments 13 Procedures 23 Analysis 25
CHAPTER FOUR: RESULTS 26 Time-on-Task 26 Verbalizations 28
CHAPTER FIVE: DISCUSSION 32 Explanation of Outcomes 32
CHAPTER SIX: CONCLUSIONS 36 Limitations 36 Future Research 36 Summary 37
REFERENCES 38
APPENDICES 42 Appendix A: Screener 42 Appendix B: Raw Data, Time 43 Appendix C: Raw Data, Verbalization 45
VITA 46
iv
LIST OF TABLES
Page
Table 4-1: ANOVA Results of Time-On-Task for Each Task 33
Table 4-2: Monad-Think-Aloud and Dyad Verbalization Counts 34
v
LIST OF FIGURES
Page Figure 3-1: Optimal Path For Task 1 18
Figure 3-2: Starting Page For Task 1 18
Figure 3-3: Search Page For Task 1 19
Figure 3-4: Optimal Path For Task 2 20
Figure 3-5: Target Page For Task 2 20
Figure 3-6: Optimal Path For Task 3 21
Figure 3-7: Target Page For Task 3 21
Figure 3-8: Optimal Path For Task 4 22
Figure 3-9: Target Page For Task 4 22
Figure 3-10: Optimal Path For Task 5 23
Figure 3-11: Search Page For Task 5 23
Figure 3-12: Optimal Path For Task 6 24
Figure 3-13: Target Page For Task 6 24
Figure 3-14: Optimal Path For Task 7 25
Figure 3-15: Home Page For Task 7 With Link 25
Figure 3-16: Target Page For Task 7 26
Figure 3-17: Optimal Path For Task 8 27
Figure 3-18; Search Page For Task 8 27
Figure 4-1: Graph of Mean Times for All Groups on All Tasks. 31
Figure 4-2: Monad-Think-Aloud Scatterplot Verbalizations and Time 35
Figure 4-3: Dyad Scatterplot of Verbalizations and Time 36
vi
ACKNOWLEDGEMENTS
As odd as it may seem, I will begin by thanking someone I’ve never met in person, and
who does not know me at all. Her name is Andrea Ames. She works for IBM as a
usability specialist. It was her stunning presentation during a Society for Technical
Communication annual conference in 2002 that convinced me that I wanted to be a
usability specialist full-time.
I would like thank in particular my thesis chair, Dr. Mark Larew, whose direction and
patience made this thesis the learning experience it was meant to be.
I would also like to thank Dr. Anthony Faiola, the dynamo who injected so much life and
energy into my program.
Further, I am grateful to Lois Ann Scheidt, whose insights into groups and technology
have enlightened, delighted, and inspired me.
There are many other influences and individuals I have to thank, including Josh Plaskoff,
who taught me the intricacies of information and knowledge; Michael Downey and
Melinda Buher, two friends who pushed me when I needed it, which was all too often;
and Ed Sullivan, Bob Orr, and William Watson, faculty, colleagues, and friends who
helped me, supported me, and took chances on me.
I also want to thank all those friends, colleagues, and utter strangers who recruited test
participants for me, and were test participants for me. And I want to acknowledge the
immeasurable assistance of my wife, who was friend, confidant, sympathetic shoulder,
budget analyst, and fellow researcher.
vii
ABSTRACT
Timothy H. Altom
THE EFFECT OF THINK-ALOUD ON TIMED PERFORMANCE
IN USABILITY TESTING
Eight tasks on a website were performed by participants in three Conditions: Monad-
Silent, in which a single participant was asked to remain silent while performing the
tasks, Monad-Think-Aloud, in which a single participant was instructed to “think aloud”
while performing the tasks, and Dyad, in which two participants were instructed to work
together to complete the tasks. Time-on-task was measured and verbalizations were
counted. Results indicated that participants in the Monad-Silent Condition completed
tasks in less time than in the other Conditions – but only for tasks that required a
relatively long time (> 60 sec) to complete. The times for Monad-Think-Aloud and Dyad
Conditions did not differ significantly. These results suggest that verbalization slows task
performance in usability testing. Implications are discussed for use of time-on-task
measures in usability tests involving verbalizations.
viii
CHAPTER ONE: INTRODUCTION AND BACKGROUND
Introduction to Subject
There is an apocryphal story about a very young but avid television fan in the late
1950s who suddenly and unaccountably developed a limp. Doctor visits and tests
revealed nothing. The boy was healthy in all respects, yet limped everywhere he went.
Finally, in desperation a doctor asked “Son, why are you limping?” The youngster
brightened and responded “My name’s Chester; I work for Mr. Dillon!” Chester Goode,
played by Dennis Weaver, was Matt Dillon’s club-footed deputy on the TV show
Gunsmoke.
The story illustrates the importance of asking why in situations where there is
misunderstanding or mystery about what. In human factors work, we often face
mysterious circumstances, and it is frequently necessary to ask just that question of our
participants: why? Letting test participants tell us why as they execute tasks has become
formalized in “think-aloud protocols.”
Think-aloud protocols have come to occupy an important place in human factors
practice and research. This is hardly surprising, given the nature of human factors work.
The Human Factors and Ergonomics Society has defined the field of human factors
broadly as:
Ergonomics (or human factors) is the scientific discipline concerned
with the understanding of interactions among humans and other
elements of a system, and the profession that applies theory,
principles, data, and other methods to design in order to optimize
1
human well-being and overall system performance. (What is human
factors/ergonomics? 2005)
Of the two components of the interactions that the Society uses in its definition, it
is the human side that is most difficult to evaluate. The “other elements of the system,”
which are engineered devices, are relatively narrower in scope and application compared
with their human counterparts in the interaction process, and display little difference in
activity from unit to unit from the same batch. Humans, by contrast, are general-purpose
devices with wide variability from person to person. Further, the non-human “other
elements” are generally easy to observe and measure. Electronic circuits have currents
and voltages. Mechanical devices have movements. These can be tracked for later study.
Essentially, designed devices are extremely transparent in their activities. They hold few
mysteries, although they may spontaneously offer surprises.
On the human side, behaviors can be observed too, but these are often not enough
to satisfy the needs of testing or design. Engineered devices have only how and what to
present to us. Humans add the dimension of why. Observing behavior alone in human test
participants tells us only what those persons did, but cannot tell us why. While
undoubtedly valuable, knowing only how and what cannot adequately help us predict
future human behaviors. It is prediction that is our holy grail, our fondest hope and
highest goal. Describing only past behaviors is like driving without looking away from
the rear-view mirror. It is imperative that we be able to say with some confidence what
will happen in the future when others manipulate the systems we are designing.
Unfortunately, recording only the behavior of evaluative test participants does not give us
thorough insight into how their successors will respond in the same situations. Humans
2
are not collections of behaviors, and observing behavior alone can take us only so far in
design.
That being said, it should be acknowledged that behaviors can take us quite far
indeed on their own. If the goal is to find major stumbling blocks in users’ ability to
manipulate an interface, behavior by itself is enormously valuable. Nielsen and Landauer
(1993) pointed out that it is possible mathematically to get by with only five users in
usability testing, if there is only a fairly small number of problems whose discovery
pattern fits a Poisson distribution. In such a case, simple behavioral observation is enough
to find the vast majority of significant problems. Although the actual number of users
needed was disputed later by Spool and Schroeder (2001), the basic premise was
unshaken, that behavior observation alone was enough to find enough problems that,
when corrected, make an interface usable.
However, simple behavioral observation has drawbacks. Although it can identify
wrong paths through Web sites, for example, it cannot explain why the users took those
wrong paths, and therefore it is difficult to know just how to fix the problem. If all five of
Nielsen’s users click the wrong button, obviously something is amiss with the button or
something concerning the button’s context. But without having participants’ thoughts to
supplement their behaviors, it can be a hit-or-miss proposition to correct the problem.
The obvious way to elicit users’ thought is to ask them to simply talk while
performing test tasks. Proponents of this “think-aloud” approach believe that it offers
great improvements on behavioral observation alone. (Wright & Monk, 1991; Snyder,
2003).
3
Yet, it must be recognized that we do not understand completely what happens
during think-aloud sessions. Psychological research is helpful in this regard. However,
even the most current psychological literature does not answer critical questions for
usability practitioners and academics, such as whether talking about a task while
performing that task slows down the user in timed tests. Can we obtain the valuable
think-aloud verbalizations while also timing users in their tasks as they solve the
problems in unfamiliar interfaces? If so, testing can become much more efficient, because
we can capture both quantitative and qualitative data simultaneously. But does the
cognitive loading of speech interfere with the user’s problem-solving capacity, such that
both cannot be invoked at the same time without an unacceptable penalty that renders the
testing itself less valuable, or even suspect altogether?
The present study addresses this question, using both single verbalizing
participants (monad-think-alouds) and paired participants (dyads), comparing them to a
Condition in which a single participant completes tasks silently (monad-silent). The
reason for including dyads is the growing prevalence of using two participants in testing
rather than one, because pairs (dyads) can produce many more verbalizations than
monads. Dyads by their nature are think-aloud participants.
Importance of Subject
There can be great benefit derived from both qualitative and quantitative testing.
Combining the two could produce greater efficiencies in design, but could also permit
enhanced interpretation of quantitative results – providing a why to go with the how
much. This can be done, however, only if it can be shown that verbalizing during testing
does not impose a time penalty.
4
CHAPTER TWO: LITERATURE REVIEW
Think-Aloud Protocols
Think-aloud protocols have become extremely popular for usability testing,
especially in prototype testing, as a way of exploring user mental operations (Wright &
Monk, 1991; Snyder, 2003). Borrowed from a psychological research technique, think-
aloud testing protocols call for the testing participant to verbalize, or speak aloud,
whatever occurs to him at that moment. The intent is to capture his thoughts, not just his
actions, and thereby contribute to an understanding of the average user’s mental model
(Davidson et al, 1997).
However, some practitioners have expressed reservations about think-aloud
protocols in testing (Ramey et. al., 2006; Guan et. al, 2006). Wildman (1995) points out
several potential problems with think-aloud: 1) The higher cognitive loading of
verbalizing in addition to problem-solving; 2) Overcoming many participants’ natural
aversion to speaking continually; 3) The wide variability of participants’ volubility,
articulateness, self-awareness, and other attributes; 4) The unreliability of think-aloud
actually reflecting participants’ thoughts.
Some of Wildman’s points are undoubtedly valid. Many participants do, indeed,
have difficulty speaking well or often. However, Wildman offers no evidence in the
article for his contention that verbalization increases cognitive loading, nor that think-
aloud is not representative of participants’ thoughts. He merely asserts the points. Still, he
speaks for many practitioners who mistrust think-aloud even as they use the technique.
5
Many claim that think-aloud produces far more usable data than behavior
observation alone, when measured by the number of problems found, and there is
empirical evidence to support the contention (Ebling & John, 2000; Haak et al, 2003).
Psychologists in the early decades of the 20th century investigated and utilized
what they termed “introspection” (Duncker, 1945) but which later investigators called
“think-aloud protocols”. Think-aloud proved to be particularly efficacious for creating
expert systems, when experts could verbalize while performing tasks at which they were
particularly skilled (Someren, Barnard, & Sandberg, 1994). Think-aloud has proven to be
useful in the study of clinical decision-making. Henry, LeBreck, and Holzemer (1989)
conducted a study in which verbalization was assessed as a possible impediment to
cognitive operations in a clinical simulation. The authors found no impairment.
Still other researchers came to doubt the usefulness of think-aloud, and some
declared it to be nearly worthless, because of the impossibility of an individual actually
knowing his own cognitive operations (Nisbett & Wilson, 1977).
In the midst of the controversy, two advocates for think-aloud protocols, K.
Anders Ericsson and Herbert A. Simon, wrote the book Protocol Analysis: Verbal
Reports as Data (Ericsson and Simon, 1984). Ericsson and Simon were primarily
interested in developing a model and mechanism for analyzing verbal reports as data, but
in the process they created a unified theory of verbalization and its validity in revealing
what is in the subject’s short-term memory. They postulated three levels of verbalization
tasks.
Level 1 tasks are often called “talk-aloud” rather than “think-aloud” because they
are not expected to involve real thought at all, but merely the vocalizing of what is
6
already orally encoded. Even problem-solving at a level 1 is not considered to be
significantly impaired by vocalizing.
Level 2 tasks involve stating thoughts associated with performing the task. This is
not simple recitation, but an actual string of statements about task performance as it
occurs.
Level 3 tasks are statements about reasoning, answering “why” instead of “how”
or “what”.
For Ericsson and Simon, level 1 and 2 task verbalizations are valid as data if
certain other Conditions are met. Based on this model, usability test participants who
merely “sing along” with their instinctive reactions will not experience substantial
cognitive loading from verbalizing on top of the cognitive loading of problem-solving. In
short, when doing simple tasks and not explaining themselves, most think-aloud
participants should not experience much, if any, time increases over silent participants.
The acceptance of Ericsson and Simon’s theories into usability testing has been
mixed at best. Boren and Ramey (2000) have criticized practitioners for ignoring
Ericsson and Simon’s admonitions about not prompting test participants so as to force
them into level 3 verbalizations. Nielsen, Clemmensen, and Yssing (2002) go further and
call the entire theory into question from a practical standpoint. The authors point out that
in the Ericsson and Simon model, test participants can be viewed only as relatively
mechanistic processing units, while in reality participants are as complex and multivariate
in their reactions as anyone else, and should be treated that way.
7
Time-on-Task as Performance Measurement
Empirical research into the performance of silent versus think-aloud monadic
problem-solving has further been complicated by the variety of ways that performance
has been gauged in various studies. For example, in one experiment by Dickson et al
(2000) looking at the effects of verbalizing during time-critical, dynamic decision-
making tasks, a group of participants was given training and practice time with a software
package called “Fire Chief”, which simulated fighting forest fires. The investigators were
examining whether the verbalization of the bases of decisions would degrade
performance, with performance defined as task completion or fewest number of errors.
They cite earlier research that showed with fair uniformity that such verbalization, which
corresponds to Ericsson and Simon’s level 3 tasks, actually result in better performance
in the laboratory. But they also note that such simple situations as were previously
investigated are not reflective of real-world circumstances, in that they were not time-
critical or dynamic (changing as the task progressed). When they used Fire Chief it
introduced those aspects, and they then found that verbalizing the bases of decisions
during time-critical, dynamic operations degraded performance, in this case measured by
the amount of forest saved at the end of the timed test period.
Few studies have focused on time-on-task as a measurement of performance
during verbalization. One study by Strayer et al (2003) that came close to using time-on-
task, indicated a direct link between verbalization and reaction time as a result of using
cell phones while driving. The study found that cell phone conversations increased
reaction time by creating large amounts of inattention blindness. This would not be
8
unexpected given Ericsson and Simon’s contention that level 3 verbalizations, as cell
phone conversations inevitably would be, involve considerable cognitive overhead.
A study done by Loeffler and Polkehn (2000) examined the use of verbalization
on monads in usability testing, but again used completion rate as a measurement. As
others had seen before, they found that in some tasks, think-aloud enhanced task
completion performance when tasks were untimed and not dynamic. However, they also
found the effect to be stronger in some tasks than in others.
This prior research would seem to indicate the basic validity of Ericsson and
Simon’s three-tier model of verbalization, as it relates to performance. In essence, if a test
participant is being asked to verbalize the reasons for his choices, performance
measurements can either increase or decrease, depending on the nature of the task,
possibly because level 3 verbalization involves cognitive overhead that interferes with
time and attention, but enhances decision-making.
This would not seem to bode well for combining think-aloud and time-on-task
protocols. In usability testing, participants can spontaneously switch between levels 1, 2,
and 3, constantly invoking and shedding higher cognitive functions. Further, practitioners
can unintentionally push participants into level 3 verbalizations (Boren & Ramey, 2000).
It is apparent that despite intensive research over the last five decades, think-aloud
is still poorly understood, even as it continues to be of enormous benefit in usability
testing. And, if Ericsson and Simon are right, then time-on-task and think-aloud may
indeed be compatible, so long as the test participant does not spend much time in Level 3
conversation, something the test facilitator may be able to control to some degree.
9
Use of Dyads
As noted previously, one way that verbalization can be enhanced is with the use
of dyads in testing. This approach was originally set out by Miyake (1982). Others, such
as O’Malley, Draper, and Riley (1984) took up the idea and developed it further. The
essential benefit to what Miyake called “constructive interaction” and what is here called
“dyads” is that both participants talk more often and in greater depth than merely asking
them to talk to themselves or to a monitor while alone.
The addition of a second participant, however, would seem to add a great many
new factors that could skew time-on-task measurements, perhaps so many that measuring
time-on-task would again become impractical, even if the participants were strictly held
to Levels 1 and 2 interactions.
The use of dyad groups has proven to be potentially helpful in a variety of
situations. For example, pair programming, where two individuals program together in an
organized fashion, produces more code overall, as well as more reliable code. And pair
programming has been shown to be advantageous in teaching programming, as well
(Williams & Kessler, 2000; Cockburn, 2001; McDowell et al, 2003; McDowell et al,
2002). There is also contradictory evidence of pair programming’s effectiveness. Hulkko
and Abrahamsson (2005) subjected four case studies to statistical analysis, and it was
found that neither productivity nor quality rose as a result of pair programming’s use. But
the authors also note that their study was limited and that more studies were needed.
There is somewhat more research being done around dyadic learning in education.
Much of this research indicates that students working together learn more and are less
likely to drop courses (McDowell et al, 2002).
10
In decision-making competitive situations, pairs have been found to be more
strategic, and to be more effective, than individuals (Cooper & Kagle, 2005).
In usability, the use of dyads has been, if anything, even less fully studied than it
has been elsewhere, although it has been championed by some writers on the basis of
personal experience (Snyder, 2003). Some research has focused on children’s dyads
(often referred to as “constructive interaction”) and the effects on usability testing. Als,
Jensen, and Skov (2005) looked into monad and dyad performance with children and
found that dyad pairs of children found more problems in products during usability
testing. Similar studies have explored children’s ability to verbalize their thoughts
concurrently (during testing) as opposed to respectively (after testing) (Robinson, 2001).
It is unclear how much this research generalizes to adult usability testing, nor whether it
applies to time-on-task measurements from usability tests.
One oft-cited study by Hackman and Biers focused on determining whether think-
aloud individuals in solitude, think-aloud individuals with an observer present, or dyads
(called “teams” in the study) would produce more and better verbalization. Although the
study was most interested in the quality and quantity of verbalization, it also recorded the
time-on-task data as a measure of task performance. It was found that the three groups
did not differ significantly in their timed performance, and that although there were more
verbalizations in the team Condition that were of value to designers, the investigators
concluded that overall the think-aloud protocol was not outstandingly productive of
useful feedback from users (Hackman and Biers, 1992). It should be noted that in this
study, there was no silent control group, so time-on-task measurements could not be
compared between think-aloud and silent Conditions.
11
At least one study has shown that monads and dyads do not differ significantly in
their time-on-task. Two groups worked on several tasks on interactive TV. One group
performed the tasks as monads, while the other group worked in dyad pairs. The results
suggest that across all tasks there were no significant differences between monad and
dyad task completion times, although one task, performing a search, did reach
significance (Grondin et. al., 2002).
It is reasonable to assume that the addition of a social element and its attendant
cognitive overhead to the dyad structure will increase time-on-task. Likewise, it is also
reasonable to presume that the cognitive load of speech, however slight, may well impose
at least a small time penalty over silent problem-solving.
Hypotheses
H1: Think-aloud monads will display longer times-on-task than silent monads.
H2: Dyads will have comparable times-on-task to think-aloud monads.
12
CHAPTER THREE: METHODOLOGY
Participants
Participants were recruited through notices placed in public areas. These included IUPUI,
two churches, and several workplaces. A screener was used to filter out those who could
qualify for testing (see Appendix 1). The participants were almost evenly split in gender,
with 14 males and 18 females, between the ages of 18 and 50. All completed the
preliminary screener, and all declared themselves to be at least moderately skilled in both
navigating the Web and performing common search functions in environments such as
Google. They also self-declared that they had not previously visited the Indianapolis-
Marion County Public Library (IMCPL) site. Two participants proved to have too low a
proficiency level with Web navigation and search, and were replaced.
Treatments
The research design was a three-group independent sample approach using
samples of eight in each of three treatment levels: monad-silent, monad-think-aloud, and
dyad. The independent variable is the treatment level; the dependent variables are the
time-on-task, recorded in seconds, and the number of verbalizations. In addition, because
Tasks are highly varied in complexity, cognitive difficulty, and demand for problem-
solving abilities, the Tasks themselves can be considered independent variables for
analysis.
Hardware. Two laptops were used. One was an IBM R51, with a TrackPoint
pointer, two mouse buttons, and a touchpad. This laptop had a built-in wireless card, a
1.7GHz Intel processor, 2.00 GB of RAM, and ran Windows XP Professional Version
13
2002, Service Pack 2. The other laptop was an IBM T22 with a Pentium III, 128 Meg
RAM, external wireless card, TrackPoint pointer, two mouse buttons, Windows 2000
5.00.2195 Service Pack 4. Both machines were set to the same resolution of 1024 X 769
for testing purposes. Sound was muted off.
Software. Both machines were equipped with Mozilla Firefox version 1.5.0.4.
Both copies of Firefox were equipped with the Web Developer’s Toolbar, but it was
switched off for testing purposes.
Testing website. An existing public website was chosen for testing,
www.imcpl.com, the website for the Indianapolis – Marion County Public Library
(IMCPL). The site was chosen for several reasons. 1) It presents a significant challenge to
first-time users; 2) It does not change dramatically in a short period of time, providing a
relatively stable testing artifact; 3) The site offers many opportunities for testing
scenarios, from simple navigational Tasks to formidable search functionality; 4) It
replicates well the type of site most commonly designed for multiple users, offering
several paths in navigation and search.
Tasks. Eight Tasks were constructed, and phrased as questions to be answered
from the site.
Task 1. Is the copy of Profiles in Courage by John F. Kennedy checked in at the
Glendale branch? (You may use the library’s book search function for this question)
This Task requires the use of the library’s search capability, which is difficult for
most users to manipulate. It requires considerable thought and planning. This Task
probably requires the most problem-solving ability. See Figure 3-1 for an optimal path.
See Figure 3-2 for a screen shot of the home page, from which the participant must
14
determine where to find the appropriate search function. See Figure 3-3 for a screen shot
of the actual search screen.
Home Page(Search library
catalog link)Search Page Results Page Target Page
Figure 3-1: Optimal path for Task 1.
Figure 3-2: Starting page for Task 1.
Figure 3-3: Search page for Task 1.
15
Task 2. What is the replacement charge for a lost library card?
This is a straightforward Task in navigation, but requires the user to scroll. See
Figure 3-4 for an optimal path to the page. See Figure 3-5 for a screen shot of the
scrolling page.
Home Page Using YourLibrary
Getting aLibrary Card
Figure 3-4: Optimal path for Task 2.
Figure 3-5: Target page for Task 2.
16
Task 3. Who is the chief executive officer of the library?
This is another straightforward navigation Task, but this one has two or three
locations on the target page where users could find the name of the CEO. See Figure 3-6
for an optimal path. See Figure 3-7 for the page on the optimal path with the information.
Home Page(Search library
catalog link)
About theLibrary
Figure 3-6: Optimal path for Task 3.
Figure 3-7: Target page for Task 3.
17
Task 4. What time does the Decatur branch library close on Saturday?
This is another navigational Task that requires scrolling and determination of
location from a map. See Figure 3-8 for an optimal path. See Figure 3-9 for the page on
the optimal path with the information.
Home Page(Locations and
Hours link)
LibraryLocations
Figure 3-8: Optimal path for Task 4.
Figure 3-9: Target page for Task 4.
18
Task 5. Is there a special event being held next Sunday?
Navigational Task that requires mentally categorizing and eliminating dates. The
target page has event search that must be properly used to complete the Task. See Figure
3-10 for an optimal path. See Figure 3-11 for the page on the optimal path with the
information.
Home Page Events andClasses
Figure 3-10: Optimal path for Task 5.
Figure 3-11: Search page for Task 5.
19
Task 6. What is the number for Call-a-Story?
This Task requires the participant to discover that Call-a-Story is a children’s
feature. This Task is made more difficult because the site co-names the function “Call a
Pacer” or “Call an Indianapolis Indian”. See Figure 3-12 for an optimal path. See Figure
3-13 for the page on the optimal path with the information.
Home Page(Just for Kids >
Call-a-Story)Just for Kids
Figure 3-12: Optimal path for Task 6.
Figure 3-13: Target page for Task 6.
20
Task 7. What is the minimum age a person must be to qualify for employment as a
library page?
Requires the user to read text on the target page and use the headings as text
markers. See Figure 3-14 for an optimal path. See Figure 3-15 for the home page with the
link. See Figure 3-16 for the page on the optimal path with the information.
Home Page(EmploymentOpportunities,
bottom of page)
EmploymentOpportunities
Figure 3-14: Optimal path for Task 7
Figure 3-15: Home page for Task 7 with link.
21
Task 8. When is the next library book sale to be held?
Requires two functions: finding events, and finding a specific date. See Figure 3-
17 for an optimal path. See Figure 3-18 for the page on the optimal path with the
information.
Home Page Events andClasses
Figure 3-17: Optimal path for Task 8.
Figure 3-18: Search page for Task 8.
The optimal paths shown above are not the only paths through the site. Choosing
the optimal path would result in the smallest possible time, and other paths would
contribute to substantially greater times.
Procedure
Testing proceeded almost identically for all participants. Single participants were
randomly assigned to monad-silent or monad-think-aloud groups by coin flip. Each group
23
was fixed at eight participants each. Assignment to the dyad Condition was made by
convenience, when two participants could be scheduled together.
The participants were seated comfortably in front of the laptops. For dyads, one
participant sat directly in front of the screen, while the partner sat just to the side. For
dyads, one participant was the “driver” who manipulated the machine, while the partner
assisted with advice and suggestions. The participant was given a familiarization period
to play with the controls and feel confident with them.
In two cases, dyad members switched roles once between Tasks. In each case the
exchange was made once during the testing session, early on. The request was made
because in each case the “driving” participant did not feel comfortable in his role. To
avoid the possible time bias of a frustrated or flustered “driver”, this was permitted.
Monad-silent participants were specifically told not to verbalize during testing.
Monad-think-alouds were told to verbalize continuously, and if their production of
verbalizations diminished, they were prompted with the term “Please keep talking”.
Utterances for monad-think-alouds were counted. Dyads were also told to keep talking
continuously while testing, and their verbalizations were counted and added together. The
verbalizations were recorded as tick marks on the data sheet. A Task verbalization was
generally defined as a verbalization that was recognizable as a semantically permissible
verbalization and was functionally related to the Task at hand. This eliminated such
verbalizations as “uh” and “hmmm”.
The researcher was positioned to the side of the participant. The stack of Tasks
was preordered into the randomized sequence specified on the materials packet for that
session. An online random number generator was employed to produce a string of digits
24
from 1 through 8, in random order. This string was noted on the outside of the envelope
containing the session materials and Tasks were presented in the order specified.
The site offered more than one search mechanism, any one of which might skew
the results for navigational Tasks. Therefore, all participants were told not to use search
capability, unless the researcher approved such use for a Task.
Each Task began with the investigator reading the Task aloud, then placing the
sheet with the Task on the table within reading distance of the participant. The timing for
each Task began when the participant made the first “progressive move”, which is here
defined as a keyclick, mouse movement, or other manipulation of the computer controls
in furtherance of the Task. Timing stopped when the participant verbalized the answer to
the Task question. Either a stopwatch or digital watch was used for timing.
Analysis
The first analysis was conducted with an overall ANOVA, using a planned
comparison of monad silent < monad-think-aloud = dyad. The comparisons were
Condition, Task, and Condition x Task. Each Task was then individually subjected to an
ANOVA, using the planned comparison of monad-silent < monad-think-aloud=dyad. The
correlation between number of verbalizations and time-on-task was examined for monad-
think-aloud and dyad, along with an F-test for significance between those Conditions.
Post-hoc tests were conducted as was deemed appropriate.
25
CHAPTER FOUR: RESULTS
Time on Task.
Figure 4-1 shows the mean Task completion times for the three treatment groups
for all eight Tasks, along with standard errors of the mean.
20
60
100
140
180
220
260
300
340
380
420
Task #
S T D
1 2 3 4 5 6
S T D S T D S T D S T D S T D S T D S T D
7 8
Seco
nds
460
Monad-silentMonad-think-aloudDyad
Figure 4-1: Graph of mean times for all groups on all Tasks.
Several patterns are notable in Figure 4-19:
There is wide variation in the completion times across Tasks. The mean
completion time across Conditions for Task 1 is 323.0 sec, whereas the
26
mean completion times across Conditions for Tasks 2, 3, and 4 are 33.0,
31.7, and 46.3 sec, respectively. The mean completion times for the
remaining Tasks range from 61.3 sec to 123.0 sec.
Although there is no apparent pattern of differences across Conditions for
the Tasks that were completed the fastest (Tasks 2, 3, and 4), the mean
completion time for monad-silent is lower than for monad-think-aloud and
dyad in four of the other five Tasks (Tasks 1, 5, 6, and 8).
Task 7 presents a unique ordering of completion times with monad-silent
taking longer than monad-think-aloud and dyad.
An overall ANOVA, using a planned comparison of monad-silent < monad-think-
aloud = dyad, indicated that the effect of Task was significant (F(7,147)=30.89; p < .01),
that the planned comparison of Condition approached significance (F(1,21)=2.79; p=.11),
and that the interaction of Condition x Task also approached significance
(F(14,147)=1.41; p=.16).
Each Task was individually subjected to an ANOVA using the planned
comparison of monad-silent < monad-think-aloud = dyad to test for significance between
treatment Condition. The results of these analyses are presented in Table 1. Only Task 1
approached a conventional level of significance with p < .10.
27
Task F (1, 21) p 1 3.24 0.09
2 1.06 > 0.30
3 < 1 > 0.30
4 < 1 > 0.30
5 1.61 0.23
6 < 1 > 0.30
7 1.61 0.22
8 1.12 0.30
Table 4-1: ANOVA results of time-on-task for each Task.
Newman-Keuls post-hoc tests indicated that Task 1 was different from all of the
other Tasks; that Tasks 5 and 6 are different from Tasks 1, 2, and 3; and that Task 6 is
also different from Task 4 (all were at p<.05; all other comparisons p>.05).
Because it appears that a common pattern of results was observed in Tasks 1, 5,
and 6 and these Tasks had significantly longer completion times than the other Tasks, a
separate ANOVA was conducted on only these Tasks with the planned comparison of
monad-silent < monad-think-aloud = dyad. This analysis indicated a significant effect of
Condition (F(1,21)=4.53; p < .05 and a significant effect of Task (F(2,42)=23.80; p < .01,
but the interaction of Condition x Task was not significant (F(2,42)=1.10; p > .20.
Verbalizations
Table 4-2 shows the mean number of verbalizations for all Tasks for the monad-
think-aloud and dyad Conditions. It can be noted that the number of verbalizations varies
across Tasks, with Task 1 resulting in a mean number of verbalizations greater than 30,
28
Tasks 3, 4, and 5 resulting in mean numbers of verbalizations less than 10, and the
remaining Tasks resulting in mean numbers of verbalization between 10 and 20.
T1 T2 T3 T4 T5 T6 T7 T8 Think-Aloud 36 8 5 8 22 13 15 15 Dyads 46 6 5 6 13 12 9 8
Table 4-2. Monad-think-aloud and dyad verbalization counts.
Verbalizations varied across Tasks, and were not consistently higher for monad-
think-aloud participants or dyads. An ANOVA of the verbalization results indicated lack
of significance between Conditions F(1,14)=0.48; p = 0.50, but significance between
Tasks (F(7,98)=18.14 p < .01. There was no interaction significance in Condition x Task
(F(7,98)=1.24 p>.20 (0.29). Post-hoc Newman-Keuls tests indicated that Task 1 was
significantly different from all other Tasks, and that Task 5 was significantly different
from Task 3. No other effects were significant (p=.05).
It can further be noted that the number of verbalizations for each Task in Table 2
appears to correspond to the Task completion times in Figure 4-1. Figure 4-2 presents a
scatterplot of number of verbalizations vs. mean completion time for each Task for the
monad-think-aloud Condition. Figure 4-3 presents the comparable scatterplot for the
dyad Condition.
29
0
50
100
150
200
250
300
350
0 5 10 15 20 25 30 35 40
Verbalizations
Tim
e in
Sec
onds
Figure 4-2: Monad-think-aloud scatterplot of verbalizations and time.
0
50
100
150
200
250
300
350
400
450
0 10 20 30 40 5
Verbalizations
Tim
e in
Sec
onds
0
Figure 4-3: Dyad scatterplot of verbalizations and time.
The scatterplots indicate a strong linear relationship between the factors of time
and number of verbalizations, and this holds for both monad-think-aloud participants and
dyads. The correlation for the monad-think-aloud Condition was r(6)=.97; p < .01 and for
the dyad Condition was r(6)=.99; p < .01.
30
CHAPTER FIVE: DISCUSSION
The results show that, as anticipated, verbalizations increased with longer time-
on-task, and verbalizing participants took somewhat longer to perform the Tasks than did
participants who did not verbalize. Also as expected, there are variations between Tasks.
Explanation of Outcomes
It should be noted that there was found only limited support for the hypothesis
that verbalization necessarily increases time-on-task. Analysis showed that the main
effect of Task is significant, while the main effect of Condition approached significance.
Similarly, the interaction of the two factors approached significance, but did not reach it.
Of all the individual Tasks, only Task 1 closely approached significance between
Conditions. Generally, then, one might conclude that the amount of verbalization
exhibited during a test might have no effect whatsoever on timed performance. But a
closer examination of three Tasks provided intriguing evidence to the contrary.
Tasks 1, 5, and 6 were distinctive because they took longer to complete, and post-
hoc analysis revealed that they were indeed significantly different from the others. Tasks
1 and 5 were search Tasks; previous studies have found that search Tasks can take longer
than navigation Tasks (Grondin et. al., 2002). Task 6, however, was a navigation Task,
and its inclusion in this group of lengthier Tasks requires some examination.
Of all the Tasks, Task 1, a search Task, was undoubtedly the most difficult for
most participants to perform. It involved the most thought, the most Web pages, and the
most key and mouse clicks. Participants frequently had to start over or take alternative
paths. In Task 1, the participant was asked to find out if a particular book title was
31
currently on the shelf at a particular branch. This was possible to achieve only through
the site’s book search functionality. However, there are various distracters that appear in
the website, such as a prominently displayed search text box on the home page that,
unfortunately, does not work for this Task. Participants had to find the appropriate search
function, then interpret the results.
Given these hurdles, it is not remarkable that Task 1 displayed by far the highest
times of any Task. However, what is worthy of a remark is that the dyad Condition
exhibited the highest average time. In previous studies, it has been fairly well established
that dyads consistently perform better in Task completion measures. Here, Task
completion was not measured, and all participants completed all Tasks, but it is
interesting to speculate that if both completion rate and time-on-task were to be combined
in a study of difficult and lengthy Tasks, the higher dyad completion rates might come at
the price of similarly elevated dyad time-on-task measures.
Task 5 was likewise a searching Task, albeit a less complex one than Task 1.
Unlike in Task 1, in Task 5, participants quickly found the correct search page, but were
then presented with several selection controls. Participants frequently tried more than one
combination of control attributes before finding a set that would provide the answer.
Task 6 was not a search Task. It was, in fact, a navigational Task, little different
in nature from Tasks 2, 3, 4, and 8. The higher times for Task 6 as compared with Tasks
2, 3, 4, and 8 may simply be due to navigational difficulties with this particular website
while performing Task 6.
What is one to make of the fact that in general longer Tasks exhibited the most
difference in Condition? It is possible that the main effect of verbalization on
32
performance is simply more salient when Tasks take longer, and that Tasks 2, 3, 4, and 8
present smaller effects simply because they did not last as long. If it is postulated that
each verbalization takes a finite and specific, though very small, time lapse that interferes
with or pauses Task operations, then it would not be unexpected that Tasks requiring
more numerous or more frequent verbalizations would accumulate more verbalization
time lapses, as well. Work by Grondin et. al. (2002) suggested this when they found that
the only Task in which monads and dyads approached significant time differences was in
a search Task that took longer than the others. The verbalization count data certainly
suggests that there is a high correlation between verbalization numbers and time. On the
other hand, participants did not pause noticeably while verbalizing, so any definable
“time penalty” for each verbalization would likely be very short. The exact time lost in
any particular verbalization would be material for later study.
Verbalization levels as defined by Ericsson and Simon (1993) may also have
taken part in the effect. There was no attempt to analyze the verbalizations as data, as
Ericsson and Simon did, and to determine the verbalization levels. Ericsson and Simon
would hold that level 3 verbalizations would increase the times, and it is possible that the
more obvious differences between the Task grouping of 1, 5, and 6 and the grouping of 2,
3, 4, and 8 are due to the fact that 1, 5, and 6 took longer by the nature of their Tasks, and
that therefore there were more opportunities for level 3 verbalizations to be inserted.
Task 7 is a distinct anomaly, reversing the pattern seen in 1, 5, and 6, in that
monad-silent participants had the highest time scores. There was no significant effect of
Condition within this Task. No good explanation presents itself for the pattern reversal of
33
Task 7, which was a navigational Task similar to Task 6, except that Task 7 may merely
be an artifact of chance. Further research might answer this question more thoroughly.
Many studies have shown think-aloud protocols to be helpful in a number of
situations, including clinical applications (Henry, LeBreck, & Holzemer, 1989) and
expert thought process elicitation (Someren, Barnard, & Sandberg, 1994). There is good
evidence that think-aloud can increase Task completion rates (Loeffler & Polkehn, 2000).
But until more is known about verbalization and its effect on performance in timed
testing, usability testing practitioners should be cautious when combining think-aloud
protocols with time-on-task measurements. Further, investigators should undertake more
research into these potential interactions.
34
CHAPTER SIX: CONCLUSION
Limitations
The present study had several potential confounding factors that may have
skewed times: 1) the nature of the website, having alternative paths for most Tasks (Task
selection); 2) participants’ apparent skills variability, despite having been screened; 3)
small sample sizes, 4) very small pilots (two monad participants).
The website itself was a commercially available one, and therefore its various
paths and controls were not under experimental control. There were a great many
alternative paths for some Tasks, which complicated the timing picture.
Participants exhibited widely variable skills. Some had better search skills, but did
not have a mental model that matched that used for navigation in the site. Others had the
opposite attributes. Likewise, there was no screening or control for various other factors
that may, in hindsight, have affected the outcome.
There were only two small pilot tests, each with a monad. Some of the Task
differences may have been capable of adjustment if more pilot tests could have been
conducted.
Future Research
Several aspects of the current study would seem to merit further work: 1) the
higher times and closer approach to significance for Task 1 would indicate that Tasks
with longer expected times should be undertaken, to see if longer times and significance
are related; 2) testing for specific Task types to determine if Task type has a direct effect
on time-on-task measurements and patterns, as in Task 7; 3) testing on other types of
35
sites, both those built especially for testing, and those in general use; 4) exploration of
possible confounding factors, such as possible differences in gender interaction or
verbalization performance; 5) determining what the size of a “time penalty” for each
verbalization might be.
Summary
This study investigated the effect of verbalization on time-on-task performance of
monads and dyads. A direct effect was found between Conditions that, while not found to
be statistically significant across all Tasks, was nonetheless significant between Tasks
that took longer to perform. This is noteworthy, and has potential for future study. A
significant direct effect was also found between Tasks, indicating a possibly fruitful area
for later work.
36
REFERENCES
What is human factors/ergonomics? (2005). Retrieved August 3, 2006, from
http://www.hfes.org/web/AboutHFES/about.html
Als, B. S., Jensen, J. J., & Skov, M. B. (2005, June 8-10). Comparison of think-aloud and
constructive interaction in usability testing with children. Paper presented at the
IDC 2005, Boulder, CO.
Boren, M. T., & Ramey, J. (2000). Thinking aloud: Reconciling theory and practice.
IEEE Transactions on Professional Communication, 43(3), 261-278.
Cockburn, A., & Williams, L. (2001). The costs and benefits of pair programming. In
Extreme programming examined. Boston, MA, USA: Addison-Wesley Longman
Publishing Co., Inc.
Cooper, D. J., & Kagel, J. H. (2005). Are two heads better than one? Team versus
individual play in signaling games. The American Economic Review, 95(3).
Davidson, G. C., Vogel, R. S., & Coffman, S. G. (1997). Think-aloud approaches to
cognitive assessment and the articulated thoughts in simulated situations
paradigm. Journal of Consulting and Clinical Psychology, 65(6), 950-958.
Dickson, J., Mclennan, J., & Omodei, M. M. (2000). Effects of concurrent verbalization
on a time-critical, dynamic decision-making task. Journal of General Psychology,
127(2), 217-228.
Duncker, K. (1945). On problem-solving. In Psychological Monographs (Vol. 58, pp. 1-
114). Washington, DC: The American Psychological Association, Inc.
Ericsson, K. A., & Simon, H. A. (1984). Protocol analysis: Verbal reports as data.
Cambridge, Massachusetts: MIT Press.
37
Grondin, N., Bastien, J. M. C., & Agopian, B. (2002). Les tests utilisateurs : avantages et
inconvénients des passations individuelles et par paires. Paper presented at the
IHM 2002.
Guan, Z., Lee, S., Cuddihy, E., & Ramey, J. (2006, April 22-27). The validity of the
stimulated retrospective think-aloud method as measured by eye tracking. Paper
presented at the CHI 2006, Montreal, Quebec.
Haak, M. J. V. D., Jong, M. D. T. D., & Schellens, P. J. (2003). Retrospective vs.
concurrent think-aloud protocols: testing the usability of an online library
catalogue. Behaviour and Information Technology, 22(5), 339-351.
Hackman, G. S., & Biers, D. W. (1992). Team usability testing: Are two heads better
than one? Paper presented at the Proceedings of the Human Factors Society, 36th
Annual Meeting.
Henry, S., LeBreck, D., & Holzemer, W. (1989). The effect of verbalization of cognitive
processes on clinical decision making. Research and Nursing Health, 12(3), 187-
193.
Hulkko, H., & Abrahamsson, P. (2005). A multiple case study on the impact of pair
programming on product quality. Paper presented at the ICSE '05, St. Louis, MO.
Loeffler, J., & Polkein, K. (2005, March 21-23). The influence of thinking aloud on the
quality of task completion in user tests. Paper presented at the GOR-General
Online Research Conference 2005, Zurich, Switzerland.
McDowell, C., Werner, L., Bullock, H., & Fernald, J. (2002, February 27-March 3). The
effects of pair-programming on performance in an introductory programming
course. Paper presented at the SIGCSE 02, Covington, KY.
38
McDowell, C., Werner, L., Bullock, H. E., & Fernald, J. (2003). The impact of pair
programming on student performance, perception and persistence. Paper
presented at the 25th International Conference on Software Engineering.
Miyake, N. (1982). Constructive interaction. San Diego: University of California, Center
for Human Information Processing.
Nielsen, J., Clemmensen, T., & Yssing, C. (2002, October 19-23). Getting access to what
goes on in people's heads? Reflections on the think-aloud technique. Paper
presented at the NordiCHI Arhus, Denmark.
Nielsen, J., & Landauer, T. K. (1993, April 24-29). A mathematical model of the finding
of usability problems. Paper presented at the SigCHI Conference on human
factors in computing systems, Amsterdam, The Netherlands.
Nisbett, & Wilson. (1977). Telling more than we can know: Verbal reports on mental
processes. Psychological Review, 84(3), 231-259.
O'Malley, C., Draper, S., & Riley, M. (1984). Constructive interaction: a method for
studying user-computer-user interaction. Paper presented at the IFIP
INTERACT'84 First International Conference on Human-Computer Interaction.
Proctor, R. W., & Zandt, T. V. (1994). Human factors in simple and complex systems.
Boston: Allyn and Bacon.
Robinson, K. M. (2001). The validity of verbal reports in children's subtraction. Journal
of Educational Psychology, 93(1), 211-222.
Snyder, C. (2003). Paper prototyping: The fast and easy way to design and refine user
interfaces: Morgan Kaufmann.
39
Someren, M. W. v., Barnard, Y. F., & Sandberg, J. A. C. (1994). The think aloud method:
A practical guide to modelling cognitive processes. London: Academic Press.
Spool, J., & Schroeder, W. (2001). Testing web sites: five users is nowhere near enough.
Paper presented at the Conference on Human Factors in Computing Systems,
Seattle, WA, USA.
Wildman, D. (1995). Getting the most from paired-user testing. Interactions, 2(3), 21-27.
Williams, L. A., & Kessler, R. R. (2000). All I really need to know about pair
programming I learned in kindergarten. Communications of the ACM, 43(5), 108-
114.
Wright, P. C., & Monk, A. F. (1991). The use of think-aloud evaluation methods in
design. SIGCHI Bulletin, 23(1), 55-57.
40
APPENDICES
Appendix A: Screener NAME: _______________________________________________________ PHONE: _____________________ EMAIL:______________________ PROFESSION: ____________________________________________ How would you rate your ability to use the WEB? OUTSTANDING _____AVERAGE _____BASIC______ NONE ______ Have you performed searches on the WEB using search engines such as Google?
YES_____ NO ___ Which of these websites have you ever visited?
IUPUI Library (www.ulib.iupui.edu) ______________ Indianapolis/Marion County Public Library (www.imcpl.com) ________ Google (www,google.com) ______________________ Centers for Disease Control (www.cdc.gov) _________
41
Appendix B: Raw Time Data
Task 1 Task 2 Task 3 Silent TA Dyad
Silent TA Dyad
Silent TA Dyad
237 595 501
23 43 19
135 48 39
170 261 651
22 29 17
13 24 50
187 326 280
6 23 45
8 60 13
170 282 246
25 12 18
34 9 133
404 234 300
75 32 27
31 8 10
161 520 512
16 28 85
45 20 5
200 58 270
15 36 100
16 5 9
453 356 379
25 64 12
16 19 11
Task 4 Task 5 Task 6
Silent TA Dyad
Silent TA Dyad
Silent TA Dyad
41 62 150
140 47 104
42 301 196
38 24 34
22 172 38
55 75 149
17 38 38
61 718 142
85 160 14
22 43 23
70 15 47
65 41 33
48 19 23
57 78 128
134 226 192
31 68 14
14 110 24
22 14 37
116 91 95
48 51 158
211 36 142
22 41 12
53 230 42
174 270 274
42
Task 7
Task 8
Silent TA Dyad
Silent TA Dyad
64 83 261 38 55 78
118 56 29 86 35 73
170 190 38 42 51 45
92 29 27 66 55 36
34 48 25 61 101 84
169 48 55 21 48 29
286 185 113 23 30 116
37 57 9 44 235 22
43
Appendix C: Raw Verbalization Data
MonadID T1 T2 T3 T4 T5 T6 T7 T8 M109 42 2 8 11 5 20 6 6M110 25 6 5 4 6 5 8 7M111 39 7 7 5 80 18 26 6M112 41 3 2 7 3 3 4 10M113 25 8 3 5 16 17 8 12M114 52 8 3 11 18 3 10 11M115 14 10 3 18 8 12 45 8M116 46 18 5 5 39 28 10 59
DyadID T1 T2 T3 T4 T5 T6 T7 T8 D101 37 5 4 13 10 12 17 10D102 78 4 8 6 4 17 5 10D103 70 10 8 12 38 6 14 14D104 29 4 11 1 11 3 5 10D105 38 8 3 4 21 17 6 10D106 59 9 3 4 6 9 12 4D107 33 5 2 8 12 11 9 5D108 27 1 2 2 3 19 2 2
44
VITA
Timothy H. Altom Education
Master’s of Science in Human-Computer Interaction, Expected August 2006 Indiana University-Purdue University at Indianapolis Advisor: Dr. Anthony Faiola Bachelor of Arts, English, May 1990 Indiana University-Purdue University at Indianapolis
Research Interests • Use of technology in facilitation of group activities • Technology as enabler of distributed cognition • Statistical techniques for determining user experience and behavior
Teaching Interests • Statistics for technology • Usability
Related Experience Employment Senior Business Consultant Perficient, Inc. July 2005-Present
• Consulting, information architecture • Consulting, Web analytics • Consulting, user-centered design
Systems Interaction Designer Indiana University January 2005-July 2005
• Represented IU to the Sakai tools team • Acted as resource and liaison to IU Support and Implementation Team
for Oncourse CL • Participated in Sakai tool redesign • Performed accessibility testing • Performed usability testing
Contract Position Eli Lilly and Company May 2004-December 2005
• Participated in software quality project • Responsible for data acquisition and analysis for change management
process Technical Communications Manager Solutions Technology Inc. June 2001-May 2004
• Responsible for all technical communications issuing from company • Initiated incorporation of usability into consulting
45
• Created online education for product familiarization Vice President and Head Technodude Simply Written Inc. January 1995-June 2001
• Head of technical development, in fields such as SGML, XML, HTML, Web delivery
• Co-creator of the Clustar System for structured writing Teaching Adjunct Faculty Computer and Information Technology, School of Engineering and Technology, IUPUI August 2004-Present
• Descriptive statistics, undergraduate • To teach inferential statistics in Spring of 2007, undergraduate
Adjunct Faculty School of Informatics, IUPUI August 2005-Present
• Usability, undergraduate Publications
• Altom, T., Buher, M., Downey, M. and Faiola, A. (2004). Using 3D landscapes to navigate file systems: The MountainView interface. Proceedings of the 8th International Conference on Information Visualization, 645-649.
• Faiola, A., Altom, T., and Groth, D. (2005) Integrating the visualization of personal histories of file usage into 3D landscapes: Enhancing desktop file searching using TerraSearch+. Publication pending.
• Altom, T. (2003). XML in motion: the scalable vector graphic. Intercom. 50 (6): 10-13
• Columnist for Indianapolis Business Journal. 2002-Present. Return on Technology.
• Altom, T. (1999). Programming with Python. Prima Publishing. • Altom, T., et al. (1999). Microsoft Office 2000 User Manual. Que
Publishing. • Altom, T., Chapman, M. (1999). Hands-On HTML. Prima Publishing. • Altom, T. (1997). Designing for dyslexics. Intercom. 44 (4): 8-10 • Altom, T. 1996. How to get along with impossible co-workers Intercom.
43 (8): 16-20 • Altom, T. 1996. The future form of online help files. Intercom. 43(7):16-
18 Professional Associations
• Senior member, Society for Technical Communication. Twice president of local chapter.
• Founding member, local chapter of Usability Professionals Association • Member, SIGCHI
46