0:00
In this video, we're gonna talk about running studies in-person. And to give you
a preview, in the next video, we're going to talk about running studies online. The
web has made it much easier to run controlled experiments and I think that
there's a lot of people who run studies that they wouldn't have been able to
previously do. But there's also a lot that you can get from watching people. Don't
miss out on that. In particular, at watching people live, you can talk to them
about what their doing, get their feedback, see points of confusion, and in
general, have a much higher bandwidth engagement with the participants in your
studies, then you would get just out of a click stream online. So, today we're going
to talk about planning, executing, and analyzing in-person studies. The first
step is to make clear goals, for example, say, you were building a room reservation
system for a Computer Science department. One strategy would be to put all the
information on one page. Another strategy would put different pieces on different
pages. And you might be particularly interested in if we split things up. Does
that change whether people will book a right size room or not? So, before running
your study, lay out its scope. You don't need to cover an entire system, especially
if you're working on a huge piece of software. It's okay to have a narrow
scope. And that scope will be guided by the purpose of your study, what you hope
to learn. I encourage you to come up with a hypothesis ahead of time and figure out
a way to know whether your hypothesis is true or not. Set up your study in way so
that you can see whether or not people's behavior supported your hypothesis. Book
an appropriate place for your study. If the interface is something that will be
used in a quiet room, book a quiet room. If, on the other hand, it's gonna be used
at a train station, test it at a train station. Figure out who you're gonna
recruit and how, and how many people. And come up with scenarios that you give
people when they come in. You want to m ake sure that these scenarios are
realistic and ideally something that these participants can be motivated to care
about. At least put themselves in the shoes of the actual user of the system.
So, here we have some questions, like will people book the right sized room and for
the right length of time. And this implies to measures. The size of room that they
booked, how long it was booked for. You can also get measures like task completion
time, how long it took and especially if you have realistic users, you can see how
this task interweaves with other things they do. So, if people are actually
booking a room for an actual meeting that they need to have, you might see whether
people makes notes to themselves to also order food or whether they need to
coordinate with people over e-mail first to, for example, get an approximate
attendance count. And if you think things like screen size might affect how users
behave, have a screen that's appropriate for what they'll have so. The simplest
thing to do obviously is to have a stock setup for everybody that is pretty normal.
One danger that you can see is that developers often have fancier computers
than normal users. So, don't test people on the latest greatest fanciest thing if
what people are going to be looking at your site with is a 5-year-old laptop. If
you can, try to get at least two people to be present for the study. That way you can
have one person who is the facilitator and another one who can take notes. It's
important to have concrete tasks among other things that lets you compare the
performance between people. And I encourage you to write them down as
opposed to speaking them so that everybody gets the same experience, making it even
easier to compare. Here's an example of a creativity task asking people to draw
creatures from a novel planet. And you can see that the instructions are pretty
clear. And you can see that the instructions are pretty specific, even
though the output of the task is really open ended. And here's an example of the
meeting room t ags. You can see how it gives some concrete details and context.
I've been amazed and heartened by how seriously most of the participants in our
studies take what they do. And the flip side of that is that sometimes people have
left in tears when they've been unable to perform well. In studies, people can feel
like they're being put under a microscope. And you have a responsibility to alleviate
that stress. For starters, make consent voluntary and make sure that consent is
informed. Avoid pressuring people to participate and let them know that they
can stop at anytime. And maybe the most important rule of all for minimizing
stress and increasing the quality of the interaction that you have with
participants is remind people multiple times that you're testing the site or the
user interface, the system, not them and that's an important frame change because
when you're testing the system, if the user is unable to complete the task, the
fault is the system's fault. If participants believe that they are the
ones being tested, then for example, if I am being unable to perform a task that
reflects poorly on me and my abilities and I may be stressed out. If, on the other
hand, I, as a participant believe that I am helping the designer test the system,
then when I encounter a roadblock, it's not a poor reflection of my abilities.
It's I've done something good. I've found a bad part of the user interface that I
can then help the designer make better. There's several important considerations
in a study like this. For example, what's going to be your order of tasks? One good
approach to this is to start out with a really simple task to get people warmed
up. And then you can move to more complex tasks as you go through the study. Other
times you may shuffle the order so there's no effect of one of the tasks on another
one. How much training, if any, are you going to provide? It depends on how the
real system's going to be used. If you have a ticket machine, not so much. If you
have a surgical system, maybe a little bit more. What are you gonna do if somebody
doesn't finish? Probably the answer shouldn't be lock them in the room until
they do. For a lot of tasks, I recommend deciding in advance what a reasonable
upper band for time is. And you can let participants know that. Another strategy
is that if somebody hits a small stumbling block, you may elect to help them along so
that they can get to the rest of the user interface so that you're not purely
stalled out on one point. And finally, pilot your study. A pilot participant is
somebody that runs through your studies before the actual participants. And the
main goal of a pilot participant is to work out the kinks in the study design.
Also for a usability study, pilots can find really obvious useability bugs that
you wouldn't want to burn real participants on. In fact, I recommend
doing two pilots. One with a colleague who may know the interface well just as a way
of getting all of your materials ready. Figuring out what the tasks are, all that
kind of stuff. Then, do another pilot with one real user and that will help you see a
bunch of these kinks that you didn't see ahead of time. I promise you that if you
run a pilot, you'll discover things that otherwise would have screwed up the study.
And how are you gonna capture the results? There's a number of ways of getting
feedback from people. At the very least I recommend having a paper notebook or a
laptop or other note taking device handy so that you can note down critical
incidents, either aha moments that you have about how to improve the design,
things that work really well, stories that you wanna share with other stakeholders or
problems that you'll need to bring back to the design team to try and figure out how
to fix. You might record video and a great reason to record video is it will let you
grab those moments of either success or real challenge and let you share those
with other people. Depending on what you're doing, screen recording can serve
much the same purpose. It's going to depend on whether you want to gather the
user's facial expression or the crisp contents of the user interface. Are you
going to interrupt participants? If you have a question to ask them about why they
are doing what they are doing, if you want to talk to them about some feature of the
user interface or if you'd like to help them get past a stumbling block so that
they can get onto the next thing. There's no universal answer to this question. It
really depends on what you're hoping to get out of the study. And finally, are you
going to ask users to think aloud while they go through the study? The think aloud
method can be a really powerful way of getting feedback about a user interface.
And here's what I mean by think aloud. Say, for example, you are studying how
somebody changes the staples in a stapler. If I were a participant in that study
[noise] and I were using the think-aloud method, I might start out by saying, well,
I'm changing the staples in the stapler because it, it seems to be out. It's not,
it's not working anymore. So, and see I haven't used this stapler before. I'm
going to, oh, there we go. Okay, so I grabbed the top. I thought that might work
and that opened it up. And I have my staples here. And I'm gonna put those
right in there. And make sure this part goes forward. Oops, that didn't go. There
we go. And then try and close it without getting a staple out. There we go. That's
think aloud. And it helps you get a window into how people are thinking about the
task they're doing. It's not something that most of us are used to doing. And so
you'll need to prompt people at the beginning of the study to think as they're
going. And then you'll need to offer prompts during the study as well. Prodding
people to remember to say what they're thinking or what they're trying to do or
any questions that arrives. Or even what it is they're reading on the screen. So,
remember to prompt people to keep talking. And decide ahead of time what things you
will help on and what things you won't. If you can, write that down. Consistency is h
elpful. Is thinking aloud always going to give you the right answers? No. Among
other things, if you gotta talk while you're doing the task, that may change how
you do the task. In some cases, it may mean that you're more effective because by
trying to bubble up what it is that's the issue, that may help you get through it.
In another case, it may make you less effective because all of a sudden you have
to do two separate things, work the interface, and speak up simultaneously.
And at the very least it's almost certainly going to slow you down. And you
shouldn't take what people say as a veridical representation of what they're
thinking. We can almost always get a rationale or reason for something even if
it isn't at all the actual reason. In studies like this, it's usually not
because people are being malicious or trying to hide anything. It's usually just
because, it's usually just because they're trying to help you out and generate an
answer. Oftentimes those answers are useful for your design process. But they
may or may not be what's actually going on in your doodle. When people arrive, the
first thing that you want to do is greet them. So, the facilitator is going to
welcome the participant, explain what the setup is, explain the overall narrative of
the scenarios that they'll do. And get them setup and trained, if necessary, with
the interface. During the studies, you wanna collect both processed data and
bottom-line data. The processed data tends to be more qualitative and tends to have
real insights to it. The bottom-line data tends to be more of the numbers that the,
you will then use for more quantitative analysis. Bottom-line numbers are great
for things like task completion time. Did my interface speed people up or increase
completion rate? But, don't combine it with think aloud. The reason, think aloud
slows you down and it slows different people down, different and so, the
variance in task completion time, introduced by think aloud means that it
just doesn't make that much sense anymore. So, if yo u care both about the numbers
and about the think aloud data, you may wanna run those on separate studies. At
the end of the study, debrief the participant, bot, so they learn what your
goals were and they can find out more information about what you're trying to
do. And also, so that you can learn more holistically from tham after the
completion of everything, what it is that they're thinking. And now that you're
through this study, in the next video we'll move on to analyzing the data.