Anda di halaman 1dari 20

The complete guide to user testing

websites, apps, and prototypes


The complete guide to user testing websites, apps, and prototypes 1
Contents
03 INTRO
04 DEFINE YOUR OBJECTIVE

05 IDENTIFY WHAT AND WHO YOU’RE STUDYING


Determine your product and devices
Select your test participants
Other considerations

07 BUILD YOUR TEST PLAN


Create a series of tasks
Write clear questions

14 LAUNCH A DRY RUN

15 BEWARE OF ERRORS

17 ANALYZE YOUR RESULTS

19 CONCLUSION

The complete guide to user testing websites, apps, and prototypes 2


Remote user Intro
research is fast, User feedback is the key to making any business successful—whether you’re launching an app, redesigning
reliable, and a website, refining product features, or making ongoing improvements to your customer experience.

scalable. So how do you go about getting that feedback?

Remote user research is a fast, reliable, and scalable way to get the insights you need to improve your
customer experience. Unlike traditional in-lab usability testing or focus groups, you can recruit participants
and get results in a matter of hours, not days or weeks. You’re not limited by your location, schedule, or
facilities. You’ll get candid feedback from real people in your target demographic in their natural environment:
at home, in a store, or wherever they’d ordinarily interact with your product.

Performing remote user research on a regular basis helps you:

VALIDATE YOUR PRODUCT IDEAS BEFORE COMMITTING RESOURCES

INFLUENCE DECISION-MAKERS IN YOUR COMPANY TO MAKE IMPROVEMENTS

ALIGN YOUR TEAM ON PRIORITIES

IDENTIFY OPPORTUNITIES TO DIFFERENTIATE YOUR COMPANY FROM YOUR COMPETITION

TRANSFORM YOUR COMPANY’S APPROACH TO CUSTOMER EXPERIENCE

In this eBook, we’ll cover how to plan, conduct, and analyze your remote user research. If you’re new
to research, you’ll learn the ropes of setting up a study and getting straight to the insights. And if
you’re an experienced researcher, you’ll learn best practices for applying your skills in a remote
study setting.

The complete guide to user testing websites, apps, and prototypes 3


Define your objective
The first step toward gathering helpful feedback is setting a clear and
focused objective. If you don’t know exactly what sort of information
you need to obtain by running your study, it’ll be difficult to stay on track
throughout the project.

Ask yourself: What am I trying to learn?

You don’t need to uncover every usability problem or every user behavior
in one exhaustive study. It’s much easier and more productive to run a
series of smaller studies with one specific objective each. That way, you’ll
get focused feedback in manageable chunks. Be sure to keep your
objective clear and concise so that you know exactly what to focus on.

Example of a complex objective:


Can users easily find our products and make an informed
purchase decision?
This objective actually contains three very different components:
1. finding a product, 2. getting informed, and 3. making a purchase.

Example of a better objective:


Can users find the information they need?

As you set your objective, think about the outcomes your stakeholders
will care about. You may be interested in identifying how users interact
with the onboarding flow, for example, but that won’t be helpful if your
team needs to identify ways to improve the monetization process.

Keeping your objective front and center will help you structure your
studies to gain insights on the right set of activities. It’ll also guide you in
whom you recruit to participate in your study, the tasks they’ll perform,
and what questions you should ask them.

The complete guide to user testing websites, apps, and prototypes 4


Identify what and who you’re studying You can user
Once you know your research objective, it’s time to determine a few specifics.
test digital
First, think about what type of product you’re studying and which devices will be used to interact with it.
products at
Are you looking for feedback on a prototype? A released or unreleased mobile app? A website? If you have
any stage of
multiple products you want to learn about, then you’ll want to set up multiple studies—it’s best to test one
product at a time so you get the most focused feedback.
development,
Keep the URL or app file handy when you set up your study. This will be where your participants will start
from wireframe
their study.
to prototype to
PRODUCT TYPE STARTING PLACE live product.
Prototype (in a tool like InVision) Shareable URL
Static image Upload to Box or Dropbox; copy shareable URL
Website URL
Released app Name of the app in the App Store/Play Store
Unreleased iOS app .IPA file
Unreleased Android app .APK file

Consider which devices and/or browsers you’ll want to include in your study. If your product is available on
multiple devices, we recommend testing your mobile experience with at least as many users as your desktop
experience.

Next, consider who you will need to recruit to participate in your study.

With remote user testing, you don’t need to search far and wide to find people to give you feedback. You
simply select your demographics or other requirements when you set up your study.

The complete guide to user testing websites, apps, and prototypes 5


In many cases, you can use basic demographic criteria (such as age, So how many test participants should you include?
gender, and income level) to recruit your study participants. If you have
a broad target market or if you’re simply looking to identify basic usability It’s been demonstrated that five participants will uncover 85% of the usability
issues in your product, then it’s fine to use broad demographics. The problems on a website, and additional users will produce diminishing
key is to get fresh eyes on your product and find out whether people returns. Resist the temptation to double or triple the number of users to
understand how to use it. uncover 100% of your usability problems. It’s more efficient (and easier)
to run a study with five participants, make changes, run another study
In other cases, you’ll want to get a closer match for your target audience, with another five users, and so on.
such as people who work in a certain industry, make the health insurance
decisions for their household, own a particular kind of pet, etc. This is Finally, if you are looking for trends and insights beyond basic usability
most appropriate if you’re interested in uncovering trends and opinions issues, you may want to include a larger sample size. We recommend five to
from your target market rather than pure usability. To recruit the right people eight participants per audience segment—so if you have three distinct buyer
for your study, you’ll set up a screener question. Screener questions are personas, you will want to include a total of 15-24 participants in your study.
qualifying questions that allow only the people who select the “right”
answer to participate in your study. Other considerations

Sometimes, you may be interested in a more elaborate study. These


methodologies are also possible with remote user research:

Longitudinal studies: Watching users interact with a product over


a period of time.

Competitive benchmarking studies: Comparing the user experience


of several similar companies over time.

Beyond-the-device studies: Observing non-digital experiences, such


In rare instances, you may need to use your own existing customers for
as unboxing a physical item or making a purchase decision in a store.
your study, such as to test out an advanced feature of your product that
a non-customer wouldn’t understand. Moderated studies: Leading users through your study one-on-one
in real time.

USERTESTING TIP: Benchmarking studies: Tracking changes in your user experience on


a monthly or quarterly basis.
The broader your demographic requirements are, the faster you’ll
get your results because more people will qualify for your study. To learn more about how and when to use each of these techniques,
check out our UX Research Methodologies Guidebook.

The complete guide to user testing websites, apps, and prototypes 6


Build your test plan
Your test plan is the series of tasks and questions your participants will follow and respond to during
the study.

The key to a successful study is a well-designed plan. Ideally, your test plan will result in qualitative and
quantitative feedback. Ask users to perform specific tasks and then answer questions that will give you the
insights you need in a measurable way.

A task should be an action or activity that you want a user to accomplish at that time.

Example of a task: Go through the checkout process as far as you can without actually making a purchase.

Use a question when you want to elicit some form of feedback from a user in their own words.

Example of a question: Was anything difficult or frustrating about this process?

Using questions to get quantitative measurements:

Rating scale, multiple choice, and written response questions can be helpful when you’re running a large
number of studies and you’re looking to uncover trends. You’ll be able to quickly glance at the resulting data
rather than having to watch every video. From there, you can zero in on the outliers and the most surprising
responses.

We recommend taking a few moments to think back to your objective and consider the best way to convey
results to your team. When the results come back, how do you want that feedback to look? Will you want
quantitative data so you can create graphs? Written responses that you can use to create word clouds?
Verbal responses so you can create a video clip and share it with your team?

Establishing the type of deliverable you need from the outset will help you determine the right way to collect
the information you need.

The complete guide to user testing websites, apps, and prototypes 7


This chart highlights various question types and shows the kind of Consider using both broad and specific tasks in your tests
results that can be expected when those questions are put to use.
Broad, open-ended tasks help you learn how your users think. These types
of tasks can be useful when considering branding, content, layouts, or any of
TYPE EXAMPLE BENEFITS the “intangibles” of the user experience. They give you the chance to assess
usability, content, aesthetics, and users’ sentimental responses. Broad tasks
Verbal Describe and demonstrate Produces qualitative feed-
what, if anything, was most back and makes great clips are also good for observing natural user behavior. If you give participants the
response
frustrating about this app. for a highlight reel. space to explore, they will!

Multiple Do you trust this company? Great for collecting


choice • Yes • No categorical responses. Here’s an example of a broad task: Find a hotel that you’d like
(These can be nominal,
dichotomous, or even ordinal.) to stay in for a vacation to Chicago next month. Share your
thoughts out loud as you go.
Rating How likely are you to Good for collecting
scale return to this site again? ordinal variables.
“Not at all likely” to “Very likely” (Low, medium, high)
USERTESTING TIP:
When you’re not sure where to focus your test, try this: Run a
Written What do you think is Good for running post-study
missing from this page, analysis & collecting quotes very open-ended test using broad tasks like “Explore this app
response
if anything? for user stories. as you naturally would for about 10 minutes, speaking your
thoughts aloud.” You’re sure to find areas of interest to study
in a more targeted follow-up test.

Create a series of tasks


Specific tasks help pinpoint where users get confused or frustrated trying to
The key to collecting actionable feedback is getting your participants to do something specific. They’re useful when you’re focused on a particular
take action by performing specific tasks. Tasks are the steps that guide feature or area of your product and you need to ensure that the participants
the user from the start to finish of your study. Along the way, you’ll ask interact with it.
them to perform a series of activities on your app, site, or prototype. As
they work, they’ll be sharing their thoughts out loud.
Here’s an example of a specific task: Use the search bar to find
When you’re creating tasks for your study, focus on the insights you want hotels in Chicago, specifying that you want a non-smoking double
to gather about the specific objective you determined. If you have a lot of room from July 10-15. Then sort the results by the highest ratings.
areas you want to test, we recommend breaking these up into different
studies. In most cases, it’s best to keep your study around 15 minutes in
length, so keep this in mind as you plan your tasks.

The complete guide to user testing websites, apps, and prototypes 8


Plan tasks using logical flow Make task instructions concise

The structure of your study is important. We recommend starting with People are notorious for skimming through written content, whether they’re
broad tasks (exploring the home page, using search, adding an item to interacting with a digital product or reading instructions for a user test.
a basket) and moving in a logical flow toward specific tasks. The more
natural the flow is, the more realistic the study will be—and the better One way to ensure that your participants read your whole task is to
your results will be. make the task short and your language concise.

Example of poor logical flow: Example of poor task wording:


Create an account > find an item > check out > search the site > “Add the item to your cart. Now shop for another item and add it
evaluate the global navigation options. to your shopping cart. Then change the quantity of the first item
from 1 to 2. Now go through the whole checkout process using
Example of good logical flow: the following information...”
Evaluate the global navigation options > search the site > find an
item > create an account > checkout. Example of good, concise task wording:
This single task should be split into at least four separate tasks:

Ask yourself whether you’re interested in discovering the users’ natural 1. Add the item to your cart.
journey or whether you need them to arrive at a particular destination. 2. Shop for another item and add it to your cart.
If it’s about the journey, give the participants the freedom to use the 3. On the shopping cart, please update the quantity of the first item
product in their own way. But if you’re more focused on the destination, you added from 1 to 2.
guide them to the right location through your sequence of tasks.
4. Now proceed through the entire checkout process.
If you’re interested in both the journey and the destination, give the
users the freedom to find the right place on their own. Then, in subsequent
tasks, tell them where they should be. You can even include the correct
URL or provide instructions on how to navigate to that location.

Also, if you think a specific task will require the user to do something
complicated or has a high risk of failure, consider putting that task
near the end of the study. This will help prevent the test participants
from getting stuck or off track right in the beginning, throwing off the
results of your entire test.

The complete guide to user testing websites, apps, and prototypes 9


WRITE CLEAR QUESTIONS

Once you’ve mapped out a sequence of tasks for users to attempt, it’s time to start drafting your questions.
It’s important to structure questions accurately and strategically to get reliable answers and gain the insights
that you really want.

Tips for gathering factual responses

Don’t use industry jargon. Terms like “sub-navigation” and “affordances” probably won’t resonate with the
average user, so don’t include them in your questions unless you’re certain your actual target customer uses
those words on a daily basis.

Define new terms or concepts in the questions themselves (unless the goal of your study is to see if they
understand these terms/concepts).

If you’re asking about some sort of frequency, such as how often a user visits a particular site, make sure you
define the timeline clearly. Always put the timeline at the beginning of the sentence.
BAD: How often do you visit Amazon.com?
BETTER: How often did you visit Amazon.com in the past six months?
BEST: In the past six months, how often did you visit Amazon.com?

After you’ve written the question, consider the possible answers. If the respondent could give you the answer
“It depends,” then you should make the question more specific. It’s best to ask about first-hand experiences.
People are notoriously unreliable in predicting their own future behavior, so ask about what people have actually
done, not what they would do. It’s not always possible, but try your best to avoid hypotheticals and hearsay.

Example of asking about what someone will do or would do:


How often do you think you’ll visit this site in the next six months?
Example of asking about what someone has done:
In the past three months, how often have you visited this site?
Example of hearsay: How often do your parents log into Facebook?
Better example: Skip this question and ask the parents directly!

The complete guide to user testing websites, apps, and prototypes 10


Tips for gathering subjective data
Subjective states are relative. “Happy” in one context can mean something
very different from “happy” in another context.
Opinions are tricky. To accurately gather user opinions in a remote
study, the name of the game is making sure that all participants are
For instance:
answering the same question. The stimulus needs to be standardized.

To make sure your participants are all responding to the same stimulus,
give them a reminder of which page or screen they should be looking
at when they respond to the question. For example, “Now that you’re in
the ‘My Contacts’ screen, what three words would you use to describe Happy Not happy
this section?”
(Happy = opposite of not happy)
You’re not judging the intelligence of your respondents when analyzing
their results, so make sure that your questions don’t make them feel
that way. Place the fault on the product, not the test participant.

Bad example: “I was very lost and confused.” (agree/disagree) Happy Neutral Unhappy
Good example: “The site caused me to feel lost and confused.” (Happy = the best!)
(agree/disagree)

Be fair, realistic, and consistent with the two ends of a rating spectrum.

Very happy Happy Neutral Unhappy Very unhappy


Bad example:
“After going through the checkout process, to what extent do you trust (Happy = better than neutral, but not the best)
or distrust this company?” I distrust it slightly ←→ I trust it with my life

Good example: Plus, emotional states are very personal and mean different things to
different people. Being “very confident” to a sheepish person may mean
“After going through the checkout process, to what extent do you trust
something very different from what it means to an experienced executive.
or distrust this company?” I strongly distrust this company ←→ I strongly
trust this company

The complete guide to user testing websites, apps, and prototypes 11


Avoid asking vague or conceptual questions. Break concepts up when you’re asking the questions and
put them back together when you’re analyzing the results. For example, imagine that you want to measure
parents’ satisfaction with a children’s online learning portal. Satisfaction is a vague and complex concept.
Is it the thoroughness of the information? Is it the quality of interaction between the teachers and students
online? Is it the visual design? Is it the quality or difficulty of the assignments posted there?

Instead of asking about overall satisfaction, ask about all the criteria independently. When you’re analyzing
the results, you can create a composite “satisfaction” rating based on the results from the smaller pieces.

Avoid leading questions

With leading questions, you influence the participants’ responses by including small hints in the phrasing of
the questions themselves. More often than not, you’ll subconsciously influence the outcome of the responses
in the direction that you personally prefer.

Leading questions will result in biased, inaccurate results, and you won’t actually learn anything helpful.
In fact, it might lead you to make flawed decisions. While it may lead to the answer you “want” to hear, it
doesn’t help your team make improvements, so don’t be tempted to use leading questions!
BAD: “How much better is the new version than the original home page?”
GOOD: “Compare the new version of the home page to the original. Which do you prefer?”

USERTESTING TIP:
If you’re asking about task success, remember to define what a success is. If the previous task instructs a
user to find a tablet on Amazon and add it to the cart and you ask “Were you successful?” be sure to clarify
whether you are asking about finding a tablet or adding it to the cart.

The complete guide to user testing websites, apps, and prototypes 12


Best practices for different question types

Rating scale questions allow you to measure Multiple choice questions are great for collecting Written response questions result in short
participants’ reactions on a spectrum. They’re a yes/no responses or answers that can’t be applied answers that can be used to collect impressions
great way to benchmark common tasks and to a scale. and opinions.
compare the results with a similar test run on
your competitor’s product. • Multiple choice responses should be • Ask questions that can be answered in a
exhaustive, meaning that every possible couple of words or sentences at most.
• Use relative extremes: Make the negative response should be included in your Typing long responses can become frustrating
feeling have the lowest numerical value response options. At the same time, you for participants, especially on mobile devices.
and the positive answer have the highest want a manageable number of responses. Good example: What three words would you
numerical value. In other words, make We recommend two to six response options use to describe this app?
difficult = 1 and easy = 5, not the other per question. If you suspect that there are
way around. just too many options, do your best to guess • Use these questions sparingly. The greatest
which options will be mentioned most, and value in remote user research usually comes
• Stay consistent throughout the test! Use the then include an “Other” option. from hearing participants speak their thoughts
same end labels and the same wording when aloud naturally. Written response questions
you’re repeating a question. • Ask only one question at a time. Don’t do this: are good for getting a snapshot of the users’
“Did you find the tablet you were looking for, impressions, but if you overuse them, the
• Consider asking “why?” after a multiple-choice and was it where you expected to find it? quality of the responses will often degrade
or ratings scale. Then, when you get your Yes/No” Instead, break it up into two separate after several questions.
results back, you can go back and hear the questions.
participants discuss their answers or responses. • Create a word cloud from all your users’
Asking “why?” also prompts people to think • Choose mutually exclusive responses since responses to quickly see which words they’re
critically about their answer. users will only be able to select one answer. using to describe their experience.
If it’s possible for more than one answer to be
true, include a “More than one of these” option.

The complete guide to user testing websites, apps, and prototypes 13


Launch a dry run
Before you launch your study to all of your participants, we recommend If you set your study up in UserTesting, you can easily run one test, review it,
conducting a dry run (sometimes called a pilot study) with one or two and then add more participants to the same study. Or, if you need to make
participants. This will help you determine whether there were any flaws changes, you can create a similar study with the click of a button and then
or confusing instructions within your test. The primary goal is to make edit the tasks or questions as needed.
sure that people who read your tasks and questions interpret them in
the way they were meant to be understood. At the end of every study, it’s a good practice to review your notes to identify
any weak portions of your test and rewrite these tasks or questions for next
Here’s a good structure for doing this: time. This will save you time in the long run, especially if you’re testing
1 Release your study to just one or two participants. out prototypes first, and then running more people through a similar test
once you’ve pushed out the new changes to your website or app.
2 Listen to them as they process each task and question, and take note of
any trouble they encounter while trying to complete their first pass. Another reason to spend time to get your questions just right is if you plan to
3 Once you’ve made adjustments and you feel pretty good about your run a benchmark study. When you run an ongoing review of your website or
test, we recommend having one more person run through the test to app over time, you shouldn’t change your questions along the way or you’ll
ensure you’ve cleared up any confusing or leading questions. risk skewing the results and ruining any chances of an accurate comparison.

The complete guide to user testing websites, apps, and prototypes 14


Beware of errors
There are several common errors that can skew your data. The good news is that if you’re aware of them, you
can avoid them. Below, we’ve outlined the most common error types along with some suggestions on reducing
your chances of running into these problems in your studies.

SAMPLING ERROR

Sampling error occurs when you recruit the wrong participants to participate in your study. When this happens,
you may end up with a bunch of opinions from people outside your target market—and therefore, they aren’t
very helpful.

For example, perhaps your target market includes SaaS sales executives, so you tried to recruit people who work
in software sales, but the actual participants ended up being electronics store retail associates.

SOLUTION:
Ask clear and precise screener questions to help qualify potential study participants. If you’re uncertain whether
your screener questions are accurately capturing the right users, do a dry run with a small handful of participants.
As the first step of your study, have them describe aloud what they do for a living (or how familiar they are with your
industry, or whatever criteria will help you determine whether they’re actually your target market).

RESEARCHER ERROR

With this type of error, participants misunderstand a task or question because of the way it was worded. Study
participants will often take instructions literally and search for the exact terminology that you include in your tasks.

SOLUTION 1:
Try out your study with several people and monitor their reaction to your questions. You’ll quickly learn whether
or not your questions are accurately communicating what you want them to.

SOLUTION 2:
Be aware of your target audience and ask questions in a manner that naturally resonates with them. Use plain
language. Slang, jargon, regionalisms, and turns of phrase can easily confuse participants during a study.

The complete guide to user testing websites, apps, and prototypes 15


RESPONDENT ERROR SOCIAL DESIRABILITY ERROR

In this case, the participants are giving you inaccurate or false information. With this error, participants feel pressured to give a response that
There are several reasons that this may occur: they think is most popularly accepted in society, even if it’s not true.
For example, if you ask people about their tech-savviness, people may
• They don’t trust you with their personal information.
over-report their abilities because they think it’s “better” than not being
• They’re uncomfortable sharing details of their personal lives. tech-savvy.
• They’ve become fatigued and have resorted to bogus responses
to get through the test quickly. SOLUTION 1:
• They don’t understand whether you’re looking for their opinion or
When you’re looking for test participants, be sure to explain that you value
the “right” answer.
the skillsets or demographic characteristics you’re requesting. Emphasize
that you hope to learn how your product will be useful or beneficial to people
SOLUTION 1:
like them.
Reassure participants that their responses won’t be shared publicly.
SOLUTION 2:
SOLUTION 2:
Reassure your participants that they’ll remain anonymous.
At the very beginning of your study, be sure to explain that if they have
to fill out any personal information, their responses will be blurred out to ACQUIESCENCE BIAS
protect their identity.
When acquiescence occurs, the participant will tell you what they think you
SOLUTION 3: want to hear out of fear of offending you. For example, they may dislike your
Keep your test short (around 15 minutes, in most cases) so you don’t app but don’t want to make you feel bad about your work. This is more
fatigue your participants. common in moderated tests than unmoderated tests.

SOLUTION 1:
FAULTY PARTICIPANT RECALL
If you’re too close to the product (for example, if you’re the designer), you
These errors occur when a participant is unable to correctly remember the may want to use an impartial moderator to moderate your tests for you.
event you’re asking about. This happens when your question asks them to Skilled researchers can help ensure impartiality, reducing barriers to
recall something too far in the past or in too much detail. sharing the truth.

SOLUTION: SOLUTION 2:
Do a gut check. Can you remember the specifics of something similar? Reassure participants that you value their truth and honesty and that none of
If not, revise your question. their feedback will be taken personally.

The complete guide to user testing websites, apps, and prototypes 16


Analyze your results
Once you’ve launched your study and gotten your results back, it’s time
to get to work on analysis. Before you do anything, start by thinking back
to your objective. The objective is important and easy to lose sight of.

Be sure not to get distracted by irrelevant findings, and stay focused.


(However, you may choose to keep a side record of interesting findings
that aren’t related to the objective but may prove useful later.) With Explore correlations
your objective in mind, you can jump into the data and videos.
When reviewing rating scale and multiple-choice responses, look for
correlations between the length of time spent on a task and negative
WHAT TO LOOK FOR responses from participants. This is a starting point for identifying areas
of the user experience that may be particularly troublesome.
When we’re reviewing full recordings of user tests, we often make
annotations along the way. These annotations are captured alongside It’s important to remember, however, that correlation doesn’t tell the
the video, so you can easily jump back to interesting moments later on. whole story. Don’t jump to conclusions based on correlations alone;
be sure to watch the relevant moments of the videos to hear what
As you review the data from your questions, keep an eye out for any participants are thinking.
responses that are outside the norm. If nine participants thought a task
was fairly easy but one thought it was extremely difficult, investigate
further by watching the video that goes along with the unusual response.
What about that participant’s experience was different?

Take note of user frustrations as well as items that users find particularly
helpful or exciting. These become discussion points for design teams
and can often help to uncover opportunities for improvements to future
releases. It’s important to identify the things that people love, too, so
that you don’t inadvertently try to “fix” something that’s not broken
when trying to improve the user experience of your product.

The complete guide to user testing websites, apps, and prototypes 17


HARNESS THE POWER OF THE SPREADSHEET
Sharing your
It’s easy to gather, analyze, and share your findings right within the UserTesting dashboard. You’ll be able to
see time on task, responses to metrics questions, and more at a glance.
research findings
However, if you find that you want to perform more detailed analysis, you can download the data from your
will help your
study into an Excel spreadsheet. This can be especially helpful for compiling findings from multiple studies
side by side, breaking down patterns in studies with lots of participants, and comparing responses from
company stay
different demographic groups. focused on the
SHARE YOUR FINDINGS user experience.
Research is meant to be shared! Once you’ve established your findings, it’s time to present them to your
stakeholders. In most cases, this means hosting a meeting with the involved departments and sharing the
relevant findings in a presentation.

Here are a few ideas for successfully relaying your research findings:

Create a highlight reel of video clips.

Use charts to represent any interesting metrics data from your questions.

Back up your claims with user quotes from the studies.

Use a word cloud to display the most common words used throughout your study.

Be careful not to place blame on any of your teammates. If you have a lot of negative findings, choose
your words carefully. “Users found this feature frustrating” is much easier to hear than “This feature is
terrible.”

Encourage team members to ask questions about the findings, but remind them not to make excuses.
They’re there to learn about the customer experience, not to defend their design decisions.

Sharing research findings with stakeholders and colleagues in multiple departments can be a great way to
promote a user-centered culture in your company.

The complete guide to user testing websites, apps, and prototypes 18


Conclusion
We encourage you to make remote user research a standard part of
your development process. User feedback is crucial for designing and
developing a successful product, and remote research makes it fast and
easy to get that feedback.

With a clear objective, the right tasks, and carefully planned and worded
questions, you’ll gather useful, actionable insights.

And whether you plan on conducting your research on your own or


enlisting the help of UserTesting’s expert research team, you’re well
on your way to improving your customer experience.

The complete guide to user testing websites, apps, and prototypes 19


Create great experiences
UserTesting is the fastest and most advanced user experience research platform on the market. We give marketers,
product managers, and UX teams on-demand access to people in their target audience who deliver audio, video,
and written feedback on websites, mobile apps, prototypes, and even physical products and locations.

2672 Bayshore Parkway, Mountain View, CA 94043


www.usertesting.com | 1.800.903.9493 | support@usertesting.com

The complete guide to user testing websites, apps, and prototypes 20

Anda mungkin juga menyukai