When I started my journey as a quantitative UX researcher, I was looking for literature that could provide me with a solid foundation in this area. The book Quantitative User Experience Research by Chris Chapman and Kerry Rodden seemed to be a perfect fit for this purpose. Today I’d like to share my thoughts on this book.
General impression
Both authors are Ph.D. researchers with extensive experience in the industry. Chris Chapman is a psychologist, currently working as a a Principal UX Researcher at Amazon Lab126. You may be aware of his other books “R for Marketing Research and Analytics” and “Python for Marketing Research and Analytics”. Kerry Rodden a Senior Principal Researcher at Code for America previously worked at Google and founded the quantitative UX research role there. Probably you’ve heard about HEART metrics framework, so he’s the one who introduced it. As you can see, the authors have a solid background in the field.
Generally speaking, the book is a good starting point for those who want to dive into quantitative UX research. It covers a wide range of topics, from overview of the field to practical advice on how to conduct research. However, I found the general parts too general while practical parts turned out to be too shallow for me. Perhaps, it’s because I already have some experience in product analytics, but I expected to get more interesting insights and practical tips from the book. I was quite upset when the authors said that this or that aspect is out of the scope of the book. Nevertherless, it was good to refresh some basics, follow the authors’ narrative, get some new ideas and references. Also, I appreciate the reproducible examples provided in the book: datasets and R code are available so you can understand each concept in detail.
The only part I’d like to focus on, part III, relates to some practical aspects of quantitative UX research. The other parts represent rather a big picture of the field and seemed not interesting to me, e.g. state of the art, introduction to statistics and programming (that was especially surprising for me: just a single chapter for such broad topics), career advices, etc.
Part III. Tools and Techniques
The part consists of 4 chapters. I’ll briefly summarize each of them sharing my thoughts on them.
Chapter 7 “Metrics of User Experience”
This chapter is mostly devoted to the HEART framework: Happiness, Engagement, Adoption, Retention, Task success. It’s a well-known topic, there are many articles on this everywhere (for example, see Kerry’s original paper), so I won’t go into details here. Their example of how the framework was applied to Gmail when the labels feature was introduced in 2009 is good (actually it’s a summary of another Kerry’s paper), but besides that, there’s nothing practical more. Unfortunately.
Funny fact. From this chapter, I finally understood how Gmail labels work. Yes, I still sort emails into folders in old-fashioned way, but I will stop doing this now, I promise. It turned out that it took me, as a Gmail user, 15 years to adopt this feature.
Chapter 8 “Customer Satisfaction Surveys”
This chapter is devoted to some practical aspects of Customer Satisfaction surveys (CSat) analysis. Since I rarely deal with surveys in my work, I was curious to extend my knowledge in this area.
The most surprising fact for me was that the Net Promoter Score (NPS) is far away from being “the one number you need to grow” as it was claimed by Frederick F. Reichheld. As a data analyst, I could expect that obviously one number can’t be enough to describe the whole customer satisfaction, but now I’m surprised that this metric is still so popular.
The chapter is provided with a nice and reproducible example of how to analyze survey data using R. The most interesting part here was yet another example of Simpson’s paradox. The authors showed a declining trend for CSat for a product rolled out in the US and Germany while the same CSat metric was flat for the US and increasing for Germany. The reason, as you may have guessed, was in the different distribution of responses in these countries: the US respondents gave higher scores on average and their share was decreasing over time.
Finally, the following couple of quotes is worth to be mentioned, no comments needed.
It is common for CSat to remain constant for an be perplexed by this and ask, “We released great feature X! Why isn’t CSat going up?” There are many possible answers to this, but two fundamental ones are that customers don’t care as much as you do about your features; and respondents often answer such questions with a high-level brand evaluation, which changes slowly. Also, when a product is performing well, an unchanging rating is a good thing. The important thing is to use CSat as an indicator of health.
Stakeholders often want to know which factors lead to or “drive” CSat. This is often approached with a large survey and a regression model to look for predictors of CSat; see Chapter 7 in 25 or in 127. Unfortunately, such data often has high collinearity (correlation among items) that makes statistical modeling difficult or impossible. We largely recommend to avoid such efforts and to focus on qualitative assessment of reasons for satisfaction or dissatisfaction. If you decide to pursue driver analysis, you’ll want to investigate collinearity and dimensional reduction (discussed in Section 8.6). This involves iterative research to understand relationships in the data and build models that isolate and estimate the important effects.
However, I’m curious why correlation among items seems to be a problem. Some models, such as decision trees or its modifications, should handle correlated features well.
Chapter 9 “Log Sequence Visualization”
I won’t lie if I say that I started reading this book because of this chapter mostly. I know how difficult it is to visualize user behavior data, so I was hoping to get some new ideas. In a nutshell, the only approach the authors demonstrate here is a sunburst diagram. I heard about it later but I’ve never used it in my work, so the comprehensive and reproducible example (again, with R code and a dataset, many thanks to the authors) was quite helpful.
However, plenty of other visualizations and approaches were not even mentioned even in “Learning more” subsection (look at this overwhelming survey for example). On the other hand, this strengthened my belief that I should keep reviewing papers and writing about them here in my blog.
Chapter 10 “MaxDiff: Prioritizing Features and User Needs”
Yet another promising chapter that my eye caught. The problem that I solve occasionally as a product analyst is how to rank product features according to user preferences. Whereas I basically work with event logs and try to interpret user behavior as a signal of satisfaction, the authors describe an approach based on survey data. The idea of the MaxDiff method is to ask respondents a simple question like “Which item engages you MOST LIKELY and LEAST LIKELY?” for a set of items split into multiple smaller sets. Applying some math, you can get a preference score for each item along with the corresponding confidence interval. As a result, you can not only rank the items but also estimate the significance of the differences between them.
The chapter provides a step-by-step guide on how to conduct such a survey and a lot of practical advice. As usual, a reproducible example with R code and a dataset is included. Frankly speaking, I didn’t dive into all the details but now I definitely know where to look for them when I need them.
Personal notes
Below I put (for personal use mostly) notes and highlights that I made while reading the book rather.
Conclusion
If you’re a beginner in quantitative UX research I definitely recommend this book. For more experienced researchers, it may only be useful if you’re interested in exploring the HEART framework, sunburst diagrams, specific survey techniques (such as customer satisfaction surveys and MaxDiff), or gaining a basic understanding of the work of a quantitative UX researcher. The book is well-structured, easy to read, and provides a lot of references for further reading. Some of them I’ll take to my reading list:
- Schwarz J., Chapman C., & Feit E. M. (2020). Python for Marketing Research and Analytics. Springer.
- Sauro J., & Lewis J. R. (2016). Quantifying the User Experience: Practical Statistics for User Research. Morgan Kaufmann.
- Gourville J. T. (2004). Why Consumers Don’t Buy: The Psychology of New Product Adoption. Harvard Business School.
- Norton D. (2020). Escape Velocity: Better Metrics for Agile Teams. Onbelay.
- Norman D. (2013). The Design of Everyday Things (Revised and expanded.) Basic Books.
- Sweigart A. (2020). Beyond the Basic Stuff with Python: Best Practices for Writing Clean Code. No Starch Press.
- Kline R. B. (2015). Principles and Practice of Structural Equation Modeling (4th ed.) Guilford Press.
- Carver R. (1978). The case against statistical significance testing. Harvard Educational Review, 48(3), 378–399.