This page is part of a global project to create a better online reviews system. If you want to know more or give your feedback, write at [email protected] and we’ll grab a beer ;)

Thresholds

Comparing average ratings is often the first thing users do during the calibration step (see “Why Do We Look at Online Reviews”). With many available options, people need to narrow down their choices before moving to the compare step. The average rating, visible in the list of options, plays a significant role in this decision-making process.

Google Maps’ listings of restaurants: the average rating, the name, and the location are the three pieces of info available.

Google Maps’ listings of restaurants: the average rating, the name, and the location are the three pieces of info available.

At this stage, all listings below a certain threshold may be automatically and almost unconsciously excluded.

Looking at the screenshot from Google Maps, you probably wouldn’t consider options below a 4 rating. This means that 4 is a threshold. Add to that that these filters are not just psychological: platforms pre-selects and removes low-rated options to avoid overwhelming users with too many choices. For example, Google only shows highest-rated spots, and Uber first propose rides to the best-rated drivers.

When it’s not automatic, platforms provide filters to remove low-rated options, and 70% of people use them $^1$.

Google Maps’ rating filter

Google Maps’ rating filter

Airbnb’s new “Guest favorites” label filters listings above 4.9 (with other few conditions)

Airbnb’s new “Guest favorites” label filters listings above 4.9 (with other few conditions)

ReviewTrackers study results $^1$

ReviewTrackers study results $^1$

In this context, it’s understandable why businesses might gate reviews to stay above the threshold (see “Review Gating”).

Moreover, thresholds and standards vary across industries, countries, and platforms (e.g., the average rating is 4.3 stars on Google, 4.25 stars on Tripadvisor, and 3.65 stars on Yelp). The same applies to customer support satisfaction and net promoter scores $^2$.

Psychological Tricks

An interesting research study $^3$ shows that people use different strategies when making choices based on customer ratings. These strategies vary with the number of reviews and the average rating range.

A personal example on Airbnb: I tend to select listings above 4.8 and only consider listings between 4.6 and 4.8 if the higher-rated ones are either booked or too expensive. Every time I do this, I feel silly because:

Another psychological bias is that while consumers expect a high average rating, a 5-star average rating can appear suspicious. A study showed that 68% of consumers either “agree” or “somewhat agree” that a high review rating is not trustworthy unless there are a lot of reviews $^4$. This skepticism arises from the abundance of fake reviews (see “Fake Reviews Flood the Web”). Human behavior is indeed complex.

The psychological impact isn’t limited to the average rating. The number of reviews also plays a significant role, and people seem to use a Bayesian approach to compare options. We’ll cover that in “Number of Reviews”.

<aside> 💡 Exploration

</aside>