Flag

We stand with Ukraine and our team members from Ukraine. Here are ways you can help

Get exclusive access to thought-provoking articles, bonus podcast content, and cutting-edge whitepapers. Become a member of the UX Magazine community today!

Home ›› UX Design ›› Cloze Test in Practice

Cloze Test in Practice

by Paivi Salminen
4 min read
Share this post on
Tweet
Share
Post
Share
Email
Print

Save

What if removing every sixth word from your mobile content could reveal how well users actually understand it? That’s exactly what the cloze test does, and when applied to real privacy policies from platforms like Duolingo and X, the results are telling. Text that looks clean and simple on the surface quickly collapses under even minor disruption, exposing a deep mismatch between how content is written and who is actually reading it. If your users can’t fill in the blanks, chances are they never truly understood what they agreed to in the first place.

In the previous article, I explored why mobile reading is fragile and introduced a cloze test as a way to measure whether users genuinely understand what they read on a small screen.

Now I’d like to move from theory to practice by using examples. What actually happens when we apply the cloze test to real mobile content?

A quick reminder: how the cloze test works

In Mobile Usability,” Jakob Nielsen and Raluca Budiu describe the cloze test as a simple empirical comprehension test.

It works in three steps:

  1. Replace every Nth word in a piece of text with a blank. A typical test uses N = 6, although a higher value makes the task easier.
  2. Ask participants to read the modified text and fill in the missing words individually.
  3. Calculate the score as the percentage of correctly restored words. Synonyms and minor spelling mistakes are accepted because the aim is to measure comprehension, not spelling ability.

If users score around 60% or higher on average, the text can generally be considered reasonably comprehensible for that specific audience.

It’s also important to distinguish between readability and comprehension. Readability formulas estimate the education level required to process a text. Comprehension reflects the interaction between a specific text and a specific group of users. A text may have a high readability level and still be poorly understood by its intended audience.

Applying the cloze test to real mobile privacy policies

To see how this plays out in practice, I applied a cloze-style modification (removing roughly every sixth word) at the beginning of the privacy policies of two well-known mobile platforms: Duolingo and X (formerly known as “Twitter”). Below are excerpts with blanks inserted.

Example 1: Duolingo

Even with blanks inserted, you can probably infer several missing words. But notice what kind of text this is: abstract nouns, legal phrasing, layered clauses, repeated structural patterns.

Before inserting blanks, this paragraph likely scores somewhere around a Grade 13-15 reading level using standard formulas such as Flesch-Kincaid, roughly equivalent to early university level.

That level isn’t unusual for legal or policy text. But Duolingo’s audience includes teenagers, school learners, and casual adult users reading on a phone in short bursts of attention. The mismatch becomes visible when you apply the cloze method: many of the missing words are only guessable if you already understand the legal structure of privacy documentation.

Example 2: X

Again, much of this looks simple on the surface. Short sentences, familiar words. But notice how heavily it depends on implied structure: “certain information,” “our products and services,” “provide,” “required,” “account.” The text is repetitive but abstract. Without the missing words, meaning quickly becomes unstable.

In full form, this passage likely also falls around Grade 12-14. That places it well above the recommended reading level for mobile content aimed at general audiences.

Why the reading level matters

In “Mobile Usability,” Nielsen and Budiu demonstrate an example paragraph from Facebook that scored at a 14th-grade reading level. They point out that while a college-educated adult might complete the cloze test, this level is inappropriate for much of Facebook’s younger audience. Even university students, when online in casual contexts, prefer text that does not feel like a textbook.

The same logic applies here.

Privacy policies are rarely written with leisure or mobile conditions in mind. Yet users encounter them on small screens, often under time pressure, and must make decisions based on them.

If a paragraph requires university-level reading ability to score well on the cloze test, it is already misaligned with younger users, international users reading in a second language, distracted mobile readers, or anyone scanning rather than deeply studying text, among others.

Mobile content should generally aim for something closer to a Grade 6th-8th reading level, especially when comprehension affects consent, privacy, or account creation.

What the cloze test reveals in these examples

What becomes clear from applying the cloze method is not just that the text is complex. It’s how that complexity manifests.

  • Abstract phrasing breaks down quickly when context is interrupted.
  • Legal repetition does not equal clarity.
  • Structural predictability helps, but only up to a point.
  • “Professional” tone often increases reading level without increasing understanding.

On a desktop, a motivated reader might persist. On mobile, comprehension collapses much faster.

The bigger lesson

The goal of applying the cloze test here is not to praise or criticise specific companies. These are only examples to demonstrate how easily meaning can unravel when text is even slightly disrupted.

Mobile usability is not only about whether users can tap a button. It is also about whether they understand what they are agreeing to, enabling, or providing.

If removing every sixth word makes comprehension collapse below 60%, that tells us something important. And if the original text already requires university-level reading ability, that tells us even more.

The question, then, is not whether users can read our mobile copy. It is whether they can understand it quickly, under distraction, and without needing to feel as though they are studying for an exam.

General AI disclaimer

ChatGPT was used to analyze the texts of the two examples and estimate their reading levels.

The article originally appeared on Substack.
Featured image courtesy: Amanz.

post authorPaivi Salminen

Paivi Salminen
Päivi Salminen, MSc, is a digital health innovator turned researcher with over a decade of experience driving growth and innovation across start-ups and international R&D projects. After years in the industry, she has recently transitioned into academia to explore how user experience and design thinking can create more equitable and impactful healthcare solutions. Her work bridges business strategy, technology, and empathy, aiming to turn patient and clinician insights into sustainable innovations that truly make a difference.

Tweet
Share
Post
Share
Email
Print
Ideas In Brief
  • The article applies a cloze test, a practical comprehension tool, to real mobile privacy policies from Duolingo and X, demonstrating how even seemingly simple text can collapse in readability on small screens, and why closing the gap between reading level and user comprehension matters for mobile content design.

Related Articles

Discover why your most irreplaceable asset isn’t the technology you use. It’s your humanity.

Article by Pavel Bukengolts
Reimagining Work: How Designing for Humanity Will Shape 2030
  • The article argues that creativity, empathy, and emotional intelligence aren’t threatened by AI but become more valuable as automation takes over routine tasks, freeing people to focus on complex, uniquely human challenges.
  • It highlights that the key to thriving in an AI-driven world is using technology to enhance human potential: optimizing environments for focus and well-being, rather than letting it overshadow the qualities that make us effective.
  • The piece emphasizes that as workplaces evolve toward 2030, empathy becomes a core leadership skill: the engine behind authentic collaboration and meaningful human connection in increasingly automated environments.
Share:Reimagining Work: How Designing for Humanity Will Shape 2030
5 min read

Discover how personalization crosses the line from serving users to silently shaping them.

Article by Tushar Deshmukh
The Ethics of Personalization: When UX Crosses the Line from Helpful to Harmful
  • The article argues that personalization walks a fine ethical line between empowering users and quietly manipulating them.
  • It exposes how over-filtering doesn’t just limit content; it limits identity, replacing user curiosity with algorithmic compliance.
  • The piece calls on UX practitioners to treat ethical personalization as a foundational responsibility: one that demands transparency, fairness, and respect for human dignity.
Share:The Ethics of Personalization: When UX Crosses the Line from Helpful to Harmful
4 min read

Learn why your users decide whether to stay or leave before they even understand your product.

Article by Tushar Deshmukh
The Psychology of Onboarding: First Impressions Rule the Brain
  • The article argues that onboarding is not where users begin; it is where they decide whether to stay or leave.
  • It shows that most onboarding failures are not design problems; they are psychological ones.
  • The piece challenges designers to recognize that first impressions are cognitive anchors and that the brain rarely revises its judgments.
Share:The Psychology of Onboarding: First Impressions Rule the Brain
5 min read

Join the UX Magazine community!

Stay informed with exclusive content on the intersection of UX, AI agents, and agentic automation—essential reading for future-focused professionals.

Hello!

You're officially a member of the UX Magazine Community.
We're excited to have you with us!

Thank you!

To begin viewing member content, please verify your email.

Get Paid to Test AI Products

Earn an average of $100 per test by reviewing AI-first product experiences and sharing your feedback.

    Tell us about you. Enroll in the course.

      This website uses cookies to ensure you get the best experience on our website. Check our privacy policy and