Flag

We stand with Ukraine and our team members from Ukraine. Here are ways you can help

Get exclusive access to thought-provoking articles, bonus podcast content, and cutting-edge whitepapers. Become a member of the UX Magazine community today!

Home ›› Accessibility ›› Voice Search and Voice Interfaces 101

Voice Search and Voice Interfaces 101

by Adam Fard
8 min read
Share this post on
Tweet
Share
Post
Share
Email
Print

Save

Typing search queries is gradually becoming an outdated artifact of the past and voice tech is becoming more useful and promising by the day. So what other prospects await for voice tech and why is it so special? We’re about to find out!

Technology is moving at a staggering rate, and failing to adapt to the ever-changing technological landscape may end up being a costly mistake for most businesses. One of the more important developments of recent years is voice search and voice interfaces—they’ve gradually shifted from a novel way of searching for stuff on Google to a standard way of human-computer interaction. 

The numbers seem to support this viewpoint. The voice and speech recognition market is slated to grow at a 17.2% CAR to reach $26.8 billion by 2025. Similarly, there’s also a significant change in user sentiment—research suggests that in 2021, consumers prefer to use voice to look for information online rather than doing it manually.

In this article, we’ll take a look at some essential voice search statistics, how people interact with voice interfaces, and why startups should care about all of this. 

Let’s dive right in. 

What is a voice UI?

Also known as voice-enabled search, voice UI allows people to request information only by speaking, without the need to type. 

As a result, this technology allows to simplify and optimize the search process, which provides for better UX and a more accessible environment. This means of human-computer interaction will enable users to perform searches much quicker while also reducing cognitive load and friction.

Some statistics on voice interfaces 

Voice interactions have always made humans unique—it’s a quicker and more efficient way of communicating meaning. In the grand scheme of things, it’s surprising that we didn’t incorporate it in our tech earlier. Back in 2016, Google’s Voice Search Statistics suggested that 20% of all searches were done by voice. A mere five years later, it seems like typing search queries is gradually becoming an outdated artifact of the past. 

Let’s take a quick look at some voice search statistics:

This overwhelming consumer shift towards voice interfaces will inevitably lead to the mass adoption of this technology among startups. Here are a few examples:

  • Fintech products will gradually normalize voice-enabled payments and enhanced security via voice biometrics. 
  • Edtech platforms will use voice interfaces to assist children with learning and motor disabilities.
  • The healthcare industry will use advanced voice recognition to automate filling in patient files, which may in some cases take up to 6 hours a day for some. 

These examples are but a glimpse into the immense capabilities of this technology. This is a great time for startup founders as well as usability specialists to do their best to familiarize themselves with voice tech and incorporate it into their products. 

What can users do with voice?

According to a survey conducted by Adobe, the most common types of voice searches performed by users are:

  • Searching for music via smart speakers (70%);
  • Requesting the weather forecast (64%);
  • Asking fun questions (53%);
  • Searching for things online (47%);
  • Checking the news (46%);
  • Asking directions (34%);

A customer intelligence report published by PwC suggests that users prefer using voice for a wide array of actions like texting friends and searching for stuff on the internet. However, they are somewhat reluctant to shop for things using voice. 

Here’s how one participant explains this reluctance: “I would shop for simple things like dog food, toilet paper, pizza… but ‘can you order me a sweater?’ That’s too risky.”

What is a voice user interface (VUI)?

Voice User Interface (VUI) is the primary or supplementary auditory, tactile or visual, interface that allows voice interaction between devices and people. These interfaces use speech recognition and natural language processing technologies to transfer the user’s speech input into meaning and, eventually, commands. 

When the Voice UI emerged

Here are some important dates and landmarks. 

  • The first-gen VUI was launched by Nuance and SpeechWorks in 1984 via Interactive Voice Response (IVR) systems. They became mainstream in 2000. 
  • Apple introduced the concept of Siri in 2006. Siri came in 2007
  • Google introduced a voice-enabled search in 2007
  • Cortana (Microsoft, 2011)
  • Amazon introduced Amazon Echo in 2014. It is a smart speaker that works with virtual assistants such as Siri, Alexa, etc.
  • Google Assistant (Google, 2016)

How to Design a Voice Interface?

Designing voice interfaces typically refers to augmenting the abilities of existing voice assistants like Siri or Cortana. 

Most assistants allow third parties to develop new capabilities that will enable to improve a customer’s experience with both the assistant and the service provider. Think of a travel agency that would allow its customers to book flights via the voice assistant of their choice. There are specific names for these extensions: Amazon and Microsoft call them “skills.” Apple refers to them as “intents,” while Google named them “actions.” 

While we’re dealing with an entirely new medium, you’ll learn that the process of designing a voice interface isn’t very different from putting together a GUI. 

Pre-design Stage

Before we dive into actual interactive design, we need to look into an array of factors that will influence the shape of our end-product. These factors can be technological, geographical, environmental, sociological, and so forth. Our goal in the pre-design stage is to dig deep into the context in which our product will exist.

One of the more critical aspects of this context is the device category you’re developing an interface for. The first question you’ll have to answer is, “Will this product be developed for phones, wearables, stationary connected devices, or non-stationary computing devices?.”

After you’ve established what device this product is for, it’s important to create a use case matrix where you’ll outline what actions the product will be used for and how often you think users will perform them. 

The Main Design

1. User research

By now, most of us will be aware of how important user research is in the UX design process, but it’s important to underline that it plays an even more important role when it comes to VUIs. Not only will it allow you to understand your users’ pain points, desires, and aspirations, it will also provide you with an in-depth understanding of how they actually speak to their assistants. 

2. Customer Journey mapping

A customer journey map with the voice as a channel assists UX researchers in identifying the requirements of the user at different stages of development. 

Fundamentally, journey maps allow designers to understand user pain points at various stages of product use. 

3. VUI competitor analysis

The idea behind this step is to understand how your competition approaches the same problem and how you can do it differently or improve on that. Here are a few essential questions you should ask at this stage: 

  • What are the primary use cases for my competitors’ apps?
  • What voice commands do they use?
  • What do their customers enjoy and dislike about their product, and how can we leverage this insight? 

4. Gather requirements

Creating clear requirements for your VUI is crucial, since this will be a guiding document for the product’s developers. Here are two essential components of your requirements:

  • Key scenarios of interaction that are also voice-compatible
  • Intent, utterance, and slot

Let take a closer look at each point: The first step here would be to define the user scenarios, which will then be translated into a dialog flow between the user and the VUI. A scenario is pretty much a user story — it consists of an agent, an action, and the transformation that occurs as a result of that action. 

To explore the second point, let’s think of a generic request people make to their assistants: “Alexa, order me a cab to Times Square.” This request features three important factors: intentutterance, and slot

  • The intent is the overarching objective behind a user’s command; it can be high utility or low utility. The former is a highly specific request like “Order me a cab to Times Square,” the latter, on the other hand, is more vague, like “Tell me more about Times Square.”
  • The utterance represents all the possible ways a request can be structured. In our case, it would list different iterations of the original request like “I need a cab to Times Square,” “Get me a cab to Times Square,” “Find me a taxi to Times Square,” etc. 
  • The slots are additional bits of information that the VUI might ask for in order to complete the user’s request. For instance, if the user said, “Siri, I need a cab,” the request is too vague in order to be successfully executed. Therefore, the assistant will ask for an address—this is a slot.  

5. Prototyping

When prototyping for VUIs, designers need to put their scriptwriter’s hat on and think about how they should structure a person’s interaction with the interface so that it satisfies the requirements mentioned above. 

Source

In this case, our prototype is called a dialog flow, and it has to describe the following aspects of the interaction:

  • Keywords that trigger a particular interaction;
  • Branches that outline the directions in which the conversation can go in;
  • Sample interactions for the user and the assistant;
Source

6. Usability Testing

Like GUIs, VUIs need to go through rigorous usability evaluation. There are many types of tests that can surface usability issues in voice interface like Wizard of Oz testing, recruited usability studies, recognition analysis, and so forth. 

Trends in Voice Interface

There is a broad spectrum of astounding developments happening in voice right now. Here are some of the hottest trends in the field: 

  • Voice assistants in mobile apps: products will gradually move to a voice-first interface and become an anticipated feature for most users. 
  • Outbound calls using smart interactive voice response software powered with natural language understanding technology: this software combination is projected to replace agents in call centers by offering a powerful and flexible solution. 
  • Voice cloning is a technology marketers, filmmakers, and a variety of other content producers will benefit from. It will allow to make synthesized speech more emotional, customizable, and human-like. 
  • Voice assistants in smart TVs: very soon, you’ll never have to look for your remote control anymore. Voice assistants will soon be used to browse channels, launch apps, and take total control over your television using your voice only. 

Final Words

Voice tech is becoming more useful and promising by the day and is gradually becoming an important part of tomorrow’s user experience standards. This is an exciting new field that designers, researchers, and developers will see more of in the next few years, as user adoption inevitably surges.

Fragmented approaches
post authorAdam Fard

Adam Fard
Adam is a senior lead UX/UI designer with more than 8 years of experience. Adam's passion for design steadily grew into his own agency, that he's currently leading.

Tweet
Share
Post
Share
Email
Print
Ideas In Brief
  • This article covers some essential voice search statistics, how people interact with voice interfaces, what can users do with voice and why startups should care about all of this.
  • Typing search queries is gradually becoming an outdated artifact of the past.
  • The author believes that the overwhelming consumer shift towards voice interfaces will inevitably lead to the mass adoption of this technology among startups.
  • How to Design a Voice Interface?
    • Pre-design Stage
    • The Main Design (user research, customer journey mapping, VUI competitor analysis, gathering requirements, prototyping, usability testing)

Related Articles

Discover how GPT Researcher is transforming the research landscape by using multiple AI agents to deliver deeper, unbiased insights. With Tavily, this approach aims to redefine how we search for and interpret information.

Article by Assaf Elovic
You Are Doing Research Wrong
  • The article introduces GPT Researcher, an AI tool that uses multiple specialized agents to enhance research depth and accuracy beyond traditional search engines.
  • It explores how GPT Researcher’s agentic approach reduces bias by simulating a collaborative research process, focusing on factual, well-rounded responses.
  • The piece presents Tavily, a search engine aligned with GPT Researcher’s framework, aimed at delivering transparent and objective search results.
Share:You Are Doing Research Wrong
6 min read

The role of the Head of Design is transforming. Dive into how modern design leaders amplify impact, foster innovation, and shape strategic culture, redefining what it means to lead design today.

Article by Darren Smith
Head of Design is Dead, Long Live the Head of Design!
  • The article examines the evolving role of the Head of Design, highlighting shifts in expectations, responsibilities, and leadership impact within design teams.
  • It discusses how design leaders amplify team performance, foster innovation, and align design initiatives with broader business goals, especially under changing demands in leadership roles.
  • The piece emphasizes the critical value of design leadership as a multiplier for organizational success, offering insights into the unique contributions that design leaders bring to strategy, culture, and team cohesion.
Share:Head of Design is Dead, Long Live the Head of Design!
9 min read

Discover how digital twins are transforming industries by enabling innovation and reducing waste. This article delves into the power of digital twins to create virtual replicas, allowing companies to improve products, processes, and sustainability efforts before physical resources are used. Read on to see how this cutting-edge technology helps streamline operations and drive smarter, eco-friendly decisions

Article by Alla Slesarenko
How Digital Twins Drive Innovation and Minimize Waste
  • The article explores how digital twins—virtual models of physical objects—enable organizations to drive innovation by allowing testing and improvements before physical implementation.
  • It discusses how digital twins can minimize waste and increase efficiency by identifying potential issues early, ultimately optimizing resource use.
  • The piece emphasizes the role of digital twins in various sectors, showcasing their capacity to improve processes, product development, and sustainability initiatives.
Share:How Digital Twins Drive Innovation and Minimize Waste
5 min read

Join the UX Magazine community!

Stay informed with exclusive content on the intersection of UX, AI agents, and agentic automation—essential reading for future-focused professionals.

Hello!

You're officially a member of the UX Magazine Community.
We're excited to have you with us!

Thank you!

To begin viewing member content, please verify your email.

Tell us about you. Enroll in the course.

    This website uses cookies to ensure you get the best experience on our website. Check our privacy policy and