Flag

We stand with Ukraine and our team members from Ukraine. Here are ways you can help

Get exclusive access to thought-provoking articles, bonus podcast content, and cutting-edge whitepapers. Become a member of the UX Magazine community today!

Home ›› Companies Aren’t Prepared for Outbound AI in the Hands of Consumers

Companies Aren’t Prepared for Outbound AI in the Hands of Consumers

by Robb Wilson
3 min read
Share this post on
Tweet
Share
Post
Share
Email
Print

Save

People tend to believe that companies are going to use AI to eliminate as many jobs as possible. It stands to reason that some businesses will try this approach—even though it’s a complete misuse of the technology. What we’re currently seeing, however, is individuals picking up generative tools and running with them, while companies are dragging their feet into integration efforts.

What might happen as a result of this is that consumers will be the ones to bring down companies. There are laws that prevent companies from spamming people with unwanted outbound messages, but there are none stopping consumers from flooding contact centers with AI agents.

It’s basically free for people to cobble together agents that can robocall service centers and flood systems with data designed to get them discounts, or worse, to confuse and deceive. Customers might start hammering a company because word gets out that they give a credit for certain circumstances. This could create a snowball effect where their call centers are flooded with millions of inbound inquiries that are lined up to keep calling, all day long.

Whatever their intentions, it’s free and easy for consumers to scale ad hoc efforts to levels that will overwhelm a company’s resources. So what are companies going to do when their customers go outbound with AI? 

I asked this question recently on the London Fintech Podcast and the host, Tony Clark, had the response I’ve been looking for: “You may have let the genie out of the bottle now, Robb,” he said, looking a bit shocked. “I’m sure the tech is available. I imagine my 14-year-old could probably hook up 11 Labs or something with the GPT store and be off on something like that.”

The truth is, most companies that are evaluating agentic AI are thinking myopically about how they will use these tools offensively. They are ignoring the urgent need for agentic systems that can provide defensive solutions. 

These systems must allow AI agents to detect and stop conversations that are just meant to burn tokens. They need human-in-the-loop (HitL) functionality to make sure agents’ objectives are validated by a person who takes responsibility for the outcomes. This environment also needs canonical knowledge—a dynamic knowledge base that can serve as a source-of-truth for AI agents and humans.

These are the base requirements of an agent runtime. A critical component to integration, an agent runtime is an environment for building, testing, deploying, and evolving AI agents.

  • Runtimes maintain agent memory and goals across interactions
  • Runtimes enables access to external tools like MCPs, APIs, and databases
  • Runtimes allow multi-agent coordination
  • Runtimes operates continuously in the background

And in terms of helping businesses use AI defensively, runtimes handle input/output across modalities like text and voice, so AI agents can spot bad actors and alert humans. In UX terms, it’s the backstage infrastructure that transforms your product’s assistant from a button-press chatbot into a collaborative, contextual, goal-oriented experience designed that can proactively protect organizations and their customers. However companies choose to frame it, there’s emerging risk in sitting back and waiting to see what will happen next with AI. It just might be the end of your company.

post authorRobb Wilson

Robb Wilson

Robb Wilson is the CEO and co-founder of OneReach.ai, a leading conversational AI platform powering over 1 billion conversations per year. He also co-authored The Wall Street Journal bestselling business book, Age of Invisible Machines. An experience design pioneer with over 20 years of experience working with artificial intelligence, Robb lives with his family in Berkeley, Calif.

Tweet
Share
Post
Share
Email
Print

Related Articles

What makes players keep coming back? This piece explores the psychology of hot streaks — how momentum, perception, and smart design can turn short wins into lasting engagement.

Article by Montgomery Singman
The Psychology of Hot Streak Game Design: How to Keep Players Coming Back Every Day Without Shame
  • The article shows how hot streaks tap into players’ psychology, turning momentum into motivation.
  • It highlights the hot-hand fallacy, where players overestimate their chances of continued success.
  • The piece argues that ethical streak design should enhance engagement without exploiting addictive behavior.
Share:The Psychology of Hot Streak Game Design: How to Keep Players Coming Back Every Day Without Shame
26 min read

Trusting AI isn’t the goal — relying on it is. This article explores why human trust and AI reliance are worlds apart, and what UX designers should focus on to make AI feel dependable, not human.

Article by Verena Seibert-Giller
The Psychology of Trust in AI: Why “Relying on AI” Matters More than “Trusting It”
  • The article argues that “reliance,” not “trust,” is the right way to think about users’ relationship with AI.
  • It explains that human trust and AI reliance are driven by different psychological mechanisms.
  • The piece highlights that predictability, transparency, and control make users more willing to rely on AI.
  • It concludes that users don’t need to trust AI as a partner — only rely on it as a dependable tool.
Share:The Psychology of Trust in AI: Why “Relying on AI” Matters More than “Trusting It”
4 min read

What if your productivity app could keep you as focused as your favorite game? This article explores how game design psychology can transform everyday tools into experiences that spark flow, focus, and real engagement.

Article by Montgomery Singman
Flow State Design: Applying Game Psychology to Productivity Apps
  • The article shows how principles from game design can help productivity tools create and sustain a flow state.
  • It explains that games succeed by balancing challenge and skill, providing clear goals, and offering immediate feedback — elements most productivity apps lack.
  • The piece argues that applying these psychological insights could make work tools more engaging, adaptive, and motivating.
Share:Flow State Design: Applying Game Psychology to Productivity Apps
12 min read

Join the UX Magazine community!

Stay informed with exclusive content on the intersection of UX, AI agents, and agentic automation—essential reading for future-focused professionals.

Hello!

You're officially a member of the UX Magazine Community.
We're excited to have you with us!

Thank you!

To begin viewing member content, please verify your email.

Tell us about you. Enroll in the course.

    This website uses cookies to ensure you get the best experience on our website. Check our privacy policy and