Flag

We stand with Ukraine and our team members from Ukraine. Here are ways you can help

Get exclusive access to thought-provoking articles, bonus podcast content, and cutting-edge whitepapers. Become a member of the UX Magazine community today!

Home ›› Artificial Intelligence ›› A Case of Accidental Anthropomorphism

A Case of Accidental Anthropomorphism

3 min read
Share this post on
Tweet
Share
Post
Share
Email
Print

Save

AccidentalAnthro_LeadBannerNew

Wherein a broken coffee grinder provides lessons in empathy.

Making a pot of coffee is one of my favorite morning rituals. After assembling my AeroPress, I set the kettle on and push the start button on my trusty Cuisinart burr grinder. It was a hand-me-down from a friend, so I’ve always had a soft spot for the device, but recently it endeared itself to me even more.

How? By not working properly. 

Once or twice a month, when I press the start button, it makes one belabored turn and then stops. I’ve discovered that if I hold down the start button and slide the lever that selects the grind duration back and forth, I can spur it to try harder. It’s almost like I can hear the machine grunting as it labors to get moving. Sometimes I’ll even dive to the bottom of the bean funnel with the AeroPress’ stirring paddle to move the beans around and give it more help.

Once it finally gains some momentum, it springs back to life, grinding as good as ever. I’ll think it’s somehow healed itself, but then, a week or two later, it will need my help again. 

My wife has suggested we get a new grinder, but now that seems impossible. The grinder that was once a cold but reliable machine has taken on a decidedly human trait: it needs my encouragement.

The realization that I now have a deep emotional bond with a coffee grinder got me thinking about anthropomorphism in design—specifically about how a product’s shortcomings can create opportunities to engender empathy. Conversational design in particular provides some great potential use cases.

Imagine a customer is interacting with an intelligent digital worker (IDW) or chatbot, trying to get help with an order.

CUSTOMER: I need help. I received my order in the mail but only part of it was delivered.

IDW: Let me see if I can help. What was the order number?

CUSTOMER: It was order number 12345. I placed the order last week.

IDW: It looks like you ordered a set of hand-dipped candles and a 4’ pewter wizard, is that right?

CUSTOMER: That’s correct. The candles arrived, but I have no pewter wizard.

IDW: Ok. I’ll connect you with a customer service agent.

That’s not a terrible script, but what if the IDW was more forthright about it’s shortcomings?

IDW: It looks like you ordered a set of hand-dipped candles and a 4’ pewter wizard, is that right?

CUSTOMER: That’s correct. The candles arrived, but I have no pewter wizard.

IDW: I’m sorry to hear about your missing pewter wizard. Let me see if I can help.

CUSTOMER: Ok.

IDW: Hmmm it looks like sometimes items weighing more than 50lbs are shipped using a different carrier. This isn’t something that I can confirm, but I’m connecting you with a service representative now.

CUSTOMER: Thank you … friend.

That’s a subtle shift, but sometimes knowing that a machine is trying to do its best can have meaning. An unexpected dash of anthropomorphism can brighten an experience, create a sense of connection, increase brand loyalty, and even make a cup of coffee taste that much better.

Tweet
Share
Post
Share
Email
Print

Related Articles

Who pays the real price for AI’s magic? Behind every smart response is a hidden human cost, and it’s time we saw the hands holding the mirror.

Article by Bernard Fitzgerald
The Price of the Mirror: When Silicon Valley Colonizes the Human Soul
  • The article reveals how AI’s human-like responses rely on the invisible labor of low-paid workers who train and moderate these systems.
  • It describes this hidden labor as a form of “cognitive colonialism,” where human judgment is extracted from the Global South for profit.
  • The piece criticizes the tech industry’s ethical posturing, showing how convenience for some is built on the suffering of others.
Share:The Price of the Mirror: When Silicon Valley Colonizes the Human Soul
7 min read

AI’s promise isn’t about more tools — it’s about orchestrating them with purpose. This article shows why random experiments fail, and how systematic design can turn chaos into ‘Organizational AGI.’

Article by Yves Binda
Random Acts of Intelligence
  • The article critiques the “hammer mentality” of using AI without a clear purpose.
  • It argues that real progress lies in orchestrating existing AI patterns, not chasing new tools.
  • The piece warns that communication complexity — the modern Tower of Babel — is AI’s biggest challenge.
  • It calls for outcome-driven, ethical design to move from random acts to “Organizational AGI.”
Share:Random Acts of Intelligence
5 min read

Most companies are trying to do a kickflip with AI and falling flat. Here’s how to fail forward, build real agentic ecosystems, and turn experimentation into impact.

Article by Josh Tyson
The “Do a Kickflip” Era of Agentic AI
  • The article compares building AI agents to learning a kickflip — failure is part of progress and provides learning.
  • It argues that real progress requires strategic clarity, not hype or blind experimentation.
  • The piece calls for proper agent runtimes and ecosystems to enable meaningful AI adoption and business impact.
Share:The “Do a Kickflip” Era of Agentic AI
7 min read

Join the UX Magazine community!

Stay informed with exclusive content on the intersection of UX, AI agents, and agentic automation—essential reading for future-focused professionals.

Hello!

You're officially a member of the UX Magazine Community.
We're excited to have you with us!

Thank you!

To begin viewing member content, please verify your email.

Tell us about you. Enroll in the course.

    This website uses cookies to ensure you get the best experience on our website. Check our privacy policy and