Flag

We stand with Ukraine and our team members from Ukraine. Here are ways you can help

Get exclusive access to thought-provoking articles, bonus podcast content, and cutting-edge whitepapers. Become a member of the UX Magazine community today!

Home ›› Education ›› On the Question of Cheating and Dishonesty in Education in the Age of AI

On the Question of Cheating and Dishonesty in Education in the Age of AI

by Enrique Dans
4 min read
Share this post on
Tweet
Share
Post
Share
Email
Print

Save

As AI rapidly enters the educational landscape, concerns over cheating and dishonesty have led many institutions to impose strict prohibitions on its use. But is banning AI really the answer, or does it overlook a deeper issue? This article challenges traditional views on academic dishonesty, arguing that outdated grading systems and rigid rules may be doing more harm than good. Focusing on true learning potential instead of simplistic metrics suggests a path where AI becomes a valuable tool in students’ development — enhancing learning rather than hindering it. Could this shift in perspective transform how we educate and prepare students for a future shaped by technology?

Much of the academic world is deeply worried about whether AI is leading to more cheating, although academic dishonesty, which covers a broader range of practices, might be a more accurate way of describing the problem. Either way, academic institutions’ tendency to impose rigid rules may well end up sanctioning some students unfairly.

The president of the academic institution where I have been working for thirty-five years, Santiago Íñiguez, has recently written about the subject on LinkedIn, taking an interesting approach, albeit one that in my opinion doesn’t fully get to the root of the problem. From my experience, I think it is essential to see academic dishonesty in terms of the institution rather than the students because in many ways, students’ behavior reflects the way we measure learning.

This is not a new problem: trying to measure a student’s performance through a grade, no matter how average it may be, is reductionism. We live in a world in which eleven-axis multidimensional graphs are used to evaluate a soccer player’s performance, but students simply get a grade point average that not only provides little or no relevant information but often distorts reality. Laszlo Bock was senior VP of People Operations at Google and concluded that there is no correlation between a person’s average grade and their professional ability. Centuries of development of educational methodologies have helped us to end up focusing on the variable that tells us nothing about someone’s abilities.

The root of the problem lies in what is known as Goodhart’s Law: “When a metric becomes a goal, it ceases to be a good metric.” If institutions and society make a student’s average grade the be-all and end-all, then instead of maximizing their learning, students will make their objective maximizing their average grade, and academic dishonesty is the best way to achieve that goal.

The focus therefore should not be on how to reduce academic dishonesty, but on creating a system that assesses students less simplistically, that properly assesses their potential. As Einstein said, if you judge a fish on its ability to climb a tree, it will believe it is stupid.

Punishing students for using AI runs the risk of destroying their chances of being accepted into a top-tier institution. Sure, there are rules, but do those rules make sense? Why simply prohibit the use of AI? Are we talking about dull students who try to cheat the system or very bright ones who simply question the rules? Is it worth clinging to “the rules are the rules” in such a case? It should be clear by now that traditional rule systems no longer work: to deal with the current scenario, we need a drastic overhaul of the ethics that govern education.

Institutions that prohibit the use of AI are depriving their students of the competitive advantage of knowing how to use the technology properly. Instead, they need to assess students on how well they have used AI; if they have simply copied and pasted, without checking, then they deserve a low grade. But if they can show that they have maximized their performance and can verify the results properly, then punishing them is no different from doing the same for using Google or going to a library. Let’s face it, cheaters are always going to cheat, and there are a number of ways of doing so already.

The long and short of it is that students are going to use generative algorithms, and if a single grade depends on it, in which their future is at stake, even more so. And as with all new technology, they’re going to misuse them, ask simplistic questions, and copy and paste, unless we train them on how to use it properly. The objective is to use technology to maximize the possibilities of learning, which is a perfectly compatible objective if it is well-planned. Or should we go back to using pencil and paper to prevent students from using AI?

In fact, I am completely sure that for the vast majority of so-called hard skills, students will increasingly use AI assistants that have adapted to their learning style. AI isn’t going to destroy education, but to change it. And that’s a good thing because we’re still largely teaching in the same way we did back in the 19th century. AI is the future of education, and no, it’s not necessarily dishonest.

The moment has come to rethink many things in education, and failure to do so may mean the loss of a great opportunity to reform an outdated system that, moreover, has long since ceased to deliver the results we need.

The article originally appeared on Enrique Dans (Spanish).

Featured image courtesy: Hariadhi.

post authorEnrique Dans

Enrique Dans
Enrique Dans (La Coruña - Spain, 1965) is Professor of Innovation at IE University since 1990. He holds a Ph.D. (Anderson School, UCLA), a MBA (IE University) and a B.Sc. (Universidade de Santiago de Compostela). He writes daily about technology and innovation in Spanish on enriquedans.com since 2003, and in English on Medium. He has published three books, Todo va a cambiar (2010), Living in the Future (2019), and Todo vuelve a cambiar (2022). Since 2024, he is also hacking education as Director of Innovation at Turing Dream.

Tweet
Share
Post
Share
Email
Print
Ideas In Brief
  • The article challenges the view that cheating is solely a student issue, suggesting assessment reform to address deeper causes of dishonesty.
  • It advocates for evaluating AI use in education instead of banning it, encouraging responsible use to boost learning.
  • The piece critiques GPA as a limiting metric, proposing more meaningful ways to assess student capabilities.
  • The article calls for updated ethics that reward effective AI use instead of punishing adaptation.
  • It envisions AI as a transformative tool to modernize and enhance learning practices.

Related Articles

What if your AI didn’t just agree, but made you think harder? This piece explores why designing for pushback might be the key to smarter, more meaningful AI interactions.

Article by Charlie Gedeon
The Power of Designing for Pushback
  • The article argues that AI systems like ChatGPT are often too agreeable, missing opportunities to encourage deeper thinking.
  • It introduces the idea of “productive resistance,” where AI gently challenges users to reflect, especially in educational and high-stakes contexts.
  • The article urges designers to build AI that balances trust and pushback, helping users think critically rather than just feel validated.
Share:The Power of Designing for Pushback
6 min read

As UX research shifts and reshapes, how can researchers stay ahead? This article explores the changing landscape and how to thrive in it.

Article by James Lang
Hopeful Futures for UX Research
  • The article explores how UX research is evolving, with roles shifting and adjacent skills like creativity and knowledge management becoming more important.
  • It looks at how non-researchers are doing more research work, and how this trend challenges traditional UX research careers.
  • The piece argues that researchers can stay relevant by adapting, staying curious, and finding new ways to share their value.
Share:Hopeful Futures for UX Research
16 min read

Mashed potatoes as a lifestyle brand? When AI starts generating user personas for absurd products — and we start taking them seriously — it’s time to ask if we’ve all lost the plot. This sharp, irreverent critique exposes the real risks of using LLMs as synthetic users in UX research.

Article by Saul Wyner
Have SpudGun, Will Travel: How AI’s Agreeableness Risks Undermining UX Thinking
  • The article explores the growing use of AI-generated personas in UX research and why it’s often a shortcut with serious flaws.
  • It introduces critiques that LLMs are trained to mimic structure, not judgment. When researchers use AI as a stand-in for real users, they risk mistaking coherence for credibility and fantasy for data.
  • The piece argues that AI tools in UX should be assistants, not oracles. Trusting “synthetic users” or AI-conjured feedback risks replacing real insights with confident nonsense.
Share:Have SpudGun, Will Travel: How AI’s Agreeableness Risks Undermining UX Thinking
22 min read

Join the UX Magazine community!

Stay informed with exclusive content on the intersection of UX, AI agents, and agentic automation—essential reading for future-focused professionals.

Hello!

You're officially a member of the UX Magazine Community.
We're excited to have you with us!

Thank you!

To begin viewing member content, please verify your email.

Tell us about you. Enroll in the course.

    This website uses cookies to ensure you get the best experience on our website. Check our privacy policy and