- AI for Admins
- Posts
- 🤖 AI cheating just got trickier
🤖 AI cheating just got trickier
Humanizers, adversarial prompts, and the search for relevant classwork
Educators are still concerned about AI’s threat to academic integrity.
In a Carnegie Learning poll, 52 percent of educators cited the biggest AI challenge in schools to be “students using AI for cheating.”
(That was ahead of “lack of training and support” and “difficulty in integrating AI tools into the curriculum.”)
More bad news: the cheating issue isn’t getting any easier.
AI detectors are still a total dumpster fire. Teachers are STILL using them.
And now, students are finding new ways to avoid detection by them.
It’s an endless spiral that takes us farther and farther from a focus on learning.
In today’s newsletter, we’ll look at those new AI detection developments — and some more productive solutions than “a better AI detector.” 🤦🏻♂️
In this week’s newsletter:
🗳 Poll: Concerns about bias
😈 The cat and mouse game of AI cheating
📺 AI and the changing landscape of education
📚 New AI resources this week
🗳 Poll: Concerns about bias
This week’s question: How concerned are you about bias in AI responses?
🟨🟨🟨⬜️⬜️⬜️ 5 (very concerned) (10)
🟨🟨🟨⬜️⬜️⬜️ 4 (10)
🟩🟩🟩🟩🟩🟩 3 (somewhat concerned) (16)
🟨⬜️⬜️⬜️⬜️⬜️ 2 (3)
⬜️⬜️⬜️⬜️⬜️⬜️ 1 (not at all concerned) (2)
Some of your responses:
Voted “somewhat concerned”: Understanding that the bias that exists in AI is founded in the bias that has already been perpetuated by authors of web-based data at large, what I find most concerning is making sure we educate teachers and students on how to identify it, and ensuring that teachers can identify and correct the bias in their own prompting when designing content.
Voted “not at all concerned”: AI is much like the introduction to the web. People freak out at first, and the next thing you know, everyone is using it and cannot live without it.
Voted “very concerned”: I worry that AI will reproduce and reinforce the hidden, structural biases we already see in so many systems in the US. We must be skeptical about and critical of AI-generated information if we want to disrupt systems of oppression that all too often are reproduced in our schools.
🗳 This week’s poll
For this poll, most of us will probably want an “E: All of the above” option … but I’m leaving that one out!
Choose the one that you think is most important right now (and explain in a comment afterward).
Instructions:
Please vote on this week’s poll. It just takes a click!
Optional: Explain your vote / provide context / add details in a comment afterward.
Optional: Include your name in your comment so I can credit you if I use your response.
How do we solve the AI / academic integrity issue? |
😈 The cat and mouse game of AI cheating
Image created with Google Gemini’s Imagen 3 model
Ever since ChatGPT came on the scene, educators have been worried about its impact on academic integrity.
How do I know? Because whenever I do workshops and presentations with educators, it’s one of the main questions …
… and my “AI vs. Cheating, Plagiarism, and Academic Integrity” session is usually one of my most-attended sessions. (Check out the slides here.)
(Spoiler alert: If you come to that session, it’s not about beating students at the cheating game — but rather how to think differently about the entire “academic integrity” situation.)
Since the beginning, the “AI cheating” game has been a cat-and-mouse game … a game of “whack-a-mole” that is impossible to win.
And even as new tools emerge, it’s still going to be impossible to win.
Here’s a new slide in my aforementioned conference session on AI cheating:
Back and forth. Back and forth.
Accusations. Threats. Technology measures and countermeasures.
Let’s address the latest step in this cat-and-mouse game — humanizers and adversarial prompting. (I share this not as a solution, but to keep you informed on what people are talking about so you’re prepared.)
Then, let’s talk about some real solutions.
😼 🐭 Text humanizers
In the natural progression of the cat-and-mouse game of “AI cheating,” the most recent step makes logical sense …
If teachers are going to use AI detectors (which are heinous and inaccurate … more on that in a moment), then students who want to use AI to avoid classwork should try to beat the AI detectors.
That’s where text humanizers come in.
They take AI-created text and make it sound less like AI — and less likely to be detected by AI detectors.
Some examples (without links, because I don’t want to give these sites extra web traffic or eyeballs):
BypassAI
Humbot
Undetectable AI
Humanizer . org
And the list goes on and on and on.
😼 🐭 Adversarial prompting
You don’t even need a text humanizer to beat AI detectors, though. Some extra prompting in ChatGPT (or your AI assistant of choice) will do it.
This paper, written by university faculty in Vietnam and Singapore, highlights AI prompting techniques that can avoid detection.
They include:
Rewriting with intentional errors
Varying sentence length
Increasing text complexity
Downgrading text complexity
Rewriting as a Non-Native English Speaker (NNES) with IELTS Band Level 6
Paraphrasing
The paper’s conclusion offers three implications:
The major implication of these findings for educators and administrators is that the use of AI text detectors should not be implemented uncritically; users must consider the impacts and limitations of AI text detection technologies before using them for assessment or educational practices.
Secondly, the results of such technologies should not be used for punitive actions or in accusations against students without a high degree of certainty, and those in the position to evaluate the results from such technologies need to consider the ease with which detectors can be evaded, and their potential to inequitably impact certain student populations.
Finally, the findings imply that educators must radically reconsider their assessment structures and practices in light of new technology, given that current efforts to detect GenAI content are unlikely to be successful.
😼 🐭 AI detectors
We’ve touched on this a LOT in this newsletter, but in case you’ve missed it …
AI text detectors are terrible at their jobs. If you go to their websites, they won’t tell you that. They’ll boast some percentage accuracy rate that’s not true (or realistic).
Published academic research has backed this up …
🤦🏻♂️ We’re on a race to nowhere
Have you noticed something about this back-and-forth battle about AI detection?
The longer it goes, the less it focuses less and less on the learning.
And, to some extent, it’s a monster of our own creation.
When we agree to play this game against our students — and try to beat them at it — we only encourage them to take further steps in their own game.
It’s a no-win situation.
The real problem is that we are actually using these AI detectors in the first place.
Lots of teachers are being lured in by the siren’s song of “everything can go back to the way it was before” and “you won’t have to change how you teach.” They’re buying a bill of goods that doesn’t produce results — accurate results, that is.
🤷♂️ So … what do we do?
We’re in a messy time of transition. Much like we did with previous innovative disruptions to the classroom (calculators, encyclopedias, search engines, 1:1 computing, YouTube, etc.), we need to evolve.
That doesn’t mean that we have to throw out all of our traditional classwork and create brand new assignments right away.
We’ll have to ask ourselves a couple crucial questions …
1. Can we save our original classwork / assignments?
I think the answer is — maybe, maybe not. Here are some things to consider:
When AI cheapens the final product, focus on the process.
Collaborative work can encourage original human thought.
Incorporate creative final products to demonstrate learning.
Have students reflect on their work (how they’d do it different, etc.).
Consider an element of AI/student collaboration to existing classwork.
Scaffold writing activities -- or chunk them into smaller assignments.
2. How can we reimagine our classwork?
This is the big question lots of us are trying to figure out right now.
Here are a few things we’ll really need to think about …
Thinking and skill development: Can AI support these instead of removing them? How can AI augment rather than replace?
Goals and objectives: Instead of saying “this is how we’ve taught and learned before,” we’ll have to go back to the basics. What are we trying to accomplish?
Experimentation: Let’s try incorporating AI into the learning process in some ways and afterward, ask ourselves: “What worked? What didn’t? How could this be different next time?”
Focus on the future: With any changes — or decisions to stay the same — we must ask: How does this prepare students for the future they will face?
🤔 What do you think?
How do we get this right?
What are some positive steps you’re seeing in your realm of education?
What else do we need to consider?
Hit reply and let me know — I’d love to engage in the conversation! — or respond to the poll above!
📺 AI and the changing landscape of education
Natasha Berg, M.Ed. works as the Multimedia and Technology Integration Specialist at a local high school in South Dakota. She has spent her career learning about and developing her skills in education and educational technology.
Berg believes that new and emerging technology should be integrated into classrooms as it fully prepares students to enter the 21st century workforce and helps make learning accessible to students of all abilities.
In this TEDx Talk, she suggests that generative AI might be the catalyst we need to push us forward.
"If students feel as though what they are learning in the classroom will benefit them in the long run, they will be invested in learning. They won't be looking for shortcuts to simply get the work done."
PS: She quoted me in this video! I had no idea until a friend sent it to me!
📚 New AI resources this week
1️⃣ Google Learn About: This conversational learning companion helps you grasp new topics and deepen your understanding, adapting to your unique curiosity and learning goals.
2️⃣ Should Human Sports Referees Be Mostly Replaced With Automated Ones? (via The New York Times): This is the latest in many, many decisions on that big question — How much AI is too much, and how much is OK?
3️⃣ AI in Education in 2024: Educators Express Mixed Feelings on the Technology’s Future (via EdTech Magazine): There is optimism about artificial intelligence’s productivity gains amid worry about inappropriate student use.
I hope you enjoy these resources — and I hope they support you in your work!
Please always feel free to share what’s working for you — or how we can improve this community.
Matt Miller
Host, AI for Admins
Educator, Author, Speaker, Podcaster
[email protected]