Opinion
Artificial Intelligence Opinion

No, AI Detection Won’t Solve Cheating

Five steps to address concerns about student ChatGPT use
By Kip Glazer — April 12, 2024 4 min read
AI Robot caught in a spot light. Artificial intelligence plagiarism, cheating and ai detection concept.
  • Save to favorites
  • Print

As a high school administrator, I am often the person who has to respond to a teacher’s request to address cheating. Since the introduction of artificial intelligence detection tools, I have watched as efforts to catch cheating often do more harm than good. Not only are these tools unreliable, but they can also significantly damage trust between student and teacher.

Although some students definitely cheat, I have also learned that some do so because they are afraid of failing or even afraid of disappointing someone—including their teachers. Sadly, I have seen many educators react to all cheating as a personal affront rather than understanding the underlying causes. Adding AI detection tools to the mix can create an additional layer of stress and disappointment for all involved, even as recent research suggests AI tools like ChatGPT have not increased the frequency of cheating from high schoolers.

Because we have come to rely on computers so much, educators might be mistakenly led to believe that AI detection tools could eradicate cheating altogether. This form of techno-chauvinism—that AI can eradicate an age-old concern—is causing additional harm and angst among students and teachers who are grappling with this new reality.

Since the explosion of ChatGPT and other AI tools beginning in late 2022, many educators—including the teachers at my school—have turned to AI detection tools to catch plagiarism. In many of the cases I’ve seen, students whose writing was flagged by AI detection vehemently denied cheating.

Unfortunately, the damage and harm of accusing students of plagiarism based on inconclusive results have been extreme in some cases. For example, when one writing assignment was flagged for a 51 percent likelihood of AI use, the teacher asked administration to send that student to another location to be individually supervised for every future test. Rather than taking the result as probability, which I would recommend, the teacher considered that 51 percent as conclusive because it was machine-generated.

I have also met with irate parents who demanded that their child be removed from a class because the teacher inquired or implied that the student might have used an AI tool to cheat. These parents and students have considered the presence of an AI detection tool as a grave threat to building a positive teacher-student relationship.

I am less interested in whether AI detection tools can detect cheating or not; what does interest me, however, are the opportunities that AI can or will afford our students to improve their learning.

I believe that AI can be used as a tool to create individualized and differentiated learning opportunities for students. Just as students are now able to use calculators when taking standardized tests, I imagine a world where AI could be available to all students and be readily used to assess students’ creativity and problem-solving skills. Teaching the students to use AI to iterate might be the best use of these powerful new tools.

To achieve that goal—and ultimately render cheating concerns irrelevant—schools should consider the following:

1. Establish a clear AI policy that clarifies what is allowed for which assessment.

One of the challenges that I faced while working to support my teachers as they address cheating concerns was that our policy language didn’t necessarily specify what was allowed. During one of our investigations into cheating, we discovered that some students were genuinely confused, because they thought that using tools like Grammarly, an AI writing-assistance platform, was allowed.

Even as the speed of advancement of these tools requires us to adjust our policies frequently, it is necessary to have clear guidelines of what is allowed. Students deserve the chance to learn how to use these tools productively without being labeled a “cheater.”

2. Involve the students in creating an AI policy.

Ultimately, school policies exist to support student learning, and involving students in the policy-creation process can build their AI literacy skills. It also gives our students the opportunity to understand why certain policies exist and how adapting to them can be to their benefit rather than purely punitive.

3. Establish a policy-review process and revisit the policies governing AI use.

To address the impact of rapidly changing technology, schools should regularly review the AI classroom policy. Once again, I would advocate soliciting student input in the process whenever possible.

4. Consider certain formative and summative assessments that are plagiarism-proof.

Teachers can use AI to expand a constructionist approach to learning and project-based learning opportunities in the classroom. I encourage teachers to use AI tools in classroom instruction that focuses on what students have learned rather than asks regurgitation of discrete pieces of information that they memorized.

5. Give students a dignified way to recover from their mistakes—including cheating.

We all recognize that making mistakes is a part of being human and that learning is a deeply human endeavor. Our cheating policies should reflect this by creating a dignified path for our students who have cheated. Considering 60 percent to 70 percent of high schoolers recently reported engaging in at least one “cheating” behavior, that path to recovery is particularly important. After all, learning requires continuous yet productive failures and recoveries.

The most important aspect of my job as a school leader is to keep our students and staff safe, and I consider their emotional safety to be just as important as their physical safety. Just like the power tools in shop class, AI can do incredible harm to our students if not used appropriately. Let’s all continue to look at ways to keep our students safe as we implement AI in schools.

A version of this article appeared in the May 08, 2024 edition of Education Week as No, AI Detection Won’t Solve Cheating

Events

This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Special Education Webinar
Integrating and Interpreting MTSS Data: How Districts Are Designing Systems That Identify Student Needs
Discover practical ways to organize MTSS data that enable timely, confident MTSS decisions, ensuring every student is seen and supported.
Content provided by Panorama Education
Artificial Intelligence Live Online Discussion A Seat at the Table: AI Could Be Your Thought Partner
How can educators prepare young people for an AI-powered workplace? Join our discussion on using AI as a cognitive companion.
Student Well-Being & Movement K-12 Essentials Forum How Schools Are Teaching Students Life Skills
Join this free virtual event to explore creative ways schools have found to seamlessly integrate teaching life skills into the school day.

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
View Jobs
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
View Jobs
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
View Jobs
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.
View Jobs

Read Next

Artificial Intelligence Letter to the Editor I’m Pro-Technology, But AI’s Role in Education Worries Me
A parent shares his concerns with artificial intelligence in K-12.
1 min read
Education Week opinion letters submissions
Gwen Keraval for Education Week
Artificial Intelligence 'Grok' Chatbot Is Bad for Kids, Review Finds
The chatbot on X suggests risky behavior, and is unsafe for teens, Common Sense Media says.
4 min read
Workers install lighting on an "X" sign atop the company headquarters, formerly known as Twitter, in downtown San Francisco, July 28, 2023. Grok is the artificial intelligence chatbot built into the social media platform X.
Workers install lighting on an "X" sign atop the company headquarters of X, a social media platform formerly known as Twitter, in San Francisco on July 28, 2023. Grok is the artificially intelligent chatbot built into the social media platform.
Noah Berger/AP
Artificial Intelligence States Put 'Unprecedented' Attention on AI's Role in Schools
Most of the bills address AI literacy and require guidance on responsible use of the technology.
4 min read
Image of AI in a magnifying glass superimposed over an aerial view of a school.
Collage via EdWeek and Getty
Artificial Intelligence 'Dangerous, Manipulative Tendencies’: The Risks of Kid-Friendly AI Learning Toys
Toys powered by AI often generate inappropriate responses to questions.
4 min read
Photo illustration of a 3d rendering of a chatbot hovering over a motherboard circuit.
iStock/Getty