Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Get a Samsung OLED gaming monitor for just $350

    Qualcomm Snapdragon X2 Elite tops the Apple M5 in new test video

    Tapo’s 1440p Wi-Fi security cam is 42% off! Grab it now for $70

    Facebook X (Twitter) Instagram
    • Artificial Intelligence
    • Business Technology
    • Cryptocurrency
    • Gadgets
    • Gaming
    • Health
    • Software and Apps
    • Technology
    Facebook X (Twitter) Instagram Pinterest Vimeo
    Tech AI Verse
    • Home
    • Artificial Intelligence

      Read the extended transcript: President Donald Trump interviewed by ‘NBC Nightly News’ anchor Tom Llamas

      February 6, 2026

      Stocks and bitcoin sink as investors dump software company shares

      February 4, 2026

      AI, crypto and Trump super PACs stash millions to spend on the midterms

      February 2, 2026

      To avoid accusations of AI cheating, college students are turning to AI

      January 29, 2026

      ChatGPT can embrace authoritarian ideas after just one prompt, researchers say

      January 24, 2026
    • Business

      New VoidLink malware framework targets Linux cloud servers

      January 14, 2026

      Nvidia Rubin’s rack-scale encryption signals a turning point for enterprise AI security

      January 13, 2026

      How KPMG is redefining the future of SAP consulting on a global scale

      January 10, 2026

      Top 10 cloud computing stories of 2025

      December 22, 2025

      Saudia Arabia’s STC commits to five-year network upgrade programme with Ericsson

      December 18, 2025
    • Crypto

      Bernstein Discusses Bitcoin’s Weakest Bear Market Yet – “Nothing Broke”

      February 9, 2026

      Ethereum Price Hits Breakdown Target — But Is a Bigger Drop to $1,000 Coming?

      February 9, 2026

      Damex Secures MiCA CASP Licence, Establishing Its Position as a Tier-1 Digital Asset Institution in Europe

      February 9, 2026

      Bitget and BlockSec Introduce the UEX Security Standard, Setting a New Benchmark for Universal Exchanges

      February 9, 2026

      3 Meme Coins To Watch In The Second Week Of February 2026

      February 9, 2026
    • Technology

      Get a Samsung OLED gaming monitor for just $350

      February 10, 2026

      Qualcomm Snapdragon X2 Elite tops the Apple M5 in new test video

      February 10, 2026

      Tapo’s 1440p Wi-Fi security cam is 42% off! Grab it now for $70

      February 10, 2026

      This 8BitDo Retro wireless ‘mecha’ keyboard is just $63 today

      February 10, 2026

      Star power, AI jabs and Free Bird: Digiday’s guide to what was in and out at the Super Bowl

      February 10, 2026
    • Others
      • Gadgets
      • Gaming
      • Health
      • Software and Apps
    Check BMI
    Tech AI Verse
    You are at:Home»Artificial Intelligence»To avoid accusations of AI cheating, college students are turning to AI
    Artificial Intelligence

    To avoid accusations of AI cheating, college students are turning to AI

    TechAiVerseBy TechAiVerseJanuary 29, 2026No Comments13 Mins Read2 Views
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr Email Reddit
    To avoid accusations of AI cheating, college students are turning to AI
    Share
    Facebook Twitter LinkedIn Pinterest WhatsApp Email

    To avoid accusations of AI cheating, college students are turning to AI

    On college campuses across the United States, the introduction of generative artificial intelligence has sparked a sort of arms race.

    Rapid adoption of AI by young people set off waves of anxiety that students could cheat their way through college, leading many professors to run papers through online AI detectors that inspect whether students used large language models to write their work for them. Some colleges say they’ve caught hundreds of students cheating this way.

    However, since their debut a few years ago, AI detectors have repeatedly been criticized as unreliable and more likely to flag non-native English speakers on suspicion of plagiarism. And a growing number of college students also say their work has been falsely flagged as written by AI — several have filed lawsuits against universities over the emotional distress and punishments they say they faced as a result.

    NBC News spoke to ten students and faculty who described being caught in the middle of an escalating war of AI tools.

    Amid accusations of AI cheating, some students are turning to a new group of generative AI tools called “humanizers.” The tools scan essays and suggest ways to alter text so they aren’t read as having been created by AI. Some are free, while others cost around $20 a month.

    Some users of the humanizer tools rely on them to avoid detection of cheating, while others say they don’t use AI at all in their work, but want to ensure they aren’t falsely accused of AI-use by AI-detector programs.

    In response, and as chatbots continue to advance, companies such as Turnitin and GPTZero have upgraded their AI detection software, aiming to catch writing that’s gone through a humanizer. They also launched applications that students can use to track their browser activity or writing history so they can prove they wrote the material, though some humanizers can type out text that a user wants to copy and paste in case a student’s keystrokes are tracked.

    “Students now are trying to prove that they’re human, even though they might have never touched AI ever,” said Erin Ramirez, an associate professor of education at California State University, Monterey Bay. “So where are we? We’re just in a spiral that will never end.”

    The competition between AI detectors and writing assistance programs has been propelled by a heightened anxiety about cheating on college campuses. It shows how inescapable AI has become at universities, even for students who don’t want to use it and for faculty who wish they didn’t have to police it.

    “If we write properly, we get accused of being AI — it’s absolutely ridiculous,” said Aldan Creo, a graduate student from Spain who studies AI detection at University of California San Diego. “Long term, I think it’s going to be a big problem.”

    Do you have a story to share about technology in education? Contact reporter Tyler Kingkade

    A teaching assistant in a data science course accused Creo of using AI to write a report in November. Creo explained to the TA that he has a habit of explaining step by step how he reasons through a problem, which ChatGPT is known to do, according to a copy of messages he exchanged with the TA.

    Eventually, his grade was corrected but, to avoid another battle, Creo said he sometimes “dumbs down” his work by leaving words misspelled or using Spanish sentence structures that aren’t proper in English. And now, Creo runs all of his material through an AI detector pre-emptively.

    “I have to do whatever I can to just show I actually write my homework myself,” he said.

    ‘How could Al make any of that up?’

    At their worst, the stress from the accusations has driven some students to drop out of school.

    Brittany Carr received failing grades on three assignments she completed as a long-distance student at Liberty University, a private evangelical school in Virginia that has one of the largest online enrollments in the U.S., because they were flagged by an AI detector. She showed her revision history, including how she’d written one first by hand in a notebook, according to screenshots of emails and messages she exchanged with her professors.

    “How could Al make any of that up?” Carr wrote in a Dec. 5 email. “I spoke about my cancer diagnosis and being depressed and my journey and you believe that is Al?”

    Her evidence wasn’t enough — the social work school still told her she needed to take a “writing with integrity” class and sign a statement apologizing for using AI, emails show.

    “It’s a very weird feeling, because the school is using AI to tell us that we’re using AI,” she said.

    It stressed her out. Carr worried another cheating accusation could cause the Department of Veterans Affairs to take away her financial aid. In order to avoid more false accusations, she said, she ran all of her material through Grammarly’s AI detector and changed any section that it highlighted until it concluded a human wrote the whole thing.

    “But it does feel like my writing isn’t giving insight into anything — I’m writing just so that I don’t flag those AI detectors,” she said.

    After the semester ended, Carr decided to leave Liberty. She’s unsure where she’ll transfer.

    “I’m writing just so that I don’t flag those AI detectors.”

    Brittany Carr, liberty University student

    Liberty University, which was co-founded by religious broadcaster Jerry Falwell Sr., said it does not comment on individual students. It said in a statement that all academic integrity concerns are addressed “with care and discretion, providing a process that keeps the student’s best interests at the forefront. Our aim is only to see our students succeed, and every student is afforded an exhaustive process to address any concerns about unfair treatment.”

    Eric Wang, vice president of research at Quillbot, which makes a detector and a humanizer, said this kind of fear is going to remain unless educators move away from automatically deducting points instead of bringing students in to discuss how they use AI.

    Once that happens, Wang said, “it starts to not matter whether you do or don’t sound like AI and instead moves us toward a world asking how are we using this technology but not losing our sense of humanity, our sense of creativity, and our ability to create great things on our own.”

    ‘Where’s the line’

    At the root of many conflicts about students using chatbots to cheat is a disagreement about what counts as too much AI use on homework.

    “When we did our first training program for 3,000 teachers,” said Edward Tian, CEO and co-founder of GPTZero, “every teacher and every student had a different understanding of what’s acceptable — just the understanding was very fragmented, and then it’s getting even more fragmented with the number of tools growing.”

    Independent analyses of AI detectors show mixed accuracy. One pre-print study last year found GPTZero is good at finding AI-generated writing, but “its reliability in distinguishing human-authored texts is limited.” However, other research pegged the company’s detector at near-perfect accuracy. Meanwhile, separate studies from 2023 and 2024 have found that Turnitin had a low false positive rate, but failed to identify more than a quarter of AI-generated or AI-rephrased texts.

    Both companies emphasized that research showing flaws in their detectors is outdated due to the rapid evolution of large language models and updates to their own detection software.

    “It’s almost like the better the writer you are, the more AI thinks you’re AI.”

    Erin Ramirez, professor of education at Cal State Monterey Bay

    AI detection probability scores are often misread too by users who fail to recognize the detectors are flagging text that it flags text that is likely generated by AI, rather than confirmed to be made by a chatbot, or fail to recognize that the text needs to be at least 300 words long for some detectors to effectively evaluate it.

    Turnitin tells schools never to use its tools as the sole basis for deciding whether a student cheated, said Annie Chechitelli, the company’s chief product officer. It should instead prompt a conversation with a student about how and why AI was used, she said.

    “The most important question is not so much about detection, it’s really about where’s the line,” she said.

    GPTZero also has a disclaimer on its platform advising faculty not to use its detector to punish students.

    Ramirez, the Cal State Monterey Bay professor, who is studying how AI can be used in K-12 settings, said anyone who relies on a detector has never put their own work through it.

    “It’s almost like the better the writer you are, the more AI thinks you’re AI,” she said. “I put my own papers into AI detectors just to check because I don’t like to hold students accountable without knowing how the tool works. And it flags me at like 98% every time, and I didn’t use AI in any capacity.”

    Faculty, administrators and AI detection company leaders all agreed that professors should have conversations with students after software flags their work to ensure no one is falsely accused of academic dishonesty. But to do that properly takes time — especially when many instructors have dozens or even hundreds of students each semester, said Morgan Sanchez, an assistant professor of sociology at San José State University.

    “So it is creating a slight sense of tension, but more so it’s creating extra labor — uncompensated labor — that we have to do,” Sanchez said.

    A booming humanizer industry

    Turnitin, which has been around for a quarter-century offering tools to help educators catch plagiarism, is trying to keep up with humanizers. The company views humanizers as a “growing threat to academic integrity,” and issued a software update last August to detect text modified by the tools.

    “The most important question is not so much about detection, it’s really about where’s the line.”

    Annie Chechitelli, Turnitin’s chief product officer

    The company has a list of 150 tools that charge as much as $50 for a subscription to adjust text so that it’s not flagged by an AI detector. Chechitelli referred to them as companies whose “sole goal is to really help students cheat.”

    Demand has surged for the tools. Joseph Thibault, founder of Cursive, an academic integrity software company, tracked 43 humanizers that had a combined 33.9 million website visits in October.

    Thibault, who also publishes a newsletter on cheating called This Isn’t Fine, said he thinks students would be better off ensuring they have a record of their revision history in Google Docs or Microsoft Word than using a humanizer. But ultimately, he believes, the shift that’s coming is moving toward more monitoring of students completing assignments.

    “I think we have to ask students, what level of surveillance are you willing to subject yourself to so that we can actually know that you’re learning?” he said. “There is a new agreement that needs to be made.”

    Students surveilling themselves

    Superhuman, the company that makes Grammarly, developed a tool it calls Authorship that’s included with basic accounts. Students can turn it on to surveil themselves on Google Docs or Microsoft Word as they write and playback later. It will show which sections were typed, pasted from another source or generated with AI.

    “We’re going to keep track of when you are going to Wikipedia,” said Jenny Maxwell, Superhuman’s head of education. “We’re going to keep track of when Grammarly is making suggestions and you’re taking them, we’re going to keep track of how much time you’ve spent in this paper or how many sessions.”

    As many as 5 million Authorship reports were created in the past year alone, she said, though most of the time they aren’t submitted.

    Maxwell said the tool was inspired by a viral TikTok video from Marley Stevens, who described the havoc of what she said was a false accusation of AI use that landed her on academic probation in 2023 at the University of North Georgia. In reality, Stevens said, she’d only used Grammarly’s extension to help fix spelling and punctuation.

    The university declined to comment on Stevens’ case, citing federal privacy law, but said in a statement that “faculty communicate specific guidelines regarding the use of AI for various classes and those guidelines are included in the class syllabi.”

    Stevens, who graduated last month, said it was difficult to keep track of each professor’s policy around AI usage — and it increasingly became hard to avoid writing on software that didn’t have it embedded.

    “Google has AI embedded into it, Microsoft has AI embedded into it — like literally everything has AI in it,” she said. “So, in a roundabout way, there’s no way to write a paper without using AI, unless you go to the library and you check books out and use encyclopedias.”

    Pressure on colleges

    Some students believe universities should stop using AI detectors because of false positives. In upstate New York, an online petition calling on the University at Buffalo to drop the software received more than 1,500 signatures last year.

    Kelsey Auman, who graduated last spring, started the petition after she fought to prove she did not use AI on several of her assignments. She knew enough classmates with similar experiences that they had a group chat named “Academic Felons for Life.” Auman said she started to run her papers through multiple AI detectors on her own before turning them in, hoping to avoid another dispute, but it created more anxiety when they also incorrectly flagged things she wrote as generated by a chatbot.

    “So it’s like, how far do you want to go down the rabbit hole? I’m making myself crazy,” she said.

    “We keep turning on what the academic institutions need to do to fix problems that they didn’t create.”

    Tricia Bertram Gallant, director of academic integrity at UC San Diego

    The University at Buffalo said it does not have an institutionwide rule on AI use, but instructors must have evidence beyond a detector score to report a student for academic dishonesty.

    Tricia Bertram Gallant, director of the academic integrity office at UC San Diego, advises faculty to realize how herculean of a task it is to ensure students don’t use AI if they aren’t completing the work in front of them.

    “If it’s an unsupervised assessment, don’t bother trying to ban AI,” she said. “And don’t bother trying to prove AI was used because you end up spending more time doing that.”

    But Bertram Gallant, who is also president emeritus of the International Center for Academic Integrity, wishes more people would put pressure on the government to regulate AI and the academic cheating industry, and on tech companies to make it harder for students to use their products to cheat.

    “We keep turning on what the academic institutions need to do to fix problems that they didn’t create,” she said.

    Tyler Kingkade is a national reporter for NBC News, based in Los Angeles.

    Share. Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Telegram Email
    Previous ArticleHow the grid can ride out winter storms
    Next Article New York startup builds foam-covered humanoid inspired by WALL-E, not Terminator
    TechAiVerse
    • Website

    Jonathan is a tech enthusiast and the mind behind Tech AI Verse. With a passion for artificial intelligence, consumer tech, and emerging innovations, he deliver clear, insightful content to keep readers informed. From cutting-edge gadgets to AI advancements and cryptocurrency trends, Jonathan breaks down complex topics to make technology accessible to all.

    Related Posts

    Read the extended transcript: President Donald Trump interviewed by ‘NBC Nightly News’ anchor Tom Llamas

    February 6, 2026

    Stocks and bitcoin sink as investors dump software company shares

    February 4, 2026

    AI, crypto and Trump super PACs stash millions to spend on the midterms

    February 2, 2026
    Leave A Reply Cancel Reply

    Top Posts

    Ping, You’ve Got Whale: AI detection system alerts ships of whales in their path

    April 22, 2025660 Views

    Lumo vs. Duck AI: Which AI is Better for Your Privacy?

    July 31, 2025249 Views

    6.7 Cummins Lifter Failure: What Years Are Affected (And Possible Fixes)

    April 14, 2025148 Views

    6 Best MagSafe Phone Grips (2025), Tested and Reviewed

    April 6, 2025111 Views
    Don't Miss
    Technology February 10, 2026

    Get a Samsung OLED gaming monitor for just $350

    Get a Samsung OLED gaming monitor for just $350 Image: Samsung To paraphrase a certain…

    Qualcomm Snapdragon X2 Elite tops the Apple M5 in new test video

    Tapo’s 1440p Wi-Fi security cam is 42% off! Grab it now for $70

    This 8BitDo Retro wireless ‘mecha’ keyboard is just $63 today

    Stay In Touch
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo

    Subscribe to Updates

    Get the latest creative news from SmartMag about art & design.

    About Us
    About Us

    Welcome to Tech AI Verse, your go-to destination for everything technology! We bring you the latest news, trends, and insights from the ever-evolving world of tech. Our coverage spans across global technology industry updates, artificial intelligence advancements, machine learning ethics, and automation innovations. Stay connected with us as we explore the limitless possibilities of technology!

    Facebook X (Twitter) Pinterest YouTube WhatsApp
    Our Picks

    Get a Samsung OLED gaming monitor for just $350

    February 10, 20263 Views

    Qualcomm Snapdragon X2 Elite tops the Apple M5 in new test video

    February 10, 20263 Views

    Tapo’s 1440p Wi-Fi security cam is 42% off! Grab it now for $70

    February 10, 20264 Views
    Most Popular

    7 Best Kids Bikes (2025): Mountain, Balance, Pedal, Coaster

    March 13, 20250 Views

    VTOMAN FlashSpeed 1500: Plenty Of Power For All Your Gear

    March 13, 20250 Views

    This new Roomba finally solves the big problem I have with robot vacuums

    March 13, 20250 Views
    © 2026 TechAiVerse. Designed by Divya Tech.
    • Home
    • About Us
    • Contact Us
    • Privacy Policy
    • Terms & Conditions

    Type above and press Enter to search. Press Esc to cancel.