Stanford study reveals concerns about bias in AI detectors for non-native english writing

The experiment revealed that these AI detectors were prone to misidentifying writing as AI-generated for non-native speakers, even when it was not the case. This bias could have serious implications for students’ academic integrity and may lead to unwarranted accusations of cheating.

Share the post
Photo via Pexels

Research conducted by Stanford computer scientists raises significant concerns about the bias and accuracy of AI detectors when evaluating writing by non-native English speakers.

The experiment revealed that these AI detectors were prone to misidentifying writing as AI-generated for non-native speakers, even when it was not the case. This bias could have serious implications for students’ academic integrity and may lead to unwarranted accusations of cheating.

The study underscores the need for continuous improvement and evaluation of AI tools to ensure their fairness and reliability, especially in educational settings with diverse student populations.

The issue of AI detectors flagging writing as AI-generated based on predictable word choice and simpler sentence structures can disproportionately affect non-native English speakers. Many non-native speakers may naturally use simpler language and sentence structures due to their proficiency level, which could lead to false positives when AI detectors misinterpret these traits as characteristics of AI-generated content.

AI models like ChatGPT learn from the data they are trained on, which includes a wide range of language patterns and writing styles. If the training data contains more examples of simpler language usage, the AI model might develop a bias toward generating simpler sentences. This leads to false positives when AI detectors interpret non-native English speakers’ writing as AI-generated due to its simplicity.

Weixin Liang, one of the Stanford study’s authors, was initially doubtful of claims of high accuracy from AI detectors. He wanted to scrutinize how these tools perform for students with linguistic backgrounds like his.

“The design of many GPT detectors inherently discriminates against non-native authors, particularly those exhibiting restricted linguistic diversity and word choice,” Liang said in an email.

Certain international students perceive added risks. Educational institutions commonly caution their international students that allegations of academic misconduct could result in suspension or expulsion, jeopardizing their visa status, making the prospect of deportation a genuine concern.

Shyam Sharma, an associate professor at Stony Brook University, is working on a book about the U.S. education system’s treatment of international students. He believes that universities often neglect to provide adequate support for this group on campus, and professors may not fully grasp their distinct challenges. Sharma views the persistent use of flawed AI detectors as an illustration of how institutions overlook the needs of international students in the country.

Turnitin made a software update in June that permits institutions to deactivate the AI writing indicator, which means the software will still evaluate writing for AI but won’t reveal its assessment to instructors. As of the end of July, only two percent of Turnitin’s customer institutions had opted for this choice, as reported by the company.

In response to low accuracy, OpenAI, Quill.org, and CommonLit ceased their AI detectors’ operations in July. They cited the complexity of generative AI tools as a challenge for detection. Meanwhile, Turnitin has maintained its stance on high accuracy despite these concerns.

Nathan Yasis

Nathan Yasis

Nathan studied information technology and secondary education in college. He dabbled in and taught creative writing and research to high school students for three years before settling in as a digital journalist.

banner place

What to read next...
Nathan Yasis

Nathan Yasis

Nathan studied information technology and secondary education in college. He dabbled in and taught creative writing and research to high school students for three years before settling in as a digital journalist.