Burger menu

Human vs. Machine Performance in Generated Text Detection

Ability to tell AI-content from human writing is crucial, as it prevents bias, academic dishonesty, fake news and defamation, and various types of fraud.

Why is AI-Generated Content Detection Important?

ChatGPT4 giving a fictional advice on how to breathe under the water

Content produced by Large Language Models (LLMs) lacks human expertise and understanding, which often makes it unreliable and even dangerous. A GenAI model, ChatGPT is known to provide false and harmful medical advice, discriminative statements, incorrect legal information, and so on. 

Besides, it can be used for making fake news, writing academic content, generating flawed computer code, or employing it as a weapon for scam activities and smear campaigns. Therefore, the ability to detect generated text content is of key importance.   

NYC Laws 144 incorrectly interpreted by ChatGPT4

Datasets Used for Generated Text Detection

Several datasets and initiatives exist to train detection models, evaluate LLM-based responses in terms of quality and safety, as well as help real people sharpen their ability to detect LLM-generated texts. 

  1. Real or Fake Text (RoFT)
RoFT website

Real or Fake Text (RoFT) is a platform, where guests are challenged to locate the boundary when human-written passages are replaced with the machine-created content in a single text. With this gamified process, authors managed to reveal that only 15,8% of all participants can correctly identify machine-content so far.

  1. Human ChatGPT Comparison Corpus (HC3)
Use frequency of speech part in human and AI’s answers

The dataset contains two major data corpuses:

  • Human-based. It features public question-answering datasets and Wiki sources.
  • Generated. Answers are provided by ChatGPT-3.5, with additional requests and conditions.

Turing test and Helpfulness value were featured as primary metrics for separating human answers from those by AI. As it turned out, ChatGPT’s answers have a pristine structure, are longer in size, and may contain fabricated facts. Human answers often shift to similar or distant topics, contain informal speech and emotional expressions, and also may seem more subjective.   

ChatGPT4 giving false info on normal body temperature of humans
  1. IDMGSP Dataset
The process of generating a scientific paper based on training input by real human scientists

The dataset includes human-authored scientific works, AI-generated abstracts, introductions, and conclusions, as well as writings co-authored by humans and artificial intelligence. The purpose of the study is to train a classifier that can detect generated articles on science with nearly perfect accuracy.

The Logistic Regression (LR) and Random Forest (RF) classifiers showed a 90% accuracy at distinguishing human and AI content when tried against the test dataset. Meanwhile, the RoBERTa model demonstrated a 75% accuracy at spotting papers co-authored by humans and AI. However, the fake materials were generated with GPT-3 and SCIgen, which lose in comparison to more advanced GPT-4 and upcoming ChatGPT-5.

Subtle racial stereotypes observed in LLMs

Human Detection vs. Machine Detection Experiments

Human prowess at detecting synthesised texts in comparison to AI-detectors has been put to test in a number of studies. 

  1. Experiments with Scientific Writing

A group of experiments were hosted with a focus on scientific papers. The results showed that human reviewers pay attention to the writing style, which is artfully mimicked by the AI. However, generated papers aren’t always clear on goals and methods of the research, which separates them from human authors. At the same time, human-written papers can be even more formulaic, as observed by volunteers.

Another study shows that humans are unable to differentiate AI and genuine writing whatsoever, while specialised models can detect at least a portion of them. 

  1. Experiments with Online Reviews

As observed by the Yale School of Management professor Balazs Kovacs, both humans and AI tend to mistake them for feedback written by actual people. Interestingly, readers see the AI-reviews as “superhuman”, which implies that they deserve more trust than real person’s comments.

It is speculated that believable AI-generated reviews can be used for manipulating market and consumers, launching smear campaigns against rivals, or stirring artificial excitement around a specific product.     

Try our AI Text Detector

Avatar Antispoofing

1 Followers

Editors at Antispoofing Wiki thoroughly review all featured materials before publishing to ensure accuracy and relevance.

Article contents

Hide

More from AI Generated Content

avatar Antispoofing Text Watermarking: Definition, General Approaches, Application

What Is Text Watermarking? Text watermarking is a technology, which allows copyright protection with the help of special metadata inserted…

avatar Antispoofing Bias against Non-Native English Writers in AI-Detectors

The problem of bias — related to age, ethnicity, or race — has been discussed in relation to AI since…

avatar Antispoofing Plagiarism in AI-Generated Texts

Can AI-Generated Text Be Detected by Plagiarism Checkers? Plagiarism concerns in regard to GenAI have been raised repeatedly, especially since…

avatar Antispoofing Methods of Natural Language Watermarking

What Are the Main Methods of Natural Language Watermarking (NLW)? NLW is a technique that protects a writing with a…