AI detectors are pattern-based AIs, just like the LLMs they are evaluating. They are trained on examples of AI-generated and human-generated text, and make their best probabilistic guess about whether a given piece of text is more like one than the other.
There are two main categories by which the writing is evaluated: perplexity and burstiness.
Perplexity is a measure of how surprising the text is, that is, how different it is from what is likely to be the next predicted word in a sequence. For example, if a sentence starts, "It's raining out, so take your . . .", the word "umbrella" rates as low perplexity, while the word "monkey" is high perplexity.
Burstiness is a measure of the variability of sentence length and structure. Like this paragraph. The sentence lengths vary; the structure does as well. That's burstiness.
In general, AI creates text with lower perplexity and burstiness than humans do. AI generators will therefore rate text with low perplexity and burstiness as more likely to be AI-created.
The short answer is: not very well.
Now that we know how AI detectors work, several problems are immediately obvious which can lead to both false positives and false negatives:
Learn more:
Why AI writing detectors don’t work - Benj Edwards, Ars Technica
OpenAI confirms that AI writing detectors don’t work - Benj Edwards, Ars Technica
We tested a new ChatGPT-detector for teachers. It flagged an innocent student - Geoffrey A. Fowler,The Washington Post
AI-Detectors Biased Against Non-Native English Writers - Andrew Myers, Stanford Institute for Human-Centered AI
The short answer is: yes, but with caution. An AI detector can be another piece of data in a larger picture when a student is suspected of using AI, but should never be the sole piece of data relied upon. An AI detectors cannot prove AI use, so treat the AI detector as a diagnostic tool, not a decision maker.
Learn more:
Be Your Own Best AI Detector - Justin Marquis, Gonzaga IDD