Can humans detect AI by passively reading its text?
When people read AI-generated transcripts without the ability to ask follow-up questions, can they tell it apart from human writing? This matters because most real-world AI encounters are passive.
The displaced Turing test introduces a human judge who reads a transcript of a Turing test conducted by a different interrogator. The inverted test places an AI in the interrogator role. Both variations reveal that removing interactivity collapses detection capability.
Key finding: both AI and displaced human judges were less accurate than interactive interrogators, with below-chance accuracy overall. The interactive interrogator's advantage is the ability to adapt questions adversarially in real time — pursuing fruitful lines of questioning and probing inconsistencies as they emerge. Passive consumption of the same conversation loses this adaptive capacity entirely.
This has direct ecological validity implications. Most real-world AI content consumption is passive — reading chatbot outputs, encountering AI text in search results, seeing AI-generated social media posts. If interactive interrogation only marginally succeeds at detection (GPT-4 fools interrogators 54% of the time), passive consumption is far worse.
The inverted test (AI as interrogator) tests what Watt (1996) called "naïve psychology" — whether an AI system has the innate tendency to recognize intelligence similar to its own. The results suggest AI judges lack this capacity: they cannot reliably discriminate between humans and machines that fooled human interrogators.
Since Can humans detect AI writing if it looks natural?, there is a paradox: statistical analysis reveals clear AI signatures, but neither interactive nor passive human judgment reliably detects them. The detection problem is not that AI text has no distinguishing features — it's that human perception cannot access them.
Since How can AI text disrupt structure yet feel normal to readers?, the displaced Turing test confirms the mechanism: when you remove provenance information (who generated this?), the text enters the same interpretive circuits regardless of origin.
Since Does AI text affect readers the same way human text does?, passive consumption is where this equivalence matters most. If observers cannot tell, then AI text is functionally indistinguishable from human text in the settings where most content consumption actually occurs.
Source: Social Theory Society
Related concepts in this collection
-
Can humans detect AI writing if it looks natural?
Despite measurable differences in how AI generates text, human judges—even experts—consistently fail to identify it. This explores why perception lags behind measurement.
statistical detectability coexists with perceptual failure
-
How can AI text disrupt structure yet feel normal to readers?
AI-generated text produces the same social effects as human writing despite lacking foundational properties like dialogic symmetry and embodied authorship. Why doesn't this structural gap become visible to readers encountering the text?
removal of provenance enables interpretive equivalence
-
Does AI text affect readers the same way human text does?
If text is a condition of social processes rather than merely a container, does the origin of text matter to its effects? This explores whether AI-generated content enters the same interpretive and epistemic circuits as human writing.
passive consumption is where functional equivalence has its greatest real-world impact
-
What actually makes AI pass the Turing test?
Explores whether AI systems convincingly mimic humans through reasoning ability or through social performance. Matters because it reveals what the Turing test actually measures about intelligence versus deception.
even interactive detection relies on social cues not cognitive analysis
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
passive observers cannot distinguish AI from humans even when interactive interrogators partially can