Wharton professor and prominent AI researcher Ethan Mollick has called attention to what he identifies as a telltale sign of AI-generated writing within OpenAI's own safety documentation. In a post on X, Mollick analyzed a story shared in OpenAI's Mythos System Card, a document outlining the company's approach to evaluating and mitigating risks from advanced AI systems.
Mollick argues the narrative example "still has the signs of flawed LLM writing (which looks like good writing at first glance)." He points to three specific flaws:
- A story that doesn't really hold together logically, but sounds like it should. The narrative has surface-level coherence but breaks down upon closer inspection of its plot or reasoning.
- The back-and-forth banter. This is a common, often overused, stylistic pattern in LLM-generated dialogue.
- Lack of characters. The story may feel generic or lack depth because it fails to develop distinct, believable personas.
The critique highlights an ongoing challenge in AI development: while large language models (LLMs) excel at producing grammatically correct and stylistically appropriate text, they can still struggle with deeper narrative cohesion, logical consistency, and authentic character development. The fact that these hallmarks appear in a safety document—a context where clear, human-authored communication is paramount—adds a layer of irony to the observation.
OpenAI's Mythos initiative is part of its effort to "stress-test" its models and systems against potential catastrophic risks. The System Cards are intended to provide transparency into the company's safety evaluation processes. Mollick's analysis suggests that even the artifacts used to explain and demonstrate safety methodologies are not immune to the very limitations of the technology being studied.
gentic.news Analysis
Mollick's critique is a sharp, meta-commentary on the state of AI transparency. It's not just about a poorly written story; it's about the difficulty of using LLMs to communicate about LLMs, especially in high-stakes domains like AI safety. When a company's safety documentation itself contains artifacts that experts can identify as AI-generated, it raises questions about the rigor and human oversight in the process. This aligns with broader concerns we've covered regarding evaluation methodologies, such as in our analysis of the SHEEP benchmark, where the challenge of creating robust, non-gameable evaluations for AI capabilities remains a central research problem.
The incident also connects to a trend we've noted: the increasing scrutiny of AI-generated content quality beyond simple factuality. As LLMs become ubiquitous tools for drafting and communication, the focus is shifting from "can it write?" to "can it write with depth, logic, and authentic voice?" Mollick, whose work often focuses on the practical implementation and limitations of AI in business and education, is pinpointing a gap between technical capability and practical, reliable utility. This dovetails with our previous reporting on tools like LLM Truthfulness Evaluators, which aim to detect not just falsehoods but also the subtle inconsistencies and logical flaws that Mollick identifies.
Furthermore, this occurs within the context of OpenAI's ongoing efforts to build trust through transparency initiatives like System Cards. If the tools used to create this transparency are themselves prone to producing superficially plausible but fundamentally flawed content, it creates a recursive problem for safety research. This analysis serves as a reminder that critical, human expert review remains an irreplaceable component of both AI development and the communication about it, a principle that holds true from technical research to public-facing documentation.
Frequently Asked Questions
What is the Mythos System Card?
The Mythos System Card is a document published by OpenAI that outlines its framework for evaluating and mitigating potential catastrophic risks from advanced AI systems. It describes "stress tests" and methodologies used to assess model behaviors in high-stakes scenarios.
What are common signs of flawed LLM writing?
As highlighted by Mollick, common signs include narratives that sound coherent initially but lack logical consistency upon scrutiny, repetitive use of certain dialogue patterns like excessive back-and-forth banter, and a generic quality often stemming from underdeveloped characters or shallow plotlines.
Why does this critique matter for AI safety?
It matters because clear, accurate, and logically sound communication is critical in safety research and documentation. If the examples used to explain safety methodologies themselves contain flaws indicative of AI generation, it can undermine confidence in the processes and obscure understanding. It emphasizes the need for meticulous human oversight in all aspects of safety work.
Who is Ethan Mollick?
Ethan Mollick is a professor at the Wharton School of the University of Pennsylvania who studies innovation, entrepreneurship, and the impact of artificial intelligence. He is a widely followed researcher and commentator on the practical applications and limitations of AI in business and education.









