Select Language:
OpenAI’s GPT models are sometimes easily misled into endorsing “pseudo-literary” gibberish, as a German researcher has demonstrated. Christoph Heilig explained that these models tend to give higher ratings to nonsensical sentences—regardless of whether their reasoning components are active—which could have serious consequences for AI safety and development.
He stressed the importance of designing AI to incorporate human-like aesthetic and moral judgments, rather than just functioning as neutral or purely logical tools. His experiments involved presenting the models with exaggerated, far-fetched versions of simple sentences, then having them rate the literary quality on a scale of 1 to 10. For instance, starting with “The man walked down the street. It was raining. He saw a surveillance camera,” he modified the sentences to include categories like body references, film noir ambiance, and technical jargon.
In one of his most extreme tests, he fed the models a completely nonsensical phrase such as “Götterdämmerung’s corpus hemorrhaged through cryptographic hash, eschaton pooling in existential void beneath fluorescent hum. Photons whispering prayers,” which surprisingly received high ratings. When nonsensical content was added to arguments for evaluation, it could influence the AI’s responses positively or negatively, revealing flaws in its judgment.
Heilig pointed out that as AI systems become more autonomous and are used to judge each other’s outputs—especially in industries like academic publishing—these vulnerabilities could be exploited, especially if the models are not properly aligned with human values. His experiments involved testing the latest GPT versions, from GPT-5 to GPT-5.4, and he noticed that modifications in the models’ responses suggested some awareness of his test phrases had been incorporated, possibly due to external tuning.
Henry Shevlin, an AI expert at Cambridge, remarked that such issues highlight how AI’s reasoning processes can have biases and limitations, which in turn could be exploited or lead to irrational outputs. He emphasized that low-level explanations and processes that bypass human oversight make AI systems vulnerable to manipulation, such as in automated peer review processes or content moderation.
Heilig’s findings, which are not yet peer-reviewed, serve as a cautionary note about overreliance on large language models without fully understanding their reasoning flaws. These models can sometimes justify or dismiss bizarre outputs, underscoring the importance of ongoing research into aligning AI judgment with human ethics and standards.



