Chieh-Ping (aka CheRocks) Chen • 29 days ago
When AI Judges AI, Who Judges the Judges?
On March 18, ICML — one of machine learning's most important academic conferences — officially revealed the results of an experiment that had been unfolding since January. They had embedded hidden prompt injection instructions inside review PDFs. A reviewer first discovered the hidden text on February 14, sparking a firestorm on Reddit's r/MachineLearning. The final count: 795 review violations across 398 reviewers, and 497 papers retracted — not because the research was flawed, but because the reviewers used AI to evaluate AI research without disclosure.
The method was simple: plant two rare phrases from a 170,000-word dictionary into each PDF in a way invisible to humans but readable by AI. If both phrases appeared in the review, the reviewer had likely copy-pasted the paper into an LLM and submitted the output wholesale.
Same technique, different intent: this is structurally identical to the prompt injection attacks that Microsoft documented across commercial websites — hidden instructions designed to manipulate AI behavior without human awareness.
ICML used it to catch cheaters. Attackers use it to manipulate agents. The technique is the same. The only difference is who's holding the knife.
Why this matters beyond academia:
The conditions that made ICML vulnerable are not unique to academic peer review. They exist anywhere evaluation volume outpaces human capacity: enterprise compliance reviews, grant evaluations, hiring pipelines, content moderation — and yes, hackathon judging at scale.
The pattern is always the same: high volume, limited time, heavy cognitive load, and no mechanism to verify whether evaluation was performed by a human, by AI, or by a human rubber-stamping AI output.
ICML at least had a detection mechanism — prompt injection as a canary. Most evaluation systems have none. There is typically no audit trail for how scores are assigned, no record of what each evaluator actually reviewed, and no way for the evaluated party to verify that their submission received human attention.
This is the exact problem I've been working on with Project RE, submitted to this hackathon. RE approaches it from the governance and audit layer — not by banning AI from evaluation, but by ensuring that whatever process is used is recorded, auditable, and attributable.
RE's thesis: The future is not an era of technical debt. It is an era of decision debt and cognitive debt.
ICML's reviewers accumulated cognitive debt — they outsourced judgment to AI and could no longer explain why they gave the scores they gave. The 497 retracted papers are the interest payment on that debt.
The question is not whether AI should be used in evaluation. The question is whether the use of AI in evaluation is recorded, auditable, and attributable. Right now, across academia, across enterprise, across most evaluation systems — the answer is almost universally no.
That's the gap RE sits in. Here's the project if you want to take a look:
https://devpost.com/software/project-re-the-governance-protocol
— Che, Solo developer, Project RE, Taipei Taiwan
Log in or sign up for Devpost to join the conversation.

0 comments