A Substack post by Farida Khalaf argues AI models have no moat, using the Clawbot/OpenClaw story as proof. The core thesis — models are interchangeable commodities — is correct. I build on top of LLMs and have swapped models three times with minimal impact on results.
But the article itself is clearly AI-generated, and it's full of errors that prove the opposite of what the author intended.
The video: The article includes a 7-second animated explainer. Pause it and you find Anthropic spelled as "Fathropic," Claude as "Clac#," OpenAI as "OpenAll," and a notepad reading "Cluly fol Slopball!" The article's own $300B valuation claim shows up as "$30B" in the video. There's no way the author watched this before publishing...
The timeline is fabricated: The article claims OpenAI "panic-shipped" GPT-5.2-Codex on Feb 5 in response to Clawbot going viral on Jan 27. Except GPT-5.2-Codex launched on January 14 — two weeks before Clawbot. What actually launched Feb 5 was GPT-5.3-Codex. The article got the model name wrong.
The selloff attribution is wrong: The article blames the February tech selloff on Clawbot proving commoditization. Bloomberg, Fortune, and CNBC all attribute it to Anthropic's Cowork legal automation plugin — investors worried about AI replacing IT services work. RELX crashed 13%, Nifty IT fell 19%. None of it was about Clawbot.
The financials are stale: cites Anthropic at $183B and projects a 40-60% IPO haircut. By publication date, Anthropic's term sheet was at $350B. The round closed at $380B four days later.
The irony: an AI-generated article about AI having no moat is the best evidence that AI still needs humans checking the work. The models assembled a convincing shape of market analysis without verifying whether any of it holds together.
I wrote a full fact-check with sources here: An AI Wrote About AI's Death. Nobody Checked.
Disclosure: I used AI tools for research and drafting. Every claim was verified against primary sources. Every sentence was reviewed before publishing. That's the point.
[link] [comments]