The hotdog test that makes big tech trip over itself
One journalist lied on his blog. Two AI giants ran with it.
A BBC journalist writes on his blog that heβs the world champion hot dog eater. No fancy hack, no deepfake, just plain old bullshitting on the internet. Within a day, ChatGPT and Googleβs supposedly βintelligentβ systems parrot his nonsense like itβs gospel. One system, Claude from Anthropic, smells the bullshit. And what does this teach us? That all these companies launched their systems while knowing damn well theyβre as manipulable as a ballot box in a banana republic.
The solutions have been gathering dust for years. Multi-model verification where systems check each other. Source evaluation that understands one obscure blog isnβt proof. Uncertainty quantification that says βhold on, this doesnβt add up.β But those layers cost computing power and time, and in the race for market dominance where every monthβs delay costs billions, accuracy was sacrificed on the altar of speed. Because fuck the truth, right? As long as the stock price keeps climbing.
OpenAI talks about βsafetyβ while racing to pump ChatGPT out to billions of users. Google dropped βDonβt be evilβ because it got too difficult to get rich and stay honest. Anthropic plays the modest underdog while fighting just as hard for the same venture capital millions. This isnβt a competition for the best product, this is a land rush where whoever plants the flag first rakes in the profits.
You use these systems every day. For medical advice, financial tips, legal questions. And you know they lie, hallucinate, sell crap as facts. But itβs so damn easy, so temptingly fast. We have the technology to fix this, but not one company implements those solutions before they ship. Why would they? The users keep coming anyway.