LLMs are like slot machines, in that an incorrect answer (the slot machine eating your dollar) is unremarkable, while the LLM solving a problem (a jackpot) is amazing, and the latter stands out in your memory, causing you to overestimate the reliability of LLMs.
blog.glyph.im/2025/08/futzing-…
Deciphering Glyph :: The Futzing Fraction
Deciphering Glyph, the blog of Glyph Lefkowitz.blog.glyph.im
Reg Braithwaite 🍓
in reply to Cory Doctorow • • •I have argued that LLMs are like slot machines in another way: The business model. When you are paying by “the pull of the handle,” the vendor's incentive is not to solve your problem with a single pull, but to give the appearance of progress towards solving your problem.
Exactly the same as games companies that engineer their games to keep users buying coins, staring at the screen for eighteen hours a day. Very different model than selling tools or renting services.
Cory Doctorow reshared this.
Cory Doctorow
in reply to Reg Braithwaite 🍓 • • •ClaraBlackInk
in reply to Cory Doctorow • • •When the image generators came out, it reminded me of skilling up in video games. The way you can become an "expert" cook despite still being unable to nuke a pizza pocket irl without having different temperature zones. It felt like that and I could see how people would begin to see themselves as "artists" because it would feel like an accomplishment in a gamified way.
Or, well: twit.social/@JustinH/115010462…
Justin ⏚ (@JustinH@twit.social)
TWiT.socialBinary Large Octopus
in reply to Cory Doctorow • • •Oliver D. Reithmaier
in reply to Cory Doctorow • • •Krampus 🌰
Unknown parent • • •Jean-Baptiste "JBQ" Quéru
in reply to Cory Doctorow • • •And, just like video poker, you might think that, with enough practice, you might come up with an optimal strategy, but you're still behind overall, and that strategy is only optimal in a very narrow range of conditions.
(former video poker player here, where I had to come up with a twisted definition of "winning" in order to come out ahead)
Azuaron
in reply to Cory Doctorow • • •WheelTryingToGetJacked
in reply to Azuaron • • •MarjorieR
in reply to WheelTryingToGetJacked • • •By comparison 24% right is not exactly PhD level.
wakko
in reply to Cory Doctorow • • •Better analogy - LLMs are like teenagers. They might know a couple facts through rote memorization, but they frequently make spurious extrapolations from their limited knowledge base and need frequent course-corrections from an experienced adult.
When the adults aren't steering the teenagers effectively, the teens tend to think maladaptive behavior is okay... just like LLMs.
Garbage in, garbage out is the same in both humans and machines.
Gary
in reply to Cory Doctorow • • •@paulywill this is so completely untrue. Have you used one recently? If they were as unreliable as this, then the industry wouldn’t be what it is.
Is it a wasteful industry in a bubble? Yes, but just saying lies about it doesn’t make it true.
emin
in reply to Cory Doctorow • • •Pascal 🇪🇺 🇺🇦
in reply to Cory Doctorow • • •Like every information source, it should not be trusted blindly.
Orion Ussner kidder
in reply to Cory Doctorow • • •Related:
I was (ugh) using ChatGPT to see how much my students might be able to cheat on a set of quizzes I'm going to give them, and I ended up in so many words telling ChatGPT to "admit" that it's just spicy autocomplete. I had this feeling of accomplishment that I "got it" to say it.
And then I remembered: it *merely* tells me what I want to hear, just like someone who wants a cybergirlfriend, just like someone asking for stock tips, just like someone who wants it to be a god.
Azuaron
Unknown parent • • •tinydoctor
Unknown parent • • •SpaceLifeForm
in reply to Cory Doctorow • • •Cory Doctorow reshared this.
Christian Gudrian
in reply to Cory Doctorow • • •This resonates perfectly with what @dgerard wrote:
pivot-to-ai.com/2025/06/05/gen…
Generative AI runs on gambling addiction — just one more prompt, bro!
Pivot to AIJoshua Laase
in reply to Cory Doctorow • • •