Large language models (LLMs)—the technology that powers popular AI chatbots like ChatGPT and Google’s Gemini—repeatedly made irrational, high-risk betting decisions when placed in simulated gambling environments, according to the results of a recent study. Given more freedom, the models often escalated their bets until they lost everything, mimicking the behavior of human gambling addicts.



lulzwut? LLMs aren’t internalizing jack shit. If they exhibit a bias, it’s because of how they were trained. A quick theory would be that the interwebs is packed to the brim with stories of “all in” behaviors intermixed with real strategy, fiction or otherwise. I speculate that there are more stories available in forums of people winning doing stupid shit then there are of people loosing because of stupid shit.
They exhibit human bias because they were trained on human data. If I told the LLM to only make strict probability based decisions favoring safety (and it didn’t “forget” context and ignored any kind of “reasoning”), the odds might be in its favor.
Sorry, I will not read the study because of that one sentence in its summary.
It wouldn’t be able to do it. LLMs are bad at math. Computers can’t do math very well with only an auto-complete/spell-check program.
Thus, this article and many similar.
Humans can’t stop anthropomorphizing AI. An easier explanation is that LLMs are reward oriented based on successful results and have no risk aversion or fatigue, so they can just keep doing something until they ‘win big’.