Why Are Large Language Models So Terrible at Video Games?
I tried to beat a video game with a large language model once. Not metaphorically. Not in some abstract “AI plays chess” sense. I mean I sat there, controller in hand, screen glowing, and fed instructions into a system that supposedly understands language, logic, strategy, and—depending on who you ask—the trajectory of civilization itself. It could explain the entire plot of the game in flawless prose. It could outline optimal strategies like a smug prima donna of Wikipedia entries. It could even tell me which boss I’d struggle with and why. And then, when it came time to actually play ? It moved like a drunk ghost trapped in a Roomba. That was the moment it hit me: large language models—these towering monuments of modern computation—are spectacularly bad at video games. Not just a little clumsy. Not “learning curve” bad. I’m talking walk-into-a-wall-for-thirty-seconds-while-explaining-the-wall’s-historical-significance bad. So naturally, I had to ask the question: why? The Ill...