Podcast Banner

Podcasts

Paul, Weiss Waking Up With AI

Understanding Grokking

On this week’s episode of “Paul, Weiss Waking Up With AI,” Katherine Forrest explores “grokking,” an emergent learning phenomenon where AI models continue to learn and begin to generalize well after training appears complete.

Stream here or subscribe on your
preferred podcast app:

Episode Transcript

Katherine Forrest: Hello, everyone, and welcome back to “Paul, Weiss Waking Up With AI.” I’m Katherine Forrest, and you’ve got me flying solo again. Anna will be back and she’s going to, I’m sure, realize I’ve gone rogue and have done all kinds of episodes in her absence that she’ll be like, “what? You did X, you did Y?” as I take advantage of this time, but she will be back. And I wanted to sort of set the scene for you. And then we’re going to talk about something today called grokking, G-R-O-K-K-I-N-G. But before we get there, I want to set the scene for you today because I know I talked or spoke a number of times during the summer about being in Maine and all of the extraordinary coffees. And then I’ve also talked about my place up in Woodstock, or maybe I didn’t say it was in Woodstock. I said it was upstate New York, but actually it’s Woodstock. But it’s just extraordinary today, oh my goodness. It is so amazingly beautiful outside. It’s warm, the trees are starting to turn. It’s just incredible. So I’m up here again trying to push through the remainder of this book that I’m finishing up called, “Of Another Mind.” And I’ve said that like a million times and you’re probably all like, “ oh she’s never going to finish the book, ’she’s never going to finish the book.” But I am going to finish the book. Everybody’s going to be so excited about that when it actually does happen. But anyway, in the process of researching this book, I came across the concept of grokking, which is what we’re going to talk about today.

And it’s really something extraordinary; it’s unusual, and we’re going to be hearing a lot more about it because it’s becoming an active area of research and it’s a learning process. So this book that I’m writing—actually I’m co-authoring it with my wife, Amy Zimmerman—and it’s a book about, well, I’ve described it in the past; I won’t go back into it all right now, but part of it deals with, you know, the cognitive abilities of AI. And when you’re dealing with cognitive abilities of AI, you’re dealing with how it learns, among other things, many other things. And so this learning process that I ran across called grokking is an emergent learning style. And we’ve talked about things that are emergent before, that’s sort of unexpected behavior coming out of models. You know, where you develop the model or somebody’s developing the model, and the model then starts to have this unexpected behavior. And chain of thought was actually an emergent behavior, and reasoning was an emergent behavior. There are a number of things that have been emergent behaviors and also some more of the unusual ones that we’ve talked about, but grokking is also an emergent behavior. So let’s talk about that. And I have to just mention that grokking is completely separate and apart from the xAI Grok model, G-R-O-K. I have no idea, and I was going to look it up before I got on air today, about whether or not the Grok model’s name came from the same place that the word grokking came from. They might have, it would stand to reason certainly, but I don’t know for sure. But anyway, the grokking process that I’m going to talk about is a learning phenomenon, and it’s not the Grok model and has nothing to do with the Grok model. Although, presumably the Grok model, like every other model that’s high-capability right now, experiences grokking.

So grokking itself, or the word grok, came from actually an old sci-fi book that was by Robert Heinlein, H-E-I-N-L-E-I-N. It was a 1961 sci-fi book that was called “Stranger in a Strange Land.” And it was about this guy whose last name was Smith. And he’d been raised by Martians, and he returns to Earth and he challenges all kinds of, you know, Earth-bound norms. And he introduces Martian philosophy that emphasizes what’s described as, in the book, radical individual freedom, a unique form of empathy and something that then is all put together and is called grokking. But apparently the word grok, G-R-O-K, was a Martian term meaning to understand something so thoroughly that you become one with it. Okay, so that’s the derivation of the word. Now the word grokking has been applied, now, more recently in the AI world to sort of describe, or become the word that describes, this emergent phenomenon of learning. So, you might be familiar with people using the word grok in a colloquial sense. Say, “oh I grok to that.” Actually, it’s often the British New Zealanders will say, “I grokked that,” meaning that I got that. And today, in the AI world or machine world, there’s this phenomenon. Now we’re getting into what grokking really is.