Podcast Banner

Podcasts

Paul, Weiss Waking Up With AI

Agentic AI: Reasoning or Only an Illusion of Reasoning?

In this week’s episode, Katherine Forrest and Anna Gressel explore the latest advancements in AI agents and reasoning models, from Claude Opus 4’s agentic capabilities to Apple’s research on the limits of AI reasoning, and discuss the real-world implications of increasingly autonomous AI behavior.

Stream here or subscribe on your
preferred podcast app:

Episode Transcript

Katherine Forrest: Hello, and welcome to today’s episode of “Paul, Weiss Waking Up With AI.” I’m Katherine Forrest.

Anna Gressel: And I am Anna Gressel.

Katherine Forrest: And Anna, okay, I feel like we should have a whole episode dedicated to interviewing you about where you have been. Where have you been? Tell us.

Anna Gressel: So I’m back in New York now, but I was previously in what was a very sunny Berlin and then a very, very, very sunny Abu Dhabi and Dubai. So it was super fun to be on the road for a little bit. I was speaking at a conference called GITEX in Berlin on a bunch of different topics, including AI agents, and then headed off to Abu Dhabi and Dubai to run some roundtables for GCs on AI agents, which is really just such an interesting topic, Katherine. It’s like, I know we talk about it all the time, but every week there’s a new development. Every week there’s something new to unpack. The UAE is just doing interesting things too, so it’s always fun to be there and listen and hear what is happening and what people are thinking about there as well.

Katherine Forrest: Yeah, there is really so much happening in AI agents, and I’m jealous that you were able to go and talk about it for these various roundtables. Today, however, I’m going to capture you as a captive audience to talk about agents with us and to talk about the Claude Opus 4 and Claude Sonnet, the Claude 4 family. And it’s an agentic, very powerful and highly capable model. So we’ll talk about that today.

And then also, just yesterday—in terms of the day we’re taping this (we’re taping this on June 10th), and so just yesterday—Apple released a paper that is different from, of course, the Claude 4 world, but also about capabilities. It’s called “The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity.” So I thought we would dive into that and take advantage of some of your agentic thoughts and words of wisdom on all of this.

Anna Gressel: Yeah, that sounds great. And I’m excited to talk about the Apple paper, because I don’t know about you, Katherine, but it is all over my social media as well. All the folks I follow on LinkedIn are talking about it. So it’s a moment to really unpack what that means, as I’m sure many of our listeners will have seen that, at least bumping around over there. But I think there’s so much to be excited about these days. And in terms of the Opus 4 model, it’s more intelligent as a base LLM, and it’s a reasoner. And in addition to that, it’s more integrated with external tools and more agentic than its predecessors, so we can unpack what that means in practice.

Katherine Forrest: Yeah, let’s unpack what it means to be more agentic. I mean, I know one thing before you get into it: one thing that Claude Opus 4 can do is it can actually use tools in parallel. It can actually send out requests for tool use to accomplish parts of a problem all at once. So it’s not doing serial tool use; it’s, with a single agent, doing parallel tool use. But tell me what “more agentic” means to you.