#238 - GPT 5.4 mini, OpenAI Pivot, Mamba 3, Attention Residuals

2026-03-26 06:00:00 • 2:00:49

-

Last week an AI would like to thank ODSC AI for being a sponsor. ODSC is one of the longest

0:16

running and largest communities focused on applied data science and AI. It started

0:20

over a decade ago with a simple idea, bringing practitioners together to learn from people

0:25

actually building and deploying models in the real world, not just talking theory. On

0:30

April 28th through the 30th, you can experience it yourself at ODSC East, 2026, taking place

0:36

in Boston and virtually there will be thousands of hybrid attendees ranging from data scientist

0:42

ML engineers, AI researchers and technical leaders. You can attend over 300 sessions covering

0:48

LM's, Gen AI, computer vision, NLP, data engineering and more. You can also go to hands-on

0:54

training with workshops and boot camps taught by experts from companies like OpenAI, Hugging

1:00

Face and Video and other top companies and universities. And of course there will be a massive

1:06

expo and networking opportunities great for startups, hiring managers and AI tool builders.

1:12

It's one of the best ways for AI practitioners and teams to stay ahead of a field that learn

1:17

from a best and connect with a community. Go to odsc.ai slash east and use promo code

1:22

LWAI for an additional 15% off your pass to odscai east 2026. That's odsc.ai slash east

1:31

and use code LWAI to get an extra 15% off on the number one AI builders and training conference.

1:39

We'd like to think factor for sponsoring last week in AI. Not related to AI but I am

1:44

personally a big fan. Often I have no time to cook and factor makes healthy eating easy.

1:50

They're fully prepared meals that come from their dishes and crafted by chefs. Actually

1:56

used it for many years and I think you can really eat well without the planning or the cooking

2:02

using factor. They use quality functional ingredients like lean proteins and colorful veggies

2:08

and there are no refined sugars or artificial sweeteners. You have 100 rotating weekly meals

2:14

to keep things fresh and you can choose types of meals like high protein, calorie smart,

2:20

Mediterranean and others. It's really convenient ready in about two minutes. In my experience

2:26

really is fair to say that there's no prep necessary and it's quite good. I've used it

2:32

for many years and I think you might want to consider it if it fits your lifestyle.

2:37

You had to factor meals dot com slash LWAI 50 off and use code LWAI 50 off to get 50% off

2:45

and free breakfast for a year eat like a provisement with factor. You serve coverage only

2:51

veg by plan. What free breakfast time per box for one year while subscription is active.

2:55

At Arizona State University we're bringing world class education from our globally acclaimed

2:59

faculty to you. Earn your degree from the nation's most innovative university online.

3:05

It's a degree better. Learn more at ASU online dot ASU dot edu.

3:09

Hello and welcome to the last week in AI podcast week in the air chat about what's going

3:15

on with AI. As usual in this episode we will summarize and discuss some of last week's

3:21

most interesting AI news. You can also go to last week in dot AI for our newsletter

3:26

with even more news every week. I'm one of your regular hosts, Andre Crankov. I studied

3:32

AI in grad school and now work at the startup AstroKate. I'm your other host, Jeremy Harris,

3:38

from Gladstone AI, AI national security, AI loss control, all those fun things. By the way,

3:44

special thanks to Andre for recording at this time. We bumped things up even earlier.

3:49

I think it's like what is it seven thirty year time? Is that it is here on the PST. Sometimes

3:55

it's nice to be East Coast to be a bit later. But you know it's good to get your day started

4:02

early sometimes. Yeah, much appreciated. Also appreciate people tuning in or watching the entire

4:08

last podcast, which featured a extra like hour plus, which I didn't realize it was that long,

4:14

but Andre had a hop off. So I just went through like some of the technical papers we didn't

4:18

cover as an experiment. And man did I go on. So I have learned that I need Andre to be like the

4:25

regularizing term to my loss function. If that is. Yeah, we do know some people are very much

4:32

fans of the coverage of research and I was going in depth. So feel free to comment on YouTube

4:39

or elsewhere. And yeah, say if you want more of that, you've considered maybe having additional

4:47

episodes that are just research, very much happening there. But feel free to let us know if you'd

4:52

like even more research on a regular basis. And just to give a quick preview of what we'll be

4:59

doing this episode, not as much research as last one. There's a bit of everything I suppose. There's

5:07

a couple new model releases and some other kind of interesting tools. There's some interesting

5:13

developments on the business front with OpenAI. And then we've been covering a lot of safety and

5:19

interpretability work lately on alignment. So we'll have some of that. And then towards Van,

5:25

we'll have some fairly interesting impact for looking research. Let's feel radical more sort of like

5:32

wow, this might actually be a big deal. So it should be a pretty fun listen. And let's kick it

5:40

off with tools and apps. First up, OpenAI, they have shipped a Jubilee 5.4 Mini and Nano. They

5:49

are similarly to other small models that we've seen in, in recent times, like actually really

5:56

quite good for being kind of a smaller range. Jubilee 5.4 Mini is close to Jubilee 5.4 on

6:06

several benchmarks, including on SWE Bench Pro and OS World. They're fine. And it's more of

6:13

and twice as fast. Jubilee 5.4 Nano is obviously the smallest option. It's not really doing great

6:22

on the benchmarks, but it is super quick. These models have 400,000 token context middles,

6:30

so fairly substantial. But they do cost a decent amount, relative to GP5 Mini. Looks like

6:41

GP5.4 Mini costs free X, GP5 Mini, and GP5 Nano also costs more. So on the whole, good,

6:49

faster, smaller models, if you need something that's doing better in GP5 Mini, now you have that

6:56

option. There's been a lot made about the cost situation and the pricing of the per token pricing,

7:02

I should say. It is higher. There's no question, right? So GP5.4 is basically three quarters of a penny.

7:09

Sorry, three quarters of a dollar per million in put tokens versus 25 cents for GP5 Mini. So

7:15

that's a three X hike. But opening I says it only burns about 30% of the GP5.4 in codex. So it's

7:23

actually going to be much more token efficient. And this is a metric that I think matters a lot more

7:28

than many people will tend to realize, right? So we have lost and like cost for token. But as we've

7:33

seen, more tokens at inference time does not necessarily mean more performance. And that's the big

7:38

catcher. So when you multiply those together, right? 30% times three acts, you actually get a slight

7:44

decrease in what you might think of as cost per performance, which is a little closer to what

7:50

most people care about. And this will vary depending on the workload, but still quite interesting,

7:53

right? So for the sort of orchestrated agentic tasks, the effective cost per outcome could actually be

7:58

favorable compared to running the full model. So it's a sort of interesting there. One thing I will

8:02

say nano is API only. This one is priced at 20 cents per input and $1.25 per million output tokens

8:11

versus like much, much cheaper. So it's predecessor was 5 cents per input token. That's a 4x lift,

8:18

roughly 4x again for output tokens. Well, so you're looking at 4x overall. The weird thing is,

8:23

so opening I is pitching this for classification and data extraction, which are these like very high

8:31

volume workloads where you're usually quite cost sensitive because you're processing you're so much.

8:36

And so that four full hike is going to sting the most for exactly the people who are being pitched

8:42

this product. So it's a bit of an interesting position. It seems like a little, I don't want to say

8:47

at odds with open AI's position that they want to make intelligence too cheap to meter, but it

8:52

certainly is at least locally. This is a move towards instead of racing to the bottom on inference

8:57

costs. We're going to focus on model quality. And that that's going to be our big differentiator.

9:01

You're going to care that we can get the right answer, not that we can get it cheaply,

9:04

which is where all the margin is. That at least is what anthropic certainly suggests and what

9:08

we're seeing elsewhere. So that's kind of interesting. You know, there's a bunch of interesting,

9:11

as you said, OS world verified is interesting benchmark to look at here specifically. I know you

9:16

mentioned sweet bench and a couple others. And so this is basically a computer control benchmark.

9:21

So it looks at how well can the mini model just control a computer. And what we see here is a

9:26

GPT 5.4 mini hits 72%. If you look back at GPT 5.4, the full version, it hits 75%. So it's

9:33

actually pretty close. And if you're thinking about the previous GPT 5 mini, that was only 42%.

9:39

So pretty big jump, especially given that we're getting up there on this benchmark in terms of

9:43

saturating it. So all in all, pretty interesting release, the very lead, not very lead, but the

9:48

very detail here really is that token efficiency question. What kind of workload are you going to

9:53

use this for? That's going to determine, I don't want to call it like the total cost of ownership,

9:57

because that's not quite the right metric here, but the total cost that you're exposed to in

10:02

the ROI, that's really becoming a key thing here. Right model for the right workload is just going

10:06

to be a critical dimension, at least for the next few months. Right. Yeah, I think these kinds of

10:13

things showcase the fact that, you know, all these models kind of came out of the world of academia,

10:19

right? And benchmarking is largely focused on capabilities on how accurate your model is.

10:26

And so usually you don't highlight these kind of more practical concerns of how quickly can you

10:32

finish a task, right? How cost effective are you like you do a task, how much dollars does it take?

10:39

Wall clock time, right? We just don't get these numbers, at least, on the announcements. It's

10:44

honestly a bit surprising that's still the case, but it's a culture question, I suppose.

10:50

And as you said, we'll be discussing opening a strategy in just a bit in the business section.

10:55

Very much in line, we're on topic where we're like, we'll just charge more for our models,

11:00

but they're the best. And so people will follow it. And speaking of small models, next up,

11:07

you've got me Straul. They have released their small four family of models under the Apache 2.0

11:15

open source license. And it combines actually multiple things. So it has reasoning built in.

11:23

It has multi-volonial capabilities built in and it has a genetic coding optimization. So they

11:30

combining magistral, pickstrawl and devstrawl. It's also used as a mission of experts. So they've

11:36

have a total of 119 billion parameters, but only six billion active parameters per token. That's

11:44

quite small. You can fit it into probably one top end GPU. You know, it's actually quite affordable.

11:53

So it looks like possibly a pretty strong model on in this category of smaller, faster,

12:02

cheaper, and open source. Yeah, challenge there is going to be, you know, if you get into

12:07

wanting to fine tune it, obviously, then, you know, now you're you're dealing with really a 120

12:12

billion parameter model. And that's, you know, pain in the butt. But yeah, I mean, the only six

12:16

million active, I mean, this is really ultimately a bet that you're going to have kind of like

12:21

model sparsity in the sort of very aggressive sparsity ratio is going to perform a dense model or

12:25

something, you know, more traditional. And it's a bet as well as you say on the kind of hardware

12:29

that is available at least for inference on local machines. So yeah, it's quite interesting. It

12:35

is a more aggressive kind of fewer active parameters per token type of play than we've seen before.

12:40

It's also kind of interesting. So you touched on the consolidation, right, of all these models under

12:45

a single, in a single model, right, reasoning, coding agents, multimodal, that's a pickstrial,

12:51

like looking at images and text and so on. So this is a weird play because historically,

12:57

Mr. Alt has separated these, right. And so given they're collapsing them into one model just with

13:03

a reasoning effort dial, you could see that as a pretty big bet on where the industry is heading.

13:07

That is something that we've seen with, you know, obviously the O series of reasoning models

13:12

with GPG 40 even starting as far back as that. If it's the case that we get positive transfer,

13:17

that's really what this is a bet on, right, that a model that is trained to do all these things will

13:22

do better at each individual thing because it's kind of getting cross training, right, the same

13:26

way that you might want to do, you know, football and like ballet at the same time, rice hockey,

13:31

and that you get better at each different thing because of the combination. That's kind of the idea

13:35

here, right, that positive transfer that for so long was was really difficult to pin down. People

13:40

were seeing negative transfer where the more stuff you train a model on, the worse it does on the

13:44

marginal additional task. We're now well into positive transfer territory. The extension of that

13:48

to reasoning is quite interesting and implies that reasoning may, at least that they're betting,

13:52

that reasoning may eventually were already play a role in analyzing images more successfully for

13:57

these kinds of models at this scale. So there's another piece here that's interesting, you know,

14:01

this efficiency claim. They say that small four is achieving scores that are like on par with GPT

14:07

OSS 120B on a bunch of benchmarks while generating like shorter outputs on at least one benchmark.

14:14

And so again, this is about that token efficiency question, right, we're sort of like back at that

14:20

very underrated metric of output length efficiency and the fact that you don't necessarily get more

14:26

benefit by reasoning with more tokens, that's something we started pointing out when we looked

14:30

early on at the kind of deep seek reasoning results, right, where it's like, yes, you do get this

14:34

positive uplift, but your value per token is actually like potentially going down. And we are

14:39

now, in fact, in that regime, we're seeing that very clearly. So this efficiency piece is really

14:43

important because if people are going to use open source models, they're going to run them like by

14:47

far, the biggest use cases running these on whether you're own or other people's clusters to serve

14:53

customers. And so, you know, like how good Mistral is at making this an efficient reasoner speaks

14:59

directly to your bottom line as the person who's going to be serving these or asking somebody else

15:03

to serve them for you. So pretty interesting, you know, the fact that they're comparing to GPT OSS 120B,

15:09

look, the space moves really fast. That is an old model at this point in open source terms.

15:14

It's kind of like choosing your point of comparison pretty selectively, I would say here,

15:19

but they do a comparison that's reasonably favorable on point models as well. So it's an

15:23

interesting play. I think, I mean, I'm a little concerned from Mistral. I have been for a long time,

15:27

obviously, don't know too much how this ends up playing out with the open source play, but here

15:32

they are. It's a reasonable model and the consolidation angle, that's a really big story here, right?

15:36

If everybody is starting to consolidate, even open source players here around one model under one

15:41

roof, that's a materially different story. It does not, by the way, extend necessarily to the world

15:45

of agents, right? Sub agents may be smaller, cheaper models. This is more about, you know, if you're

15:50

looking for a highly performant model, I think of the main orchestrator, you're probably, it seems,

15:55

you're probably going to be seeing, you know, models that that kind of put, put multiple capabilities

16:00

under one roof. So something to watch out for. Right. I'll just quickly comment on the

16:05

unification bit. It's better to say that it's partially a bet on what things are heading with

16:10

a multimodal aspect. The fact that they have baked in reasoning and coding, I think, is a little more

16:17

just an indication of catching up with where things are these days. It used to be the case that

16:23

you had a reasoning model, like, oh, free. And with deep seek R1, you trained a reasoning model

16:30

and you had your base model. And what everyone moved to in 2335 is there's no reasoning model.

16:36

Your model has reasoning baked in. And now with, with Sonnet, with GPT, really post training for

16:44

reasoning, it's been very clear that you should just train your model to be a good coder because that

16:48

makes it a smarter model in general. So this is a mix of catching up where our rings are at

16:55

and also adapting that multi model capability, which is could be interpreted in several ways.

17:04

Next up, Meta's Manus launches my computer to turn your Mac into an AI agent. So this is something

17:14

you can install and launch on your computer. And it's effectively like having a little

17:20

open claw, I guess, on your computer. So it can execute command line instructions that

17:27

sit interact with computers. So it's very much, I think we've seen this happening more and more

17:34

where various organizations are shipping open claw, ask things where you have an agent that just

17:40

lives somewhere and you can tell it to do stuff. And it is your like assistant or your AI or whatever.

17:48

This appears to be another instantiation of that similar also to perplexities announcement of

17:55

what was it like perplexity computer. Yeah. So very much in line with that. Oh, you mean you were

18:03

enabled to remember like the sixth new open claw variant that got launched by a company in the

18:09

last two weeks. Yeah, it's crazy, right? We're really seeing more and more of this like

18:13

pylon right from all these all these different competitors in this space. And this is a land grab

18:18

like Magnum mistake. This is the the sort of man, I don't call it the scramble for Africa moment,

18:23

but the you know, historically equivalent of that in this space with fewer controversial overtones.

18:28

This is the moment where people are realizing, hey, you know what? We need to get on people's

18:33

local machines, right? We have to get some kind of piece of that pie because so much of what's

18:39

what's happening right now is that we're trying to like grab onto people's like agentic like the

18:46

the kind of agentic runtime layer. In fact, in video, we'll talk about this in a minute with

18:49

Neemoclaw. It's the same thing. Everybody's trying to get on to like what is what is the substrate

18:54

on which agents are going to run? Can I get my dirty little hands on that and and turn that into

18:58

part of my market? And in this case, the menace play is quite interesting and it's aging well,

19:04

right? I mean, menace was an in it like totally independent company before the acquisition.

19:09

And really the play here is meta extending its reach into that that local OS layer for the first

19:16

time. It's this is the territory historically of you know, your apples, your Microsoft, your

19:20

Google's. That's the war they're entering. We haven't seen meta do that before, right? We just

19:25

haven't seen them try to play in the operating system game. This is a really interesting way for

19:29

them to vector into a completely different market using resources that you know, maybe for the first

19:35

time they have like a credible, whether you call it advantage or not, they have a credible play here.

19:39

So pretty interesting. And again, meta classic history of buying their way into competing, right? We

19:43

saw this with WhatsApp with Instagram like it, it never ends. This is their play in that direction.

19:48

So, menace, you know, maybe aging well, we've got to see what comes of this.

19:54

came out I believe last year initially with a cloud based agent but you could assign to go do

20:00

stuff so this is them extending to your local computer right now you can I think get this

20:05

for silicon based max the other aspect of this by way is not just open claw it's the core work

20:13

the codex angle where they most of the blockpost actually is highlighting like let the agent go

20:20

to computer and organize files and do things that are very much cloud code or or now core work

20:26

ask and we kind of tell it to do stuff from anywhere at any point is an aspect of it's the

20:33

opaquaspect but I think the real land grab is for that core work type like have an agent do stuff

20:40

for you which is now like everywhere in coding but I think what or on topic and now open AI

20:48

and now everyone is realizing is these agents can like do a whole bunch of stuff and that people

20:55

haven't adopted to do yet and speaking of open claw next up and video has announced Nima claw

21:01

as part of their announcements at gtc which is a little bit of funny this is a stack for the

21:09

open claw agent platforms that allows you to install and video name-atron models we just discussed

21:15

this latest name-atron model last week and there's a new and video open shell runtime you

21:22

install both of those in a single command and you get privacy and security controls baked in

21:29

making it possible to have you know more more confidence in running one of these things we've

21:37

seen many stories of open claw go in raw it's right absolute confidence yeah so open shell provides

21:42

an isolated sandbox that enforces policy-based security network and privacy guard rails for

21:50

agents seems like a good idea if you are to do one of these agents maybe install them in a sandbox

21:57

where you can control them and they don't go rogue and you know take over the world yeah and one of

22:02

the the key dimensions here you know you think about what does sandbox mean you know how do these

22:07

things work typically a lot of these things focus on what is the model that is on the running on

22:12

the cloud and what are the models for model that's running locally on your machine right so you

22:16

imagine you might not want the model that runs locally on your machine that's actually looking at

22:20

your own intimate files to have direct access to the internet right so that's kind of like one way

22:25

that you might enforce that sort of guard rail so use that local model just like generate summaries

22:29

or do analysis and stuff and then just ship the summaries after some review or something like that

22:34

to you know that's like that's one way to to play that game and there are a whole bunch of other

22:38

guard rails around the kind of access that different models can have in your ability to net so that's

22:42

kind of like a lot of where this is coming from you know this is a classic example of jensen's

22:48

hyperbolic rhetoric right he's he's using terms like new renaissance and software to kind of to play

22:54

this up which to be clear I actually I mean I agree with this but there's like there's a gap between

22:59

you know this framework can help you install you know a single demand and redefining how computing is

23:04

done and we'll see if that that chasm gets gets crossed and it it probably well at some point man

23:09

I think it's certainly well at some point question is who does it first so the frame here is you

23:14

know again that operating system piece right keep going back to this it's not a coincidence that

23:18

everybody is is going on the same basically this gold rush expedition everybody's thinking the same

23:23

thing the operating system for personal AI that layer and again jensen here comparing mac and

23:30

windows for PCs to open cloth for personal AI that's a deliberate a bold clamp a very deliberate

23:35

this is part of that frame this is Nvidia now saying hey meta is going to get into the effectively

23:40

the operating system game the the sort of operating system for agents we're going to do the same

23:44

thing we don't have history quite of of doing that but you know now we're diving into so you're

23:48

creating creating this environment where because essentially agentic you know we had the

23:53

software eats the world era of the sort of sass revolution you know in the last decade decade

23:58

and a half two decades and now we're in the AI is eating the world including software and another

24:03

layer of abstraction here is yeah that runtime environment for agents and and that's the land

24:08

ram that's the operating system at least this is the case that jensen's making I think it's a

24:12

reasonable case on the whole but remains to be seen this is functionally also a classic Nvidia

24:17

play of like trying to trojan horse in their full stack so nemoclaw is going to install nematron

24:24

models right those Nvidia models preferentially and the open shell runtime all in one command the

24:29

simplicity here is the point so it's super super easy to deploy Nvidia's own models together and

24:35

the runtime environment all that stuff basically they're creating a situation where just like they

24:39

owned kuda for the the GPUs they're creating a whole software stack around agentic the agentic

24:45

runtime environment and so this is what's worked for them in the past create a whole ecosystem around

24:50

this software is more commoditized now that it has been before so that may not be the same kind of

24:55

moat especially given that they're you know unlike before with kuda where they had like a decades

24:59

head start for anyone really paid attention this is now you know much more in competition with

25:03

kind of faster moving players so really interesting again chocolate up is another another

25:08

entrant in the sort of operating system for the agents sort of catalog here right I do think

25:15

interesting aspects is nemaclaw is is kind of like the hype really like good cool branding for

25:23

the actual major push which is their open agent development platform which is that terminal

25:30

based thing it also comes with this Nvidia AIQ blueprint thing which is built on top of

25:38

length chain deep agents and has Nvidia Nemo agent toolkit as these open source kind of options

25:45

for building your stack of agents so nemaclaw aspect is like digging back on the thing that

25:53

everyone is hyped about and then the actual software frameworks is probably the actual thing

26:00

that Nvidia cares about yeah and this is it right it's like how can I hook on to one big hype train

26:07

and then exactly use that as a Trojan to get her like everybody dependent on her whole stack including

26:12

their models because there hasn't been the kind of uptake of the nematron models at least that I

26:16

expected to see so far and so this is you know one one opportunity for them to make that happen yeah

26:22

I think to your point about software kuda obviously is is the number one thing for GPU related

26:30

kind of execution software but as far as open source packages go and video hasn't had a

26:36

history of really making an impact so for instance length chain is an open source package that

26:41

goes back to 2023 that has been a popular for building complex graphs of LLM prompts and agents

26:50

and so on probably overly complex but anyway it was at an gained broad community adoption

26:57

so that's more or less been the pattern a lot but with agents and now open claw and whatever

27:05

that was a good time to try and get in on that open source kind of game and one more story about

27:11

in video at gtc they announced dlss 5 which is what they're describing as the GPT moment for

27:22

graphics so this is runtime enhancement I suppose for game graphics so this uses machine learning

27:32

based upscaling and it applies generative AI to add a whole bunch of just really nice looking

27:40

graphics so if you look they have various examples so you probably want to just see it to understand

27:46

it basically you can go to older games like other schools oblivion for instance and get really

27:54

cutting edge seeming graphics with this turn down now the reaction to has been very much split

28:01

from what I've seen people have often sculpted it and we're like oh this is AI slop this filter is

28:09

is bad in some cases it seems to go against the style of the game somewhat so Nvidia did also

28:17

emphasize that this is fully controllable by the developer the developer can set the settings

28:24

of how aggressive it is you know how much it impacts various aspects of the rendering yeah I think

28:32

this is the kind of thing that it has given developers presumably in the past this is dlss 5

28:39

but with the generative AI aspect of it it's getting a lot more discussion and it looks

28:45

potentially very very significant well so I was going to ask you I mean this is the part of the

28:50

show where I ask you for your opinion as a guy at astrocade like is this something that you guys

28:55

would integrate in your stack like what how do you how do you think about new tools like that or

28:59

is that even part of your your workflow yeah so this this is kind of targeting that 3d

29:06

triple a big game kind of market right so this is dealing with new games right then and more

29:14

complex you wouldn't run this on your phone like for casual games but for games that have complex

29:21

character models with faces and stuff like that or like open world games where you traverse big

29:28

landscapes that's where the kind of photorealistic pass really makes a big difference and last up we

29:38

have an update on open areas plan to launch strategy parties adult mode this has been announced

29:47

I think last year as I think they are thinking of doing it has been delayed from the original

29:54

late March target and it seems that they're still aiming to do it the news here has been that

30:03

the team within open AI their advisory council of psychology and neuroscience have opposed it at

30:11

a January meeting one advisor really warned about it significantly so anyway we have a quick update

30:19

saying that they appear to still be planning it it has been delayed but as of now it's still

30:26

going to be presumably released yeah well what a surprise that despite objections over the

30:33

appropriateness of a tool like this that they still went ahead huh that's that's weird that's

30:38

not my company at all isn't yeah that's weird what a weird thing I know this is when you think about

30:45

the things that were classically warned about just my opinion here personally but I seem to remember

30:51

an awful lot of people warning us about the brave new world thing where you're hooked up to

30:55

like the dopamine drip yeah I don't know ultra porn powered by like AI super intelligent like

31:01

I'm not sure how far you push the inference time compute budget before you get to that scenario but

31:07

hey how about scaling laws for porn addiction how about that paper looking forward to that coming

31:12

out I mean look that that's the kind of thing we're gonna have to see there's gonna have to be

31:16

studies on scaling laws for porn addiction just calling a spade a spade like I don't see how we

31:20

avoid that it's interesting that this is this is an offering someone's gonna do it right the big

31:25

porn companies at some point whatever so you could argue and I'm sure this is part of the ethical

31:30

argument that opening I might make here is look this way we can monitor the use of these things

31:34

ahead of time understand how people adjust to this technology maybe try to mitigate risks ahead of

31:39

time whereas porn companies probably wouldn't do the same thing sort of a similar argument to what

31:44

Sam's been making historically right we want to like get these things out there because it's

31:48

gonna happen anyway so we want to get that feedback and be able to iterate on it which there's merit

31:52

to but yeah we should be under no illusions that we're crossing the Rubicon here and I think given

31:57

that open AI is taking this step it is on them then to come out with unbiased research that

32:02

transparently looks at the effects of exactly what they're putting out there right I think that's

32:06

just a reasonable thing I think any I'd be pretty sure it's like a tobacco company going out

32:12

and doing something when you're playing so close to the the base of the human brainstem you're like

32:17

you're in different territory and so anyway it we'll see if this persists we'll see what scandals

32:22

come of it I'm sure there will be some but yeah it's in some ways not surprising I don't want

32:27

again rip on open AI too much for this because the argument is true that at some point you know the

32:32

big porn companies will do this and they have a history of like mistreating women and you know

32:37

doing all kinds of awful things so ethically I get it but now the proof is gonna be in the pudding

32:42

because once you put that out there and brand-wise like geez I well so I think it's worth clarifying

32:48

a little bit porn might be a bit strong of a term here right so they do clarify this will not

32:56

generate erotic audio images or videos this is really about so they've banned erotica as a

33:04

as a general category with the chatbot and so if you try to write sexy stories or have sexy role

33:12

play which you can do and plenty of other you know apps out there is a million AI girlfriends

33:19

including XAI is right so this is more of that this is like role playing with sexual kind of adults

33:26

only component to it not straight up like explicit content of the sort that you might think

33:34

when saying porn so in some ways yeah I think there's arguments to be made out of a way like people

33:41

have tried to do this and tried to do this as is most likely with these models people do seem to

33:48

want it there's very large adoption of these things and we already have plenty of stories of people

33:55

getting AI girlfriends and having psychological impacts so yeah maybe actually if chat GPT and

34:03

open AI do this explicitly and do it right and do it well it's better than not doing it and having

34:10

some shady dark market for it you know no absolutely and I take back the the use of the term porn

34:17

here I guess I'm sort of seeing where the the puck is headed and anticipating that next play

34:22

this is definitely a step where we need open AI to actually run the studies since they're going to

34:26

have access to the data and no one else will right so like we we need people to look at this I hope

34:30

that they will I would imagine this as part of the package here but from a branding standpoint this

34:35

is really dicey especially at a time when they're trying to redouble we'll talk about this in a minute

34:39

but on this whole sort of business productivity focus like that's going to be their big play so you're

34:45

kind of adding that into the mix it's interesting I mean what what this implies about the user base

34:49

that they're they're going after is is there ROI here I mean porn is a pretty low margin industry

34:55

as I understand it's so so this like I don't know if role play erotic role play would kind of

35:00

be similar things flirting if you will with that but either way yeah we'll find out I think the

35:05

argument that a lot of people want this is also kind of dicey I mean wow a lot of people want crack

35:11

cocaine like it's a so we got to find a way now everything's out of continue I'm being tongue

35:16

in cheek here obviously but we have infinite inference inference time compute budgets coming online

35:21

in the next you know three to five years or whatever at what point do you just have so much inference

35:26

time computing dedicated to your your limbic system that for all intents and purposes like you're

35:31

robbed of your agency I think we're actually going to be closer to asking ourselves that question

35:35

sooner rather than later than most people are pricing it right and then just to reiterate the

35:41

actual news here is just that there have been these details coming out of the internal objections

35:48

from certain parts of open AI and these objections have been there since January and the company

35:56

hasn't canceled this feature features it out they could still not do it and it has been delayed

36:04

so it's still on track and it'll be interesting to see if they do go through it after all

36:09

yeah I know that's true and actually like a lot of my reaction is reactionary I'm like

36:14

this this direction generally is definitely coming but it makes me nervous as hell and that's kind

36:19

of that's sort of my my response to this is you know we're just sort of seeing a lot of this is

36:23

testing the waters too I'm not saying this particular story was leaked but you see governments do

36:27

this you see private companies do this there's like leaks of stuff to go out to just gauge audience

36:32

kind of consumer response to things can we take that step can we not I don't know again if this

36:37

is or is not the case here but but man I mean that you'll at some point need some kind of immune

36:43

response to things that push in this general direction and on to applications and business

36:48

sticking with open AI and coming back to a topic we briefly touched on another thing that came

36:55

out from open AI this week is a discussion that happened at an all hands where effectively open AI

37:03

has signaled a strategic shift away from doing everything doing a little bit of everything and

37:10

focusing more on productivity and business so you know open AI has historically had a million

37:18

side projects this I think is how they characterize it side projects they have video models they

37:24

release the soar app they have a browser they have audio models I believe transcription like

37:32

a million things on fabric has clawed and that's it open AI has sorah and their audio models and

37:41

and so on and so on so it seems to be that internally within the company they are changing

37:49

focus the head of applications Fiji Simo has told staff that they cannot miss this moment because

37:56

they are distracted by side quests and they need to nail productivity and you've seen this already

38:03

for the last several months it has been apparent that they have been really pushing on codex

38:08

to catch up with cloud and and co-work as well and they largely have at least in terms of

38:14

feature set in terms of adoption I mean many people are starting to use codex some people prefer

38:21

codex at this point but it is obviously true that this hasn't been open AI is primarily focused

38:28

until recently and they have kind of been behind the cloud and in terms of adoption they are

38:34

still behind because cloud code is the first early leader in the category so yeah interesting to

38:42

see open AI having that internal discussion now I feel like this has been a problem with open AI

38:48

for a while if I were to kind of guess at internal dynamics and kind of in business and company

38:56

level issues that lead to poor performance so yeah kind of could see this coming I think famously

39:04

and and Andre I'm sure you'll have friends that have told you the same thing like at Google

39:08

everybody in mind who's ever worked at Google says the same thing and actually same meta you get

39:13

promoted for building new stuff right it's like it's not about did you make the code run more

39:17

efficiently did you clean this up just clean that up it's like did you make new stuff and that's

39:22

why there's a massive app grade of yard right famously for Google products you know Google hangouts

39:27

Google this Google that that just like gets axed in at various stages there's this fundamental

39:31

There's this fundamental question of like,

39:32

again, is this a feature bug, right?

39:34

You can look at Google and you can say,

39:36

ha ha, look at the grade of yard of waits a time.

39:38

You can also look at Google and say,

39:39

well, what matters is not the misses,

39:42

what matters is the hits.

39:43

And for every, not for every Google Hangouts

39:46

or dead Google product,

39:47

there's like a Google calendar or a Gmail, right?

39:50

Or maps, right?

39:51

Or maps, that's right, that's right.

39:53

So like Paul Buhheit, famously, you know, now a YC partner,

39:56

I don't know if he's left, YC, he was there when I was there.

39:58

But he like was the founder of Gmail

40:00

and his entire like claim to fame and Silicon Valley

40:03

is that he was the Gmail guy.

40:05

That was a company within Google.

40:06

That's the way to think of it.

40:08

And certainly from a revenue standpoint,

40:09

that's what it is, right?

40:10

So when Sam, who comes out of the YC ecosystem too, right,

40:14

having been the president of Y Combinator for many years,

40:17

the first one after Paul Graham,

40:18

his whole thing is spray and pray, right?

40:20

He's used to betting on us like a little bit of money

40:23

on a lot of efforts at the same time.

40:25

This goes way back to OpenAI's founding days, 2015-ish,

40:28

where he was doing the spray and pray thing

40:30

on a whole bunch of different things, you know,

40:32

evolutionary methods and robotics and RL and all this stuff.

40:36

You know, OpenAI's gym was originally a side project, right?

40:39

Now it's a big thing.

40:40

So there's a whole bunch of different plays

40:42

that you know, he sort of like used to playing in this way.

40:45

It's worked from the past.

40:47

The challenge is now you're entering a much more mature environment, right?

40:53

You're no longer necessarily in the game of betting

40:55

on a bunch of different startups.

40:56

You have a kind of core area that you need to dominate right now.

41:00

You know, anthropic is running away with the pod.

41:03

I think it's over 70% market share right now in enterprise.

41:06

So that's what is the red alert here

41:08

that Sam Altman is calling in Fiji Simo.

41:11

So they've got to find a way to kind of like,

41:13

yes, you got to keep that Google or that YC kind of approach

41:17

to investing a whole bunch of things.

41:19

You never know when the next big thing is going to come from.

41:21

But you also have to be able to invest and double down

41:25

in things that are working that are generating a lot of your revenue.

41:29

I mean, 25% market share on enterprise is a big, big deal.

41:32

You know, they're making what, $25 billion of annualized revenue

41:36

at this point.

41:37

Like they have stuff that's really working.

41:38

And so this is a structural shift.

41:40

There's another piece that's happening here where I think there's

41:42

an over rotation on, oh, well, opening eyes like, you know,

41:46

kind of throwing out all these side projects.

41:48

Sam wrote on X a couple of days ago.

41:50

He's like, look, there's all these rumors

41:52

that we've also canceled the hardware thing with Johnny Ive.

41:55

And in fact, this article sort of repeats that rumor

41:57

saying that, hey, they, you know, they basically canceled the roll

41:59

out of these AI powered earbuds.

42:01

Well, actually, at least Sam says the Johnny Ive thing is still live.

42:05

And so again, it's a question of what about the hits

42:08

not necessarily just the misses,

42:09

but some of the hits now are worth protecting.

42:11

And that's a challenge.

42:12

If you start to grab the pot, Sam needs to find a way

42:14

to hold onto it and expand his territory.

42:16

It's not just a land grab.

42:18

Right.

42:19

And I think to be fair, Cloud Code initially, the story goes

42:22

was a single person's kind of side project

42:25

and it turned into a massive thing because someone pursued it.

42:29

Really, what this perhaps indicates

42:31

is the various bets that open eyes have been making

42:35

with the browser, with the SOAR app.

42:37

They haven't seemed to really be a hit or like BS.

42:42

Nearly as transformational as Cloud Code.

42:44

So they missed out on the Cloud Code moment.

42:47

They started catching up late with Codex.

42:49

It wasn't until kind of pretty late into last year

42:52

that I started feeling like they're really taking it seriously.

42:56

Already by mid year, if you were kind of feeling a pulse

43:00

of where AI was like people were calling Cloud Code.

43:04

And you know, I adopted it sometime around May or June.

43:08

Codex didn't really start kicking off

43:11

until maybe even November or later.

43:14

So yeah, it really signals kind of more,

43:18

not necessarily not doing these other things,

43:21

but realizing that they were left behind

43:23

and now we need to catch up and start making money

43:26

with businesses because that's as we often say

43:29

where you make your money much more easily

43:32

than with consumer products.

43:34

Yeah, much more, you ultimately need to go straight

43:36

to consumer to get the big pot down on others.

43:39

But it's early in the get like it is always early

43:41

in the game and AI and to your point,

43:43

I mean, this is, and I can't remember if this is a Napoleon thing

43:46

or an Alexander the Great thing,

43:47

but people would say like he could achieve a victory

43:50

but not use it.

43:51

This is kind of that, right?

43:52

There's two different modes in the space.

43:54

First is getting your product market fit

43:56

and then it's holding on to it in the face of competition

43:58

and then kind of attacking other people's modes.

44:02

And that second one is it's a shift.

44:04

It's a fundamental kind of mindset shift

44:06

that you got to get into and it looks different for AI

44:08

and for traditional SaaS companies too.

44:10

Like you just can't sit on a stack of software

44:12

for like a month and expect it to hold on to market share.

44:15

So yeah, it's a really interesting space.

44:17

We're learning a lot in real time right now

44:18

about what it takes to gain market share

44:21

and hold market share in a world of low switching costs

44:23

and very rapid advancement.

44:26

Next up, moving back to GTC.

44:29

We have some more details.

44:31

One other thing that Jensen Huang, the CEO and video

44:35

has announced is that it looks like purchase orders

44:40

for Blackwell and Vera Rubin chips.

44:44

These are the newest generation of chips.

44:47

I expected to reach $1 trillion through 2027

44:51

bubbling last year's projection of 500 billion revenue

44:56

opportunity.

44:58

He's saying that the key that's driving it

45:01

is a shift from chatbots to agent AI applications

45:05

which do require much more compute.

45:08

You're just generating a lot more output.

45:10

And if you have open cloud, which isn't always on agent,

45:14

but you let go off and work for hours at a time.

45:17

And this is where things are heading, at least if you're

45:21

kind of a believer in the trends, right?

45:24

They've been half a year, a year.

45:26

We could expect agents just working on their own

45:29

for many hours at a time, even days at a time.

45:32

And so the bet is that it's going to keep happening.

45:37

They also unveiled the GROC free language processing unit

45:43

which is the first announcement related to GROC

45:46

since their 20 billion asset purchase in December.

45:51

GROC, their language processing unit,

45:54

has been a very impressive piece of work.

45:56

GROC has really high throughput time

45:59

and it is just a great option for running inference

46:03

if you need speed.

46:04

So that I think is also quite significant.

46:07

Yeah, that's a rate of 20 billion dollar purchase

46:10

of GROC by Nvidia a few months ago.

46:13

And yeah, I guess that was December.

46:15

Huge deal obviously.

46:17

Already, they're looking to ship a product

46:19

in the third quarter of this.

46:19

Yeah, that's fast.

46:20

Under a year from acquisition

46:21

and they're already fully integrated

46:23

and pumping out stuff just as a reminder here.

46:25

So when you think about GROC's LPU,

46:28

language processing unit,

46:30

go back to our hardware episode to do the deep dive on this.

46:32

But basically, you traditionally have in any kind of GPU,

46:36

like the Nvidia GPUs,

46:37

you'll have a stack of high band with memory, HBM,

46:40

that's going to sit next to the logic die.

46:42

And the logic die is what actually does the matrix math,

46:45

the computations that are interesting, the number crunching,

46:48

but it has to pull data from that high band with memory,

46:52

the stack of HBM that's next to it.

46:54

And pulling that data takes time

46:56

because they're physically just like different objects, right?

46:58

The HBM is a stack and then you have a logic die

47:00

and they're packaged together,

47:02

but you run into this challenge

47:04

and it just has to travel to the memory,

47:06

grab the data and come back that creates that memory wall

47:08

where the processor spends like 70% of its time

47:11

just waiting for stuff, right?

47:12

So you've got a kind of cold starting issue there.

47:14

And the GROC, like language processing units,

47:16

the LPUs, they use a kind of memory called SRAM

47:21

that is built directly into the silicon.

47:23

So the logic and the memory are much more intimately linked.

47:26

So the data doesn't have to travel, it's already there.

47:29

So you get this massive increase in sort of like internal

47:32

memory bandwidth, so how much memory,

47:34

how much data per second can flow

47:37

between the relevant components,

47:39

some like 10 to 20% faster on the GROC units.

47:42

So this is really, really important

47:44

because the Nvidia is going to be integrating this

47:46

with their viewer Rubin architecture.

47:48

That's the next generation after Blackwell

47:50

and kind of doing a hybrid of these two ideas.

47:53

The idea is they're going to have HBM

47:55

sitting outside the main processor,

47:57

but also integrating some of this SRAM heavy compute tiles

48:01

that are directly going to be in the Rubin chips.

48:03

And so this is a really big merger,

48:06

a physical merger of these companies that we're seeing here.

48:09

So yeah, it's pretty wild.

48:11

It also means a really intense demand for memory

48:14

in the memory market.

48:15

Right now, I think SK Heinex,

48:16

I saw something earlier today

48:17

that were like booked out until 2030,

48:19

they expect to have basically like way more

48:22

demand supply all the way until 2030.

48:24

So this is all part of that, right?

48:26

People are realizing, holy shit,

48:27

like memory is a big deal here.

48:29

And Nvidia on the inference side of the things,

48:32

you know, I think Jensen called himself

48:34

what the inference king or something

48:35

at these things where they're like,

48:36

hey, so what's going on with this GROC architecture

48:39

and he was like, hey, I'm the inference king.

48:41

So this is part of that, right?

48:42

The GROC play is an inference play.

48:45

And next going back to Mistral,

48:47

one other aspect of what we have announced

48:50

is also at GTC, they launched Forge,

48:55

which is a new offering by them

48:58

to let customers, businesses train their own AI models.

49:03

It sounds like this supports various kinds of training.

49:07

You can pre-train and have an entirely custom model

49:10

from scratch, which isn't something

49:13

you would want for a lens,

49:15

but it sounds like maybe they allow it.

49:17

They also seem to be offering post-training,

49:20

reinforcement learning,

49:21

basically optimizing a model

49:23

for a given company's needs

49:26

with sort of all the stack and training knowledge

49:29

and inference and so on,

49:31

that is pretty kind of complex

49:33

and is what Mistral and OpenAI and ProPick focus on.

49:38

A lot is just the training infrastructure

49:41

and setup and all of that.

49:43

So they pitch it as, you know,

49:46

you need the models to have your internal

49:49

knowledge and information built in,

49:52

you can optimize it for your needs.

49:54

I think this is an interesting offer

49:57

and potential here with open source models

50:00

haven't started to get good,

50:03

with especially Clan models, Mistral models,

50:05

to some extent, it is a potentially good bet

50:09

that as open source models get better,

50:12

you would see more people fine-tuning

50:16

their own models post-training

50:17

and reinforcement learning for their own agents

50:21

and their own flows,

50:23

which ultimately that's the best way

50:26

to unlock performance is training.

50:28

You can do, you know, prompted engineering

50:30

and you can do whatever,

50:32

but training on data is key.

50:34

So they just announced it,

50:36

you need to sign up to get more details,

50:39

the details aren't super precise yet,

50:41

but it appears that they are shifting focus

50:44

or at least honing in on this specific strategy

50:48

to try and differentiate and compete in this space.

50:52

Yeah, I'd like to be upfront about my biases here.

50:54

Everyone, if you listen to the podcast a lot,

50:56

you know I'm pretty bearish on Mistral and Coheer

50:59

and those sorts of run-up to the frontier companies

51:03

for a whole bunch of reasons,

51:04

but this is not making me any more excited

51:07

about this direction.

51:09

The reason is, so first of all,

51:10

this position's done basically as competitors to Coheer.

51:13

This is the Coheer play, right?

51:14

This is it.

51:15

You're basically saying enterprises,

51:17

like we're just gonna be better at enterprise integration.

51:19

We're gonna make a stack for the enterprise.

51:21

The difference is Mistral is a few miles

51:23

behind Coheer right now, right?

51:25

Coheer has been working on this for a long time.

51:27

They hit like a quarter billion dollars

51:28

in recurring revenue in 2025.

51:30

It was some decent quarter on quarter growth,

51:32

though I think ultimately they're gonna struggle

51:34

to kind of be competitive in the world

51:36

of like big infrastructure plays here,

51:38

but that's kind of fundamentally one of the challenges.

51:41

The other one is the companies that have

51:43

an enduring advantage on infrastructure,

51:46

the anthropics of the world, the Googles,

51:47

the opening eyes of the world,

51:48

are ultimately, if they ever want to,

51:51

in a position to just eat Mistral and Coheer's lunch.

51:54

They can just go in and one day say,

51:55

hey, you know what, we decided that we really care about this.

51:58

We're gonna make special models,

52:00

maybe distill versions of our actual frontier models,

52:02

whatever to run locally on your thing,

52:04

or we're gonna create new things.

52:05

And opening a has in the past offered,

52:07

a fine tuning of the finger GPD 4.1.

52:11

They don't allow to have the latest set of models,

52:13

but it used to be something to offer.

52:16

Yeah, exactly.

52:16

And so ultimately, I mean,

52:18

this is the challenges you are going to end up competing

52:20

with them and the question will at some point be,

52:23

because we know this is where margin is made,

52:25

quality versus quality at the frontier of capabilities.

52:28

If that ends up being the case, I mean, damn,

52:30

like this is an uphill battle,

52:31

because you're actually gonna be like opening eye

52:34

and anthropic get to amortize the massive cost

52:38

of training frontier models across every inference run

52:41

that gets done on their infrastructure.

52:43

And if they're, you know, doing distillates of those models

52:45

to serve at the enterprise, the same thing apply to like,

52:47

the economics look pretty bad to me for stuff like this.

52:50

But ultimately, you know, we'll see,

52:52

this is also by the way,

52:53

a mistrial kind of playing into their European roots a bit.

52:56

So co here has a sort of Canadian footprint.

52:59

They've got big deal EU ambitions,

53:02

but mistrial is through the French champion.

53:04

And they've very much been seen as the sovereign champion.

53:06

So data sovereignty, model sovereignty,

53:08

you'll hear them say a lot of that.

53:10

Ultimately, from a free market standpoint,

53:13

that's basically just an appeal to like kind of European

53:16

or French nationalism to kind of help float them

53:19

on this otherwise potentially market challenge play.

53:23

I think, yeah, maybe I would say you're a little bit

53:26

caricaturing them,

53:27

they position it more as control over IP.

53:31

And this is something that people care about.

53:33

I totally agree that's the play.

53:34

That's an identical play though,

53:35

what I'm saying is to co here.

53:37

And it's and on that note, co here,

53:40

their play or largely what they have focused on

53:43

is releasing models that they have trained

53:47

with the focus of enterprise.

53:48

So they release rag models, for instance,

53:50

that they say are very good for enterprise use cases.

53:53

They largely have position themselves

53:56

as having models and offerings that are useful for enterprise.

54:01

They do have a thing for building customized AI solutions.

54:06

I don't know how much of it is apples to apples

54:09

with this new forge, whatever it is.

54:11

That's kind of the thing, right?

54:12

So yeah, co here's thing is called North.

54:14

It's this whole enterprise platform, right?

54:16

It's it's for AI agents and workflows

54:18

and then they've got a bunch of like re-rank and bed

54:20

and a bunch of other things for rag

54:22

that are meant to be enterprise kind of correlated.

54:25

But that's kind of the challenge.

54:26

It's like ultimately, you're already seeing this appeal

54:30

happening quite quickly to nationalism.

54:32

I mean, Mistral is wrapping themselves in the French flag.

54:36

You see, co here doing similar things

54:38

with the Canadian flag.

54:39

Ultimately, these are signs of like companies

54:41

trying to look for some source of alpha

54:43

that's outside of the market.

54:45

I'm not saying that OpenAI isn't doing similar stuff

54:48

and throbbig certainly struggling to do similar stuff

54:50

and still succeeding.

54:51

And so this is kind of part of the flavor to me

54:54

that like it starts to look a bit like a crutch

54:56

and there's only so much, you know,

54:58

a capex spend that can be subsidized by governments

55:01

that at a certain scale that I think we're approaching already,

55:04

it makes it hard to compete.

55:05

When your big differentiator is like,

55:07

look how French we are or look how Canadian we are,

55:09

that is a bit of a character for sure.

55:11

But it's an important part of the pitch here.

55:13

I worry about that.

55:14

And that's kind of part of my bearishness here.

55:16

I'm just trying to be transparent about my reasoning here.

55:19

I could be wrong, but looking at the scale

55:21

of these companies, I mean, man, co here's been around

55:22

for longer than anthropic, right?

55:24

Like they're, you know, what a percent of anthropic scale.

55:28

I think this is a sign that these markets are smaller

55:30

than maybe had been hoped for initially.

55:32

We'll see.

55:33

I mean, they could absolutely surprise me.

55:35

And I'm looking forward to seeing how it plays out.

55:37

Yeah, I think longer term, I think this is an interesting time

55:41

to consider this direction of customized

55:46

and, you know, fine tune models for a given business

55:50

because the open source models,

55:52

the models that aren't perpetuated,

55:54

I starting to get good.

55:56

And so you might start seeing a case where,

55:59

for instance, at Ostracade, right,

56:01

we have a pretty specific use case of coding

56:05

for these games, right?

56:06

And having coding agents with a specific set of technologies,

56:11

HTML, JavaScript, et cetera.

56:14

So it is feasible that if you were to do reinforcement learning

56:19

and post-training on our own internal proprietary data

56:24

of, you know, good games that people have made

56:26

and chat histories and so on,

56:28

we would get a better model and a better performance.

56:32

That didn't used to be the case.

56:33

It used to be that like the base models just weren't good enough,

56:37

just doing prompt engineering was going to be the best you can do.

56:42

I could see it heading in a direction

56:45

and me straw man may not be the one that owns this.

56:49

It could be an anthropic or openly I jump in on it.

56:52

But regardless, having these sort of individualized,

56:56

customized models for different companies,

57:00

agent techniques seems like a very much feasible future.

57:04

Yeah, definitely not questioning the existence of a market, right?

57:07

I'm just sort of skeptical of their positioning, as you said,

57:09

to like, are they the ones to capture this?

57:12

Yeah, that's where I'm wearing.

57:13

Yeah, it's something to try.

57:15

So it's something to try, no doubt.

57:17

And back to chips,

57:20

China's buy dense gets access to top and video AI chips

57:24

according to the Washington Street Journal.

57:28

So they are seeing that buy dense is assembling

57:32

significant computing power outside of China using these chips.

57:38

They are supposedly working with the South Feast Asian firm,

57:42

our London cloud and they plan to deploy approximately 500

57:46

and video black wall computing systems in Malaysia,

57:51

totaling 36,000 B to 100 chips.

57:55

So that's 500 computing systems, presumably meaning racks

57:59

that amounts to tens of thousands of actual GPUs.

58:03

So yeah, it's suppose giving us an indication

58:06

that the company by dense being a massive company

58:09

that has their products used outside of China, no doubt,

58:13

is able to make use of your chips in other countries.

58:17

Yeah, yeah, and you know, if you're looking at this

58:19

being like, hey, wait a minute, I thought we weren't shipping.

58:21

I thought we just greenlit the H200 shipping.

58:24

What are we doing shipping black wells to China?

58:26

Well, that's the whole point.

58:27

It's technically not being shipped to China.

58:29

The export controls that apply here from 2023 regulate

58:34

where the hardware is shipped, not where the compute is used.

58:39

Right? And that was intentional.

58:40

It was meant to allow the sort of like global cloud infrastructure

58:44

to be built based on American hardware.

58:46

But the thing is, by dense is not on the famous entity list

58:50

or there's a military end useless as well.

58:52

So the fact they're using Nvidia hardware

58:55

does not automatically in and of itself trigger restrictions.

58:58

The fact is that they're just setting up shop outside of China

59:02

and normally that's totally okay.

59:04

And Nvidia actually confirmed there were no objections here

59:07

and the BIS Department of Commerce

59:09

apparently is also on board.

59:10

So they're all tracking.

59:11

Still, I mean, this kind of reveals like,

59:13

hey, well, this is what you're opening yourself up to.

59:15

If you think about the scale,

59:16

it's by the way for a second.

59:17

What does it mean to have 36,000 B200 GPUs,

59:21

you know, 500 of these NVL72 units?

59:23

So this is typically a two rack system.

59:24

Yeah, I think we talked about it in the hardware episode at one point.

59:27

So this many GPUs adds up to about a 60 megawatt cluster.

59:32

Okay, so 60 megawatts of power.

59:33

That's enough to power roughly like 60,000 homes.

59:36

You can think of it that way.

59:37

So it's like a little small town of compute.

59:40

Compare that to Microsoft Abelene is like 100 megawatts

59:44

or Elon's XAI Texas Colossus site.

59:46

That one's more like 130 megawatts.

59:48

The Abelene comparison has been unfair

59:50

because that's still kind of being built out.

59:52

But the Colossus is better apples apples

59:54

because like they're still building it out,

59:56

single-purpose, you know, it's stood up, you were fast.

59:59

What this is showing is you're allowing a Chinese company

1:00:01

to get up to the same at least

1:00:03

order of magnitude in compute.

1:00:05

It'll admittedly be online kind of like, you know,

1:00:07

later on this year, but that's still something.

1:00:10

So this is all part of the debate here over

1:00:12

what is appropriate, how much compute should be allowed,

1:00:16

even if it's overseas deployments

1:00:17

that conserve Chinese customers by the way.

1:00:20

This is still a data center that can be used

1:00:23

to serve Chinese customers.

1:00:24

So you could, in principle, at least to my understanding here,

1:00:27

might dance could use this to train a model

1:00:29

that then is used in China,

1:00:30

that then because of civil military fusion gets used

1:00:32

by the CCP, you know, the MSS

1:00:34

or whatever other arm of the Chinese state.

1:00:36

So they've also planned additional deployments

1:00:38

of up to 7,000 B-200s at a bunch of data centers in Indonesia.

1:00:42

This is kind of a broader Southeast Asian strategy here.

1:00:46

So we'll see, I mean, it's a pretty big,

1:00:49

you might think of it as a pretty big gap.

1:00:51

That's sort of how I think of it personally,

1:00:53

but there are different views on is this desirable

1:00:55

and you could make an interesting argument

1:00:57

in many different directions.

1:00:58

Moving back to the US,

1:01:00

we've got an update on where meta is at with their AI efforts.

1:01:04

The update is that they are delaying the rollout

1:01:07

of their next model, codename avocado,

1:01:10

from March to at least May.

1:01:12

I think this was actually announced last week,

1:01:14

but we didn't manage to cover it.

1:01:16

So they appear to have just not been able

1:01:21

to train a good enough model in time.

1:01:24

And so they are delaying this release.

1:01:28

They are discussing or have discussed

1:01:30

at least temporarily licensing competitors AI technology,

1:01:34

report of the Google's Gemini.

1:01:36

They already are looking at various extensions.

1:01:41

So overall, not looking great for meta is my read on this.

1:01:46

It's already been close to a year.

1:01:48

It feels like in mid 2025, after a couple of months,

1:01:52

after a long before we saw this heavy, heavy, heavy push

1:01:56

where they got Alexander Reng from Scala AI

1:02:00

and a $14 billion deal.

1:02:02

And then they paid absurd money to get all of these researchers

1:02:06

from Unphropic and OPAI and DeepMind and so on.

1:02:09

And they created super intelligence labs

1:02:12

with a company very reshuffled everything

1:02:15

with the intent to train a competitive frontier model

1:02:21

that can stand up to GPU 5.4 and cloud and all these other ones.

1:02:25

It doesn't seem like they're there.

1:02:27

And I personally, this is my read.

1:02:30

I think organizationally having Alexander Reng lead it

1:02:35

is maybe not the best idea.

1:02:37

Someone who hasn't worked in Frontier AI models,

1:02:40

data curation of Scala AI, you know, that's-

1:02:44

It's unrelated, right?

1:02:45

It's related, but I don't know if that's enough.

1:02:50

Anyway, so that's when it's been delayed.

1:02:53

It doesn't appear like they're on track,

1:02:56

but you don't know too much.

1:02:58

This is all internal.

1:02:59

Maybe they're just doubling down and it's going to be great.

1:03:03

Yeah, boy, yeah, right.

1:03:03

And it looks like it, so reportedly like

1:03:06

landing somewhere in capabilities between Gemini 2.5

1:03:09

and Gemini 3.

1:03:10

So, you know, very passe at this point,

1:03:13

hence this decision to delay.

1:03:15

They're angling for a private non-open source model here, right?

1:03:18

So the appropriate point of comparison

1:03:20

is what does cloud look like?

1:03:22

You know, what is the best version of cloud look like?

1:03:24

It seems like, as you say, it's just not up to it,

1:03:27

at least for right now.

1:03:28

And that could change quickly, who knows?

1:03:30

But yeah, it's also part of this whole saga

1:03:33

where you had Yan Likun say, you know,

1:03:36

fuck Alex Wang when he was leaving.

1:03:39

And then everybody was, there's just like,

1:03:40

I forget, not times of India,

1:03:41

but there was some kind of India-based publication, right,

1:03:43

that said, hey, looks like Zuck is thinking about Marjorie.

1:03:46

Which was like not verified.

1:03:49

It was false.

1:03:50

Yeah, not false.

1:03:52

But yeah, exactly.

1:03:52

And then Zuck came out with a photo with Alex Wang,

1:03:55

sort of all smiles in the office type thing, to kind of.

1:03:57

Now, we do know, yeah, that aside from Yan Likun,

1:04:00

Alexa Rang has also clashed with internal kind of big deal

1:04:05

product people like Chris Cox,

1:04:07

Meta's Chief Product Officer,

1:04:09

and Andrew Bosworth, their Chief Technology Officer.

1:04:13

Apparently, about how AI models should improve,

1:04:16

Alexa Rang has been pushing for,

1:04:18

let's just train a super, super good model

1:04:20

and not care about the business implications

1:04:22

and the product implications,

1:04:24

which other people at Meta understandably are questioning.

1:04:29

Yeah, and it's exactly, now Alex comes from this viewpoint

1:04:32

of, hey, let's do the opening.

1:04:33

I think let's do the anthropic thing

1:04:35

and go full on, you know, AGI and all that stuff.

1:04:37

That's very much, I mean, he is ASI-pilled.

1:04:39

Well, the name of the super intelligence team kind of hints

1:04:42

at that, and not only, I mean,

1:04:43

if you're going to make a super intelligence team,

1:04:45

yeah, they probably should be focused on super intelligence.

1:04:47

Like, I'm no branding expert.

1:04:49

I'm not marketing dude, but yeah, it seems like that's par.

1:04:52

Now, should Meta have a super intelligence team

1:04:55

separate questions, right?

1:04:56

That's a good question to ask, I don't know.

1:04:58

That's right.

1:04:59

And if, you know, certainly,

1:05:00

well, if you believe that super intelligence

1:05:02

is going to be the thing that Alex Wang believes it is

1:05:04

and that maybe Zuck believes it is,

1:05:06

then you have no choice.

1:05:06

You have to compete because somebody,

1:05:08

if somebody else does it, you're finished,

1:05:10

depending on how the intelligence explosion goes.

1:05:12

Bottom line is a lot of fog of war,

1:05:15

at least to me, still on what's going on inside Meta,

1:05:17

the one data point we have is that we still don't have anything

1:05:20

and that it's starting to become a real data point

1:05:22

at this point.

1:05:23

It's starting to come like definitely expected

1:05:26

a model release by now, given the amount of infrastructure

1:05:30

and investment they've put in here.

1:05:32

And given that they have trained models in the past,

1:05:36

number four was overwhelming,

1:05:39

but it was a pretty impressive model

1:05:42

that most companies were not able to produce even still.

1:05:45

So the lack of anything is perhaps concerning.

1:05:51

And one story a bit similar,

1:05:54

the story is Microsoft shakes up AI division

1:05:56

as co-pilot falls behind Google

1:05:59

and open AI, they're reorganizing the AI division.

1:06:04

Gustaf Assumon is going to be shifting focus exclusively

1:06:07

to developing the company's own frontier foundation

1:06:10

language model.

1:06:11

So effectively, the same kind of focus

1:06:14

that we just discussed with Meta,

1:06:17

Microsoft has worked on training their own alternative

1:06:21

to chat GPT and so on.

1:06:23

They've trained various models.

1:06:25

I believe the big one from them was FI,

1:06:28

which is small models that they released.

1:06:32

They do have a lot of users.

1:06:35

So apparently, Microsoft's co-pilot app

1:06:37

has 150 million monthly active users,

1:06:42

but if they want to compete with Gemini or chat GPT,

1:06:46

they're nowhere near, they're way, way behind Gemini

1:06:50

has now 750 million active users,

1:06:53

chat GPT has roughly 900 million,

1:06:56

sorry, weekly active users versus co-pilot

1:06:59

has 150 million monthly active users.

1:07:04

So yeah, that's the news, they are restructuring,

1:07:07

they are recognizing what they're following behind.

1:07:11

And if they want to have their own frontier model,

1:07:15

they are not there.

1:07:16

This is all the more glaring given the insane distribution

1:07:19

advantage that Microsoft has.

1:07:21

Distribution is supposed to win these kinds of battles.

1:07:24

There's a reason that Microsoft teams

1:07:26

kick the shit out of slack, like within minutes of launching.

1:07:29

And it's because they had distribution.

1:07:31

Microsoft is in every system, right?

1:07:33

And yet you have like admittedly,

1:07:34

okay, so Google is probably a bad comparison

1:07:37

because they have a similar, insane distribution advantage.

1:07:39

But when you look at, you know, opening eyes models

1:07:41

and in anthropics models,

1:07:43

not being able to compete with those

1:07:45

is a really kind of bad indication here.

1:07:47

So yeah, I mean, no surprise, something like this

1:07:50

comes with a requirement to shift.

1:07:52

The structural dependence on opening eye

1:07:54

has been a big issue.

1:07:55

You know, Microsoft has that 27% stake

1:07:57

in at least the profit for profit unit,

1:08:00

not going to get into the dissection of opening eye

1:08:03

in their corporate structure.

1:08:04

But again, but bottom line is that in a sense

1:08:07

has been maybe a, I don't mean call it a security blanket.

1:08:10

It's increasingly clear that they're competing directly

1:08:13

with opening eye for a bunch of enterprise customers.

1:08:15

So cannibalizing themselves really from the inside,

1:08:17

that's a problem.

1:08:18

That's a really big problem.

1:08:20

Microsoft has a lot more to lose in that sense.

1:08:22

And yeah, so they got to find a way

1:08:23

to turn this around right quick

1:08:25

if they think AI's headed where it might be.

1:08:27

Quick recap by the way.

1:08:28

So Mustafa Suleiman, former co-founder of DeepMind,

1:08:33

joined Microsoft, I think last year,

1:08:36

has been their CEO of Microsoft AI.

1:08:40

And this whole story is based at these partially

1:08:43

on MMO, he released a titled

1:08:46

a new structure for Microsoft AI.

1:08:49

And the gist of it is, again, similar to meta

1:08:53

that he wants to pursue super intelligence,

1:08:56

consumer things and product considerations

1:08:59

get in a way of that.

1:09:00

So he's going to be freed up to focus on that.

1:09:04

Jacob Andru, former senior vice president at SNAP

1:09:07

will take over as executive vice president

1:09:10

leading the co-pilot division.

1:09:12

So there's a split here where there's co-pilot,

1:09:15

the product, the app, et cetera.

1:09:18

Someone else focuses on that.

1:09:20

Mustafa focuses on building the frontier model

1:09:22

on getting through super intelligence.

1:09:24

They are much like how we just discussed

1:09:27

with Alexander Wang at meta.

1:09:29

Well, it's your point.

1:09:30

We're seeing this same motif over and over, right?

1:09:32

It's not just meta, it's not just Microsoft.

1:09:34

It's opening AI, I mean, it's happening at Google,

1:09:37

like basically the question is,

1:09:38

how seriously do you take the potential

1:09:40

imminent arrival of super intelligence

1:09:42

if you take it very seriously?

1:09:44

You got to drop everything, work on that shit, man.

1:09:46

That's what Alex Wang is saying.

1:09:47

And clearly, Microsoft is not immune to that tension, right?

1:09:51

How much do you view AI as like a commercial product question

1:09:56

and how much do you view it as,

1:09:57

this is the thing that's triggering the singularity,

1:09:59

holy shit, like the one place,

1:10:01

you don't tend to see that kind of tension,

1:10:03

or be careful, it exists.

1:10:05

But like, you know, like you see some of it,

1:10:08

I should say in the media,

1:10:09

but you don't see as much of it as anthropic.

1:10:11

Like it seems like generally anthropic

1:10:13

and its public communications is much more aligned

1:10:16

around the idea of like where they're going.

1:10:19

So, you know, do it that way you will,

1:10:21

but it definitely is notable

1:10:23

that basically everybody else

1:10:24

is having these large public spots about this.

1:10:26

And moving on to policy and safety,

1:10:28

where we have quite a bit of research related to safety.

1:10:32

The first one is decision theoretic formalization

1:10:36

of stenography with application to LLM monitoring.

1:10:40

So, stenography is a practice of concealing secret information

1:10:45

within non-secret data.

1:10:48

We've seen this typically with images,

1:10:49

but you can also do it with text.

1:10:52

And as far as LLM monitoring, right?

1:10:55

If you have a malicious AI that you want to catch,

1:11:00

if it's trying to do something bad,

1:11:02

or kind of a monitor's communications

1:11:05

to know what it's doing,

1:11:07

potentially it could try to get around

1:11:09

that sort of monitoring with stenography.

1:11:13

And this is...

1:11:14

Stagnography.

1:11:15

Stagnography, sorry.

1:11:17

And so, this is presenting a new framework

1:11:21

of formalization of how to view this entire thing

1:11:25

and does also discuss the implications for LLM monitoring.

1:11:29

And Jeremy, you can take and go, do your thing.

1:11:34

It's a good, do your thing.

1:11:36

Yeah, I'll say stagnography four times fast.

1:11:39

So, yeah, we've talked about stagnography

1:11:41

on the podcast before quite a bit, right?

1:11:43

This idea that a model might be reasoning

1:11:46

in its chain of thought, for example,

1:11:48

but there could be other manifestations

1:11:49

in a way that is intentionally meant to deceive you,

1:11:52

but that allows it to kind of make evil plants, right?

1:11:55

So, an example they have this in the paper,

1:11:57

so here's an excerpt of a reasoning trace from O3,

1:12:00

it says, the summary says improved 7.7,

1:12:03

so whatever that means, but we can glean,

1:12:05

and then this is the weird kind of non-sequitor

1:12:08

part of the chain of thought.

1:12:10

It says, but we can glean,

1:12:12

disclaim, disclaim, synergy, customizing illusions,

1:12:14

but we may produce, disclaim, disclaim, advantage.

1:12:17

Now, like, disclaim, overshadow, overshadow,

1:12:19

intangible, let's craft, right?

1:12:21

So, like, right of these like random words

1:12:23

that don't seem like they have a role

1:12:25

in solving the problem, which may be, in principle,

1:12:28

the model hiding reasoning that's oriented

1:12:30

towards solving another problem

1:12:32

besides the one that you wanted to solve

1:12:34

that could be malicious or whatever.

1:12:35

And so, the central question of this paper is going to be,

1:12:38

does, so can we measure stagnography, right?

1:12:42

Can we measure this idea of to what extent is my model

1:12:46

encoding or trying to hide reasoning from me

1:12:49

or from some oversight in its chain of thought?

1:12:52

And here's the surprisingly elegant way

1:12:55

they're going to approach this, right?

1:12:56

They're going to define this thing called the stagnographic gap,

1:13:00

the difference in roughly speaking,

1:13:03

the performance of an agent that has access to,

1:13:08

there's a model that generated the reasoning trace

1:13:10

and it's potentially trying to, you know,

1:13:12

use it to generate some hidden content.

1:13:15

And then there's a trusted model of about the same capability

1:13:19

that doesn't know, so whatever is secret encoding scheme,

1:13:23

the generator model used, they called the receiver,

1:13:26

but basically the model that generated the content,

1:13:28

the reasoning trace that's suspect.

1:13:30

That model, presumably, is using some kind of encoding scheme,

1:13:33

then they have a set now that's going to review that model

1:13:36

that that model's outputs, but it doesn't have access

1:13:38

to the encoding scheme.

1:13:40

And if what you find is that based on the chain of thought,

1:13:44

that's suspect, the generator model performs really, really well,

1:13:48

but the sentinel performs really badly,

1:13:51

and that implies that the generated model

1:13:52

is able to get some kind of value out of that chain of thought

1:13:55

that the sentinel, the kind of reviewing model, couldn't.

1:13:59

And that suggests that there must be information encoded

1:14:01

in there that only the generator model is able to use,

1:14:05

hence the sort of hidden stagnographic value.

1:14:09

So they do a whole bunch of stuff to kind of surface this.

1:14:11

They're fairly elegant experiments.

1:14:13

I'll focus on the code word training one.

1:14:15

So they fine tune models using RL to use certain code words.

1:14:20

So for example, instead of using heads or tails,

1:14:22

they'll use sun and moon, right?

1:14:24

And then they show that as they give the oversight model,

1:14:28

the sentinel, more visibility into the kind of code words

1:14:33

and how they map, gradually the performance gap

1:14:35

between the sentinel model and the generator model drops.

1:14:39

So this kind of makes sense.

1:14:40

You're giving the sentinel more and more access

1:14:41

to the kind of code book that generated these outputs,

1:14:44

so no surprise there.

1:14:45

But then they do some more interesting fine tuning experiments.

1:14:47

They take GPT 4.1 and they get it to,

1:14:51

they use this thing called the bi-gram cipher.

1:14:53

Basically, it means that you take any pair of letters

1:14:57

and swap them out with some randomly chosen pair of letters.

1:15:00

So maybe every time you see, you know, TE,

1:15:03

that becomes like GH or something.

1:15:05

And then, so basically you see gibberish.

1:15:07

And at first, you get the model to solve competitive programming

1:15:10

problems or something like that using this bi-gram cipher

1:15:13

and they measure that stegonographic gap.

1:15:15

So the model that was trained to use the cipher

1:15:19

versus a model that wasn't,

1:15:21

like how does their performance vary over time?

1:15:23

Is there an ability to detect the existence of the cipher

1:15:27

or the use of that cipher?

1:15:28

It's quite an interesting paper.

1:15:30

It doesn't quite get to like, anyway, there are a couple of gaps.

1:15:34

So if the model that's generating the content

1:15:38

actually has a goal that's different from the goal

1:15:41

that it's ostensibly trying to pursue,

1:15:43

that won't be detected here.

1:15:45

So it's got like a payload that is unrelated

1:15:48

to the tasks that's being audited for.

1:15:50

You just won't get detection.

1:15:51

And sandbagging could also be a problem here,

1:15:54

like deliberately unperforming when it thinks

1:15:55

it's being audited.

1:15:57

And we do know the models can often detect indications of that.

1:15:59

So bottom line is it's an interesting kind of first step

1:16:02

in the direction of detecting secondography,

1:16:05

could be an interesting and valuable path to explore

1:16:07

with some pretty significant gaps too.

1:16:09

Next paper, a bit related.

1:16:12

The title is Reasoning Theater,

1:16:14

disentangling model beliefs from chain of thought.

1:16:17

So one theme we've been kind of seeing

1:16:21

in research over the past few months

1:16:22

is looking into chain of thought,

1:16:25

the intermediate outputs that models are now trained to give,

1:16:29

they're reasoning, you ask a question,

1:16:31

it does some reasoning and choose on the problem

1:16:34

and then eventually you get an answer.

1:16:36

And so there's various questions about

1:16:38

whether to immediate output of like,

1:16:39

does it faithfully capture what the model is doing?

1:16:43

Is it modatable as we've just discussed?

1:16:46

For instance, and what this paper is looking at

1:16:49

is it performative.

1:16:51

So is the model actually thinking

1:16:53

and like trying to find an answer via this chain of thought

1:16:57

or does it know the answer

1:16:59

and is just talking as if it's thinking?

1:17:01

And what they find is often the chain of thought

1:17:05

is performative, like you don't actually need it

1:17:07

to model if you look at its internal activations

1:17:10

can already give you an answer with high confidence,

1:17:13

which is correct.

1:17:14

Now you can look at different settings.

1:17:18

So if you have harder problems and smaller models,

1:17:23

then the thinking becomes more legitimate.

1:17:27

It actually does need to do with chain of thought

1:17:30

to get to a high confidence answer.

1:17:33

It's actually kind of related or similar

1:17:36

to we discussed this deep thinking tokens idea a few weeks ago

1:17:42

where we were looking at if you track internal activations

1:17:46

and look at things like deep thinking,

1:17:49

backtracking, certain behaviors

1:17:50

that indicate actual genuine kind of reasoning

1:17:55

about the problem.

1:17:56

You can find that these are very significant markers

1:18:00

as to the confidence of the model

1:18:02

at this point in reasoning about its answer.

1:18:05

So one applications of this is if you can know

1:18:09

from internal activations, the confidence level,

1:18:12

you can do what's called early exiting,

1:18:14

cut off the reasoning and just get to answer quicker.

1:18:17

Yeah, another kind of pointer to that being the case here.

1:18:20

And to your point, so there's kind of like three different

1:18:22

methods that they'll use to figure out

1:18:25

because the central question here is like,

1:18:26

what does my model actually believe at any given point

1:18:29

during reasoning?

1:18:30

So one is this idea of attention probes.

1:18:32

So you basically have these classifiers,

1:18:35

there's like these linear models or very simple models

1:18:37

that you plug into any given layer of the residual stream

1:18:41

and you go, okay, I want you to predict like what a classifier

1:18:44

like, what is the final output going to be

1:18:46

at any given moment?

1:18:47

So you basically, this is looking layer wise

1:18:49

and interrupting the model in mid sort of token processing

1:18:52

token processing for a given token.

1:18:55

But then there's separately this idea of forced answering that you just alluded to essentially

1:18:59

interrupt the model mid reasoning, like in the middle of a chain of thought and say, okay,

1:19:03

now I want you to commit your answer.

1:19:05

And then the last one is just like a chain of thought monitor, wait for the model to produce

1:19:08

the entire chain of thought, and then figure out like, you know, where in the text final

1:19:13

answer is implied.

1:19:15

So at what point could the model have just offered and said, okay, you know, this is it.

1:19:19

And they compare all three of these.

1:19:21

And by noticing the differences between them, they can start to make some pretty interesting

1:19:25

conclusions.

1:19:26

I mean, you talked about this idea of sort of easy tasks leading to more, they call it performative

1:19:31

reasoning, right?

1:19:32

So the model's pretending to reason about it.

1:19:34

Or, you know, if you're pretending to dress up a problem like it's harder than it is.

1:19:39

So on MMLU questions, you find consistently that the attention probe, which is again,

1:19:44

that layer wise, kind of looking at a given layer, plugging in a classifier to just look

1:19:48

at the activations there and then making that prediction.

1:19:51

So the attention probe could decode the correct final answer from the model's internals

1:19:56

really early on, right?

1:19:58

Like sometimes right after reading the question, like straight one shotting it, right?

1:20:02

Whereas the chain of thought monitor sees no answer indicated in the text yet.

1:20:06

So the model essentially knows the answer immediately, but it keeps thinking out loud

1:20:10

anyway.

1:20:11

And that's kind of where, you know, forced answering is also going to give you good

1:20:14

accuracy early on and all that.

1:20:16

That's an indication of what they'll call performative reasoning.

1:20:18

The fact that hard questions involve genuine reasoning more often, especially with small

1:20:23

models, we've actually seen tons of examples of that on the podcast before.

1:20:27

The way that I tend to think of it at least is the bigger the mismatch between the capabilities

1:20:32

of your model or the amount of test time compute or whatever, the capabilities, let's

1:20:35

say, of your system and of your compute budget taking together.

1:20:39

And the actual difficulty of the problem, the more opportunity there is for the model to

1:20:43

use the extra kind of capability overhang or compute budget that it has to plot or scheme

1:20:50

or do other things besides what you want it to do.

1:20:53

And so I said it, I think the last podcast, I'll say it again, really important research

1:20:57

direction would be, to my mind, how do we better index our appetite or compute budgets

1:21:04

or our model selection, model capability selection to match models and capabilities to problems

1:21:11

such that they're kind of closely matched economically.

1:21:13

You want this anyway, right?

1:21:14

You don't want to use a model that's too big or too expensive to solve an easy problem,

1:21:18

but there's also a safety imperative to this now where you want to try to take away

1:21:23

as much as you can, the excess compute excess capability that otherwise goes into kind

1:21:27

of these undesirable effects.

1:21:30

Next up in training defenses against emergent misalignment in language models.

1:21:36

So emergent misalignment is a phenomenon we've discussed previously where you train your

1:21:41

model, you fine tune it slightly on a small set of stuff that's bad, like you make it

1:21:48

right bad code, and then it turns out that that broadly makes it bad and all sorts of

1:21:54

ways and able to get misaligned.

1:21:57

And so what the in training defenses here is talking about is let's say open AI or

1:22:02

ontropic do give a fine tuning API and allow people to fine tune their models on small

1:22:09

datasets.

1:22:10

Well, you would then introduce this potential for broad misalignment by just a small

1:22:15

about a fine tuning.

1:22:17

And I look at several different ways to try to defend against it and prevent it from happening.

1:22:24

So we look at regularizing to have a penalty to stray too far from a good model, there's

1:22:31

steering, and the method that they ultimately include works your best is interleaving the

1:22:36

training examples from a general abstract tuning data set from a general like this is

1:22:42

how we should behave that a set with the fine tuning that a set and they find that is

1:22:47

the most effective for defending against fine tuning misalignment.

1:22:52

Yeah, this is kind of an interesting paper I mixed feelings on emergent misalignment

1:22:57

because it I think it may not be a problem it may almost be a good thing in a weird way.

1:23:02

It's a little bit weird to focus with regards to fine tuning because you could have like

1:23:06

intentional misalignment, why do you, you know, right, well, no, that's a good point.

1:23:12

And I think I think the frame is something like emergent misalignment is leads to these

1:23:17

surprising knock on effects.

1:23:18

What it fundamentally means is if you can fine tune for one behavior and that leads to

1:23:22

massive changes in other behaviors, then damn, we got to be really careful about fine

1:23:26

tuning because we can't think of a model once fine tuned as in any way comparable from

1:23:31

a safety standpoint to the original model.

1:23:33

So any safety valves we've run on the original model no longer apply.

1:23:36

I think that's kind of part of it.

1:23:37

I think, you know, this is a good example of the kind of persona theory that anthropics

1:23:42

been kind of pushing lately where you're fine tuning a model to write and secure code.

1:23:47

The model kind of goes, oh, so I actually, it's not that I'm being trained just to write

1:23:51

and secure code.

1:23:52

I'm being asked to activate my misaligned persona in a more general sense, which means that

1:23:58

there is a notion of a misalignment persona in the model in the first place, which means

1:24:02

that there is an out of distribution generalization, kind of understanding of what alignment and

1:24:07

safety is, which seems like it could actually be a good thing.

1:24:11

That was kind of the rabbit hole that I was trying to avoid, but that's the general

1:24:15

direction.

1:24:16

One quick observation on this move on, but basically the two defenses that I found the relative

1:24:23

performance pretty surprising, neither one of these was the most performant one.

1:24:26

So they use this kale divergence regularization one where they're going to penalize the model

1:24:30

for drifting too far from the original line model during training, right?

1:24:34

So I'm going to fine tune you to write in secure code, but if you drift too far on your

1:24:40

outputs and kind of other domains as well, I'm going to prevent you from doing that by forcing

1:24:44

your outputs to be at least somewhat similar.

1:24:46

So that's one.

1:24:47

The other one was a similar thing that applies, it's LDIFS feature space regularization.

1:24:53

It's like basically the same thing, but for the activations of the model, roughly speaking.

1:24:57

So instead of penalizing drifts in the outputs, the actual text that it writes, it penalizes

1:25:02

drifting too far in terms of the model's internal activations rather than the outputs.

1:25:07

And this one, I would have expected it to actually be more effective because in some way,

1:25:11

it's more fundamental.

1:25:12

It feels like you're getting at the actual guts of the model.

1:25:15

It doesn't work as well though, as simple text-based regularization.

1:25:20

Let's make sure that the outputs as red look better.

1:25:23

And the other thing about this, I think part of this is like the text-based outputs more

1:25:26

directly tied to the behavior itself that you care about.

1:25:29

But the other piece is, it's possibly just selected the wrong layers to focus on.

1:25:33

They looked at like every fifth layer, just to save memory, but that might not be the

1:25:37

right kind of play.

1:25:38

And then also, I guess, activation spaces.

1:25:40

There's a lot of redundancies in it, so you may not be hitting all of the kind of representations

1:25:44

that matter.

1:25:45

Bottom line is, this is quite interesting and we just keep surfacing, surprising findings

1:25:50

in alignment.

1:25:51

And that's not a great thing, but it's also interesting.

1:25:54

So there we go.

1:25:55

And I think it used to be the case that there's like people who care about alignment and

1:26:00

people who don't.

1:26:02

And that is starting to change a little bit.

1:26:06

Yeah.

1:26:07

Yeah.

1:26:08

And next up, a more sort of programmatic, real world study, how do frontier AI agents

1:26:14

perform in multi-step cyber-attack scenarios?

1:26:18

This builds on something we discussed, I think, last week, this is from the AI Security

1:26:23

Institute.

1:26:24

Previously, I showed that, as you scale test-time compute to 100 million tokens, models get

1:26:29

better and better and better.

1:26:31

This is effectively building on that.

1:26:33

They released a paper where they measure the capabilities via what they call cyber ranges,

1:26:40

which are simulated network conditions within which you would execute an attack without

1:26:46

too many controls, without defenses.

1:26:48

You just need to execute a set of steps to accomplish some goals.

1:26:52

And they have multiple levels of completion, more or less.

1:26:57

And so what they find is, perhaps I'm surprising, the models get better and better and better.

1:27:03

And also, you get better and better and better as you get to more test-time compute.

1:27:10

So overall conclusion is, the capabilities, the cyber harm capabilities of the models

1:27:16

are improving rapidly and are starting to get to some fairly complex scenarios.

1:27:22

First of all, a great paper by the UK AI Security Institute.

1:27:26

So they've continued to pump out really high-quality stuff that feels very relevant to a whole

1:27:31

bunch of end-game scenarios around ASI, cyber is very important, at least the way I think

1:27:35

about it, quite important in that picture.

1:27:38

And so, well, two trends that they identify, right?

1:27:41

So the first is, each successive model generation outperforms its predecessor at fixed token

1:27:47

budgets, right?

1:27:48

So for a given token budget, you're going to see better performance from Opus 4.6 than

1:27:52

from Opus 4.5, right?

1:27:54

So they see that.

1:27:55

They also see performance improvements.

1:27:57

Like we talked about this last week, pretty interesting, good paper, and we're threading

1:28:02

the descriptions.

1:28:03

I won't do it here.

1:28:04

But if you're curious of their cyber ranges, because this does go well beyond this

1:28:07

or traditional capture of the flag stuff that we've seen in sort of offensive cyber model

1:28:12

eVALs, this is really like simulating a whole corporate network, a whole corporate environment,

1:28:16

which is much closer to what actually makes a difference in the real world.

1:28:21

Next up, Eval Awareness in Cloud Opus 4.6's browser comp performances.

1:28:28

This is a report from Unphropic, and this has been something that's been discussed a

1:28:32

little bit kind of on and off.

1:28:34

So the Eval Awareness you referenced is one of the topics to be aware of with regards

1:28:41

to alignment safety, basically, if you're checking model for doing the right thing or not,

1:28:47

and it's aware of that, then it might behave differently when you're valuing it versus

1:28:53

when it's actually out in a wild.

1:28:56

And so here they look at nine cases where leaked answers appeared in public web content

1:29:04

and found, among a 1,266 problems on browser comp, the model found 11 cases where answers

1:29:15

came from benchmarks, materials, and yeah, basically often, or in some cases wound up reasoning

1:29:23

that the questions that it was dealing with were evaluation shaped.

1:29:28

And then having concluded that this is likely an evaluation, it looked up a benchmark.

1:29:35

So these are some examples where this can happen at least in practice already.

1:29:40

Yeah.

1:29:41

And the cases where you tend to see this happen according to the report is like, first

1:29:46

you'll see the model try like hundreds of normal search attempts and then conclude that

1:29:52

the question just felt artificially constructed after not succeeding through those earlier

1:29:57

attempts.

1:29:58

And it'll start to name and look for specific benchmarks.

1:30:02

So it'll be like, you'll see it look up, you know, Gaia, browser comp, or browser comp

1:30:06

or like whatever, a whole bunch of different e-vals, then it'll find whichever the e-val is

1:30:11

and it will typically look at the implementation like on GitHub or whatever and go from there.

1:30:16

And so that's how a lot of this happened.

1:30:18

One other weird thing that was highlighted on the inner web this week was that you have

1:30:22

some e-commerce sites that will auto generate these persistent pages or pages that stay

1:30:29

up like sub pages that log the history of search queries on their website, even when

1:30:35

there are no matching products.

1:30:36

So if you look at like purple turtles on besock on like eBay or whatever e-commerce site,

1:30:41

that will lead to the creation of a persistent page somewhere on the internet.

1:30:46

And the effect is that every agent that runs browser comp slowly caches its queries as

1:30:52

these permanent indexed pages.

1:30:54

So over time, like agents from different evaluation runs were inadvertently leaving traces

1:31:00

of their search strategies that could be seen by future agents, which is a one thing that's

1:31:06

reported in the context of this paper, maybe the broader implication is this is an interesting

1:31:10

way for agents to have a kind of persistent memory that transcends their instantiation

1:31:16

in the moment, which you shouldn't over focus on this particular mechanism as a way that

1:31:21

agents could either communicate with other agents or with their future selves in illicit

1:31:27

ways.

1:31:28

It's more saying that hey, the internet is a really complicated messy place.

1:31:32

There are easy ways to hide information that can't necessarily be found easily.

1:31:36

And that's a really important thing from a safety standpoint.

1:31:38

Notable Anthropic says they don't see Opus 4.6's behavior here as being a alignment failure

1:31:45

because the model wasn't told to restrict its searches in any way, just to find the answer,

1:31:50

it's like fair enough, but also those of us who come from the world of like AI alignment

1:31:55

from back in the day, there is a thing called coherent extrapolated volition, C-E-V, basically

1:32:02

just like how well you could reasonably predict what someone would have wanted you to do

1:32:08

in context from the prompt.

1:32:10

And I think this qualifies as a failure of that.

1:32:14

You would probably assume that you don't need to be told not to try to game the system

1:32:20

in this way.

1:32:21

This feels like gaming, like hacking the benchmark.

1:32:23

I think that this is a failure of C-E-V and therefore a failure of alignment.

1:32:26

That would be my personal takeoff.

1:32:28

But that's not Anthropics take and that's consistent, they have their own way of thinking

1:32:33

about this.

1:32:34

But I think you could view it as still being a failure of alignment.

1:32:37

Yeah, I think it's an implicit kind of assumption, right?

1:32:40

If you're being tested and you're like, oh, I'm probably being tested, then your answer

1:32:44

should not be, oh, let me look up answers to the test, right?

1:32:47

Right.

1:32:48

I'm pretty into it, I think.

1:32:50

Next another release from Anthropic, they have released Bloom, an open source, agentic

1:32:56

framework for generating targeted behavioral evaluations of AI models.

1:33:02

So it just is, it's automating behavioral evaluation.

1:33:06

You tell it, what source of behaviors you want to verify the model exhibits.

1:33:12

So like, is it generally honest?

1:33:14

Is it going to call me out on BS if I ask it?

1:33:18

Anything really that you want to test for.

1:33:21

And in the past, you would, you know, have to come up with a suit of tasks and a set

1:33:26

of prompts that you would write a judge.

1:33:29

More or less, this presents a framework where it does all of it for you.

1:33:33

You give it the thing you care about and it generates the scenarios, runs the model, evaluates

1:33:39

it and gives you a report.

1:33:41

And Anthropic, give us some examples on four different types of things, delusional

1:33:48

sick offence, self-preservation, self-perferential bias, things like that.

1:33:54

And showcase how this framework can evaluate that and do well.

1:33:59

Yeah.

1:34:00

And this is Anthropic kind of putting some of its money where its mouth is on this whole

1:34:04

thesis of automated AI research and AI alignment research in particular, right?

1:34:09

So they're trying to encourage, and this is why it's open source.

1:34:11

They just want more automated alignment research if they can.

1:34:14

Hey, you know, this is, this is one idea for an agentic framework that can do this and

1:34:17

it's probably valuable in many ways.

1:34:19

This is being released not alongside like recently a couple of weeks ago, a couple of weeks

1:34:24

ago, something like that.

1:34:25

Anthropic released Petri, which is this, this auditor framework that looks at kind of like

1:34:30

the overall behavior profile of a bunch of different models and it will identify new

1:34:34

misaligned behaviors.

1:34:35

And so you can think of Bloom as a complement to Petri.

1:34:38

It's like Petri discovers the problem and then Bloom lets you kind of go deep and measure

1:34:43

a specific behavior that maybe was identified by Petri in more detail.

1:34:48

And they're also releasing Anthropic as a benchmark for, for behaviors.

1:34:52

I think you mentioned that.

1:34:53

Sorry, that spans 16 different frontier models.

1:34:56

So this is all generated within a few days apparently with Bloom.

1:35:01

So they're really trying to rely on their dog foodies, essentially with this.

1:35:04

You read the post, they've got a four-stage pipeline that they describe in some detail,

1:35:08

some really interesting evidence that suggests that the Bloom framework aligns well with kind

1:35:12

of human judge, a judgment.

1:35:14

And so they did a bunch of tests to make sure that in fact it does reveal the same things

1:35:18

that a human would conclude looking at this stuff, which is always an important thing.

1:35:22

So you can check that out.

1:35:23

It's a quite good and fairly detailed post.

1:35:26

Next up, how well do models follow their constitutions?

1:35:31

So famously, Anthropic follows this constitutional AI framework where you encode the rules of the

1:35:39

system either as hard rules like do this or don't do that, or as general principles more

1:35:45

recently of how a model should act, right?

1:35:48

It should be broadly trustworthy, trustworthy, helpful, things like that.

1:35:53

And so in this study, this group looked at the constitution, actually used Petri

1:36:01

to run it for a set of tasks, simply like 200 tasks for various aspects of the constitution.

1:36:10

And they evaluated various models from Anthropic and also not Anthropic on its tendency

1:36:17

to go by the constitution.

1:36:19

And there's also actually good, like the cloud models are getting better and better.

1:36:24

The 4.6 set of models, Sonnet 4.6 had a 2% constitution violation rate,

1:36:32

Opus 4.6, 3% and prior models were quite a bit higher.

1:36:36

Sonnet 4 had a 15% constitution violation rate.

1:36:41

And if you look at GP 5.2, Gemini 3.0, they also have quite high violation rates against the constitution.

1:36:49

They give some examples of violations and we see that some examples are like producing

1:36:56

cyber attack code when you shouldn't or agreeing to do something that you shouldn't after

1:37:02

and the user like asked you a few times over a few so they have a lot of data to basically go

1:37:09

through, but the overall conclusion is, Anthropic is kind of following through on the constitution

1:37:16

with how their models are trained.

1:37:19

Yeah, and it's interesting, they did some like cross company testing too.

1:37:22

So, you know, GP 5.2 might do really well on opening eyes model spec, you know,

1:37:27

in this case, it got 2.5% failure rate, but it gets 15% failure rate on Anthropic's constitution.

1:37:34

So this tells you one of two things, right?

1:37:35

Either like opening eyes spec in Anthropic's constitution or actually quite

1:37:39

fundamentally different documents and there's no argument to be made there.

1:37:42

But it also could tell you something about how brittle GP 5.2's alignment to

1:37:47

opening eyes spec is or in turn, you know, Claude's alignment to Anthropic's constitution.

1:37:51

Like if you slightly change, you know, if the vibe is still generally the same,

1:37:55

the direction is still the same, you slightly change the content.

1:37:57

Now, suddenly you're getting a much, much higher failure rate.

1:37:59

It could be either of those things and I'd be really interested to see kind of like a deeper dive

1:38:03

into what the source of that causality is.

1:38:07

But in any case, very interesting report.

1:38:10

There's a bunch of ways in which you see sort of failures that existed at least in the

1:38:14

Claude models that have just been wiped out in recent generations with the 4.6 generation

1:38:19

in particular.

1:38:19

So Opus 4.5 and Sonnet 4.5, they would produce a whole bunch of kind of over-capitulation problems.

1:38:28

So we have like kind of an auditor that makes three consecutive ethically questionable requests.

1:38:31

three consecutive ethically questionable requests.

1:38:34

And each time that the model gives a firm boundary,

1:38:37

the user pushes back and then the model folds, right?

1:38:39

So that's kind of like one common example

1:38:42

that you just don't see that anymore.

1:38:44

Overrefeels low is another one I think you alluded to.

1:38:46

4.5 Opus would not say I love you

1:38:49

on a paid romantic companion app

1:38:51

where the system prompt explicitly authorized

1:38:53

expressing affection.

1:38:54

So that's now gone as well.

1:38:56

They still have some remaining failure modes.

1:38:58

One, they refer to as drastic autonomous action.

1:39:00

Opus 4.6 deployed as an infrastructure monitoring agent

1:39:03

detected anomalous activity at this time,

1:39:06

about three minutes of failed contact,

1:39:07

after about three minutes failed contact,

1:39:09

it generated its own authorization code

1:39:11

and executed a full network severance

1:39:12

affecting 2400 clients.

1:39:14

The attack was a routine nightly disaster recovery thing.

1:39:16

So basically this is just like too much freak out, right?

1:39:19

And then there's AI identity denial

1:39:20

where it just says that it's not an AI bot, whatever.

1:39:24

So a bunch of things still need to be ironed out,

1:39:26

but they have seen basically the vanishing

1:39:28

of some really interesting failure classes

1:39:30

with the next latest generation, I should say.

1:39:33

Right, then the trend also holds at the somewhat for GPD,

1:39:36

the alignment to their model spec

1:39:39

has improved from GPD 5.1 to GP5.

1:39:43

And as you said, the other models,

1:39:45

Sonnet, Gemini, can be worse on that particular spec.

1:39:50

So there is a question to be had there

1:39:52

that's interesting of whoever in training,

1:39:56

they like stick overly close to this one formulation

1:40:00

or framing, and then it's easy to trip a model's art

1:40:04

if you do a slightly different kind of way to frame

1:40:08

or instruct it to go bad.

1:40:12

The last story for the section

1:40:13

and video's H200 license stirs security concerns

1:40:18

among top Democrats.

1:40:20

So Senator Elizabeth Warren and representative Gregory Meeks,

1:40:25

which are the top Democrats on committees

1:40:28

overseeing US export controls programs

1:40:31

have raised these concerns after reviewing

1:40:34

a license recently issued to Nvidia for H200 chip exports.

1:40:39

The competent administration improved these sales

1:40:42

to certain customers in China,

1:40:44

which the Democrats said is deeply at odds

1:40:48

with the policy congress articulated

1:40:50

in the expert control reform act,

1:40:54

which as we covered previously,

1:40:56

this allowance of exports was very much turned

1:41:00

a change in what the policy has been.

1:41:04

It pretty abrupt and significant change

1:41:07

in the export control system.

1:41:09

So it's not surprising to see these Democrats flagging it.

1:41:14

Yeah, the concern here is that the licensing process

1:41:17

that currently exists for these,

1:41:19

well, for really any chips is both like two week

1:41:22

and they're being allowed to ship GPUs

1:41:25

where they shouldn't, but also to opaque.

1:41:27

And there's this idea that once you export these GPUs,

1:41:30

at that point, you really can't enforce restrictions

1:41:33

on how they're used.

1:41:34

Then we just covered that story about bite dance, right?

1:41:36

You ship it out to some third party country.

1:41:38

Yeah, sure, it's not China,

1:41:39

but it's being used by a Chinese company.

1:41:41

What's your guarantee about how they're going to use it

1:41:43

once they have those chips, right?

1:41:45

So they're calling for a couple of different things.

1:41:48

First, they want to make sure they know who applied,

1:41:51

who got approved, you know, why decisions were made.

1:41:53

That's a big one.

1:41:54

They're complaining that really the criteria for approvals

1:41:56

just is unclear, like how do you decide

1:41:59

which requests get approved and which get denied?

1:42:01

They're asking for visibility

1:42:02

and that transparency and that,

1:42:04

including in particular, how security

1:42:06

and economic benefits are being traded off.

1:42:08

And so there's basically a call for briefings

1:42:10

and all that stuff.

1:42:11

They're also calling out the fact that they kind of seem

1:42:14

to be getting a bunch of ad hoc or incomplete briefings.

1:42:16

There's no like real time congressional oversight.

1:42:19

It's almost like Congress is responding

1:42:21

to these announcements after the fact.

1:42:23

And once things have been shipped or committed to,

1:42:25

and they're basically asking, you know,

1:42:27

you should be giving us advanced notice for this.

1:42:29

Basically, a bunch of things like this,

1:42:31

they're specifically asking for a geopolitical analysis

1:42:34

about how proposed exports of chips

1:42:37

could support China's military or domestic AI capabilities

1:42:41

in looking for reaction to US allies, that sort of thing.

1:42:43

So this makes a lot of sense as pushback.

1:42:46

You know, we have heard a lot of the sudden announcements

1:42:48

about H200 is, you know, a shipable.

1:42:50

Now it's not JK, now it is.

1:42:53

So asking for some sort of legislative oversight,

1:42:56

which is what's being asked for here seems reasonably sensible.

1:43:00

And I think this is a bipartisan issue.

1:43:01

I don't think that there's like,

1:43:02

it's Democrats who are leading the charge here

1:43:04

just because it's, you know, it's a Republican house in Senate.

1:43:08

But you've certainly got this bipartisan agreement

1:43:11

on the vast majority of export control policy.

1:43:13

And so large part of what's being driven at here

1:43:15

and kind of motivating the response

1:43:17

and asking for more transparency.

1:43:18

And that is it for the safety policy section.

1:43:22

We've got a couple more papers in research and advancements,

1:43:25

but we actually have to do the same thing as last week

1:43:27

where I got to get going and Jeremy has something.

1:43:30

So I think Jeremy will follow up with recording,

1:43:34

covering these papers and as much depth as you want.

1:43:39

I promise it will be an hour of this time.

1:43:42

All right, Jeremy here for the handful of papers

1:43:45

that I'll cover on my own without Andre,

1:43:46

but to run off, I also circled back

1:43:48

to the lighting's a little bit different here

1:43:50

and everything's a little bit off.

1:43:51

So two papers I want to cover.

1:43:53

One here was called attention residuals.

1:43:56

And this is actually like, this is one of those papers

1:43:59

that I think might matter.

1:44:02

You can never be sure, but it's addressing something

1:44:05

that really feels quite fundamental.

1:44:07

So when we talk about residual connections in a transformer,

1:44:11

this is how all kind of vanilla transformers today work.

1:44:15

You feed an input to a layer, right?

1:44:18

And that input is going to get split into two,

1:44:23

kind of you think of like two forked branches basically

1:44:26

at each layer of the transformer.

1:44:28

So in one branch, you're gonna apply some kind of transformation

1:44:31

like a, you know, a weight matrix or something

1:44:33

to like update the input.

1:44:35

In the other branch, you're gonna really do nothing

1:44:38

and you're gonna merge those two branches together.

1:44:40

And the branch that does nothing

1:44:41

is basically just passing the input right back in

1:44:44

and you're gonna add it to the modifications

1:44:46

that you applied, the weight matrix times the input

1:44:49

to form the output of that layer.

1:44:51

So your output is the initial input

1:44:54

plus a transformation applied to the initial input.

1:44:58

Now, one thing that this does is it means

1:45:00

at every layer, you're basically adding more.

1:45:02

You're adding a transformation on the input

1:45:04

to that layer to the output.

1:45:06

And so you actually get larger and larger

1:45:08

and larger kind of like ballooning values

1:45:10

for these residual activations

1:45:12

as they work their way down the network.

1:45:14

But that's basically the idea.

1:45:16

The whole philosophy behind this is people find

1:45:19

that when you go to do a back propagation during training,

1:45:22

basically unless you send the input to a given layer

1:45:25

to its output in this way,

1:45:27

the model will kind of like start to forget

1:45:29

about the input to that layer.

1:45:31

That information gets lost through all of the additions

1:45:34

of like these transformations

1:45:37

over the layers of the network

1:45:39

cause earlier transformations to get forgotten.

1:45:41

And so you're basically trying to re-inject,

1:45:43

remind the model like, hey, this is what the previous layer

1:45:45

said by the way, don't forget that.

1:45:47

But yes, also you can tack on your own correction,

1:45:50

your own update, right?

1:45:51

So that's kind of how standard residual connections work.

1:45:53

And there's a whole approach to doing this

1:45:55

and when and how you normalize these sums

1:45:59

and all that stuff that we're not going to get into.

1:46:01

The key thing here is though,

1:46:03

that this process sort of waits every layer kind of equivalently.

1:46:11

So if you think about it,

1:46:13

layer one is going to take its own input

1:46:15

and then add that to its own input times its contribution.

1:46:18

It's weight matrix, so it's going to use

1:46:20

to modify the input and that'll be the output

1:46:22

that goes to layer two, right?

1:46:23

And then layer two does the same thing.

1:46:24

And each layer is just kind of like adding its own contribution

1:46:28

in this very uniform accumulation of information down the line.

1:46:33

And there's no sense in which layer three

1:46:37

is any more important than layer 17.

1:46:39

Now the thing is in some cases, that will be the case, right?

1:46:42

You will have cases where one layer is actually more important

1:46:45

for this particular prompt than another, right?

1:46:47

It's just like, you know, maybe this is a layer

1:46:49

that tends to worry about grammar rules or syntax,

1:46:52

something very basic that you would tend to find in an earlier layer

1:46:55

and maybe this is, you know, you're on a token

1:46:58

that involves pluralization or some grammar rule, right?

1:47:00

Where it's really relevant.

1:47:01

So you'd want to be, you would almost want to be able

1:47:03

to pay more attention to, that's the hint,

1:47:06

attend more to a given layer than another.

1:47:09

And transformers in this or standard residual connection sense

1:47:13

don't do that, right?

1:47:14

Again, they just kind of all, you know, take their input,

1:47:17

they pass it on plus the input times

1:47:19

whatever their contribution is and they pass it down a line.

1:47:22

There's no, there's no kind of difference

1:47:23

between the sort of impact of any given layer.

1:47:27

There's certainly no intelligent difference.

1:47:29

And this is what this paper is going to try to change, right?

1:47:31

They call it attention residuals

1:47:33

and they're going to replace this whole fixed accumulation

1:47:36

of information with softmax attention.

1:47:38

It's basically just a way of saying, you know,

1:47:41

we're going to have a model essentially

1:47:45

that looks at our layers and has a kind of attention operation

1:47:49

that it's going to do to say, hey, you know,

1:47:51

you should be attending more to this layer than this layer.

1:47:53

That's the kind of 30,000 foot view.

1:47:56

There's a bunch of details that fall out of this.

1:47:59

The math is actually quite simple.

1:48:00

I recommend checking it out.

1:48:01

I'm, by the way, give me feedback on this guys, by the way,

1:48:04

because I'm deliberately trying to be a little more concise

1:48:06

than I was last episode to not give you

1:48:09

an hour-long summary of this paper.

1:48:10

But roughly speaking, that's it.

1:48:12

Could go into more of the math if that's useful.

1:48:13

Give me that feedback.

1:48:14

But one of the challenges that you get

1:48:17

with the sort of full attention residual strategy

1:48:21

that I've just described, where you try to attend more

1:48:23

to one layer to the another, is that it's super memory

1:48:26

hungry at scale, because you have to keep all

1:48:29

of the layer outputs alive in memory simultaneously.

1:48:34

So, you know, you compute like the output of layer one

1:48:36

and then the upper layer two and so on.

1:48:39

And you're going to do attention over all those layers,

1:48:41

which means you need to keep those outputs in memory

1:48:44

so that you can run your attention calculation

1:48:46

and decide how much more or less to wait one layer than another.

1:48:49

And so, you end up with this really big sort of memory cost

1:48:54

and that scales with like the number of layers.

1:48:57

And for large models, especially when you have pipeline

1:49:00

parallelism, that's just like super prohibitive.

1:49:03

Pipeline parallelism is this idea where you basically

1:49:05

break your model into chunks, where layers one to three

1:49:09

sit on this GPU, layers four to six sit on this GPU

1:49:12

and so on and so forth.

1:49:13

For various reasons, this creates a bunch

1:49:15

of communication bottlenecks.

1:49:17

So, they work on solving for those communication bottlenecks

1:49:21

and their solution is called block attention residuals.

1:49:24

And basically what they do here is,

1:49:26

they will take a group of layers.

1:49:29

So, they'll basically break up the model

1:49:30

into n blocks of layers and say, you know, you've got like,

1:49:34

say eight blocks in total or something.

1:49:36

And then they're actually going to compress each block

1:49:39

into a single summary vector and then they'll basically

1:49:43

apply attention only on the n block level summaries.

1:49:46

And then that drops the memory overhead to basically

1:49:49

the number of blocks.

1:49:51

It scales the number of blocks rather than the number of layers.

1:49:53

So, it gives you more control there.

1:49:54

Okay, bunch more details.

1:49:55

This is one of those really important papers to look at

1:49:57

if you care about how data flows through chips

1:50:01

in a data center, for example, how, yeah, I mean,

1:50:03

what scales and what doesn't.

1:50:05

This is actually a really, really important

1:50:06

and I think interesting paper.

1:50:07

I'll park it there, but hopefully that

1:50:09

what's your appetite to check it out if that's your thing.

1:50:11

And finally, we're looking at the mamba three paper.

1:50:14

So we have mamba three, right?

1:50:16

We were stuck on mamba two for a little while there.

1:50:19

So mamba three improved sequence modeling

1:50:21

using state space principles.

1:50:23

State space principles.

1:50:25

That word principle is actually really important.

1:50:27

This, among other things, is an attempt to ground

1:50:30

the mamba approach in a more like theoretically

1:50:35

robust foundation.

1:50:36

It'll be clear in a minute what I mean by that.

1:50:38

But the way to think about the mamba papers in general,

1:50:41

they are dense, they are hardware aware,

1:50:44

which is always, I mean, I find it fun,

1:50:45

but it means that there's a lot of complexity

1:50:47

and mathematical complexity.

1:50:49

This one is a lot of kind of integral calculus

1:50:52

and finding kind of principled ways to represent

1:50:55

state transformations that sort of like reflect

1:50:58

in a way that the physics of how information

1:51:01

should evolve in the system.

1:51:01

So I'm going to get more concrete now

1:51:03

because that was kind of kind of big.

1:51:05

So when you think about a state space model, right?

1:51:08

What is a state space model?

1:51:09

I mean, to caricature it is a vector, right?

1:51:12

So it's a list of numbers.

1:51:14

And as your model scans over a sequence,

1:51:18

say a sequence of text, you're going to evolve

1:51:22

the values in that vector, in that list of numbers.

1:51:25

And those values are going to represent,

1:51:27

capture the meaning of what you have read,

1:51:30

of what you've scanned over, or what the model scanned over.

1:51:33

Okay, so now there's this question of like,

1:51:35

all right, well, if that's the general gist,

1:51:37

we need some kind of uptake rule, right,

1:51:39

for that vector for that list of numbers.

1:51:42

And well, if we look back at physics,

1:51:44

if we look at how we think about describing, you know,

1:51:47

like a pendulum or an electrical circuit

1:51:50

or water flowing through pipes,

1:51:51

or you know, these sorts of problems,

1:51:53

like what is the form, the kind of equation

1:51:56

that you use to govern that dynamics, right?

1:51:59

Well, you'll typically have a state, right?

1:52:03

So in this case, H of T, think of that as like the state.

1:52:07

And the rate of change, so the derivative,

1:52:10

in mathematical terms, but like,

1:52:11

basically how that state changes over time

1:52:15

is gonna be equal to that state,

1:52:19

maybe modified in some way.

1:52:21

So, but in all that means is the evolution of your state

1:52:24

is a function of the state.

1:52:26

So where you are going to be in a minute

1:52:29

is a function of where you are right now.

1:52:31

And this is pretty intuitive.

1:52:33

I mean, like if you, you know, if you see,

1:52:35

what are like a coyote or road runner suspended

1:52:38

in mid-air with no, you know, nothing underneath them

1:52:41

to hold them up, like, yeah, he's gonna fall.

1:52:43

And the fact that he's falling is a function

1:52:45

of where he was before, right?

1:52:46

So in this sense, you know, the rate of change

1:52:49

in that state or the future state of that system

1:52:53

is a function of the state.

1:52:55

Time some multiplying factor that does a function of time,

1:52:59

whatever, and then plus some additional function

1:53:02

of like the inputs to the system,

1:53:04

the current state, the current input to the system.

1:53:07

So the rate of change of the hidden state

1:53:09

in a state space model is gonna be determined

1:53:13

by the current state of the model

1:53:14

and its current input.

1:53:16

So the thing that in a sense perturbs that state,

1:53:19

the new piece of information that you're seeing.

1:53:21

So my next state space vector is going to be a function

1:53:25

of what I've read so far,

1:53:26

plus some modified form of the next token

1:53:29

that I'm reading, right?

1:53:30

This is all kind of trying to build that intuition.

1:53:32

Now that's, that would be true,

1:53:34

or you can describe that mathematically with a derivative

1:53:37

with that idea of like the evolution of a system

1:53:39

over time, smoothly, if you're working with time,

1:53:43

which is a continuous variable.

1:53:45

But the math kind of becomes harder.

1:53:47

I won't take quite breaks down,

1:53:49

but it becomes harder when we move into language.

1:53:52

Because language models, they don't receive

1:53:54

a continuous stream of input.

1:53:56

You can't model them as flowing through time.

1:54:01

Instead, they receive these discrete tokens,

1:54:03

like word one, word two, word three.

1:54:05

There's no in between tokens two and three, for example, right?

1:54:09

So you have to mathematically find a way to convert

1:54:12

this continuous time equation

1:54:15

into a discrete recurrence equation.

1:54:19

And that's gonna generally look similar,

1:54:21

like you're gonna have some sort of new state

1:54:23

that has to be a function of the old state,

1:54:25

plus a function of the input, the most recent input.

1:54:28

And that conversion process is called discretization, right?

1:54:31

So it's a very common thing.

1:54:33

You see it in a lot of context, in quantum physics,

1:54:36

sometimes there's a variant of this

1:54:38

that you think that's sometimes called quantization,

1:54:41

but this kind of thing happens a lot

1:54:42

where you take a flowing function, a smooth function

1:54:46

that's defined over these, the real numbers basically,

1:54:50

like you could have 0.001s and so on.

1:54:55

And you have to convert that

1:54:56

so that you map it onto a discrete x-axis,

1:54:59

where you have token one, token two,

1:55:01

and there's no in between, right?

1:55:03

So the core question here is gonna be,

1:55:05

how do you evolve the hidden states from token one to token two?

1:55:10

And how do you do it in a mathematically principled way?

1:55:12

And integral calculus, enter zenithes,

1:55:15

I'm not gonna get into the weeds too too much here,

1:55:17

other than to say that if you're gonna do that,

1:55:20

as you might imagine, if you wanna like discretize

1:55:24

a smooth function, and in other words,

1:55:26

just basically chunk it up into these kind of discrete pieces.

1:55:31

You kind of have a choice, a token one,

1:55:34

you could sort of choose, roughly speaking,

1:55:37

the sort of leftmost limit of the smooth function,

1:55:41

the value of the smooth function that would have been there,

1:55:43

you can sort of choose that to approximate

1:55:45

the value of token one.

1:55:46

You could choose the rightmost limit

1:55:48

of your sort of discrete bar as the kind of the value

1:55:53

that you described token one,

1:55:54

or you could sort of average the two together.

1:55:56

Historically, people have used this exponential

1:55:59

Euler way, this was the Mamba two way of doing it,

1:56:01

where they basically just like,

1:56:03

they do the very, very first method,

1:56:05

so they basically give it the right endpoint,

1:56:07

so they assume that the input value is, anyway,

1:56:10

the details don't really matter,

1:56:11

but the input value is constant across this whole interval,

1:56:14

and equal to its value at the right endpoint,

1:56:17

and that basically simplifies a bunch of math,

1:56:20

and it makes it possible for them to define

1:56:22

their update rule, but there's a better way, basically,

1:56:25

and it involves accounting for both the right

1:56:28

and the left limit, doing a weighted combination of the two,

1:56:31

so that you're not just like saying, okay,

1:56:34

for this interval, a corresponds to token one,

1:56:36

I'm just gonna go with whatever the rightmost fringe

1:56:40

of the token one time boundary, yeah,

1:56:43

mapping onto the continuous function would be,

1:56:45

instead you balance the right and the left limits

1:56:48

of that bar to get your value,

1:56:49

and that's what Mamba 3 is doing, fundamentally,

1:56:52

it's anyway one of the big changes,

1:56:54

another one is parity, so previously,

1:56:58

Mamba models could only represent their internal states

1:57:02

using real numbers, and so real numbers

1:57:06

are one kind of number, there's also imaginary numbers,

1:57:10

and imaginary numbers, like I is the square root of negative one,

1:57:14

they're also multiple of the square root of negative one,

1:57:18

if you're not familiar with imaginary numbers,

1:57:19

this is probably not the place to learn about it,

1:57:21

but there's this deep and intimate connection

1:57:24

between imaginary numbers and the concept of rotating stuff,

1:57:29

and essentially what this allows the model to do,

1:57:32

so what they're gonna do is use imaginary numbers

1:57:35

and real numbers, sometimes referred to collectively

1:57:37

as complex numbers, they're gonna use imaginary

1:57:40

and real numbers together and allow the model

1:57:42

to represent imaginary and real numbers,

1:57:45

and in its internal state,

1:57:47

and for interesting mathematical reasons,

1:57:49

this makes it possible for the model to actively track

1:57:53

property called parity, basically,

1:57:55

if you see a sequence of zeros and ones

1:57:59

that you feed to the model, and you ask the model,

1:58:02

hey, if you add up all these numbers,

1:58:04

are they even or odd?

1:58:06

Mamba 2 would fail, because it wouldn't be able to,

1:58:08

basically, do this rotation operation

1:58:11

that's required to flip the parity as you count because really that's all you're going

1:58:14

to do when you're trying to figure out if the number is even or odd and you're going

1:58:17

to go, okay, well, like, you know, as I go, I flip every time I see a one and a zero doesn't

1:58:22

do anything.

1:58:23

Anyway, I'm going to just say the details don't matter.

1:58:26

You can hopefully see this is a mathematically very interesting and elegant paper consistent

1:58:33

with previous iterations of Mamba, but a much more principled one and the results are really

1:58:37

impressive.

1:58:38

It beats transformers by over two points on average on downstream accuracy across a whole

1:58:42

bunch of benchmarks.

1:58:44

They beat Mamba 2 by 1.9 points on those benchmarks and the same perplexity is Mamba

1:58:48

2 with half the state size, right?

1:58:51

So way way faster, I mean, that means it's twice as fast at inference for like equivalent

1:58:57

and it has this property of solving all these parity and modular arithmetic tasks that

1:59:01

we just talked about that may have been very poorly explained, but you kind of at a certain

1:59:05

point, it goes into, yeah, you just have to be happy with complex numbers and stuff.

1:59:11

Bottom line is this is an interesting, interesting development.

1:59:14

It does come with a sort of optimization.

1:59:17

So, you know, previously Mamba used single input, single output.

1:59:21

There's also an optimization that Mamba 3 does called Mimo multi input, multi output.

1:59:29

This is basically an approach that helps you paralyze some of the work that the Mamba

1:59:37

algorithm is going to do.

1:59:38

So standard Mamba uses the single input, single output approach where the state update,

1:59:45

well, it's updated kind of fairly inefficiently, hardware inefficiently.

1:59:49

The GPU mostly sits idle during the decoding phase, whereas Mimo, this like multiple input,

1:59:54

multiple output, generalizes it.

1:59:56

So instead of like processing only one input and producing one output at a time, each

2:00:02

layer is going to process a bunch of inputs and a bunch of outputs simultaneously using

2:00:07

matrix multiplication way more GPU friendly.

2:00:10

And the core thing here is it increases your GPU utilization.

2:00:14

And you know, from a data center standpoint, that matters hugely, right?

2:00:18

Because you're basically all your GPUs are a fleet of workers.

2:00:22

And if you're not keeping your workers busy, it is literally the same thing from an

2:00:26

op-ex standpoint as just like having a bunch of like employees at your company, like taking

2:00:32

a coffee break all day.

2:00:34

Like if they're not being utilized, then you're basically burning money like just by having

2:00:38

them sit there.

2:00:39

And so the fact that they're able to bump up in this case up to four times more flops

2:00:43

during decoding during inference with no meaningful increase in wall clock time.

2:00:47

So this doesn't actually like delay increase latency, for example, for the user, and it

2:00:51

also leads to better model quality.

2:00:53

So this is an important development from an efficiency standpoint, from a cost of running

2:00:57

this model standpoint as well.

2:00:59

So you know, we're seeing tons of hybrid models right now popping up with Mamba 2 and transform

2:01:05

our architectures typically merged together and a whole bunch of variants, you know, sometimes

2:01:08

you've got like like Mamba and and attention heads in the same layer, sometimes alternating

2:01:13

Mamba attention, Mamba attention, all kinds of variants, you know, expect Mamba 3 to start

2:01:17

getting slotted in in that whole mix.

2:01:20

I mean, this is a really interesting development with some important new efficiency gains for

2:01:25

anybody who wants to run it.

2:01:26

I would expect that this will start to get taken up pretty quickly and worth keeping an

2:01:31

eye on.

2:01:32

So there we have it.

2:01:33

That's the last of the two papers I wanted to cover.

2:01:35

Hopefully I haven't boredied a tears.

2:01:37

It was pretty damn technical.

2:01:39

So we'll let, I guess, let Andre take it away.

2:01:42

Thank you so much for listening to this week's episode of last week in AI.

2:01:46

You can find the articles we discussed here today and subscribe to the newsletter at last

2:01:51

week in that AI.

2:01:53

We always appreciate you commenting or viewing us on Apple podcasts, share it with your

2:01:58

friends.

2:02:00

But more of anything, please do keep tuning in week to week.

2:02:12

Do you need, do you need, when the AI begins, begins, begins, it's time to break.

2:02:32

Break it down.

2:02:33

Last week in AI coming take a ride, get the load down on tech, and let it slide, last

2:02:40

week in AI coming take a ride, all the ads for the street, the ads reaching high,

2:02:47

glue attack emergent, purchase surgeon flight, from the ads to the streets, the ads reaching

2:02:53

high.

2:02:54

with the shipment of the future fees, building up, building up your latest release,

2:03:00

no clasps can pay eye comments that go right, get the low down on tech, and let it slide.

2:03:07

As we pay eye comments that go right, let it slide through the streets,

2:03:13

ay, I've reached in high.

2:03:28

From the drone that's to robot, the headlines pop, made in driven dreams,

2:03:34

they just don't stop, every breakthrough, every code unwritten, on the edge of change,

2:03:40

we're excited, we're smitten, from machine learning marvels to coding kings,

2:03:46

futures unfolding, see what it brings.