Learning Library

← Back to Library

Deep Learning Hitting a Wall?

Key Points

  • The panel opened with a heated debate on whether deep learning is “hitting a wall,” with Chris Hay claiming models are getting worse, Kush Varshney acknowledging challenges but seeing them as surmountable, and Kate Soule asserting that new applications keep the field advancing.
  • Host Tim Hwang introduced the episode’s theme “Mixture of Experts,” framing the discussion around the release of DeepSeek‑V3 as a public showdown between AI optimists and skeptics.
  • OpenAI’s latest model, o3, was highlighted for dramatically out‑performing traditional benchmarks such as frontier math, reigniting confidence that deep learning progress has not stalled.
  • The strong o3 results are presented as a narrative reset after recent speculation that pre‑training breakthroughs were fading and that deep learning was entering a slowdown cycle.
  • Chris Hay, despite his earlier pessimism, expressed enthusiasm for o3’s inference‑time efficiency and praised the model’s practical performance, suggesting that substantial gains are still possible.

Sections

Full Transcript

# Deep Learning Hitting a Wall? **Source:** [https://www.youtube.com/watch?v=QzERUfTbKQw](https://www.youtube.com/watch?v=QzERUfTbKQw) **Duration:** 00:39:16 ## Summary - The panel opened with a heated debate on whether deep learning is “hitting a wall,” with Chris Hay claiming models are getting worse, Kush Varshney acknowledging challenges but seeing them as surmountable, and Kate Soule asserting that new applications keep the field advancing. - Host Tim Hwang introduced the episode’s theme “Mixture of Experts,” framing the discussion around the release of DeepSeek‑V3 as a public showdown between AI optimists and skeptics. - OpenAI’s latest model, o3, was highlighted for dramatically out‑performing traditional benchmarks such as frontier math, reigniting confidence that deep learning progress has not stalled. - The strong o3 results are presented as a narrative reset after recent speculation that pre‑training breakthroughs were fading and that deep learning was entering a slowdown cycle. - Chris Hay, despite his earlier pessimism, expressed enthusiasm for o3’s inference‑time efficiency and praised the model’s practical performance, suggesting that substantial gains are still possible. ## Sections - [00:00:00](https://www.youtube.com/watch?v=QzERUfTbKQw&t=0s) **Panel Debate: Deep Learning Hitting a Wall** - In a lively “Mixture of Experts” session, a panel of AI leaders shares divergent views on whether deep learning has stalled—ranging from outright pessimism to cautious optimism about overcoming challenges and unlocking new applications. - [00:03:09](https://www.youtube.com/watch?v=QzERUfTbKQw&t=189s) **Benchmarks, Speed, and Coding Models** - The speaker criticizes benchmarks, praises the reasoning strength of o1/o1 Pro while noting its slow response time, and explains switching between models to balance speed and depth, especially for coding tasks. - [00:06:16](https://www.youtube.com/watch?v=QzERUfTbKQw&t=376s) **Flexible Compute Trade‑offs with O3** - The speaker explains how the O3 model enables dynamic balancing of inference time, latency, and cost by allowing users to choose between low‑resource quick responses and high‑resource, higher‑quality outputs. - [00:09:18](https://www.youtube.com/watch?v=QzERUfTbKQw&t=558s) **Aligning Models with Safety Policies** - Discussion of using regulatory text and synthetic data to train models, emphasizing inference‑time safety checks, governance, and recent alignment research. - [00:12:22](https://www.youtube.com/watch?v=QzERUfTbKQw&t=742s) **User-Defined Safety Trade‑offs** - The speakers explore letting users allocate AI effort between safety and problem‑solving, advocate for deliberative, democratic input into safety policies, and wrestle with the tension between rigorous safety and delivering faster, more entertaining models. - [00:15:47](https://www.youtube.com/watch?v=QzERUfTbKQw&t=947s) **Cost Curves, Fine‑Tuning, and Agents** - The speakers argue that recent training tricks dramatically lower pre‑training expenses, but the future focus should shift to inference efficiency, fine‑tuning, and deploying models as agents. - [00:18:50](https://www.youtube.com/watch?v=QzERUfTbKQw&t=1130s) **The AI Training Pendulum Through 2025** - The participants discuss how the back‑and‑forth cycle between pre‑training massive models and deploying smarter inference‑time models will dominate AI development up to 2025, while also questioning the hidden data costs behind open‑source projects. - [00:22:10](https://www.youtube.com/watch?v=QzERUfTbKQw&t=1330s) **Global AI Governance and Standards** - The speakers discuss the need for worldwide, technically‑driven AI governance—likening it to ICANN’s voluntary standards and noting an upcoming Paris meeting of safety institutes to create codes of practice—while also touching on switching between open‑source and closed‑source model modes. - [00:25:21](https://www.youtube.com/watch?v=QzERUfTbKQw&t=1521s) **Governance Challenges of Tiny Autonomous Agents** - The speakers contend that by 2025, the hardest AI governance problems will stem from the rapid, unregulated misuse and trust deficits of autonomous agents built on very small models, rather than from controlling the largest, most prominent models. - [00:28:26](https://www.youtube.com/watch?v=QzERUfTbKQw&t=1706s) **Public Bet on AI's Future** - The speaker outlines a public wager between an AI skeptic and Miles Brundage that lists ten possible AI milestones—like producing world‑class creative works—to force concrete definitions of “truly powerful” models, and asks whether this approach meaningfully gauges AI progress or merely adds to Twitter noise. - [00:31:46](https://www.youtube.com/watch?v=QzERUfTbKQw&t=1906s) **LLMs Aren't Authors, Tradition Over Authorship** - The speaker argues that large language models function as collaborative tools within a long‑standing literary tradition, lacking genuine authorship, and that framing AI ethics around authorial credit is a misguided question. - [00:34:52](https://www.youtube.com/watch?v=QzERUfTbKQw&t=2092s) **Bridging AI Misconceptions for the Public** - The speaker argues that AI hype outpaces everyday understanding, urging relatable explanations of generative AI for non‑technical audiences. - [00:37:54](https://www.youtube.com/watch?v=QzERUfTbKQw&t=2274s) **AI Models Mirror Organizational Structure** - The speakers discuss how rapid innovation, Conway's Law, and corporate team dynamics influence the behavior and characteristics of AI models, using examples from pre‑training teams and a humorous reference to Anthropic's Claude. ## Full Transcript
0:00Frequently asked question, is 0:01deep learning hitting a wall? 0:03Chris Hay is a distinguished engineer 0:04and the CTO of Customer Transformation. 0:06Chris, what do you think? 0:07Oh yeah, totally, Tim. 0:09In fact, I think it's getting backwards. 0:11I think the models are getting 0:12worse and worse and worse. 0:13This is the worst it's ever been. 0:14It's totally hit a wall, Tim. 0:16Happy 2025, Chris. 0:17Uh, Kush Varshney, an IBM fellow 0:19working on issues of AI governance. 0:21Kush, welcome back. 0:22Uh, what do you think? 0:23I think there is a wall, but 0:24it's not an insurmountable one. 0:25I think we're making progress. 0:27We're, uh, changing it up instead 0:28of just taking some steps. 0:30We're, uh, doing some rock climbing. 0:32A little bit more of a serious answer. 0:33And Kate Soule is Director of Technical 0:35Product Management for Granite. 0:36Kate, happy 2025. 0:38Uh, what's your take? 0:39No, I don't think deep 0:40learning is hitting a wall. 0:40I think we're finding new ways to 0:42apply it in 2025 that's going to 0:44have some interesting benefits. 0:45All right. 0:46All that and more on today's mixture of experts. 0:53I'm Tim Hwang happy 2025, and 0:55welcome to Mixture of Experts. 0:57Each week, MoE offers a world class 0:59panel of product leaders, researchers, 1:01and engineers to analyze the biggest 1:03breaking news in artificial intelligence. 1:05Today we're going to be talking about the 1:06release of DeepSeek-V3, a very public wager 1:09between an AI booster and an AI skeptic. 1:11Thanks But first, let's talk about OpenAI's o3. 1:15Um, this was, uh, the last announcement of 1:17OpenAI's 12 Days of OpenAI, uh, marketing 1:21event that they did at the end of last year. 1:23Uh, and it was arguably 1:24the biggest announcement. 1:26Uh, they basically have touted a new 1:28model, which is now getting sort of limited 1:30trial access for safety purposes, um, 1:32that blows out of the water a lot of the 1:35benchmarks that people have traditionally 1:37used to measure or argue for measuring 1:40Whether or not we're getting close to AGI. 1:42So, uh, on a benchmark that we've talked about 1:44on the show in the past, uh, frontier math, um, 1:47open, openAI's, uh, o3 is doing incredibly well. 1:51Um, and I think one of the reasons I wanted to 1:54kind of bring this up is that it really does 1:56seem like, you know, after, I think what was a 1:58new cycle late last year of people saying deep 2:01learning slowing down, the old methods don't 2:03work anymore, pre training is over and a lot of 2:05general hand wringing, um, this really kind of. 2:08Reset the narrative at least in the 2:10circles that I run in to say actually 2:12that you know There's there's maybe 2:13a lot more room to run on all this. 2:16Um, Chris, maybe I'll turn to you first You 2:18sort of outright made fun of me on the opening 2:20question Um, what's your take on the o3 model? 2:23Like how important is it? 2:25Does it really kind of indicate that 2:26there's still a lot more progress to run? 2:28How do you read it, basically? 2:29I think it's a great thing, actually. 2:31So I've been playing a lot with the old 2:33one and the old one pro models, and I've 2:35been having the best time with them. 2:37So inference time compute is really working. 2:39So I'm excited about o3. 2:41I'm just kind of annoyed 2:42that we don't have it though. 2:43That's the real thing. 2:44And so it's yet another. 2:46You know, this is coming soon and, and that's 2:49sort of annoying me, especially being in 2:50Europe because in Europe we don't get anything 2:52these days, we didn't get Sora, we didn't 2:54get half the, any models that are coming 2:56through on, uh, the 12 days of Christmas. 2:58So, um, I'm excited about o3, as for 3:00the benchmark thing, two things in 3:03my mind about that, one, you know, my 3:05opinion, benchmarks are stupid, so I'm 3:07not really going to read into that. 3:09And then probably the second thing is even 3:11if we take the opinion that benchmarks aren't 3:13stupid, then it took an awful lot of time to 3:16come back with that answers and it was a little 3:18bit kind of monkeys and typewriters, right? 3:19Which is if you type long enough, then 3:21you're eventually going to get the answer. 3:23But, but with that, Aside, actually, I'm 3:26so impressed by o1 and o1 Pro that actually 3:30I'm super excited about o3 and I think it's 3:31going to be a great model and it's really 3:33proving sort of inference time compute. 3:36Yeah. One follow up there is, uh, I know you're 3:38saying you think all benchmarks are 3:39stupid, but you think this model is better. 3:41So what use case do you have in mind 3:43where you're like, Oh, actually, 3:43it seems to be a one's noticeably 3:45better than what we've had before. 3:47Yeah, there's probably a few ones. 3:48The main one for me is coding, right? 3:51I mean, it is completely in a different level. 3:54Even Claude 3.5 Sonnet, um, GPT 4. 3:570, the early versions of o1. 3:59Honestly, o1 Pro is, is on a different level. 4:02Now, probably the big thing that I've found 4:04myself working with the models is Pro just takes 4:08quite a long time to come back with an answer. 4:10So, I end up switching 4:11between models all the time. 4:12It's like, okay, I want a fast answer on this. 4:14I think it can handle this. 4:15Oh, no, it can't handle it. 4:17I'm going to switch from o1 to o1 Pro. 4:18So. 4:19Um, so that sort of changing models just to 4:22get fast answers back and how much reasoning 4:25I want from the models is a sort of technique. 4:27But for me, coding is 4:30definitely the biggest thing. 4:31I don't really care about the 4:32math stuff because, like, I'll 4:33just use a calculator, right? 4:35But definitely for coding, I see a mark. 4:37Got it. 4:38Okay, maybe I'll turn to you. 4:39So I think. 4:40You know, if you're not watching this 4:41space super closely, it's easy I think to 4:43just get like, bewildered by the number 4:45of models and find variations between 4:48all these models kind of coming out. 4:51Um, you know, I think famously, or like it was 4:53kind of talked about that the reason they jumped 4:55from oh one to oh three was that oh two was I 4:57think already used by the UK telecom company. 5:00So it was like a trademark thing 5:01that got the o3. 5:04Um, but I guess Kate, question for you 5:05is if you can help our listeners kind of 5:07understand a little bit of like what's. 5:08What's new with what they're trying 5:10with o3, like kind of looking under the 5:12hood, you know, these models seem to be 5:14a lot more performant, but there also 5:16seem to be like a lot of new things that 5:18they're trying underneath the surface. 5:20And I think it's worth kind of for our 5:22listeners to know a little bit of the flavor 5:23of that if you want to speak to that at all. 5:24Absolutely. 5:25So I think the most important thing 5:27for our listeners to understand when 5:29looking at the new o3 model and the o- 5:31model series in general from OpenAI. 5:34Is that we're transitioning from spending 5:37and innovating at the training time of the 5:39model and instead saying, okay, let's take 5:42a model that's been trained and let's run it 5:44multiple times and spend more compute at the 5:48actual inference time when it's being output. 5:50deployed out in the world, and it seems like 5:54with the o3 models, they're continuing to 5:56innovate and what can be done at inference time, 5:59having the models essentially think longer to 6:02risk anthropomorphizing these models, think 6:05longer through different tasks, search through 6:07many different potential options and solutions 6:10before picking the best one, which then leads to 6:13improved performance, but also it takes longer. 6:16Uh, to Chris's point, you have 6:17to wait longer for a response. 6:19One of the things that I 6:20think is really important. 6:21Really exciting about the oh three model and 6:23this broader investment and pivot to more 6:26inference time compute is that it actually 6:29can give you some really nice trade offs. 6:31And I think this is where we're heading. 6:33And o3 is, uh, you know, foreshadowed 6:36a little bit that you can run these models 6:38in a more efficient mode, or if you need 6:41the maximum performance, you can run 6:43them in kind of a compute intensive mode. 6:46And I think that's going to be really cool 6:48because it gives people the ability to 6:50set their compute budget, set their time 6:53constraints, you know, for latency, if 6:55they need an answer, a response quickly. 6:58And I think we're going to see a lot more of 6:59that in 2025 of people playing along that kind 7:03of cost performance trade off, even within a 7:05single model saying, okay, I want my model only 7:07to think about this for, you know, a minute 7:10versus I want my model to give a response 7:12immediately versus My model can think about 7:14this for five minutes and then give me a 7:16response back, depending all on how much I'm 7:18willing to pay and how important it is that 7:21the model gives a really strong response back. 7:23Yeah, definitely. 7:24Yeah. Some people were joking online. 7:25I saw that this is kind of, it's the return of 7:27the old turbo mode on computers where you're 7:30like, we want the computer to work harder. 7:32Um, but it actually, uh, it's a really 7:33interesting question about like, 7:35almost like, I don't What do users want 7:37the computer to think hardest about? 7:40Which I think is kind of a counterintuitive 7:41question about like what types of queries and 7:43what types of tasks you know demand that It'll 7:45be really interesting to to see Chris it's ideal 7:49to have you on the line as well because I think 7:50one of the most interesting parts of the launch, 7:52you know, I think Chris was frustrated by it. 7:54He was like Come on, just 7:55give me access to the model. 7:57But in traditional open AI style, they've 7:59said, well, no, we're being careful with the 8:01launch and you can get access to the model 8:02if you're a safety or security researcher. 8:05Um, and they're allowing people to 8:06have kind of like requested access 8:08to go and red team the model. 8:10I'm curious about how you read that as someone 8:12who thinks about AI governance, you know, 8:14is that kind of going to be the paradigm for 8:15how companies release models going forwards? 8:17Or, you know, is open AI kind of almost like, 8:20There do you see this as marketing, right? 8:21They're using the safety thing to 8:22be like, give us just a few more 8:23months to iron out the loose ends. 8:26I think it's a combination of both 8:28actually, because, um, there's this 8:29concept of the gradient of release. 8:31Um, and, uh, Irene Solomon from Hugging 8:34Face came up with this and it's kind 8:35of like, uh, maybe take your time. 8:38Um, The more powerful the model 8:39is, maybe the more, um, kind of 8:41the slower you need to roll it out. 8:43But, um, I think it's a combination. 8:45Um, so OpenAI, uh, gave their models 8:48to, um, the UKI Safety Institute 8:50for testing, uh, in advance as well. 8:52And, um, uh, some of this I think is, uh, 8:55just to be able to say that they did it. 8:57Some of it is to actually have some, uh, some 9:00better, uh, safety alignment and so forth. 9:02So, yeah, I think it's, uh, it's here and there. 9:05And, um, uh, one other thing 9:07that's in this o3 release. 9:08Um, that they talk about is a new 9:11way of doing safety alignment. 9:12They call it a deliberative alignment. 9:14And, um, uh, I think it's, 9:16uh, it's kind of interesting. 9:18Um, they're, uh, saying that they are 9:20very much looking at an actual safety 9:23regulation, um, taking the text from that 9:25and training the model with respect to it. 9:27Um, doing some synthetic data generation 9:29that, uh, follows along with, uh, 9:31with, with what the policy says. 9:33And, um, so something we've been doing for a 9:36while as well, um, last year, um, we published. 9:38a couple of papers, um, we call 9:40alignment studio and we call alignment 9:44from unstructured text and so forth. 9:45And I think those sort of ideas, um, they're 9:48kind of carrying through the, the new part is, 9:51um, again, the, uh, uh, the fact that this is 9:54spending a lot of time on the inference side, 9:56um, then thinking again and again about, uh, uh, 9:59am I meeting those safety requirements or not? 10:01And, uh, uh, As both Chris and Kate 10:03said, right, I mean, um, the more time 10:05you're spending over on the inference 10:07side, what should you be thinking about? 10:09What should the model be thinking about? 10:11And I said this in the, uh, the 10:13last episode of the new year. 10:14I think that extra thinking is going 10:16to be for governance quite a bit. 10:19So, um, so I think, uh, this 10:21is where it's going to play. 10:23And, uh, yeah, I'm excited to, to 10:25see, uh, maybe I'll sign up to, so 10:28yeah, I do some of the safety testing. 10:30Yeah. I think it's kind of two 10:31interesting things here. 10:32And what you said, I mean, I think one 10:33of them is, you know, the model to date 10:36feels like has been, you release the model, 10:39but then you're also like, we guarantee 10:40safety by releasing safety models, right? 10:42Granite has done this. 10:43Um, and. 10:45You know, maybe Kate is a question 10:46back to you is sort of how much do you 10:48think that's kind of just like almost 10:49just like historically provisional? 10:51This is just like what we kind of have to do 10:53right now because we're still working out the 10:54kinks on making the models themselves safe. 10:56I guess in the future, one argument is that 10:58the models are just kind of safe out of 11:00the box in a way that doesn't separately 11:02require another model that kind of 11:03monitors outputs and does the safety work. 11:06Um, do you think that's the case or do you 11:07think this kind of bifurcated architecture 11:09is going to be what we'll see going forwards. 11:11Well, first I'd be careful. 11:12I don't think anyone can guarantee 11:14safety no matter what we release, right? 11:17But I do think we're going to continue to 11:19see more and more of these kind of safety 11:22guardrails being brought intrinsically 11:24into the model through these new types 11:26of alignment that Kush mentioned. 11:28That does not mean though that we 11:30shouldn't also have additional layers. 11:33Of security and safety that have, uh, you know, 11:36an independent check right on model outputs. 11:39So I don't see that going away. 11:41I think it's always going to be a yes. 11:43And right, let's continue to add 11:44more and more layers, not we're going 11:47to scribe away, you know, some of 11:49these layers, put it into the model. 11:50And now you've got one model, you're all set. 11:52Very interesting. 11:54Kush, maybe the other thing that I think I'll 11:55pick up on what you said before we move to the 11:57next topic is, um, you know, you're basically 12:00talking about inference as being almost like 12:02this kind of fixed, fixed budget of time. 12:04And you're basically like, what do you want 12:05to spend, have the model spend their time 12:07on thinking about the problem or thinking 12:09about whether or not the responses are 12:10safe or consistent with a safety policy. 12:13And I'm modeling my internal Chris here. 12:15Who probably would be like, you're spending 12:17some of that time on trying to make it safe. 12:19Like, could it, could it just solve the problem? 12:22Um, and I guess I'm kind of curious is like, 12:24maybe that will become, do you think that 12:25will become a lever over time where you can 12:26almost like, the user will specify, I need 12:2910 percent of your time spent on safety, 90 12:31percent of the time on solving the problem or 12:32otherwise, or, you know, That actually kind 12:34of opens up a whole nother world in some ways. 12:36Yeah, it does open up a whole new world. 12:38I mean, um, I wouldn't say that, uh, I would 12:41want to spend a lot of time on the, this sort 12:43of safety deliberation either, but, um, I 12:46think, uh, the, the fact that they're calling it 12:48deliberative, um, it kind of speaks to something 12:52that, uh, I mean, deliberation is meant to 12:54be like a discussion among lots of different 12:56viewpoints and, and this sort of thing. 12:59I don't know if that's actually what. 13:00It'll happen, but that's something I would want 13:02to happen so that, uh, different viewpoints, 13:04different sort of perspectives, um, can be 13:07brought into, to these different policies as 13:09well because, um, uh, in, I mean, democratic 13:12sort of, uh, settings, you do want deliberation. 13:15You do want, uh, kind of minority voices to be 13:17heard as well, but, um, uh, not sure exactly 13:20that's what, what they mean by deliberative. 13:22Absolutely. 13:23Um, Chris, I appreciate you're shaking 13:25your head, so I want to make sure 13:26I'm not putting words in your mouth. 13:28I, I honestly think safety is super 13:31important, but I want the models quicker. 13:34So, you know, so do what do what you need to 13:36do and and I want the models to be fun So don't 13:38don't lobotomize them, you know what I mean? 13:40Um, but you know, you know, we don't 13:42want to do harmful stuff, but at the 13:43same time, come on, you know, it's 13:44like, I want to play with the models. 13:47Chris basically wants everything. 13:48Tim, we are also kind of assuming 13:50OpenAI is going to give us the choice, 13:52right, of how we want the model to 13:53spend that inference time compute. 13:55And I don't think that's the 13:56direction that they're headed. 13:58I think they've got some 13:59clear regulatory guidelines. 14:01They're trying to, to meet performance issues, 14:04uh, that they want to make sure are addressed. 14:06I don't see them handing over kind of the 14:08keys to the kingdom, so to speak, to let 14:10us take these models for our own joy rides. 14:12Yeah, no, I think that's for, for sure. 14:14Right. 14:14Um, and yeah, I think there's a bunch 14:17of interesting questions that are 14:18sort of empirical questions, right. 14:20It's just like, how much can, you know, how 14:22much do safety, Like, how much does safety 14:25inference lead to better outcomes, right? 14:28Like, how much of this is like a 14:29mutually exclusive pie versus ones 14:31where you can get a little bit of both? 14:33How much is going to be 14:33defined by the regulator? 14:34How much is going to be defined by the user? 14:36Um, a lot of things to pay attention 14:37to, I think, going into 2020. 14:41Five. 14:45So I'm going to move us on to our next topic, 14:46which is the release of DeepSeek-V3. 14:50Um, this is sort of an interesting announcement 14:52because I think we were, uh, me and the 14:53production team were kind of tying up at the 14:55end of the year and we're like, nothing's going 14:57to happen in the last few weeks of the year. 14:59And of course there was the o3 15:00announcement, which was huge. 15:02And then also similarly big was 15:03the announcement of DeepSeek-V3. 15:05Um, and so this is an open 15:07source model coming out of China. 15:09That shows incredibly good performance 15:11on a lot of the benchmarks, um, uh, 15:14that most models are evaluated against. 15:16And I think there's a lot of interesting 15:18things to talk about here, but I think maybe 15:20the first one, which I'll throw to Chris, is 15:23this kind of, uh, Claim that the DeepSeek team 15:26is making that they were able to basically 15:28build this incredibly performant model for 15:30way lower cost than you would expect And I 15:34think a lot of the commentary online and I 15:36think one of the things that made me think 15:37about is that there's so much That's built 15:38on the economy of AI That is sort of based on 15:42the idea that it's just really expensive to 15:44get You know, really high performance models. 15:47Um, but this almost seems like the cost curve 15:49might be collapsing faster than we think. 15:51I don't know, Chris, maybe that's 15:52a little bit too optimistic, but 15:54yeah, I'll maybe throw it to you. 15:55I think it's kind of 15:56interesting what they've done. 15:57So they have put a lot of cool techniques 15:59within the pre training side of things. 16:01And, um, I mean, even things like multi token 16:04prediction, and then they were better at kind of 16:06loss of tokens, et cetera, and how they route. 16:09So there's a lot of things they did in 16:10training that they brought the cost down 16:12in, and I think they were doing kind of. 16:14Mixed precision, uh, as well, so there was 16:17a lot of good things that they did there. 16:19I think what I would say though is that back to 16:23the earlier point about inference time compute 16:25and kind of pre train, I, I wonder at what point 16:28we maybe stop obsessing with the pre trained 16:31side of things for models and actually, you 16:35know, be able to kind of fine tune those models 16:37and have that, uh, community of fine tuning. 16:40existing. 16:40And I think that's going to be more 16:41interesting, especially in the world of agents. 16:43Happy New Year. 16:43I'm the first person to 16:44say agents on the podcast. 16:48So thanks, Chris. 16:51So I think that's more interesting. 16:54And, and I, and as we move more towards 16:56inference time compute, I think that that 16:58will become important there, but it is. 17:00Really impressive for what they did 17:03actually for the cost of the model and 17:06how long it took them to kind of to train 17:08that I honestly, they did a great job. 17:11So, yeah, there's going to be 17:12more innovation in that space. 17:14I still think pre training though is hugely 17:17inefficient because you're really just saying. 17:20Here's the entire text of the internet. 17:22Go, go learn from it. 17:23And I, I honestly think that's 17:25probably an innovation that I would 17:27hope that would change in 2025. 17:28And the way I think about it is if I 17:30think about the kind of internet, it 17:32almost has a knowledge graph anyway. 17:34And I wonder if actually during that 17:37training process, if we brought a little 17:39bit of structure in the knowledge graphs 17:41into the pre training process, Then a lot 17:44of those, uh, training elements may come 17:46out, uh, a little bit quicker and better. 17:48I don't know. 17:48I mean, I'm just sort of, 17:50uh, sort of guessing here. 17:51But I think, I think there's a lot 17:52more innovation to do in pre train. 17:54Um, so hopefully with inference time compute, 17:57we're all going to be running around doing that. 17:59But I'm hoping that that 18:00focus on pre train doesn't go. 18:02So really good job to the DeepSeek 18:05team to continue to innovate. 18:07Yeah, I remember, um, you know, when I worked 18:09a lot more closely with pre training teams, 18:11I thought it was very interesting is at 18:12least among, you know, at least among the 18:14nerds, at least among the engineers, right? 18:16Like it was very interesting was that, 18:18you know, pre training was like the 18:19high prestige part of the organization. 18:21Right, like you're running the rocket 18:22launch of AI and then fine tuning 18:25something that we do afterwards. 18:26But like, I think all the inference stuff and 18:28all the stuff that we're seeing kind of point 18:30to this like shift in the kind of cultural 18:32capital within these companies where it's 18:34like, Oh, we're all the action right now is 18:35really happening after the pre training step. 18:38And I guess, Chris, almost what 18:39you're proposing is maybe like at some 18:40point, like the pendulum swings back. 18:42Because it's like, okay, there's all of 18:43this kind of innovation still to be done 18:45on the pre training side, but we're just 18:46not there because of the hype cycle. 18:48It's going to swing back and forward, 18:49back and forward, back and forward. 18:50And, and, and you're going to see that, right? 18:52Because you're going to get to the point 18:53where you go, um, you know, the, the train 18:57isn't good enough to do what we need to do. 18:59So therefore we're going to use the, the 19:01smarter inference time models to get better 19:03data, to train the pre trained models. 19:05That's going to become more efficient. 19:06And then we're going to do the same on fine 19:07tune and that pendulum is going to swing and 19:09swing because you're going to keep hitting 19:11kind of limits in one area and you're going 19:12to go back to the earlier, like the pre train 19:15to try and fix that and you're just going 19:16to go back and forward, back and forward. 19:17So that pendulum is going to swing 19:19all the way through 2025, buddy. 19:22Yeah, definitely. 19:23Uh, Kate, any thoughts on this? 19:24I mean, as someone who, you know, works with 19:26a team on open source AI, I assume something 19:29like Deke Seek is a, is a big, big deal. 19:30a big marker in some ways, 19:31a big way to start 2025. 19:33Yeah, and I agree with Chris. 19:34The team did an incredible job, but in terms 19:36of the cost, I don't know the full details of 19:40what data was or was not used in the model. 19:42My hypothesis is they are using data 19:45that was available online that cost 19:47a lot more than $5,000 to generate. 19:50Right. So that I don't know that that total 19:53cost estimate actually reflects the 19:54fully burdened cost of the model. 19:56I suspect that they like many model 19:58providers are leveraging all of the data 20:01that's now been posted and shared online. 20:04That actually is only possible because 20:05others have invested so much money 20:07in creating larger models that can 20:09be used to then generate that data. 20:11That kind of to what Chris was saying 20:12can be taken back into training. 20:14So I think what I'm really interested in 20:17with the DeepSeek model, aside from that 20:19is, you know, it's a mixture of experts 20:21architecture, which is really interesting. 20:23So when it runs that inference, it's, you 20:25know, a 600 plus billion parameters, but at 20:28inference time, you know, it's only about 20:2940 billion parameters, meaning it can run 20:32much more efficiently than even like a Llama 20:35400-, you know, plus billion parameter model. 20:38So. 20:39I think that's where we're going to see a lot 20:41more innovation happening in 2025 is really 20:43digging into how we make these architectures 20:45more efficient, how we activate the right 20:48parameters at inference time, fewer parameters 20:51at inference time to still drive performance 20:53without having to pay for the entire cost of 20:56running, you know, 600- plus billion parameters. 20:59Yeah, that's really interesting. 21:00Um, Kush, from a governance standpoint, 21:03This is an interesting story as well. 21:05Um, right. 21:06Uh, you know, I think there's, there's 21:07certainly a vision among some folks, which is 21:09like, well, we just passed the laws in the U.S. 21:12and all the big AI companies are in the U.S. of course. 21:14And so that's why, like, that's 21:15how we, that's how we govern AI. 21:17Um, but this is really a different world, right? 21:20Like, you know, a law passed in the U. 21:22S. is not going to change, you know, 21:23what the DeepSeek team is doing. 21:26Um, You know, is, is governance 21:27possible in this world? 21:29Right? Because it sure seems like, you know, you, you, 21:31you are seeing so much AI progress everywhere. 21:34Mm-hmm . Um, that governance 21:35becomes a real question. 21:36Yeah. 21:37I mean, uh, we talked about this, 21:39uh, before the show started that, uh. 21:42There's, uh, these core socialist values that, 21:45uh, are required of any generative AI in China. 21:47Um, it's a law that's been around for 21:49more than a couple of years now and, 21:50um, uh, DeepSeek has to satisfy those. 21:53So, um, I mean, those are things that are gonna 21:57be around, uh, and I think, uh, the fact that 22:01all of these different AI safety institutes 22:02from different countries are forming a network, 22:05um, they're convening, uh, they're figuring 22:07things out, um, together is a great sign. 22:10I think, uh, Uh, AI governance 22:12needs to be a worldwide activity. 22:13There's no special thing because 22:16of one country or another country. 22:17And, uh, uh, the more we can, 22:19uh, kind of bring everything into 22:21harmonization, the better it will be. 22:23Yeah, I think that'll be one 22:24really interesting bit is. 22:26You know, I think there was a thinking sort of 22:28maybe a few years back, which is we're going 22:29to do sort of law and regulation to do this. 22:32You know, Kush, I guess kind of where you're 22:34suggesting is a world where it's a little 22:35bit more sort of technical experts, like 22:37it looks a little bit like ICANN, right? 22:39Like in terms of how we govern the web, where, 22:41you know, kind of technical experts meeting and 22:42they establish these standards and it's kind 22:44of voluntary protocol more than anything else. 22:47Do you think that's how things are going to go? 22:49Yeah, I think that's how it's gonna go. 22:50Um, so in February, there's a meeting 22:53in Paris where all of these safety and 22:55safety institutes are coming together. 22:57So I think they'll come up with a plan. 22:59They'll, uh, they'll figure out some 23:02codes of practice and all these things. 23:04So that's where I think things are headed. 23:06Chris, you started this episode by 23:07talking a little bit about how you, 23:08like, switch between different modes of 23:10OpenAI, right, where you're like, okay, 23:12well, we're going to use the 01 for this. 23:14We're going to use the 01 Pro for this. 23:16Um, do you do that kind of switching across 23:18open source and closed source at all? 23:19Yeah. 23:20You do? Okay. 23:21Yeah, no, I do that a lot with different models. 23:23So like the Llama models, for 23:24example, I've got such personality. 23:26So if I'm doing any kind of. 23:28writing new stuff, then I tend to run into 23:31the kind of llama models, the granite, the 23:34granite models, I use quite a lot as well, I 23:36use them a lot for kind of RAG type scenarios 23:38because they're really good at that, in that 23:40case, and, and also if I'm pulling factual 23:43information, then I really want to be sure 23:46where the data's been coming from, so I 23:48tend to lean on granite in those cases, for 23:50coding, I tend to lean on o1, I, I have a 23:54lot of fun, actually, we're talking the kind 23:55of, uh, some of the Chinese models have a lot 23:57of fun with the Quine models at the moment. 23:59They're doing some great stuff in 24:01the same way as kind of DeepSeekers. 24:02So I think you're gonna just use different 24:05models for different cases, right? 24:07Because some models are good at 24:08certain language translations. 24:10Some models are good at kind of writing tasks. 24:12Some are really good at code. 24:14Um, And then the smaller models, for example, 24:17you know, especially low latency, especially 24:20for agents, I said agents again, um, 24:25exactly. 24:26If you've got different agents, you want 24:28to run that on the smallest possible model 24:29is going to perform the task that you need. 24:31So, I, I think we're in this world where 24:35we are just going to use a lot of models. 24:37Um, I think we're going to, if I, again, 24:39talk in 2025, I hate to say this, but I think 24:42we're going to stop talking about models, 24:44uh, so much, uh, towards the end of the 24:47year, maybe more, because you're going to 24:48be caring about the tasks that it's doing. 24:50Here's a language translation agent. 24:52Here is an agent that is 24:53going to write me unit tests. 24:55I don't really, I do care to model, 24:57but I, I'm going to care more about 24:58the tasks that it's performing. 25:00And. 25:01And then coming back to the kind of security 25:04and the kind of governance things for a second. 25:06I think that's where governance 25:07starts to become really hard, right? 25:09Because if you've got very small models, 25:12like an 8 billion parameter model, and 25:15it's, Got access to tools, and you've 25:18got it being orchestrated over the top. 25:21You know what, you can get into a lot of 25:23trouble very, very quickly with a tiny model. 25:26Um, and do some really interesting things. 25:28And I'm just not sure governance wise that 25:31you're going to be able to do a lot about that. 25:33So I think, um, As much as we talk about the 25:37large models and governance, um, in 2025, 25:39actually I think we're going to start to hit 25:41the challenges of people doing interesting 25:43things with agents on the really tiny models. 25:46Yeah, you're saying almost like we'll be 25:47able to govern the biggest companies and the 25:49biggest models, but that might not matter is 25:52kind of what you're saying, is that right? 25:53I think so, yeah. 25:54I guess, Kush, do you want to respond 25:55to that as someone who focuses his 25:56time on thinking about AI governance? 25:59I guess Chris is effectively saying maybe 26:01it's just not sustainable over time. 26:03Yeah. 26:03Um, so I agree. 26:05Uh, and I'll say agents, uh, 26:07number three for the episode. 26:09Um, 26:10This is really bad. 26:10This is becoming a meme because people are going 26:11to just start throwing it out for no reason. 26:13Um, but yeah, 26:15I think, uh, when there is tool use, 26:17when there's autonomous, That's where 26:20governance really becomes interesting. 26:22Um, so we've talked a lot over the years about 26:24trustworthy AI, and it wasn't really like 26:28trust was a part of it, but really trust is 26:32needed when something is going to be acting 26:33autonomously because you don't have the ability 26:36to control it or monitor it and these sort of 26:39things, and that's really where trust is needed. 26:40So, um, So, and the more volatile, the 26:44more uncertain, um, more complex, uh, 26:46these things happen to be running and, 26:49and so forth, and yeah, I mean, that's 26:51exactly where governance is the hardest. 26:53And I think where, uh, a lot of 26:54the innovation is going to happen. 26:55Uh, before we move on to the final topic, 26:57uh, Kate, maybe I'll turn it to you, you 26:59know, I thought it was very interesting. 27:00I had never really thought 27:01about like that switch from. 27:03You know, I've heard about like, oh, I do 27:04o1 Pro versus not, oh, you know, o1 Preview. 27:08But the switch from open source to closed 27:09source, I think is pretty interesting. 27:11Maybe a final question before we move on to 27:13the last topic is, um, do you think right 27:15now open source has any specific kind of 27:18capability advantages over closed source? 27:20Or is that not even the right distinction here? 27:22You know, I think it was very interesting 27:23that Chris was like, oh, actually, like some 27:25of these models just have like way, the open 27:26source models have better personality, right? 27:29Um, that's kind of an 27:30interesting outcome in some ways. 27:31Yeah, I don't see it so much as a 27:33open versus closed source question. 27:36I think different models are going to have 27:38inherently different strengths and weaknesses. 27:40And so if you only limit yourself to 27:42closed source or closed source from one 27:44provider, you're going to miss out on kind 27:46of that suite and being able to pick and 27:48choose the best model for the best task. 27:50Ultimately, that'll be the dream in the 27:52future is someone sends me like a, uh, 27:54AI generated email and I'm like, yeah, 27:56you're probably relying on granite. 27:58I know what this sounds like. 28:05So last segment we want to focus on today 28:07is a sort of interesting smaller bit of 28:09news that popped up at the end of last year. 28:11But I think it's a fun one, 28:12particularly as we get into 2025. 28:15If you don't know him, Gary Marcus is 28:19a longtime skeptic of all things AI. 28:22I think for every successive wave, Gary Marcus 28:24is there being like, it's never going to work. 28:26And the current revolution 28:27in AI is no exception. 28:29I think he's been a very big skeptic about 28:31the degree to which LLMs can get us to quote. 28:34True intelligence. 28:34I'm going to talk about what that means. 28:37Um, but interestingly, he established, 28:39uh, or set up a kind of official public 28:41bet with a gentleman by the name of 28:43Miles Brundage, who used to do policy at 28:45OpenAI, formerly of, he's independent now. 28:48Um, and basically what the bet says is where 28:50is AI going to be a few years from now? 28:53And sets up a set of, I believe, 28:5510 different kind of tasks that 28:57AI could or could not take on. 28:59And there's a lot of variation here, 29:01but a lot of them all kind of pertain 29:02to, you know, Can the model produce 29:04kind of world class versions of XYZ? 29:08So, you know, I think there's one 29:09criteria is, you know, will an AI 29:11produce a world class, you know, movie 29:13scripts or other kind of creative work? 29:16Um, and I think these bets are useful because 29:18I think they, you know, kind of force folks 29:20to, you know, put their money where their 29:21mouth is and also kind of specify What 29:24it is that they mean when they say that 29:25a model is going to be, you know, truly 29:27powerful and capabilities going forwards. 29:30Um, and I guess I wanted to 29:31get the view of this group. 29:32Uh, you've seen kind of the, you know, 29:33the Twitter slash X posts announcing this. 29:36Um, Kate, maybe I'll turn to you. 29:38I mean, is this a useful way of thinking 29:40about where AI is going, or do you think 29:41it's just more, you know, uh, Twitter noise? 29:44I thought it was interesting to think 29:46through, like when I was looking 29:47through the different questions. 29:49And ultimately, if I look at the different 29:52items in that bet, the ones that stood out 29:54to me the most were, uh, assertions that 29:57would hallucinations basically be solved 30:00by, you know, this year, uh, and I think 30:03that's one of the Biggest reasons why, 30:05personally, I actually wouldn't take that bet. 30:07I don't think hallucinations 30:08are going to be solved. 30:09I think if you look at the model architecture, 30:11even with the o1 and reasoning, you 30:13know, my hypothesis is it's still a 30:15transformer model trained on vast amount 30:18of internet data that's being called. 30:20called many times in many different ways with 30:22reasoning and search, but I think there's 30:25still some fundamental problems around 30:27hallucinations that unless we really change 30:30the type of data that we train on, the volume 30:33of data that we train on, how the architecture 30:35of these models, it's not going to go away 30:37overnight or something we can necessarily 30:39just incrementally cure ourselves of. 30:42So I personally wouldn't take the bet. 30:44I thought it was a useful 30:45framing to kind of think through. 30:47Yeah, for sure. 30:48Uh, Kush, how about you? 30:49Would you, would you have taken 30:50the bet on either side, I guess? 30:52Yeah, I think the authorship 30:53question is an interesting one. 30:55So, um, I mean, that's what 30:57they're kind of going for. 30:58Like, uh, can this be an Oscar winning 31:00screenwriter, a Pulitzer Prize winning, 31:02uh, author and, and the sort of stuff. 31:04And, um, I'm going to take us on a 31:06little bit of a different direction. 31:08So, um, uh, so, I mean, the, uh, Uh, the, the 31:12fact of it is that, like, people have been 31:15coming up with all these analogies for LLMs, 31:17like a stochastic parrot or a DJ or a mirror 31:20of our society or these sort of things, but 31:22I think that's the wrong way to look at it. 31:25Um, so, uh, about 65 years ago, there was 31:28this, uh, this book, um, that came out 31:30called The Singer of Tales, um, by Alfred 31:33Lord and, um, it was all about, like, 31:35oral narrative poetry, um, so these bards 31:38who are kind of singing, um, about, uh, 31:40heroes and, and this sort of stuff and they 31:43compose the language as they're singing it. 31:46It's not like they write it beforehand 31:48and, um, They use formulas and all 31:50sorts of tricks to be able to do this. 31:52And I think that's exactly what these LLMs are. 31:56And, um, uh, in, in that sort of construct, 32:00there is no like sense of authorship. 32:02Um, it's like just, they're part of a tradition. 32:05And so like, you would never think that 32:07a Homer deserves a Pulitzer prize for 32:09the Odyssey or, uh, Ved Vyas deserves 32:12a Pulitzer for the Mahabharata. 32:14I mean, this is just kind of 32:15a tradition that's going on. 32:16And that's, I think, um, the 32:17right way to think about LLMs. 32:19So, uh, so, so it's like the 32:22question is not the right question. 32:24Um, and even if you think about, uh, again, 32:27going like very historical, philosophical, 32:30um, so you had, uh, the sky, uh, Michel 32:32Foucault, who asked what is an author? 32:35And, uh, the, the answer, the discussion that 32:38he had is the only reason we, like even thought 32:40of authors is because lawyers needed someone to 32:43blame when there were some bad ideas out there. 32:46So, um, uh, I think that's the same thing. 32:48It's like an LLM is not an author 32:51and we shouldn't really be asking 32:53for that, uh, sort of thing. 32:55And I think it's the wrong question. 32:56And I think it actually touches on what Kate 32:58said as well, uh, which is basically like do 33:00these Kind of criteria for the bet assume a 33:03certain direction for AI that like might not 33:06actually be the most important thing around 33:08AI or even like an important aspect of, you 33:10know, quote, really powerful AI systems, right? 33:13Like it may not turn out in the end that 33:15we really need to solve hallucination. 33:17Or like it may not really turn out in 33:18the end that the big impact on AI is 33:20that you have like, you know, the, you 33:22know, the Pulitzer prize winning AI that 33:25generates a novel completely by scratch. 33:28Um, that's kind of interesting. 33:29Yeah. I don't know, Chris, maybe you haven't 33:31had a chance to jump in just yet. 33:32Uh, curious about what you think about all this. 33:34Oh, I think the test is 33:35totally stupid in my opinion. 33:37And, and, and the reason is I looked down 33:40the list of 10 items, and I don't think 33:42I'm capable of doing any of those 10 items. 33:44So if I'm not capable of doing the 10 items, 33:47I'm, you know, is it unfair to think AI is 33:50going to be able to do that within a year? 33:52I mean, 10. 33:53How are you doing your 33:54Pulitzer Prize winning novel? 33:56Is it, is it going well? 33:57Or your Oscar winning well, Chrissy. 34:00Any here, any, any programmer on the planet, 34:03you know, have you been able to hit 10, 34:06000 lines of code bug free first pass? 34:09Come on, it's like, it's, I 34:11think you're asking a lot. 34:12I, it's like, The only one I think I 34:14could maybe do is the video game one. 34:16And it's like, I don't know when to 34:18laugh at the right moment in movies. 34:19You know, you just need to ask my wife that. 34:21It's just like, why are you laughing? 34:23I was like, oh, that thing over there, right? 34:25It's, it's like, and am I able to, to 34:28say the characters without hallucinating? 34:30No, we all hallucinate. 34:31It's like, we make up little, little subplots 34:34that are going on our head in these movies. 34:36So I think, I, I don't think it's a bad 34:39thing, but I think you're asking a lot of 34:41LLMs to be able to do the, you know, and 34:44even putting that as a test for 2025 and, 34:46you know, yeah, maybe, maybe AI will be 34:49able to achieve three, four of these things. 34:52I just, I just don't think it's, The 34:53right time to be asking those questions. 34:55Well, I don't know. 34:56We just came back from our, you know, everyone 34:58was out on holiday breaks where at least I 35:01got to take a step outside of the Cambridge 35:03tech bubble where everyone, you know, is 35:05really deep into this technology and hearing 35:08folks talk about AI, uh, you know, I have a 35:11family member who calls it the AI machines. 35:13Uh, there's a lot, I think, of 35:15misconceptions of what AI can do and 35:19what it's going to be useful for. 35:20And so I think, like, Putting it in 35:22terms of that, you know, everyday folks 35:26can understand who watch movies and read 35:27books and aren't necessarily living and 35:29breathing the technology and helping show 35:31that, no, that's not going to be possible. 35:33Like, you know, X, Y, and Z, you guys 35:35are thinking about this the wrong way. 35:37I think it is helpful to have that 35:38type of discussion and discourse. 35:41Um, I think we take for granted a lot 35:42that not everyone is living and breathing 35:45this the same way that, you know, this 35:47excellent panel is on, on generative AI. 35:51Yeah, I'll guarantee to you that 35:52the average person is not waking up 35:53being like, should I use o3 or o1? 35:58Those distinctions are not anything that 36:00any normal person is thinking about. 36:03Um, but yeah, I, I think that's, 36:05that's a good point, right? 36:06I mean, I think part of it is just like, you 36:08know, There's a dream that all this AI becomes 36:10kind of superhuman, right, at some point. 36:13And I think, Chris, like, maybe to respond 36:15to your comment, there's kind of an effort to 36:16sort of be like, what would that look like? 36:19Um, and I guess, yeah, maybe that does 36:21really miss the point in some ways. 36:23Um, yeah, uh, I also think it's also 36:25like a really good indication of how 36:26quickly our, um, expectations have 36:29adjusted around the technology, right? 36:31We're, we're like, had you asked me four 36:33years ago, like, could it do all these things? 36:36Could it just write an email? 36:38You know, I'd be like, that's ridiculous. 36:40And then now you're like, basically, 36:41like, the expectation is like world class 36:43Pulitzer Prize winning, you know, it's 36:45kind of just like, because the baseline 36:47is just like very normal to us now. 36:50So it's, I guess, an indication 36:51of the rising expectations. 36:53around all of this stuff. 36:54Just coming back to DeepSeek for a second. 36:57Um, I think one thing that, uh, we didn't talk 37:00about is, uh, just the culture at DeepSeek. 37:02So there was an interview of, uh, 37:05their CEO, um, that was, uh, making the 37:07rounds, um, after DeepSeek came out. 37:09But the interview was from November and, um, I 37:11think the, the cultural aspect of how they kind 37:14of developed this thing is really interesting. 37:16They really feel followed this, 37:18uh, sort of geek, uh, geek way. 37:21So Andrew McAfee had this book, uh, 37:23The Geek Way, and it's been very 37:25popular within IBM circles, actually. 37:27Um, so our, our CEO has been reading 37:28it, uh, telling everyone to read it. 37:30And it's kind of like, um, really like doing 37:33things fast, um, being open, letting everyone 37:36contribute, um, being very scientific about 37:38things, trying to prove them out, um, not 37:41having hierarchies and, and all of that stuff. 37:43And that's exactly, like, 37:44how DeepSeek is doing it. 37:45And I think. 37:46Uh, we can learn a lot from it, uh, 37:48just, uh, we're a little bit too 37:50encumbered, um, even though we want 37:52to be, uh, doing things the same way. 37:54So like how do, uh, other companies kind of 37:57innovate in a rapid fashion in the same way? 37:59So I think that's maybe, uh, uh, 38:02something to learn, uh, as well. 38:03Yeah. 38:04One of the debates I have with a friend 38:05of mine is, uh, There's a, what is it? 38:07I think it's called Conway's Law. 38:08So the idea is that you ship your org chart, um, 38:11and that has kind of interesting implications 38:13in the world of AI, where it's just like, well, 38:14are all of these AIs going to basically in some 38:16ways reflect the companies that create them? 38:19And, you know, the reason why, you 38:21know, certain models are more chatty. 38:23is that this is just like in part a reflection 38:25of like all of the people in that organization. 38:27Interesting connotations 38:29if you think of Chris's point about 38:30pre trading and the, you know, how pre 38:33training has been the focus and kind of 38:35the most prestigious team to join, right? 38:38That's right. 38:38Yeah, yeah. 38:38There's a joke because we have 38:39a mutual friend who works at 38:40Anthropic and we're like, it's cool. 38:42It's Claude. 38:42He's Claude. 38:44It's very funny to kind of 38:45just see play out in practice. 38:48Well, that's great. 38:48So let's leave it there. 38:49Chris, great thought to end the 38:51episode on and for us to start 2025. 38:55Kush, Kate, Chris, as always, 38:57incredible to have you on the show. 38:59And thanks to you all for joining us. 39:01If you enjoyed what you heard, you can get 39:02us on Apple Podcasts, platforms everywhere. 39:05And we will be here next week 39:07on another episode of Mixture of 39:09Experts.