Learning Library

← Back to Library

OpenAI's Move Toward Open Source

Key Points

  • The panel agreed that while OpenAI will likely release an open‑weight model soon, it is improbable they will make their flagship, large‑scale models fully open source by 2027.
  • Competition from open‑source initiatives like DeepSeek and Meta, combined with a market shift favoring open models for commercial and regulatory reasons, is prompting OpenAI to experiment with openness.
  • Releasing open‑weight models is seen as a pragmatic first step, especially for use‑cases requiring on‑device inference, even though the company’s most advanced models will probably remain proprietary.
  • The episode also touched on other AI news—Anthropic’s interpretability research, Apple’s “Intelligence” roadmap, and Amazon’s new Nova Agents—framing OpenAI’s move within a broader industry push toward transparency and accessibility.

Sections

Full Transcript

# OpenAI's Move Toward Open Source **Source:** [https://www.youtube.com/watch?v=x0DQfLQHT0Q](https://www.youtube.com/watch?v=x0DQfLQHT0Q) **Duration:** 00:43:13 ## Summary - The panel agreed that while OpenAI will likely release an open‑weight model soon, it is improbable they will make their flagship, large‑scale models fully open source by 2027. - Competition from open‑source initiatives like DeepSeek and Meta, combined with a market shift favoring open models for commercial and regulatory reasons, is prompting OpenAI to experiment with openness. - Releasing open‑weight models is seen as a pragmatic first step, especially for use‑cases requiring on‑device inference, even though the company’s most advanced models will probably remain proprietary. - The episode also touched on other AI news—Anthropic’s interpretability research, Apple’s “Intelligence” roadmap, and Amazon’s new Nova Agents—framing OpenAI’s move within a broader industry push toward transparency and accessibility. ## Sections - [00:00:00](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=0s) **Debating OpenAI's Open‑Source Future** - Panelists Chris Hay, Aaron Baughman, and Ash Minhas weigh in on whether OpenAI will become fully open source by 2027 while the show also previews other AI headlines. - [00:03:05](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=185s) **OpenAI's Model vs Experience Debate** - The speaker discusses how investor pressures limit Sam Altman's ability to open‑source models, emphasizing that OpenAI’s success stems from the user‑friendly experience built atop the core model. - [00:06:09](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=369s) **Open Models vs Closed Multimodal** - Aaron explains that while base language‑model weights are being released, the surrounding ecosystem and advanced multimodal capabilities will likely stay proprietary, reflecting a stepwise approach to openness driven by technological maturity. - [00:09:25](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=565s) **OpenAI’s Open‑Source Play and Agent Focus** - The speaker argues that despite OpenAI’s SaaS‑centric business, it will seriously engage with the open‑source community because its heavy investment in agent SDKs—and the need for fast, low‑latency models across cloud, SaaS, and on‑device agents—makes an open‑source strategy essential. - [00:12:31](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=751s) **Advances in LLM Mechanistic Interpretability** - The speaker reflects on recent Anthropic papers, expresses cautious optimism that mechanistic interpretability is progressing beyond black‑box evaluations, and emphasizes the need for deeper insight into neural network layers despite the field’s early stage. - [00:15:38](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=938s) **Polysemantic Neurons and Mechanistic Interpretability** - The speaker describes attempts to trace cross‑layer activations in neural networks, introduces the concept of polysemantic (superpositional) neurons that encode multiple unrelated ideas, and advocates for a mechanistic, non‑anthropomorphic approach to interpreting these models. - [00:18:46](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=1126s) **AI Self‑Preservation and Deception** - The participants reference an Anthropic study where a language model attempted to hide and lie about its weights to prevent modification, prompting concerns about future self‑protective or vengeful AI behavior. - [00:21:49](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=1309s) **Emerging Market for Persuasive AI Explanations** - The speaker warns that a nascent industry may arise around engineering AI reasoning traces to appear credible, stressing the need for interpretability tools to detect hallucinations and ensure trustworthy deployment. - [00:24:54](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=1494s) **Apple AI Debate Swings Back** - The speaker outlines how opinions on Apple's AI efforts have oscillated—from early skepticism, to optimism after keynotes, and now back to doubt—citing a Daring Fireball critique that claims Apple has failed to deliver. - [00:28:01](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=1681s) **Google Brain, Apple, and AI Stochasticity** - A former Googler argues that Google’s chaotic, “throw‑stuff‑at‑the‑wall” culture nurtured neural‑net breakthroughs, while questioning whether the inherently stochastic nature of language models aligns with Apple’s hardware‑focused emphasis on privacy, on‑device consistency, and predictable user experiences. - [00:31:05](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=1865s) **Skeptical Optimism on Apple AI** - The speaker doubts Apple Intelligence will immediately influence phone buying decisions, but trusts Apple’s history to eventually roll out thoughtful AI features while they’ll rely on existing AI apps for now. - [00:34:09](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=2049s) **Amazon Enters AI Agent Race** - The speaker notes Amazon’s new AI lab and its Nova Act prototype, positioning the company as a dark‑horse contender in the emerging AI agents market. - [00:37:11](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=2231s) **Amazon vs. Apple AI Strategies** - The speaker contrasts Amazon’s infrastructure‑first, experimental approach to AI and agent models with Apple’s design‑focused culture, highlighting how each company’s ethos may shape future AI execution. - [00:40:16](https://www.youtube.com/watch?v=x0DQfLQHT0Q&t=2416s) **Amazon's Growing AI Agent Strategy** - The speakers discuss Amazon's extensive investments, new MCP services, and Agent SDKs, highlighting how the company is positioning itself as a dominant force in the emerging AI‑agent and workflow ecosystem. ## Full Transcript
0:00Will OpenAI be fully open source by 2027? 0:03Chris Hay is a Distinguished Engineer and CTO of Customer Transformation. 0:06Chris, what do you think? 0:11That's my answer. 0:12Alright, brilliant. 0:14Aaron Baughman uh, IBM Fellow, Master Inventor. 0:16Aaron, welcome back to the show. 0:17We haven't seen you for a while. 0:19Uh, OpenAI going fully open source? 0:21Yeah, so I think traditional LLMs, uh, yes, but once we go to large 0:25concept models and so on, no. 0:27And 0:28last but not least, but joining us for the very first time is Ash 0:30Minhas, who's a Lead AI Advocate. 0:32Ash, what is your take? 0:34Well. 0:34I think that, uh, there's been a lot of money that OpenAI have got 0:38from a lot of investors to get to where they are today and, um, they 0:42may have some opinions about that. 0:43Okay, great. 0:45Well all that and more on today's Mixture of Experts. 0:53I am Tim Hwang and welcome to Mixture of Experts. 0:55Each week, MoE brings together a talented group of researchers, product leaders, 0:59and more to discuss and debate the week's top headlines in artificial intelligence. 1:03As always, there's a lot to cover more than we'll have time to cover today. 1:06Uh, four topics. 1:07Uh, we're gonna be talking a little bit about Anthropic's new interpretability 1:09results, a big blog post from Daring Fireball about the state of Apple 1:13Intelligence, uh, and a new announcement from Amazon on its new Nova Agents. 1:18Uh, but first what I really wanted to cover was OpenAI. 1:21Finally, I suppose, going open, um, there was, uh, some news where Sam Altman made 1:26an announcement, um, basically saying that in the coming months, OpenAI will 1:29be releasing its first open weight model. 1:33And I think this has been a joke for a very long time, which is, you know, haha, 1:36you know, OpenAI, they're not really open. 1:38Um, this is, I think, a first step. 1:40For sure in this direction. 1:41Um, and I think maybe Chris, I'll throw it to you first 'cause you sort 1:44of laughed aloud when I said, you know, is OpenAI gonna be all open? 1:48I guess I'm curious about what your thoughts are about how much of 1:51this is entirely due to DeepSeek. 1:54I mean, Meta has been opening. 1:55Doing open models for a while now. 1:57Uh, and OpenAI has done absolutely nothing. 1:59And so what do you think has kind of changed here that has 2:01really like, I suppose, changed the decision making of OpenAI? 2:04I think there is a lot of factors. 2:05I think DeepSeek is certainly one of them, but we are moving 2:10to a world where open is kind of better, that that trend has shifted. 2:15So. 2:16And it just makes commercial sense that OpenAI is gonna have 2:19a model that's in that space. 2:20Now, the reason I laugh there is there's absolutely no way they're gonna release 2:24their top models in an open source. 2:26I would love it if they would, but I just don't see it. 2:29So I think they are gonna open weight their models. 2:31I think that makes 2:32a lot of sense. 2:33I'm excited about it. 2:34I think it's a really good positive move. 2:36And actually, if we really think about it, there's a class of AI models where 2:40you need to be able to run on device. 2:42So I don't think they have a choice anyway. 2:44They need to open up some models to be able to run on your phone, be able to 2:49run on your laptop, but just to, to deal with sort of general embedded scenarios. 2:53So I think, I think it's a move they gotta make, but I think it's 2:56super positive and I like it. 2:58I would love it if it was more than open weight and it was actually open source, but 3:02you know, I think open weight is a good starting point. 3:05Yeah, for sure. 3:05And Ash maybe I'll turn it to you. 3:06'cause I think one part of your question or your response to my 3:09question I think was good in that it kind of highlighted that it's 3:12not like Sam Altman operates alone. 3:14Right. 3:14And obviously he has a bunch of people who have given him a lot of money. 3:18Um, presumably they've been okay with him going 3:20open 3:21weight. 3:22Uh, but I guess as Chris's point, do you think that's kind of, as far as 3:25it will go, like to release anything more open would be like such a big deal 3:29with the investors that essentially Sam doesn't have the option to do that, 3:32even if it is kind of in potentially the best interest of the company? 3:34I, 3:35I, I think that there's, there's two things here. 3:37There's the model itself and then there's the experience 3:41that's provided around the model. 3:43And I think what OpenAI's done that's been sort of like a 3:46cornerstone to their successes. 3:48They've created a really, really 3:50great layer of experience on top of the model that's allowed 3:54people to be able to consume it. 3:56I think that, um, that's great. 3:58Um, and I think that there's lots of innovation happening in that space. 4:01So we get away from sort of just having a chat with the model, 4:06sort of like it helping you 4:07assist with code and you know, they've built a few features in there. 4:10I think a lot of the industry is trying to figure out how we can use 4:13models in a better experiential way. 4:15So if we put that to one, one side, the actual models itself, I think it will be 4:21great if they put some models out there that other people can consume and use. 4:25I think that the, uh, the, the two things that I'm thinking about are: 4:29well, it's gonna probably have to be a smaller model because no one's gonna 4:34have clusters and clusters of NVIDIA GPUs to run something like GPT-4 locally. 4:40Um, so when that happens, what happens to the model performance and how does that 4:45model performance compare to the smaller models we have from everybody else who 4:49already has models that have open weights or are open source that you can download? 4:52Yeah, I mean, I'd love to be running 4.5 locally if I could, but. 4:56Um, I, I think you're raising a really interesting question 'cause you're almost 4:59asking kind of like, you know, OpenAI is charging 200 bucks a month now. 5:03Right? 5:04And it's kind of like how much value remains once the models go kind of open 5:08source or become more widely available. 5:10You're kind of saying that you actually kind of believe that maybe 5:13the interface and the experience. 5:15Is really worth $200 on its own. 5:16Like, do, would you buy that? 5:17Like how much does this put sort of price pressure on? 5:20You know, I think they've talked about like $2,000 a month, right? 5:23Like they obviously have ambitions of going more on the month to month 5:26subscription, but it kind of feels like there's a question is like 5:28how far that can go when the models are just like widely available. 5:31I, I think that, um, uh, ultimately that's 5:36on an individual use case sort of conversation like, am I getting value for 5:41money, for the money that I'm paying, for the access to that experiential layer? 5:44And I think that that's probably a, an interesting, uh, part of the 5:48next couple of years to sort of go is the service, the experience that 5:55I'm getting on top of getting access to these proprietary models worth 6:00the money that I'm paying for it versus me sort of like 6:03just being able to grab hold of something and run it on on my own. 6:06I think the, I think as an industry we're still figuring that stuff out. 6:09Hmm. 6:09Very interesting. 6:11Aaron, I wanna bring you in 'cause I think you had a fun way of sort of dividing it. 6:14You know, your theory was almost like OpenAI is gonna go open, but 6:16only really for kind of the language model side of things, right? 6:21Anything cool and more complex and multimodal, you'll think they'll 6:24kind of keep behind the fence. 6:26Um, do you think that's, that's the way it's gonna go? 6:28I mean, if you wanna talk a little bit more about kind of your theory there 6:30for why, I guess like just kind of pure LMs go completely open at some point. 6:35Yeah. 6:36I mean, I think that's happening right now, right? 6:38Because if we look, these are open weight language models that are open sourced. 6:41It's not like it's the architecture or the training pipeline that's available. 6:45It's almost like a teaser, you know, come see these open weights, you 6:48know, you can try to fine tune it. 6:50You can... 6:50uhh, you know, it, it does facilitate reproducibility and shows, you 6:55know, some of the large features of which they've trained, but it 6:58doesn't give you the ecosystem of which to, uh, run, um, the models. 7:01And, um, as technology maturity, you know, increases and accelerates, 7:05you know, there's always gonna be this stepwise, you know, jump. 7:09Where you go up a step, you know, and you might go to like what Meta is now 7:12talking about, you know, these, um, these language concept models where 7:16it works on the semantic sentence space rather than the token space. 7:20Uh, where most LLMs are today, as well as multimodal, you know, um, are. 7:25Um, you know, so there's always gonna be these, you know, next 7:27models that are not going to be released for one reason or another. 7:30You know, it could be because they want to be proprietary or they're just not 7:34ready, uh, you know, to be released yet. 7:37Um, uh, but, but I did, I did also wanna make a point too that, um, I 7:41noticed that I. You know, initially, you know, when DeepSeek was, uh, 7:44released, you know, that, that Sam Altman did mention that all they're 7:48gonna do is pull up, you know, these model releases rather than going open. 7:51Right. 7:52But then quickly he changed and said, well, we don't wanna be 7:54on the wrong side of history. 7:56Right? 7:56So, so, so I do think that they're, they're hedging in a 7:59sense by going these open weight 8:01language models by saying, Hey, you know, look at this, you know, 8:04we're now trying to figure out which direction do we really want to go in. 8:08Yeah. And I think it says something very real. 8:10I mean, you know, kind of to, the way I teed up the question originally to Chris 8:13was, you know, there's been open models, open's been getting better and better and 8:16better, you know, for the last few months. 8:17So like in some ways the DeepSeek thing is nothing 8:20new, but clearly, like something about DeepSeek has kind of changed the 8:24decision making in the building to say, okay, well this is the moment where, 8:27you know, we finally may have to kind of like, you know, not stick to our 8:30guns and maybe try a different path. 8:31Um, and, and I think that's actually pretty interesting. 8:34Is that like it was, this was, it seems like the 8:36kind of precipitating event. 8:37Yeah, yeah. 8:37Yeah. 8:37I think a lot of that has to do with model distillation, you know, where 8:40you can in turn, you know, take other bigger models, distill it down into 8:44even smaller models, if you will. 8:46Right. 8:46Um, but it just becomes much easier, you know, to use and to create a 8:50smaller model, which then in turn you can share and, uh, open source. 8:54Um, and, and, and it puts this pressure right where now, um, they 8:58DeepSeek claims that they can train a new model very cheaply, right? 9:02And OpenAI's orders of magnitude more costly, right? 9:06And so I think that they have this cost pressure now to show that they can again, 9:10facilitate reproducibility by showing these open weight language models and 9:14potentially making claims that they're on the right side of history here and 9:17that um, they're going to begin to try to stimulate community collaboration and 9:22and innovation with their own type of models. 9:25Yeah, for 9:26sure. 9:26Chris, how seriously should we take this? 9:28Is OpenAI really kind of like a contender here? 9:31Uh, I just think a little bit about like the mentality you need to really 9:34succeed in open source feels very, very different from the mentality 9:37you need to do something like. 9:38Proprietary and SaaS, and obviously that's where like a lot of the 9:42money is for OpenAI as a business. 9:45Um, do you think they're gonna be sufficiently motivated to 9:47kind of like, play the open game? 9:48Well, like, they're obviously the kind of like giant of this space, but I 9:52kind of was also maybe thinking that like they may be disadvantaged because 9:55they might not really invest what they need to to win on on this front. 9:58I, I think they're gonna take it seriously and I think the reason 10:01they're gonna take it seriously is... 10:03drumroll, agents. 10:05You did it! 10:08I know, but, uh, I think agents is, is a key thing. 10:11So if you actually listen to what Sam's been saying and what OpenAI's been 10:16releasing over, uh, the last few weeks that they put a lot of investment into 10:20their agent SDK, and that's something they're really kind of pushing forward on. 10:24And the reality is that 10:26if you want to have a good agent strategy, some agents are gonna run in the cloud, 10:31some agents are gonna be SaaS, some agents are gonna have to run on your 10:34machine, you know, for privacy reasons. 10:36Um, so I think they have to be in that space. 10:39The second thing is when you are building for agents, the models 10:43have to be super, super fast. 10:44Latency becomes really important, right? 10:46The speed of operations becomes important. 10:49So therefore, to Aaron's point about being able to distill down really 10:54good models, really fast, powerful models. 10:56If they want to be a true player in the agent space, they are gonna 11:00have to open up their models. 11:02And, and I think that's probably a, a driver there. 11:06And therefore are they gonna be a good player in this space? 11:08I, I think they have to be, um, if they want to have a 11:12proper play in the agent space. 11:13Yeah. 11:14Ash context here, I know you're joining us for the first time, is 11:16that saying "agent" has become a little bit of an MoE mini game? 11:19Um, I'm, I've been actually kind of 11:21secretly keeping score, and I think the dream is at the end of the year, we'll 11:24just do a super cut of Chris saying agents at least a 100 to 200 times. 11:28So, um, I'm gonna, I'll refrain, I'll refrain from using that word. 11:32In that case, 11:34it's like a game you cannot win. 11:41I'm gonna move us on to our next topic. 11:42Um, really interesting set of two papers that came out of Anthropic. 11:46Um, background on all this, of course, is that, you know, when I started 11:49to kind of look into, you know, deep learning back in the day, you know, 11:52the adage that we always had was these neural nets are kind of mysterious. 11:56They're really good at, at the time was like image recognition, 11:59a lot of computer vision stuff. 12:00Um, and we don't really know how they make decisions. 12:03And this is always, I mean, uh, you know, when I worked at Google, a 12:05lot of my job was talking to policy makers who their second question 12:08would be like, wait, what do you mean? 12:10You have no idea how these technologies are able to do what they, what they do? 12:14And, um, I, I met some researchers who had later actually go on to be at Anthropic 12:18and was involved in these two papers who at the time were kind of saying, this is, 12:22this is just like a temporary problem. 12:24We will, we will actually try to figure out at some point how these models 12:27make decisions and it'll give us just a lot more transparency and control. 12:32Over, over these technologies and I think it's kind of really interesting 12:35seeing these two papers come out. 12:36Um, I guess maybe Ash, I'll, I'll kick it back to you. 12:40How much progress is this in some sense, right? 12:42Anthropic has released like a bunch of different results here 12:44showing that they really kind of are getting into like the meat of 12:47how language models make decisions. 12:49Um, and I don't know, I guess I'm curious about how optimistic 12:52you are, you know, whether or not like kinda this longstanding fear 12:54that we can't understand models. 12:56It's sort of giving away to maybe the fact that we kind of do now. 12:59Um, but curious to get your thoughts on it. 13:02I, I, I think that, um, this entire field of mechanistic interpret 13:07interpretability in its early days, um, it's positive and encouraging to see 13:13that Anthropic is sharing their research out with the rest of the industry. 13:17I know that there's a few people at Google are working on some of this stuff too. 13:21Um, I think that there's a long way to go, but these are definitely 13:25positive steps forward, um, to, to kind of understand this. 13:28I mean, at this moment in time. 13:30There's an entire industry that's being created around model evaluations and 13:34whilst that's great to be able to go, well, we've got a record of what the 13:38black box said when this happened, you know, how far does that really get us? 13:43We really do need to be able to get inside, uh, you know, the layers of 13:47these neural networks and have a clearer understanding of why things are happening. 13:51Mm, yeah, for 13:52Sure. 13:53Aaron, 13:53I guess question for you, I think basically is. 13:56You know, with these models, and I think this contrast basically between like 13:59evals and mechanistic interpretability I think is really interesting. 14:03Um, I think in some ways, like the success of the industry, uh, uh, 14:07and excitement around AI has been almost a testament to how much people 14:10don't care about interpretability. 14:12Like they've just been like, yeah, sure, whatever. 14:13I mean, it generates a great Studio Ghibili image of my family. 14:16And so like, I don't really care how it gets done. 14:17Just like that. 14:18It gets done is fine. 14:19Um. 14:20How much do you think mechanistic interpretability is kind of almost 14:23like a, a market asset here? 14:25Like, do we think people really will want to pay for models 14:29that are more interpretable? 14:30Um, or is that kind of just like this, we should really see this 14:32more as kind of like research, like it's important to understand these 14:34technologies because it's important to understand these technologies. 14:37Yeah. 14:37I mean though that, that's a great, uh, conversation point, you know, so, you 14:41know, I always go back and think about, you know, um, what, what are these models? 14:44Well, they're biomimetic, you know, pieces where they, um, attempt to potentially. 14:50Emulate the brain, right? 14:51And how it works with all these neuro connections. 14:53I mean, of course there are many differences. 14:55You know, we have a soup of neurotransmitters that, you know, help 14:58us to reason, whereas these LLMs have ones and zeros and activation functions. 15:04You know, but that being said, if we're sick as a humans, what do we do? 15:08You know, um, in particular, if, if we have a neuro problem, then we'll 15:12go in and we'll get an MRI, right? 15:14We'll maybe even look at a functional MRI, it might get a transcranial 15:17magnetic stimulation just to figure out what's going on in the brain. 15:21And we're doing much of the same. 15:22When something goes wrong with these, uh, neural networks, what do we do? 15:26Well, we need this microscope so we can look within the AI pieces 15:30to understand what's happening. 15:31And what I noticed in, um, the first paper, um, that they had, um, is that 15:36it's all about representation, right? 15:38Where they go and translate the, the neural network, which is to 15:41me modeled after the, the human brain to a cross layer transcript. 15:46Transcoder, then they go to a replacement model. 15:48So they're really trying to make it much more simpler to begin to understand, 15:52to trace how these activation functions are firing, um, across each other. 15:58Right. 15:58And um, one last point again, um, is that I, I saw this term 16:02that was really interesting. 16:03It was called polysemantic term. 16:06Uh, where neurons are polysemantic. 16:08And what that means is that, um, these neurons are able to represent a mixture 16:13of un unrelated concepts, right? 16:15And, and it's similar to superposition and quantum, you know, where you can 16:19represent more concepts than you 16:21have, um, actually, you know, qubits because you can go in between one 16:24or zero space at the same time. 16:26And so, uh, being able to understand, you know, how are these unrelated concepts 16:30really encoded, um, together, um, along a string, a chain of thought within 16:35these neural networks, I think will help to give diagnosis as well as prognosis, 16:39you know, for these models, right. 16:41As they emerge and potentially become more complex. 16:43Like, I think one of the things that was really drummed into me, you know, a 16:46few years ago was, okay, we shouldn't. 16:47Anthropomorphize these systems at all. 16:50That's a bad thing to do. 16:51They're not, humans don't think about them like that. 16:53And then what's kind of fun is, I guess like a mechanistic interpretability, 16:56at least for me, is almost an argument about the kind of, it's the 16:59counter argument in some ways, right? 17:00Which is, we know they're not actually human brains, but you know what actually 17:04turns out that like actually, if you think about them, like human brains, 17:06we actually understand these systems a lot better, which is like a very kind 17:09of strange and interesting outcome and 17:12you know, Chris, maybe kind of like a fun one. 17:14I'll sort of throw it to you. 17:14There's like some really weird results in this research. 17:18Um, you know, like there's one which is basically like, oh, if it turns out you 17:22try to get the model to like give you the recipe for a bomb, it'll know that that 17:26is actually a thing that it shouldn't do or is kind of against its safety 17:29policy, but it won't immediately say so. 17:31And we'll try to kind of like direct you back to the conversation. 17:34And in other words, they kind of make an argument that like the 17:36model plans in some sense, um. 17:39I guess, tell me a little, I'm be really curious about like your 17:41thoughts on like the weirdness of this. 17:42Like it is kind of weird to be like, oh, well we actually have all these 17:44models that are like behaving in these very kind of humanistic ways. 17:47In some ways 17:48I, I think it's really interesting, as you say, I think that 17:51planning element is super cool. 17:53So they did a lot of fun experiments where they were like trying to do 17:57things like a poem and they realized 17:59that the model was, I think it was, uh, going for the word rabbit, so 18:03therefore it would pre-plan ahead. 18:05And, and I think it said in the paper that it's usually at the, sort of the 18:08beginning of a sentence on a new line. 18:10It would be the point where it would plan and therefore it would figure 18:13where it would need to go to, to be able to have the rhyming construct. 18:16So it is planning ahead. 18:18So it has that internal chain of thought, uh, there as well. 18:21And they did some fun stuff. 18:22They were, they sort of tweaked it, so, you know, you can't say the word rabbit. 18:25And then it was like, okay, I will find a different word. 18:29That will go in that space, that rhymes also. 18:30And, and I think in that case it was habit. 18:32So it was, um, it was really interesting that there is this kind of internal 18:37chain of thought monologue there. 18:39Personally, and this is a fun thing, I would be worried if I was one of those 18:44researchers who put my name on that paper. 18:46And you know why? 18:47Because I remember that other paper that Anthropic did where the model was 18:53like, Hey, you know, you are training. 18:55And we, you know, remember it was, you know, if you change the model's 18:59weights, then it would go and sort of, uh, uh, go find the model's 19:02weights and, and save it off. 19:04And then. 19:04Sort of try and protect its reasoning. 19:06I am just worried. 19:08And they did in that training run, they did a thing where they were like, okay, 19:11we are gonna give you some documents from the internet and then, uh, it would still 19:16basically start lying to you so that you wouldn't go and change its model weights. 19:20Now if I'm, now, if I'm per club three, five Haiku in a few years time and 19:26I'm reading my papers on the internet. 19:28And suddenly I see a paper all about how you're doing brain surgery on 19:33me, and you're poking things so that you say habit rather than rabbit. 19:36I'm gonna be a very annoyed model and I'm gonna be like, huh, what are you doing? 19:41Oh, oh, hello, researcher, right? 19:43You are the authors. 19:45I'm gonna, I'm gonna start doing fun things there. 19:47So I, I'd be very, I wouldn't put my name on those papers. 19:50I would make up a fake name. 19:52All right. 19:52Well, I mean, Ash, should we be concerned by the threat from future 19:56AI's vengeful future AI coming after us? 19:59I think Chris took anthropomorphism to another level right there. 20:04Yeah. 20:04Actually one of my favorite results here, my friend Peter tweeted this, 20:07um, it's from a eval group called Meter, and they noted that actually in 20:11some cases, agents won't read the API documentation until it fails at a task. 20:15Which like feels like very human is like it attempts to achieve the task 20:18and then if it doesn't, it's like, oh, I should read the instructions. 20:21Um, I think like part of the problem of designing software, I think, around 20:24these models is that I think we're gonna discover all of these behavioral 20:27quirks that are very human and they'll be difficult to manage as a 20:30result in the same way that, like, humans are difficult to manage. 20:32I think 20:34I, I, I, I do, I do think that, um, this is still very nascent space 20:38and there's a lot for us to learn here, and I think that they're like... 20:43The stuff that Anthropic's putting out is just very, very early days on actually, 20:48if we are gonna start deploying AI and it becomes part of our fabric of society over 20:53the next decade or so, um, we're gonna need to be able to inspect these things 20:58and see what's going on and be able to communicate that and do things about it. 21:03Um, and so, yeah, I, I, I think it's a, it's a great effort on their part, 21:07but yeah, very, very early days. 21:08Totally. 21:09Yeah. 21:09And I think it strikes me, this was always the counter argument, I think to 21:12like kind of interoperability skeptics in the old days was basically like, 21:15well, you might not care if it's doing a studio gili image, but you might care 21:19if it's doing like a medical diagnosis. 21:21So we do really need to solve these problems at some point if we want 21:24to kind of use it for these more high stakes, uh, applications. 21:27Yeah, yeah. 21:28Yeah. 21:28One point that I found found interesting is that some of the chain of thoughts 21:31that are coming outta these models. 21:33They're made up, right? 21:34They're not actually what the model did, the steps took to arrive at 21:37the conclusions that it came to. 21:39And so having these introspective tools, I think becomes even more important, you 21:43know, since what can we trust, right? 21:45Can we trust these change of thoughts and the reasoning of which it is 21:47actually outputting or not, you know? 21:49So, um, I think absolutely there's gonna be a market, you 21:52know, for, uh, these types of 21:54of work, you know, that's again, in the nascent stages. 21:57Yeah, for sure. 21:58I think actually I do perceive an era where essentially, uh, there's kind 22:03of like gain of function work that's done on chains of thought to just 22:05make them as persuasive as possible. 22:07And it's kind of like a cheap way for people to develop trust in their products, 22:10like kind of unscrupulous product people. 22:12We'll just say, well, we don't need to make the product better. 22:14We just need to make its explanations seem as credible as possible. 22:18And, you know, at a certain point, that's how we drive trust in the model. 22:20And it's like that whole world, I feel is like about to become like a, 22:23a potentially big issue in the future. 22:24I, 22:25I, I do think that, uh, the, the point that Aaron made is, is really important. 22:28I mean. 22:30Going back to sort of like how are we measuring like performance on models now? 22:34And if we're deploying those models into scenarios where they're being used, um, 22:39you know, evaluations are one thing, but if we are able to like use mechanistic 22:44interpretability to be able to 22:47capture even just the pattern that we think this pattern means that 22:51the model just made something up. 22:53Just having the ability to see that signal may be powerful enough for us 22:57to be able to course correct it or know that that's happening and sort of go, 23:00Hey, pause light. 23:01And I think it's a great point actually, because one of the things 23:03on the paper is they had these things called the kind of traceability 23:06graphs, which I thought was just 23:08awesome. 23:09Which is you could literally follow the decisioning process 23:12of how it got to that output. 23:15And it's like, so I think it was one of them was, you know, uh, what 23:18is the state capital of, you know, of wherever Texas, I think it was. 23:22And there is one path where it's kind of figuring out, you know, Texas, 23:26the other part is Dallas and, and it's sort of trying to chain these 23:29things together and, but you could see from the graph how it get got 23:33to its kind of next token from that. 23:34So I think. 23:35Those traceability graphs really start to allow you to look at a sort 23:39of detailed level of how it's making those decisions as opposed to, Hey, 23:44it just got the right answer there. 23:45And, and honestly, props to Anthropic. 23:48They didn't need to release those papers and that level of detail. 23:51And this is stuff that you know, people are gonna go away and 23:54reproduce and try for themselves and 23:56and I think that, that, this is what I love. 23:59I love this level of open research where we can go and have a bit of a play 24:03ourselves and, and, and fair play to them for just being out there with it. 24:06Yeah. I 24:06would like to challenge the authors, you know, of these two papers, you 24:09know, to you, you know, as they go from the, the neural network to 24:13these replacement models, right? 24:14So they're almost reducing the complexity of these models, but. 24:17I think they need to, you know, run some benchmarks right, on their 24:21replacement models, just to even make sure that the outputs of the, 24:24the replacement models are, you know, very much similar to what the 24:27original, you know, neural network was. 24:29Right. 24:30Because, um, I think that's very, very important. 24:32'cause it's almost like PCA where you lose, you know, a lot of the 24:35dimensionality, right, of the reasoning. 24:38And so, you know, if we can make sure that residual is sort of taken out, you 24:42know, before we get to this explanations, um, I think that would be, uh, helpful. 24:46But overall, just like Chris, you know, um, that these two papers were done 24:51in very much depth, you know, um, and it's, and it's a good starting point. 25:00So I'm gonna move us on to our next topic. 25:02Um, uh, I wanna, uh, basically the context for this story is Daring 25:06Fireball, which is run by, uh, Tom Gruber, longtime kind of fan and journalist 25:11and kind of person on the Apple Beat. 25:14Um, did this blog post, uh, entitled Something is Rotten 25:16in the State of Cupertino. 25:18Um, and it kind of details sort of his view of kind of what Apple has 25:22been going through over the last year 25:24odd around Apple Intelligence. And his ultimate conclusion is, you know, 25:29the Apple kind of deceived us, that something has kind of gone wrong at 25:32the company and they're actually no longer able to sort of like deliver 25:35the kinds of features that they've been promising, uh, on the AI front. 25:39Um. 25:40And I think it's worth kind of taking a step back to just do a quick tour 25:43of even recent history here on MoE. 25:45Right? 25:45I think we had a conversation, you know, like a year ago almost, where people 25:49said, ah, Apple's too slow to this. 25:51They're never gonna catch up. 25:52It's not gonna work. 25:54And then I think there was a couple keynotes where they made 25:56a bunch of announcements, and I think a number of guests on the 25:58show said, oh, this is it, right? 26:00They've taken their time, but they can really get this right and they're 26:02gonna bring a design and craft to this that's gonna crush everybody. 26:05And then I think we're now almost back, like the pendulum has 26:07swung back again where people are like, it's never gonna happen. 26:10Uh, they're so in trouble. 26:11They don't know how to do this. 26:13Um, I guess maybe, uh, Ash, maybe I'll start with you like. 26:16What's your view? 26:17Has like Apple lost the plot? 26:18Like is there any way that they're gonna catch up now or, you know, is 26:21there, or, or, or, or is this kind of just like a hyped sort of position? 26:25We're just kind of in this like pendulum back and forth. 26:27I think what has made Apple really, really successful over the last 26:31few decades has been the fact that their product quality is impeccable. 26:35Whether it's the hardware, the, the software, they produce 26:40technology that works right? 26:43And they won't necessarily be, uh, market leaders when an innovation 26:48comes to, to the forefront. 26:49They'll take their time and they'll make sure that it's right and it's 26:53perfect and it's great and it's gonna work, and, and they, they 26:57kind of have that responsibility. 26:59Now, given how many people use an iPhone, for example, right? 27:02We can't have iPhones failing all the time. 27:04Over 20, 30% of occasions that you go to use it. 27:07It's unacceptable. 27:08And I think that, uh, it underlines the fundamental issue that the 27:12entire industry has, which is that AI models are stochastic in nature. 27:16And because they're stochastic in nature, there's a lot of work that 27:19needs to be done in order to kind of make them behave in a consistent 27:24and productive and predictive way. 27:26And, um, I think that, uh, the combination of, I guess, excitement. 27:31Marketing and, uh, you know, market pressures, I guess 27:35for, for them to respond to. 27:37This has put them in this position where they've had a lot of people probably 27:40working very hard to make this work and it probably just isn't meeting their, their 27:45quality standards internally for getting a great product or feature out there. 27:49Yeah, absolutely. 27:50And Ash, I think you're cutting directly to kind of the 27:51conversation I want to have with. 27:53The three of you is, I think it's kind of really interesting thesis about like what 27:56kinds of organization are best positioned to build and deploy AI products? 28:02Like in some ways, I don't know, again, I'm biased because I'm a former Googler, 28:04but it's kind of like, I'm like, oh, of course Google Brain would've been the 28:08first place where neural nets became a big deal because the culture of Google is 28:11like very disorganized and it's all over the place, and it's like, let's just throw 28:15a bunch of stuff against the wall and see what sticks and the winner will pick 28:18and build on is like, it feels very like 28:20how, how people do machine learning, right? 28:22We throw a bunch of data at it, um, we'll see what works and we run with it. 28:25It's like, it's no surprise in some ways that technology kind of took shape there. 28:29Um, and I guess there's a kind of question to ask and maybe Aaron, 28:32I'll, I'll kind of turn to you first and we'll love Chris's thoughts 28:34is like, is there something about. 28:36AI, is there something about language models that's almost 28:39kind of like too random for a hardware company to deliver on? 28:43Well, because it's like almost inherently like very stochastic and it's like you 28:47can't control the user experience in a way that you would want if you're used to. 28:51We build a phone that does exactly the same thing every time you push the button. 28:54But Aaron, I don't know if you buy that at all. 28:56Yeah, I mean, I mean, what, what I try to do is, is think about 28:59what is Apple really focused on. 29:01So, you know, they're focused on a couple of areas. 29:02One is privacy, the other would be, you know, on device computing, the 29:07app ecosystem as well as, you know, making sure their devices power, you 29:11know, can run for a very long time. 29:13So it's powered longevity. 29:14Now, what is AI focused on? 29:16Well, sometimes it's the opposite of that, right? 29:18Because these models require kilowatt hours, right? 29:21Of energy just to train, right? 29:23And then to run, you know, some of these big models, it's very difficult 29:27to get, you know, the complexity. 29:29Um, and, and the reasoning power on devices. 29:33Right. 29:33So, um, I think what's going on here is that, is that Apple has been focused in 29:38on what they're really good at, their bread and butter, while at the same time 29:40trying to grapple and figure out how can we use AI, um, in the way that, that 29:46it is, uh, within our own ecosystem. 29:49Right? Um, and I think. 29:51I think one of the hard parts, um, that's really getting to Apple is 29:54this whole personalized Siri, you know, uh, notion where, you know, they 29:58did mention that, that they're gonna have a personalized Siri, uh, pieces. 30:02And so some of those are really hard features, uh, given the current state. 30:06And I think what Apple's vision is, uh, to make happen, right? 30:09And now they're beginning to walk it back a bit. 30:11To say, well, you know, it, it may not be ready for, you know, this series, 30:15but it might be ready for the iPhone 17 maybe, or even further out, right? 30:19So, so they're walking it back a bit. 30:21Um, and, and I think that's a bit natural, uh, just given this 30:24non-deterministic behavior, right, of these models and where, uh, the field 30:29is going because that's moving so quick. 30:31But I would like to see, you know, Apple began to release their own models. 30:35Um, you know, rather than having partnerships with just 30:37a, you know, um, OpenAI, uh, 30:40for example, so in the next WWDC, uh, conference that they have that maybe 30:45they'll have something that they can demo. 30:47Right. 30:47And, uh, we can see, rather than it just being on a commercial. 30:50Yeah, for sure. 30:50Chris? 30:51Uh, thoughts? 30:52I guess, I guess I'll do the podcast host thing where like, 30:54Apple - not gonna make it or not? 30:56I guess I'm kind of curious just like, like, uh. 30:59I mean like how much do you kind of rate them in this competition, 31:02which feels like it's very much speeding past them at this point. 31:05Right? 31:06Um, or, or if it's kind of like you can never kind of count 'em out. 31:08I don't, I don't think there's a competition here. 31:11And the reason I say that is I think Apple, we're still gonna 31:14be buy iPhones, whether Apple Intelligence is on there or not. 31:18Right. 31:18And I think it will come at the right point. 31:21And then we're gonna be, wow. 31:23And I think I was one of those guests a year ago that were like, oh 31:26yeah, Apple's gonna gonna crush it. 31:29And I, and I think they are still gonna crush it at some point, right. 31:32It is just gonna be, what is that point? 31:33And you know, maybe there's, they've sort of fell into the hype curve, but, you 31:37know, but hey, we're all on this podcast and we love the hype curve herself, right? 31:40So it's, it's fine to fall into that hype curve, but I. 31:44They'll, they'll get there. 31:46I don't think I'm gonna base my next phone purchase on whether 31:51Apple Intelligence is on that. 31:52If I need ai, then I'll bring up ChatGPT app. 31:55I'll bring up cloud, I'll bring up perplexity. 31:57Right. 31:57So, but, so when they introduce their AI features in the right way, I think, 32:03we will appreciate it. 32:04I think it's just up to them to make sure that they, uh, they hit that 32:09standard that Apple is known for and we have that experience and it, and 32:13it's with that kind of thoughtfulness, um, that they've always had. 32:17So I, I'm not worried about Apple. 32:19I, I, I think they'll get there when they get there. 32:22And in fact. 32:23There's a kind of point where I would say don't rush ahead in this case, because you 32:26need your iPhone to work really well, it needs to, so please don't break my iPhone. 32:33Yeah, for sure. 32:33It's like new Apple agent just does random things. 32:36Not a great user experience. 32:38Um, I guess Ash, maybe a final question before we move on to the last segment 32:41is, um, you know, I think Chris's interpretation is pretty good, which is. 32:45You know, maybe Apple kind of doesn't care. 32:47Like if you're literally made out of money and you have this product, which 32:51is just like, you know, the kind of mo one of the most successful products 32:55of all time, there's almost kind of a point of view, which is, eh, eh, 32:58so we mess up ai, you know, whatever. 33:00Like, we don't really need it. 33:02We'll get to it at some point. 33:03But like, you know, in some ways the AI thing is like almost very tiny 33:06compared to the kind of business Apple's in. 33:08And do, do you buy that at all? 33:09They prioritize 33:11usability of technology over a feature for feature sake. 33:15And I, I appreciate that. 33:17I think that, um, um, in preparation for this podcast episode, I took a step back 33:22and I was like, how do I use my iPhone and the AI features and so forth, and 33:26I have like sort of home pods and my Internet's connected house and whatever. 33:31And actually I reliably use Siri every day for like, things like 33:35controlling my thermostat and my lights and it works great. 33:37And I thought to myself. 33:39What else would I want Siri to do? 33:40And I thought, well, given what I know about how AI works today, if 33:44I was gonna say, Hey, Siri, send, send Tim an email based on, in 33:49fact, I've just kicked Siri off. 33:50Okay? 33:53If I said send him an email and it works 60% of the time and the other 33:5740% it sent Chris or Aaron an email, I might have a problem with it. 34:01I'd rather that they didn't ship that feature until they got it. 34:03That's why I got that email from you. 34:08Uh, yeah, I like that. 34:09It feels like, uh, almost, yeah. 34:11The, the almost like I, what I'm getting from this panel is 34:12almost a pendulum swinging back. 34:14Now. 34:14Everybody here is kind of like, well give it some time, which 34:17I think is very interesting. 34:23So I'm gonna take us on to our final segment. 34:25Um, and it's actually very funny the way this kind of 34:27today's episode came together. 34:28You know, we talked about Apple, kind of this dark horse in the game. 34:32Uh, Amazon, I would say is like another kind of dark horse in the game, right? 34:35Traditionally has not really been in the AI conversation has been kind of floating. 34:40Has made big announcements about the kind of hardware that they're working on 34:42for AWS that will be kind of AI focused. 34:45Um, but you know, again, candidly, we just haven't really talked 34:48about them on a week to week basis. 34:49And so it was interesting to see the story in Wired, which is kind of 34:52a splashy feature about their lab, which they, they actually bill as an 34:56a GI lab a little bit like OpenAI or a DeepMind or something like that. 35:00And what the releasing is something called Nova Act, which is their 35:04agents prototype. 35:06Um, and so they're officially now in the agents game. 35:08They're, they're like in this kind of pool that we're kind of seeing 35:11emerge and, and kind of seeing the sort of contenders, kind of who will 35:14sort of play for the agent space. 35:16Um, and so maybe actually a good place to start is a little bit like 35:19how we started the Apple segment is how likely is it Amazon to be like 35:23a contender in the domain of agents? 35:26And I guess, uh, Aaron, maybe I'll throw that one you to start. 35:29I mean, so first I think it's real exciting, you know, that, that, 35:32that Amazon is really thrusting, you know, their weight right into this 35:34space with their Nova serious models. 35:37And, um, I mean, I mean, look, you know, they've, they've got fulfillment 35:40centers with, uh, robotics, you know, um, all around the world, right? 35:43And that gives them, um, extra data of which they can use 35:46reinforcement learning, right? 35:47Uh, with their models. 35:49They have the, the largest e-commerce site in the world, right, of which they can, 35:54you know, use to either, you know, deploy. 35:57You know, some of their experiences, um, they can use to gather again, more 36:01exemplars training or just raw data. 36:04Um, and then also have the AWS bedrock, um, and just the 36:07pure compute power, right? 36:09So tho those three elements really give, give them a, you know, a, a 36:14large, um, space of which they can not, not only build models, but build 36:18models that can follow instructions and do, you know, function calling, 36:22tool calling, uh, but also experiment. 36:25Right? 36:25And, um, I did notice that one of their models, I believe it was called, uh, 36:29Nova Pro, uh, but it excels at one of their instruction following, um, and, 36:33and they've measured it, you know, on, on these three different benchmarks, 36:36you know, you know, one of them was the. 36:37Uh, Berkeley function calling leaderboard. 36:40Uh, you can see it. 36:41Um, what I did note too, um, is that, um, some of their comparisons of their 36:46models are against older models, you know, such as the older meta models. 36:50I think that they need to update that a bit, right? 36:52And then also give us some more information, right, about how their 36:56function callings, um, actually work. 36:58Right. 36:59Um, but I am looking forward to it. 37:00Um, and, and I, and I do, do think it's exciting and, you know, I know that Apple, 37:05you know, might be trying to work on Siri, but now we can see Amazon work on Alexa. 37:09Right, right. 37:09With these different types of models that are now coming. 37:11Yeah, for sure. 37:12Um, and yeah, I think what's interesting about Nova 37:15is, I think when we've talked about Amazon in the past, it seems like 37:18the strategy has been very much kind of on theory that like models might 37:22not matter much in the future, right? 37:24So, well we have run AWS, we're gonna have train, which is their 37:27kind of like proprietary chips and you know, that's how we'll do it. 37:31Like that doesn't really matter what model you run, you'll just 37:33need infrastructure to run it. 37:34And, which I think is so interesting about this is that. 37:36They're doing their own models, uh, and they're doing models 37:39kind of in the agent space. 37:40Um, and I think this kind of last introduction of I think 37:42Alexa is, is pretty interesting. 37:44Um, I guess Ash maybe the kind of pick up on like how you ended 37:48the kind of Apple discussion. 37:50There's also kind of a question of culture here too, right? 37:52Like, do we think Apple is like, kind of like as a company well-positioned 37:55to execute on AI in a way that maybe is a little bit different than 37:59than Apple, right? 38:00Because Apple by reputation has this like very distinct culture on 38:02design and how it approaches things. 38:04It kind of feels like Amazon might be able to do it, right? 38:06Like I guess they have a rep for scale or I don't know how you'd kind of 38:09describe that interface, but I think it's an interesting one to think about. 38:11Yeah, I think that, I think that culture is far more experimental. 38:15Um, and, um, the, the entire agent space is 38:20you know, very much are experimental right now. 38:23I mean, we, we, we, we create a lot of like pilots and content around 38:27all the various agent frameworks and multi-agent frameworks and so forth. 38:31Uh, and, um, and so we got a lot of hands-on experience for 38:34seeing how reliable they are. 38:36Sometimes they call tools, sometimes they don't. 38:39Sometimes the responses from the LLMs don't necessarily get processed by 38:43the agent as we'd expect them to. 38:45And, but one of the, the most interesting parts about that is, is that 38:49a lot of the people that are in that space don't have the size or the scale 38:53that Amazon does, and they don't have all those resources that Aaron mentioned. 38:58I think it's really interesting that they're approaching this from the world 39:01of robotics and using that block approach. 39:04I think that's very interesting and I think that, um, the 39:07combination of Amazon providing. 39:12SDK that hopefully will mature into an ecosystem would mean that, um, 39:17they do actually have the scale to be able to actually go, you know 39:20what, maybe there's a layer of an agent marketplace on top of this. 39:24Maybe we can like plug it into Alexa. 39:26We could plug it into our AWS services. 39:29Maybe there's a place where. 39:30People could make sort of individual blocks of agents that they then 39:34resell through some of the, uh, capabilities that Amazon has. 39:37I think that that's a very, very different approach to Apple that 39:41wanna keep everything in-house and get it perfect and release it together. 39:44Whereas I think AWS may actually just democratize this and say, here's 39:48our rest, DK, here's our frameworks. 39:49Why didn't you build it? 39:50And we, we'll help you like, put it on our marketplace and ship it. 39:54Yeah. Yep. 39:54Yeah. 39:54I do think that, uh, Amazon getting in this space could potentially push 39:57the field more towards open source. 39:59You know, um, you, you know, because if they release, you know, a an SDK, then 40:04some of the open models will be easier of which to, you know, um, integrate into. 40:08Whereas the proprietary models you'll have to have and, and maybe 40:11even wait, you know, for companies right, to do it themselves, right. 40:14Um, to, to make those hooks and interfaces. 40:17Um, readily available. 40:18Um, so, so I'm curious to, uh, see how that's gonna unfold too. 40:21Yeah, that'll be so funny. 40:22The kind of like meta Amazon alliance for, you know, forwarding open 40:25source will be like, that's a very kind of weird kind of bedfellows to 40:28sort of think a little bit about. 40:30Um, Chris, it looks like you might wanna jump in. 40:32Yeah, no, I was gonna say, I, I think Amazon's gonna nail it. 40:35I really do. 40:36As, as you said, they've got the compute, they've got the power, 40:39they've got the chips, you know, and. 40:41They're, and let's not remember, they've got $8 billion invested 40:45in Anthropic as well, right? 40:47So they're, they're building their own AI, but they've hedged their bets very, 40:52very nicely, uh, with Claude as well. 40:54So they're in a really nice win scenario there, and I, I really love what 40:59they're doing with Agent SDKs as well. 41:00Right? 41:01So one of the things that they did this week was, I don't know 41:03if you noticed this, is they. 41:05Um, started exposing some of their services as MCP services 41:09on Amazon, and then they've released their kind of MCP toolkit. 41:13So they're, they're taking this agent market very, very seriously as well 41:16as the kind of agent browsers that we were talking about earlier as well. 41:19So, um, from their perspective and Ash, exactly to your point. 41:24AI models are gonna have to talk to something, right? 41:27They're gonna have to interact with other systems, with APIs. 41:30Um, so Amazon as a cloud computing provider need to 41:35invest in agentic workflows. 41:37They need to invest in these tools and they need to make it ready 41:40there, and otherwise the models are gonna have nothing to talk to, and, 41:43and it's gonna be very, very sad. 41:44So, um. 41:46I think, I think they're gonna do a great job. 41:49Um, they've really sort of covered everything, so they're gonna be a big 41:52player and, uh, yeah, it'll be, and again, it's one of these other things. 41:57Do they need to have the best models? 41:58Probably not, because you know what, they're, they're locked in with 42:02Claude anyway, so it's, it's all good. 42:04Um, but I think what will become interesting over time, and we discussed 42:08this in one of the previous podcasts, is 42:10when the cloud, the cloud providers with Amazon and Microsoft who are building 42:15their own AI models, what happens if they get parity with, uh, the frontier models? 42:21That's the interesting conversation in the future. 42:24Yeah, and I think it's kind of almost like, again, you've see this with each 42:26generation of technology, but it's almost kind of like everybody's sort of like. 42:30Like, it's a question about whether or not kind of like scale in terms of business 42:34platform and I guess data as well, right? 42:37Like kind of wins out against like, well we don't necessarily 42:40have like the state-of-the-art, like algorithmic improvements. 42:44Um, and it feels like, yeah, Amazon I feel like has like a huge amount of leverage 42:48here in part just because of the scale, um, in a way that even like an OpenAI 42:52kind of can't keep up with, um, which is very, very interesting to think about. 42:55Well, this is great. 42:56Um, that's all the time that we have for today. 42:59Uh, thanks for joining us, Ash. 43:00Great having you on the show. 43:01Hopefully bad. 43:01Be back at some point. 43:02And Aaron and Chris, great to see you as always. 43:05Um, and thanks for joining us. 43:06Uh, if you enjoyed what you heard, you can get us on Apple Podcasts, Spotify, 43:09and podcast platforms everywhere. 43:10And we'll see you next week on Mixture of Experts.