Learning Library

← Back to Library

Exploring Open‑Source Mixture‑of‑Experts AI

Key Points

  • The show opens by questioning the notion of truly autonomous AI, emphasizing that models only predict tokens and require external control to act.
  • Recent AI news highlights include OpenAI’s $1 trillion data‑center plan, Alibaba’s partnership with Nvidia on robotics and self‑driving cars, IBM’s PDF‑decoding model topping Hugging Face downloads, and Meta’s AI‑powered digital dating assistant.
  • The main discussion centers on Tongi Deep Research, a new open‑source LLM with 30 B total parameters but only 3 B activated per token, optimized for long‑horizon information‑seeking tasks.
  • Panelists note that Tongi’s “task‑trained” approach and efficient activation make it a significant step forward for agentic models that can handle extended reasoning.
  • The episode also promises to cover additional topics such as AP2, AI‑related safety concerns, the latest AirPods, and a massive $100 B investment in AI initiatives.

Sections

Full Transcript

# Exploring Open‑Source Mixture‑of‑Experts AI **Source:** [https://www.youtube.com/watch?v=h1YqbXQN5N8](https://www.youtube.com/watch?v=h1YqbXQN5N8) **Duration:** 00:52:38 ## Summary - The show opens by questioning the notion of truly autonomous AI, emphasizing that models only predict tokens and require external control to act. - Recent AI news highlights include OpenAI’s $1 trillion data‑center plan, Alibaba’s partnership with Nvidia on robotics and self‑driving cars, IBM’s PDF‑decoding model topping Hugging Face downloads, and Meta’s AI‑powered digital dating assistant. - The main discussion centers on Tongi Deep Research, a new open‑source LLM with 30 B total parameters but only 3 B activated per token, optimized for long‑horizon information‑seeking tasks. - Panelists note that Tongi’s “task‑trained” approach and efficient activation make it a significant step forward for agentic models that can handle extended reasoning. - The episode also promises to cover additional topics such as AP2, AI‑related safety concerns, the latest AirPods, and a massive $100 B investment in AI initiatives. ## Sections - [00:00:00](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=0s) **Debating AI Autonomy & Industry News** - The segment opens the Mixture of Experts podcast by questioning autonomous AI, introduces a panel of experts, and previews a rapid‑fire news roundup that includes OpenAI’s trillion‑dollar data‑center plan. - [00:04:32](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=272s) **Closing Gap Between Proprietary and Open‑Source AI** - The speakers note that open‑source models are rapidly catching up to proprietary LLMs, but the primary differentiation now lies in specific use cases such as embedded devices, privacy‑focused deployments, and costly deep‑research agents. - [00:07:40](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=460s) **Diverging Paths of AI Ecosystems** - The speaker argues that open‑source AI will continue to split from proprietary, product‑centric solutions by focusing on narrow components within broader agent architectures, and wonders if upcoming research papers will spark new trends like model distillation and continual pre‑training. - [00:12:38](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=758s) **Layered Protocols and Business Viability** - The speakers debate the difficulty of evolving a niche “punk‑rock” lab into a market‑ready, self‑sustaining company before pivoting to discuss Google’s new AP2 protocol, which stacks atop existing agent frameworks to standardize authentication, authorization, and payment handling for AI agents. - [00:16:48](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=1008s) **AI‑Driven Shopping Meets Crypto Payments** - The speakers evaluate a camera‑based auto‑shopping concept, argue that major players like Google have a strategic edge in integrating crypto payments and AI, and note how prior smart‑contract knowledge becomes valuable in this emerging market. - [00:20:05](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=1205s) **Hidden Protocols, Liability, Market Share** - The speakers argue that while everyday users remain unaware of and indifferent to underlying security protocols such as AP2, developers and liable institutions will care, and Google’s integration of an ATA‑based secure payment system could force widespread adoption and boost its market dominance, with rivals like Anthropic possibly working on comparable technology behind the scenes. - [00:23:32](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=1412s) **Debating AI Risk Book** - The hosts introduce the provocative AI safety book “If Anyone Builds It, Everyone Dies” and examine its claim that super‑intelligent AI poses existential danger, questioning whether AI CEOs are being sufficiently cautious. - [00:26:54](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=1614s) **Balancing AI Autonomy and Safety** - The speakers argue that strict human oversight and limited use‑case deployment are essential to prevent self‑propagating AI scenarios, citing GPT development safeguards and autonomous‑vehicle mishaps as cautionary examples. - [00:30:08](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=1808s) **Untitled Section** - - [00:33:21](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=2001s) **Fiction, Reality, and AI Wearables** - The speakers explore how AI blurs the line between fictional storytelling and actual capability, then shift to highlight emerging AI‑enabled wearables such as Meta’s Ray‑Band and Apple’s real‑time translation AirPods. - [00:36:40](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=2200s) **Apple’s AI Comeback Debate** - The speakers speculate on Apple’s potential turnaround in artificial intelligence, compare its progress to Google’s, consider ecosystem integration, and question the underlying models behind features such as AirPods translation. - [00:40:41](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=2441s) **Seamless AI Integration via UX** - The speakers argue that AI must become invisible to consumers—abstracting tokens and model details—by delivering flawless, stylish user experiences, a strategy they see Apple poised to capitalize on. - [00:43:51](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=2631s) **Nvidia's $100 Billion OpenAI Investment** - The speakers debate the ramifications of Nvidia's announced $100 billion infusion into OpenAI, likening it to stock buybacks and exploring its financial and strategic implications. - [00:47:01](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=2821s) **AI Alliances, Energy Demands, and Tribal Dynamics** - The speakers debate upcoming AI partnerships and the massive 10‑gigawatt power needs of new facilities, likening the shifting loyalties to school‑house rivalries. - [00:51:12](https://www.youtube.com/watch?v=h1YqbXQN5N8&t=3072s) **Debating OpenAI's $100B Allocation** - The panel speculates whether the massive funding will primarily expand data‑center capacity or fund more efficient next‑generation chips, framing it as a push toward greater raw compute and a hint of forthcoming AGI. ## Full Transcript
0:00premise of a an AI that is capable of 0:05acting autonomously 0:08is one that I struggle with. AI models 0:12fundamentally do nothing except predict 0:14the next token. AI systems 0:18do things and somebody has to turn them 0:21on and turn them off. All that and more 0:23on today's Mixture of Experts. 0:26[Music] 0:31I'm Tim Huang and welcome to Mixture of 0:33Experts. Each week, Moe brings together 0:35a panel of the sharpest minds and 0:37quickest wits to help you digest the 0:39week's news in artificial intelligence. 0:41Today, I'm joined by a championship 0:43lineup crew. We have Mihi Crevetti, 0:45who's a distinguished engineer, Aentic 0:46AI, Gabe Goodart, uh, chief architect, 0:49AI, open innovation, and Sandy Besson, 0:52AI research engineer. Welcome to you 0:54all. We've got a very, very packed 0:56episode. But I think I was just informed 0:57by one of our producers. We're going to 0:59try to wedge another story in. We're 1:01going to talk about Tongi Deep Research. 1:03We're going to talk about AP2. We're 1:05going to talk about AI's killing us. 1:07We're going to talk about the new 1:07AirPods. And finally, we're going to 1:09talk about a small hundred billion 1:11investment. But first, we've got Eiley 1:13with the news. Eiley, over to you. 1:16[Music] 1:19Hey everyone, I'm Eiley McConnan. I'm a 1:22tech news writer for IDM Think. I'm here 1:24with a few AI headlines you might have 1:26missed this week. Another news story 1:28involving trillion with a T. Yes, that's 1:31trillion. Open AI has unveiled plans to 1:33build one trillion dollars worth of data 1:36centers across the US and abroad. Can 1:39tech companies be frenemies? Chinese 1:41tech giant Alibaba is partnering with 1:43American chipmaker Nvidia. Alibaba will 1:46use Nvidia's hardware and software to 1:48develop robotics and self-driving cars. 1:51Turns out there's a world of useful 1:52information in those PDFs that most 1:55people never read. Granite Dockling, a 1:57small but mighty new model from IBM, can 2:00decode those PDFs and has jumped to the 2:02top of the most downloaded models list 2:04on hugging face this week. Are you 2:06suffering from swipe fatigue from online 2:08dating apps? Well, Meta has added a new 2:11digital dating assistant to Facebook so 2:13you can use AI to identify your special 2:15somebody more easily. Want to dive 2:17deeper into some of these topics? 2:18Subscribe to the Think newsletter. It's 2:20linked in the show notes. Now, back to 2:22the episode. 2:27So, for our first segment, I really 2:28wanted to talk about Tongi Deep 2:30Research. Um, so this is a model that 2:33has zoomed to the top of the hugging 2:34face leaderboards. It's an agentic large 2:37language model that quote features 30 2:39billion total parameters with only three 2:41billion activated per token. And it's a 2:43lab model that's specifically designed 2:45for long horizon deep information 2:48seeking tasks. Um, and best of all, it's 2:51all open source. So, Gabe, given our uh 2:54you're our open source guy, maybe I'll 2:55kick it over to you first. Initial 2:57impressions about this model. What do 2:58you think? I think it's a really cool 3:00step forward and I think it the novelty 3:06in this model is that it was taskrained 3:09specifically for this long horizon and 3:12paired with software that was designed 3:15to implement the long horizon search in 3:17an iterative recursive way. Um and I 3:20think that's really cool. So you know at 3:22this point generating tokens is not 3:24novel. um your mixture of experts is not 3:26novel but the shape of those tokens that 3:29are generated and the patterns that it 3:31was trained on are novel and that's 3:33pretty cool. Um I specifically think 3:36their heavy mode uh is a really 3:38interesting doubling down on a specific 3:40agentic architecture around state 3:43management and consolidation that we 3:45haven't seen. We've certainly seen 3:47individual agents built to do this 3:49pattern but we haven't seen them paired 3:51with a purpose-built model. And I think 3:54um you know there are a lot of great 3:56public uh implementations of some flavor 3:59of deepness in research out there that 4:02can run on your laptop. I use them 4:04regularly. But I will say, you know, 4:06they don't measure up to uh public 4:10Frontier model implementations. Uh, and 4:13I think it's really cool that this team 4:15is is pushing the boundary there and 4:16trying to get something that you can run 4:18on a powerful workstation to measure up 4:20to, um, you know, a frontier research 4:23system. So, uh, kudos to, you know, 4:25really pairing the model implementation 4:27and the software ecosystem around it. I 4:29think that's awesome. 4:30>> Yeah, I think it's a really big kind of 4:32development. And Mihi, maybe I can call 4:34on you to talk a little bit about 4:35trends. I think one of the things we're 4:36always watching ate is like sort of this 4:39frontier right between sort of like the 4:41proprietary model and open source 4:43catching up and it feels like over the 4:45last few months that just continues to 4:47narrow and narrow and narrow and narrow. 4:49Um, and so like I guess maybe this is 4:52always a good data point to check in. 4:53Like do you feel pretty soon like this 4:56this margin is going to disappear, 4:57right? Like it really will be that every 4:59time a proprietary company proprietary 5:01model comes out we're going to see a 5:02open source implementation that's almost 5:04as good like almost immediately. 5:06>> I think it's likely although if you look 5:08at the use cases that's where we see the 5:11biggest differentiation between these 5:13models. A lot of the use cases we have 5:15for these smaller models, especially the 5:17open source models that can compete in 5:18the space, have to do with embedded 5:20devices, with local laptops, with 5:22privacy, being able to run these models, 5:24you know, behind a firewall or running 5:26them at a cost. If you look at the use 5:28case of deep research, since we're 5:30talking about agents and deep research, 5:32that's basically a large language model 5:34that's doing planning to use tools like 5:36both internet search and internet search 5:38as well as searching all of your 5:40document collections and really going 5:42through potentially millions of tokens. 5:45That can get expensive and that can get 5:47very slow if you're using one of the 5:49frontier models. So there's a lot of 5:51organizations who are looking for doing 5:53this cheaper, faster with smaller 5:56models. preferably behind their firewall 6:00especially if you have some of these 6:02models work with financial data or HR 6:04data or internal data. So I definitely 6:06see a sweet spot for many of these uh 6:09smaller 6:10I would say purpose-built models. 6:12>> Yeah, for sure. I think almost you're 6:14saying like we think about it as like oh 6:16is open source catching up? I guess Mi 6:18are you do I hear you right in kind of 6:20saying almost like we should think about 6:21these as like almost two different 6:22markets like what open source trying to 6:24do actually pretty different from like 6:25what like I don't know open AI is trying 6:27to do 6:27>> I believe so and look open AI also has 6:30their you know GPT5 nano which is you 6:32know too cheap to meter 35 cents for a 6:35million token or I'm probably leaving a 6:37comma left or right in their in the 6:39price but again too cheap to meter I 6:41don't care if I'm using it for agents I 6:43know it's fast I'm going to know it's 6:45going to be cheap enough and I can throw 6:46it at use cases like deep researcher and 6:48say, "Yeah, it's fine to crawl through a 6:50hundred of these websites and send every 6:52one of those tokens." But maybe the CIO 6:54office is not going to agree with me 6:56sending all of that internet data to one 7:00of these large language models. So, if 7:02there's a model that can run it 7:03privately or a model I can run on my 7:06laptop or maybe even in the future on my 7:09phone, that's awesome. 7:10>> Sandy, I want to bring you into this 7:11conversation. One of the things that we 7:13talked about on the last episode was 7:15that there was this paper that came out 7:16that was some called something like what 7:18are people using chatbt for? Um, and I 7:21thought the funny part about it was like 7:22everybody looked at it and I think our 7:24guests were all like it's search. People 7:26are using chatpt for search. Um, and I 7:30guess to Mihi's comment, it almost kind 7:31of feels like what we might see if we 7:33had a paper that was like what are 7:35people using open source models for is 7:37like it would look very very different. 7:39And I guess kind of question for you is 7:40like do you feel like these ecosystems 7:41are going to just keep diverging over 7:43time like that over time we should 7:45actually really think about like open 7:46source almost solving like a completely 7:48different set of problems from what you 7:50know the like a chat GPT is trying to 7:52solve. I think it's the the 7:54juxtaposition of productizing versus 7:56plugging into a broader more like bigger 7:59problem, right? Typically in open-source 8:02solutions, they're not solving like the 8:05whole suite. They're solving like a very 8:07narrow piece of the puzzle. So, I could 8:10very much see using this uh deep 8:12research agent as part of a broader 8:16agent team or broader agent 8:18architecture, but it's probably not 8:20going to be like the full picture that 8:22stands alone and is like a hosted 8:24solution somewhere, right? Um, and so I 8:27I kind of see it diverging in that way. 8:30You know, like I also wonder slightly 8:33tangentially, but like I wonder we we 8:36see we saw was it this year or last 8:39year? with deepseek. I can't even 8:40remember timelines anymore. But uh we 8:44see um like once in a while these big 8:48papers come out and I don't know whether 8:50this one will will be that kind of big 8:53paper for this type of of trend or 8:55pattern but 8:58I think just like we saw Deep Seek kind 9:00of emphasize distillation and now 9:04distillation became a big deal after the 9:06Deep Seek paper came out. I wonder 9:09whether this paper would trigger some 9:10sort of trend in terms of that triathlon 9:13of training where you do like continual 9:17pre-training then fine-tuning and then 9:20on policy RL like I wonder whether this 9:23will now become a common pattern because 9:27of this paper. 9:28>> Yeah, that's right. And that's actually 9:30a fun way of thinking about it. I we 9:31haven't really talked about it like that 9:32on the show before. I think frequently 9:34we're very like benchmarked, right? 9:37We're like, "Oh man, this model is like 9:39so much better on this benchmark and not 9:40as good on this benchmark." Sandy, 9:42you're almost saying like we should 9:43almost measure how important a model is 9:45by almost like how influential it is, 9:47right? In terms of like how it kind of 9:49shapes how people are doing things. 9:51>> Totally. Sometimes it's not the first 9:52model that comes out that's actually 9:54like quote unquote the the the best. It 9:56might be soda in some ways, but like 9:58it's actually the trend that it drives 10:02that like points us in a different 10:04direction. Gabe, maybe a final question 10:06on this before we move to our next 10:07topic. I wonder a little bit about 10:09business model here. Uh so if you're 10:12Tongi Labs, right? You've just done this 10:15open source thing. Um at some point, do 10:18you also want to go closed source? Like 10:19I think we've been talking a lot about 10:20the big kind of like open AI and 10:22anthropic, you know, kind of all sort of 10:24leaning and thinking more about open 10:26source. But I wonder if the ecosystem 10:28goes the other way at some point where 10:29like leaders in open source eventually 10:31start to go more closed as well or if 10:32you know almost like constitutionally 10:34it's like a different direction. 10:35>> So I I actually really like something 10:37that you said Sandia and I'm going to 10:38pick on it here which is like it's 10:40really this two directional like the 10:43closed source products are trying to 10:45give something to you like here's a 10:47thing you should have it the open-source 10:50tools software and models are trying to 10:52plug into an existing ecosystem. So, do 10:56I think that labs that are primarily 10:59open source are eventually going to want 11:01to go closed source, too? I think that's 11:03a possibility. I think if they come up 11:04with something genuinely novel or 11:06something that they really think, you 11:08know, we could actually put this in 11:10front of people and people would be 11:11willing to pay us for it and they can't 11:13get it anywhere else, we have a shot at 11:14taking the crown on some novel 11:16capability, they might. I think the 11:18other route and especially that I've 11:19seen when talking with other uh small 11:22organizations that have an open core uh 11:26the other route is really leaning into 11:27that plug-in portion and that by that I 11:30mean an enterprise tier of some variety 11:32right and so engaging with your same 11:35open source componentry possibly with 11:37some additional componentry around it uh 11:39for meeting enterprise needs like better 11:42authentication management better data 11:44sovereignty etc um and engaging in 11:46almost like a consulting type of view 11:48with a large enterprise client that 11:50allows them to have a revenue stream for 11:53their software while continuing to push 11:54the envelope in the open. So, I think 11:56that's the business model that I have 11:58seen more often for these types of 12:00shops. Um, and the other thing you said, 12:02Sandy, that I also loved was like 12:03there's also just the influence game 12:05like especially depending on the life 12:06cycle of these startups. Um, a lot of 12:09times the metric isn't dollars, it's 12:12likes, clicks, retweets, uh, download 12:15and download, stars, all all of the 12:18social metrics, right? And I think, um, 12:21I don't frankly know quite where uh, you 12:25know, we're at here, but I think 12:27depending on the journey of whether or 12:29not they're being measured on dollars, 12:31um, it may still just be in the can we 12:34move the needle on the influence game. 12:36Um, and I think that in and of itself is 12:38a business model that has a decent 12:40chance of panning out either in an 12:43acquisition or in enterprise deals. Um, 12:46I think the leap to being a private 12:50selfpropelling entity that has products 12:53you offer to the market is a tough one 12:54to make. 12:55>> Yeah, for sure. Yeah. I want to create 12:57like the most like punk rock lab which 12:59is like a disaster from a business 13:00standpoint but is just like incredibly 13:02influential. 13:04>> Exactly. Exactly. most stars of any lab. 13:06You know, 13:06>> you're not alone there. There are labs 13:08that are literally doing that. 13:10>> That's right. 13:15>> I'm going to move us on to our next 13:16topic. Um I always joke that, you know, 13:18a space is really maturing when people 13:21are introducing protocols to sit on top 13:23of other protocols. Um and we have a 13:27good example of that from this week. Uh 13:29Google announced a new protocol that 13:30they are calling AP2. Um, and AB2 builds 13:34upon a lot of existing agent frameworks, 13:36uh, like model context protocol and 13:38other things that we've we've talked a 13:39lot about. Um, and what AP2 is 13:41attempting to do is basically set up a 13:43common structure for agents to engage in 13:47commerce and payments online. Um, and 13:50sort of, you know, in the blog post, 13:51Google sort of describes what they're 13:53really trying to do is make sure that 13:54agents can, you know, have proper 13:56authorization, be authentic, right, and 13:59be sort of accountable to the fact that 14:01they're going to move money around for 14:02people in the future. And the most fun 14:04part about it is that AP2 explicitly 14:08contemplates a situation where you are 14:10not there when your agent goes out and 14:12does its stuff on the internet. Um, and 14:15what they're proposing is a thing that 14:16they call the intent mandate, which is a 14:18cryptographically signed record of what 14:20you want the agent to do. And that's 14:22kind of how they instantiate like, okay, 14:24what's the scope of authority you're 14:25giving the agent to do when it goes out 14:27there and, I don't know, you know, puts 14:29a bunch of books into a shopping cart 14:30and, you know, checks out for you. So, 14:32Mihi, I guess you're our agents guy, uh, 14:35I think on this panel. Um, how big of a 14:37deal is AP2? Like, do you think it's 14:38going to get big up adoption? Do you 14:41think people should be paying attention 14:42to it? I think it's got huge potential 14:43because it kind of tries to solve a 14:45problem that neither A2A or MCP solve 14:49today. And the reality is if you look at 14:51the ecosystem, MCP is solving the 14:54problem of building your tool once and 14:56having that tool work with any kind of 14:58agentic framework. So you're kind of 15:00decoupling the agent and the tool. But 15:03all of the respective security, 15:05authentication, authorization mechanisms 15:07that you're going to use, how you're 15:08handling secrets, secrets management, 15:10data management, certificate management 15:12is really left up to you in terms of how 15:15you implement it or how you make it 15:16work. And there's already like five 15:19versions of MCP and they all implement, 15:21you know, more and more and more around 15:23the area of security. But I don't think 15:25we're at a point where we can say, "Yep, 15:27this is something I can easily build a 15:30payments processing system on top of or, 15:32you know, have an agent go off and book 15:35my travel to some island and, you know, 15:37make all the payments and trust it with 15:40that information. Not even from the 15:42perspective of trusting it to get it 15:44right, but from a perspective of how 15:46it's handling that data. You can't put 15:49your payment information, your credit 15:51card information in the text that goes 15:54into the model. There needs to be some 15:56kind of a side mechanism on how the data 15:58is transferred cryptographically 16:01verified, not up to the interpretation 16:03of a large language model. So I think 16:05what Google did here was very smart. Um 16:08they tried to leapfrog the effect that 16:11MCP got because MCP has been adopted 16:14virtually everywhere. Microsoft, OpenAI, 16:16IBM and so on. 8way probably less. So, 16:20so now they went to all of the banks, 16:22all the financial institutions and 16:24especially some of their core client 16:26base both in the finance industry but 16:29also in the advertisement industry cuz I 16:31can see this take up for example for you 16:34know those single person shops who are 16:37using Tik Tok as their mechanism of 16:40advertisement and say hey talk to your 16:42personalized shopping assistant. He's 16:43going to pick the right clothes for you. 16:46you take a picture with your camera and 16:48it's going to buy you a new, you know, 16:50kind of dress or pants or whatever every 16:53single month and ship it back to you and 16:55you give it the authority to make those 16:57purchases on your behalf. So, I I think 17:01it could potentially open up a new 17:03market, a market they're very competent 17:05in, where they have potential both 17:07clients and relationships in where I 17:10don't think Entropic, for example, has 17:12the same kind of leverage within the 17:14same institutions. they're not a payment 17:15provider like Google has Google Pay. So 17:18I think it's a brilliant move. I do 17:21think some of the items are so so 17:23they're also releasing X4 for two 17:26extensions for 8way. 17:27>> So this is going back to crypto 17:29payments. 17:30>> So we're going back to web free and 17:32MetaMask and the Ethereum foundation and 17:34coin bags. It's like wait I thought we 17:37were done with the web free stuff or 17:38we're now doing AI. Come on one trend at 17:40a time. So, I think it'll be 17:43interesting. Uh, it'll be even more 17:46interesting if we see somebody like 17:48Facebook and their metaverse get excited 17:50about it, but time will tell. 17:52>> I'm so glad I spent those two weeks 17:54about a year and a half ago figuring out 17:55how to write smart contracts. That might 17:57be useful now. 17:58>> Yeah, exactly. Exactly. It was all worth 18:00the investment of time and effort. Well, 18:02so there's two directions to go and I 18:04think I want to hit on both of them. I 18:06guess one of them that just to build on 18:08Mah what you just said was thinking a 18:10little bit about like I hadn't really 18:11thought about basically like this is 18:13sort of competition between um kind of 18:16anthropic and Google in some sense. Um 18:18if you're anthropic like what is the 18:21counter move here? Do you really care? 18:22>> I mean we'll have to see if entropic 18:24adopts it. 18:26>> That would be an interesting move 18:27because you know we've seen that even 18:29OpenAI for example has adopted MCP now 18:31and you can go in chat GBD and use that. 18:34So we're going to see if we see similar 18:36adoption from Entropic. Maybe the 18:38protocols are going to merge and we're 18:39going to have A2A plus MCB plus AP2 all 18:43in one big 18:44>> one protocol to them all. Yeah. 18:46>> One protocol to roll them all for AI 18:47agents securely doing anything. But 18:51we'll see. 18:51>> Sandy, can I ask you a question about um 18:53so I used to do a lot of work in the 18:55privacy space. Um and I think like one 18:57of the things that we always thought for 18:58privacy is surely everybody will wake up 19:01one day and care about their privacy. 19:02And so we're going to create lots of 19:04things for people to kind of like tweak 19:05like the data that they share and then 19:07every time there's a large data breach 19:08you'd say surely now people will want to 19:11care about their privacy. And so I guess 19:13I have a question for you on just like I 19:15see some of this stuff and maybe I'm 19:16just like very skeptical at this point. 19:18I'm kind of like do agents need this 19:20kind of thing in order to succeed? like 19:22implicitly we're always like oh well you 19:24need trust in order to transact online 19:26but like isn't one counter-argument that 19:28like kind of people don't care and so 19:30you know this will be maybe good at the 19:31margin but it's not really like 19:32necessary for agents to start really 19:35interacting in the economy 19:36>> I think in a if you want to it really 19:39shifts the liability right like it 19:41shifts the liability from the security 19:45aspect which hopefully uh AP2 has 19:49covered right to the ability of the 19:53model itself to make the right choices 19:55like Mihi said. And I think people don't 19:59care as long as it goes right. Right. 20:01Like if if AP2 is already in place, 20:05yeah, people won't really care. They 20:07won't even know it's in place. That's at 20:09the protocol level. Most people don't 20:10even know what's happening, right? They 20:12you're using an agent and interacting 20:14with uh an agent on a platform. you 20:17don't know whether it's an Ato agent or 20:19it's using MCP servers or what's going 20:22on behind the scenes, right? 20:23>> In fact, it's good that you don't know, 20:24you know, 20:25>> and and so I think in a way people won't 20:27care, but the people developing the 20:31tools and the institutions that are 20:32responsible for the liability will care. 20:35Um and and so everyday consumer, no, I 20:39don't think they care. um they like to 20:43be told that it's trustworthy and that 20:45their stuff is secure, but they don't 20:46care how. Uh the people that are liable, 20:49I think they will. And I think they like 20:51Mihi was saying it's such a smart move 20:54for Google because this will force 20:58adoption of ATA as well because in order 21:02to use it it's built on top of ATA and 21:04so 21:06it it will become the de facto that if 21:09you want to use this secure payment 21:11protocol you're also making it an ATA 21:13compatible agent. Um and so I think that 21:16they will gain a lot of market share 21:18there and there was some chatter to your 21:20comment about Anthropic developing a 21:23similar agent agent type protocol. Um 21:26but we haven't seen it yet and so maybe 21:28they will develop it but they started 21:31talking about that like a good 6 months 21:33ago. Maybe they're developed media edit 21:35internally and they haven't released it. 21:37Or maybe they've kind of taken the 21:38OpenAI approach where they were like, 21:40"Yeah, MCP like OpenAI was like, "Yeah, 21:43we'll just adopt MCP. Maybe that will 21:44happen too." 21:45>> Yeah. I think one other item here is the 21:48market is targeting as well because I'm 21:50probably not the demographic or the 21:51market who's going to set up a 21:53personalized shopper on TikTok based on 21:55some ad I've seen scrolling down and go, 21:56"Oh, sure. a $100 every month, you can 22:00go off and do some random shopping for 22:02me. Or I'm not going to delegate my 22:04vacation to some assistant. I'm going to 22:06be there. I'm going to check every 22:07single hotel. I'm going to do the 22:08numbers. I'm going to ed it up. I'm 22:09going to go through the payment. I'm 22:11going to look has the payment gone true. 22:13So, I I I think consumer habits are 22:16changing as well. And there's certainly 22:18a market for it. It's it's emerging. 22:20It's not quite there yet, but maybe in 22:2210 years everybody's going to just go 22:24off and say, you know, hand it over to 22:25the AI to do my shopping. you'll figure 22:27out what I need. 22:28>> For sure. And I think it's like it's not 22:30too far away, you know, like people set 22:32up recurring payments and the whole 22:34point of that is to offload, you know, a 22:36transaction each month. The question is 22:38just like how much broader are you happy 22:40with that going, you know, over time? 22:41>> I I've never set up a single recurring 22:43payment ever. Whenever I see one of 22:45those like I don't trust it. 22:47>> Yeah. Okay. Something that I do think 22:50will like this will propel is the use of 22:53computer use. You know, until now, 22:55computer use has been um kind of like 22:59researchbased, right? Just collect 23:01information in a way that you can't 23:02access other ways. But once we have this 23:07protocol in place and it's able to 23:09actually check things out for you 23:11securely and manage secrets and 23:13credentials, and I think that that will 23:15propel the actual usability of computer 23:17use where right now it's like a cool 23:19thing, but honestly, no one's using it 23:22that much in practice. Yeah, that's 23:24really interesting. It's kind of like 23:24the downstream effects. It maybe turns 23:26out like once you get payments right, 23:27like lots of other things kind of come 23:29out of it. 23:33I'm going to move us on to our next 23:34topic. Um, we're going to do a little 23:37bit of a book review uh for our next 23:38segment. Um, so a book came out that has 23:41been getting a lot of chatter online. 23:42It's dramatically titled If Anyone 23:44Builds It, Everyone Dies. Um, and it's a 23:48book authored by two, uh, long-standing, 23:50you know, people who I think have 23:52participated in the AI policy, AI safety 23:54discussion. Um, Elzar Udicowski and Nate 23:57Sorz. Um, and, uh, I guess Gabe, maybe 24:00I'll kick it over to you first on this. 24:02Um, the core of the book is the idea 24:05that at some point we may build super 24:07intelligent AI and in which case the 24:09minute we build it, it becomes really 24:11dangerous and everybody might die. Hence 24:13the title, if anyone builds it, everyone 24:15dies. uh we can talk about the validity 24:17of that claim but I think I actually 24:19want to start somewhere different which 24:21is I think part of the argument of the 24:22book which I found quite interesting is 24:25just to kind of take the CEOs of all 24:27these AI companies at face value where 24:30we have had kind of a cycle of CEOs 24:32coming out and saying AI is probably the 24:34most powerful most dangerous most risky 24:38most uh most promising technology um and 24:43you know I think almost an argument of 24:44the the book is basically like if you 24:46buy what they're saying, shouldn't they 24:48be a lot more careful than they are 24:49right now? And I'm curious about what 24:51you think about that as as an argument. 24:53>> Shouldn't they be a lot more careful 24:54than they are right now? It's a very 24:56good question. Uh, you know, certainly 24:59some some introspection here as we are 25:01also hopefully helping to build some 25:04technology in this domain. Um, 25:09I think there are a couple of aspects to 25:13this that 25:17the 25:20premise of a an AI that is capable of 25:24acting autonomously 25:28is one that I struggle with. Um, because 25:32somebody has to write a for loop. Now 25:33that somebody might be another AI model, 25:35but somebody had to write the for loop 25:37for that AI model, etc. Um, 25:42AI models fundamentally do nothing 25:45except predict the next token. AI 25:48systems 25:49do things and somebody has to turn them 25:52on and turn them off. And so yes 25:56absolutely at a meta level care needs to 26:00be taken to ensure that the systems we 26:03are building do not have the aggregate 26:07set of capabilities to hit any of these 26:09escape points that we have considered. I 26:12think from a hypothetical 26:16what could go wrong perspective, 26:18there are real risks here, right? 26:21Fundamentally, the issue here is a a 26:24much more holistic one than some 26:26simplistic evolution of a sensient AI. 26:29It is a system of humans, companies, 26:32software, hardware, and yes, a whole 26:36pile of floatingoint numbers coming 26:38together. and the humans are the input 26:42to that system, right? And I think one 26:45of the big worries is obviously that the 26:47models themselves, the system itself 26:49becomes self-propagating. 26:52Um, and that only happens if you give it 26:54the tools and the capabilities to become 26:56self-propagating. And so this 27:00fear about that if somebody did it, it's 27:03possible, it could happen. But the 27:06safety procedures, the use cases that 27:09these systems are being set up for um 27:11can and should be limited to avoid these 27:15scenarios. Um and I think at least what 27:17we're seeing currently, you know, I 27:20highly highly doubt that open AAI is 27:25letting GPT5 write code that goes into 27:28GPT 5.1, at least not cart blanch 27:32without any engineers looking at it. So, 27:34it's really the humans in the loop that 27:36we have to be careful of and as stewards 27:38of this technology, we have to make sure 27:40that we're being diligent about that. 27:42Um, but I don't think we're anywhere 27:44close to escape velocity on doomsday. 27:46>> Gabe, I saw you wrapping up and then 27:48Mihi and Sandy immediately went off 27:49mute. Um, so, uh, Sandy, how about you 27:52go first? 27:53>> Well, I was just I was I had a more of a 27:55question to pose to Kim's opinions. Do 27:58you think we're going to get close 28:00enough where we go oopsie and then like 28:05backtrack and be like, okay, maybe that 28:07was like a a little far, that was a 28:09little scary. Uh, we learned a lesson 28:11from that. 28:12>> Maybe one example and then Gabe would be 28:14interested in your response. You know, 28:15in the autonomous vehicles context, it's 28:18true that like, you know, there was a 28:20company called Cruz, right, that was in 28:22San Francisco operating. they had a 28:24number of accidents and had to pull back 28:26and then you know but you know I guess 28:29luckily there was like the technology 28:30retoled and now Whimo and a number of 28:32other companies are kind of running in 28:33the space and so I guess we like at 28:36least maybe by way of responding to your 28:37question it's like almost like we we do 28:39see that pullback and iteration I guess 28:42the question is whether or not like you 28:44know AI broadly writ I guess is maybe 28:47different in that respect um Gabe do you 28:48want to jump in? Yeah, I mean I think 28:50you know we're still holding the 28:51steering wheel. So absolutely we will 28:53definitely see cases where we have real 28:55world harm done um and have to dial it 28:58back. In fact we're seeing that now. Um 29:01you know there are whole huge 29:04discussions to be had around the 29:06negative role we've had it on this 29:07podcast to some degree of these models 29:10in the context of educating children um 29:13or mental health in terms of folks going 29:15down rabbit holes that they wouldn't 29:17with another human. uh and have, you 29:20know, an infinite self-reassuring uh 29:22hallucination machine in front of them. 29:24So, there are real real risks to these 29:26things. And I do not mean to minimize 29:28those. Um but it's it's particulate. I 29:31it's um here's an area where these 29:34things have risk. Now, let's go address 29:36it by trying to tackle this risk. um the 29:40sort of runaway acceleration um you know 29:44end of the world scenario still feels um 29:47I I don't really see us sort of knocking 29:50on that door short of some like 29:51genuinely bad actors entering and trying 29:54to break things uh and trying their best 29:57to uh you know be a villain in a deep 29:59underground layer and launch the 30:01apocalypse. So 30:02>> in other news I've heard that cloud code 30:04is writing 95% of the code for cloud 30:07code. 30:07>> Yeah. Yeah. 30:08>> So AI is writing. 30:09>> All right. Bye. Okay. You're right. 30:11>> Like I think it is writing itself at 30:13this point. 30:14>> Yeah. On a more serious note, I think 30:16this has been a trend in every single 30:18piece of literature, media, movies, and 30:21games. So I I personally really enjoyed, 30:23for example, some of the games uh like 30:25Mass Effect where you have the Reapers, 30:27which is like a sentient species of AI 30:30where humanity built AI at some point 30:32they realized it was sentient and then 30:34they went, "Whoops, 30:36we're going to stop that." And when the 30:38AI saw that it was being stopped, it 30:41reacted and said, "Oops." And you know, 30:43humans are our enemies. Um, there's also 30:46Dune, like if you remember in Dune, one 30:49one of the premises was thou shalt not 30:51make a machine in the likeness of a 30:53human mind because that's how it the 30:55whole thing started. There was the 30:57Berian jihad and then when you know AI 31:01and humans fought and since then you 31:03were not allowed to build AI. Um there's 31:06showdown in system shock if you've 31:07played like you know DOSS game 199596 31:10whenever that came out uh where again 31:13somebody built a super intelligent AI 31:15and it turned on mankind or matrix so I 31:17think in every popular media games 31:19movies books this has been a trend 31:22whether it's going to happen we don't 31:25really know should we be more careful 31:28well if we're careful then our 31:31competition is not going to be careful 31:32and they're going to get ahead of us So 31:34everybody's pushing to create more and 31:36more innovation. I think at some point 31:38there need to be guardrails in terms of 31:42how these AI systems are allowed to 31:44connect and interact with the real 31:46world. Should you have an AI system, for 31:48example, in charge of medical equipment 31:51or make life like and life and death 31:53decisions. So for example, if you have 31:55an accident and you're in the hospital, 31:56the AI can go in and say, "We're going 31:58to turn off your life support because we 32:00can use your organs." And that decision 32:02is made by an AI. no longer by your 32:04family and the doctor. So, it can take 32:06take some very dark turns if we don't 32:09create rules and regulations in terms of 32:12what systems were allowed to to connect 32:16these AI these AI applications to. But 32:18as long as you're connecting it to your 32:20GitHub repo tools or your, you know, web 32:23search tools, I think we're going to be 32:24fine. 32:25>> Yeah, exactly. It doesn't feel like deep 32:26research presents this threat. 32:28>> Mihi made a made a good point. He 32:30referenced a lot of uh like fictional 32:33fears of humans. Um and I think this is 32:36what this plays into whether it's true 32:39or not. It it is it is clearly a a fear 32:42of humanity is that we will be overcome 32:45or become extinct or taken over. Um and 32:48then if you think about okay well what 32:49is what what is artificial artificial 32:52intelligence trained on? Okay, all of 32:54the data comes from humans, at least 32:58right now, right? Like all of the data 33:00comes from the internet and our own, you 33:04know, fears and our own publications. 33:07And so clearly, like maybe in some ways 33:10we're uh it's chicken and egg. We're 33:12kind of like enabling this aspect of it 33:15not wanting to be shut down because it's 33:17fed on all of our fears. 33:20Yeah, I love that because I was about to 33:21say like I was like, "Well, Miha, you're 33:23just talking about fiction. This book is 33:24supposed to be a piece of non-fiction." 33:26I guess Sandy, you've already kind of 33:28anticipated me because you're kind of 33:29saying, "Well, there's sort of a weird 33:31part to this which is regardless of what 33:33AI is, it's informed by all this data 33:36that has this fiction." And so like it 33:39itself is kind of reenacting this in 33:41real life. And so there's kind of this 33:43weird kind of muddling between like the 33:45fiction of what these AIs do and the 33:47reality of what they will do under a 33:48number of conditions. 33:53>> Cool. Well, we've got, as I promised, a 33:55lot to cover. So, I'm going to move us 33:56on to our our next story of the day. Um, 33:59this was a fun one and also building 34:01again from the discussion that we had 34:03last episode. So last episode we talked 34:05a little bit about a startup called 34:07Alterra Ego um and uh Meta doing a bunch 34:10of demos with wearables, right? They 34:12launched a new Ray-B band that has a 34:14bunch of AI features built into it and 34:15they had some problems on the demo, but 34:17the features are are pretty cool. Um and 34:20I guess not to be left out, um Apple is 34:23dropping its new uh AirPods, which I'm 34:26wearing one of them here today. The cool 34:29thing about the sort of new AirPods 34:30though is that they're going to have 34:31this built-in feature for uh real time 34:35translation. Um, which I think is like 34:38in some ways been like one of the really 34:40cool dreams of AI that I've been waiting 34:42for for many years is the idea that like 34:44you can hear audio and it just 34:46translates into a language you know 34:47autonomously. Um, and so I guess uh Gabe 34:52maybe I'll kick it over to you. you 34:53know, this is this is pretty cool and it 34:56feels very different from a lot of the 34:58other wearable AI features that have 35:00been launched lately. Um, but I'm having 35:04a hard time articulating why this feels 35:06so different. Um, and so I'm wondering 35:07if you could help me kind of like think 35:08a little bit about like why is this 35:10maybe different or maybe it's actually 35:12quite same to like what Meta is 35:13attempting to kind of portray as like 35:15the future of how AI is going to be with 35:17you. I think this is a 35:20solution to a problem instead of a 35:23solution in search of a problem. And I 35:25think that's why this feels different. 35:27You know, we work for an international 35:28company. Um I I think most discourse 35:33that I engage in happens in English, but 35:35uh there are many folks for whom that's 35:37not their first language. Um, and uh, as 35:41a native English speaker, I feel both 35:43privileged and a little bit uh, uh, bad 35:46all the time about forcing everyone to 35:48conform to my language, right? Um, I 35:51think 35:52language translation, whether it's for 35:54personal travel, whether it's for 35:56international commerce, whether it's for 35:58just about anything, is a real real 36:00issue in a global world and people face 36:02it every day. And so, yes, um, lots of 36:07cool possible features out there that 36:10could use AI in real time on some kind 36:12of a device that's attached to my body 36:14that I might like try out the demo and 36:16think, man, this is super cool, and then 36:19put away, but I could see, certainly not 36:22for everyone, not everybody lives in a 36:23multilingual world, but for folks that 36:25do, a a earpiece that you stick in that 36:28can give you that translation with 36:31minimal effort and a smooth UX is a real 36:33game changer. So, it's an actual problem 36:35and an actual solution to the problem. I 36:37think that's why this feels different. 36:38>> Sandy, is this the uh beginning of Apple 36:40making a comeback on the AI side of 36:42things. Um, you know, I think the most 36:45interesting reversal of fortunes we've 36:47been observing for the last 12 months 36:48has been, oh my god, Apple's going to 36:50kill it on AI. Oh man, they seem so far 36:52behind. And also Google, they seem so 36:54behind to, wow, they seem to be really 36:56doing all good all of a sudden. Um, I 36:58guess the question is if we're about to 36:59like go through another inflection point 37:01where by December next year, we're going 37:03to be like, "Oh, Apple's got this." We 37:05should have never have doubted them. 37:07>> You know, I've been getting a lot of ads 37:09for um Google's new phones on my TV. I'm 37:13being targeted for that. So, and and I 37:15was just talking to my husband 37:17yesterday. I was like, I wonder whether 37:18they're going to take over um Apple's 37:20device sales very shortly because 37:22they're so integrated in with the entire 37:27ecosystem, right? That that's Google's 37:29MMO is to integrate everything together. 37:32Um I do we know the model actually being 37:36used for the translation in the AirPods 37:39is has that been released? The article 37:41mentioned Apple Intelligence, which is 37:42their like multi-layered ondevice plus 37:45escalate to a secure cloud plus escalate 37:48to a frontier model at at all costs. So 37:50I I doubt we know specifically, but I 37:53suspect there's an ondevice component 37:54for the latest iPhones. 37:56>> Sure. Like with the latency that makes 37:58sense, but but I I also wonder like how 38:01much they're partnering to make this 38:03happen because you know there's a 38:05history of Apple partnering which I 38:07think is a good thing, right? if if you 38:09don't have it in house at the moment. 38:10You don't want to like just not do it, 38:13not keep up, right? And so they've 38:15partnered with um Google before to make 38:18Chrome the default browser rather than 38:21Safari. Um they've partnered with um and 38:25I they think they have a big partnership 38:26right now with Gemini, right, to bring 38:28it potentially and to catch them up a 38:30bit um since they delayed their Apple 38:34Intelligence release for the next 38:36iPhone. And so, you know, I I want to 38:39say yes. I I am hopeful because I think 38:41more players in the game is a good 38:44thing. Um I don't think we should have 38:48incumbents that capture like 75% market 38:52share when it comes to things like 38:54devices. Um which maybe is what Apple 38:57has kind of done in the past. Um maybe 38:59this is allowing them to share that a 39:01little bit more. this probably not a 39:03favorable if an Apple exec is listening 39:05is probably like I don't agree but but I 39:09think if in terms of a fair market that 39:11is reality is that we should have 39:13competition and maybe this is allowing 39:15some competition to shine through. So 39:17I'm not sure it's the worst thing. Look, 39:19I'm just kind of um going to make a 39:21reference back to a book again, 39:22Hitchhiker's Guide to the Galaxy, 39:24Douglas Adams, 1979. 39:26And I think he predicted a lot of this 39:29um with the Babelish, which was a fish 39:31you put in your ear from the 39:32hitchhiker's guide and it would 39:34translate between languages in real 39:35time. I think he's credited with even 39:38creating the concept of a tablet because 39:40the hitchhiker's guide was, you know, 39:42the size of a small pocket book. It had 39:44all the words information. and you could 39:46put the babel fish in your ear to give 39:48you real time translation. So I think 39:50these are things that humans have wanted 39:52for a very very long time. It came up in 39:54fiction, it came up in movies. Uh we've 39:56seen various implementations over the 39:58years. I remember this feature was 40:00present in Skype which is no longer a 40:02thing right now. But in Skype you could 40:04do real time translation with your I 40:07would say you know if you have a family 40:08and they speak a different language you 40:10can do realtime translation. I think 40:12it's down to the user experience. Making 40:15it part of something like your 40:16headphones is going to be a lot more 40:18conducive to folks who are not used to 40:20technology. They're not going to go and 40:22download the model. They're not going to 40:23go into Skype. Um I think the first 40:26couple of iterations are going to have 40:28some interesting challenges. So I expect 40:31a lot of funny situations in a lot of 40:33countries where you think I can now 40:35speak Italian. It's like no not quite 40:37>> and you can cannot. 40:39Um, so, so I expect there's going to be 40:41a couple of challenges, especially if 40:42you're going to the local model first. 40:45Uh, but I see it as a good thing. So, I 40:47see it as AI becoming integrated into 40:50technology to the point where from a 40:52consumer spe perspective, it disappears. 40:55You don't care what model is being used. 40:57You don't care. You What's a token? I 40:59don't care what the token is, right? I 41:01just want this thing to translate what 41:02you're saying in real time. Done. 41:04>> I I I think your your point about UX is 41:06spot on, Nihi. And I think um it really 41:10it to your earlier question Tim about 41:12what feels different. Um I think that's 41:15what Apple is getting right with this 41:16and to the question also about whether 41:18or not Apple has a comeback in store. I 41:20think that is their avenue is UX that as 41:24you aptly put it Mihi uh has no 41:27reference to the fact that there are 41:28tokens anywhere. It just fits into 41:30something that's already part of your 41:31ecosystem and it makes it better and it 41:33solves a problem. Um, you know, one 41:36other incumbency that Apple has, uh, is 41:39sort of style and popularity. You know, 41:41frankly, Tim, seeing you in those nice 41:43white headphones, um, I I see people 41:45walking down the street with those 41:47headphones in all the time, especially 41:48all the teens in my neighborhood. It's 41:50it's a look. And so now that those solve 41:54an additional problem, that is not 41:56saying I have to now put on a clunky 41:57pair of additional glasses or uh some 42:00heads up display that makes me look like 42:02a geek uh to actually, you know, bring 42:05this technology into my life. It's 42:07fitting into something that's already an 42:09accepted thing in everyday life. And so 42:11the user experience is really spot-on. 42:13So I think that's the avenue um that has 42:16some some real traction here. Yeah, just 42:18look up on the internet United Nations 42:20translation device because you see them 42:23at the United Nations and there's like 42:24this awkward cheap looking thing with 42:26the antennas and I think there's a 42:28person behind it but still it's it's not 42:31a look. 42:32>> Yeah, for sure. Um Sandy, final thought 42:35on this? 42:35>> Yeah, I I I'll wrap this up with a a 42:38quick story. Um earlier this year I was 42:40in Japan doing an awesome solo trip, 42:42getting lost in so many train stations 42:44and having to ask for directions, right? 42:47And um and so I pull out my and I was 42:50using GPT because it just is 42:52semantically better translation um and 42:55voice mode in a way, right? And so I was 42:57doing the translation, but then I would 42:59hand it back to like the older gentleman 43:01and he had no concept how to like 43:06how to participate, right? And so it 43:08wasn't he would speak but not press the 43:10button and then I'd press the button too 43:11late and it would just it was it didn't 43:13work, right? because of the learning gap 43:16and the skills gap and so this I think 43:20is a if they can nail a way to solve 43:22this right that would bridge that gap 43:25where you don't have to change your 43:28learning pattern you just change the 43:29experience 43:30>> yeah this is going to be a super 43:31interesting interface and I want to kind 43:33of keep coming back to it as we go about 43:34like where where the AI will be in this 43:37wearable ecosystem I think it's really 43:39interesting and I think yeah as Sandy 43:40you're pointing out in some ways like 43:42even sort of quote unquote seamless 43:44interface faces have a lot sort of built 43:46into them. Um, so a lot to talk about 43:48there. 43:52Uh, I'm going to end this on a final 43:54story which is breaking as of this week. 43:56Um, it's a headline that could only come 43:59from 2025. And that headline is that 44:02Nvidia is investing $100 billion. I'm 44:06laughing as I say it because it's such 44:08an absurd number in some ways. A hundred 44:10billion dollars in open AI. Um, and so I 44:14want to take the maybe like last five 44:16minutes of this episode to just talk a 44:17little bit about this news. Obviously 44:19the topline number is mind-boggling. Um, 44:23but I guess Mihi, maybe I'll get you to 44:25respond to one thing that occurred to 44:26me, which is Nvidia's about to get or 44:29OpenAI is about to get all this money 44:31from Nvidia. Isn't OpenAI just turning 44:33around and giving it back to Nvidia? 44:38>> It's like stock buybacks. It's perfect. 44:42I mean, it's it's kind of strange, 44:44right? Like like that actually we should 44:46read this as open AI gives Nvidia 44:48hundred billion dollars back, right? 44:52I mean, some of it will be spent on 44:53personnel and that kind of thing, I'm 44:55sure. But yeah, um what do we what do we 44:57make of that? That sounds that seems 44:58weird. It does, but then you look at 45:01things like stock buybacks and 45:02everybody's doing it and then you're 45:04making billions of dollars. If you look 45:05at the Oracle stock, which kind of blew 45:08up as well from a very similar event, um 45:11you're going to see it has a big 45:12financial impact. Then second, they're 45:13sponsoring their biggest client. You 45:16know, if OpenAI goes down, they're 45:18they're going to be in trouble. I think 45:20there was an article going around that 45:21Nvidia has like, you know, three, four 45:23large clients and then the rest. And 45:26then there's a couple of gamers in there 45:28with GPUs, but those are not as 45:30important as those three, four, you 45:32know, whale of customers. And 45:36there's also the opportunity for AMD or 45:40Intel or some other company to come in 45:43and say we have found a solution and 45:47OpenAI is going to diversify. So by 45:49making the investment they're locking in 45:51their biggest potential target. There is 45:54no indication that the OpenAI is going 45:56to reduce the number of GPUs they need. 45:58They're still making their models, you 46:00know, CH GBT freely available for a 46:03billion users or whatever the numbers 46:06are up to right now and they're going to 46:08continue doing it on Nvidia GPUs. So, I 46:11think it's a brilliant move. I think 46:13it's going to 46:15have massive effects on the industry. I 46:18would have wished to see a bit more, I 46:20would say, diversity in terms of what's 46:23supported for inference providers. So 46:26AMD and Intel give us more GPUs, make 46:28them better. I've got an AMD GPU right 46:31there. I love it. But when it comes to 46:33the software, 46:35that's where I think CUDA and what open 46:38what um Nvidia is doing still still has 46:41the lead. 46:42>> Yeah. Sandy, is um is Enthropic in 46:44trouble? They're sort of left out of 46:46this, right? Like 46:47>> there seems to be unlimited money right 46:49now. I I I'm I'm not sure. They they 46:51plan to raise I think a $8 billion round 46:55just recently and raised a $13 billion 46:58round. 46:58>> They'll be okay is what you're saying. 47:01>> Maybe they are in trouble. Maybe this 47:03will be a catalyst for them to to in 47:06order to compete and stay in the ring to 47:08need to raise even more um than 47:10anticipated. But 47:13I I I feel like we're just going to see 47:15more partnerships emerging. We're going 47:16to see more alliances like clubs. You 47:19know, something I did want to comment on 47:21though is the amount of energy that the 47:25facilities that they're investing in are 47:27putting out is enormous, right? Like I 47:31think typically Meta has notoriously 47:34some of the bigger biggest scale 47:36projects in this industry and this like 47:39out competes that tenfold. Um I think 47:42there I think the number was that they 47:45expect 10 gawatts of power at least from 47:50the facility which is like a billion 47:53light bulbs at a time. 47:55>> It's a lot of light bulbs. 47:56>> It's kind of insane. 47:58>> Yeah. Well, and Gabe maybe you can take 48:00us home for the final comment. I mean 48:02this kind of idea of tribes I think is 48:04really interesting because it does sort 48:05of feel like we're getting like it's 48:07going to be Gryffindor and Ravenclaw. 48:10it's going to be, you know, Haronin and 48:12what have you. Like I I like it feels 48:14like initially it was oh okay well 48:16Anthropic is going to kind of pair off 48:17with say Amazon and OpenAI is going to 48:20pair off with Microsoft and that's for 48:22the the sort of cloud right and then now 48:24it feels like okay well the next step in 48:25the game is open AI is going to go with 48:27Nvidia and I don't know maybe anthropic 48:29is going to go with AMD right and then 48:32like I guess the next step in that game 48:33would be Sandy what you're talking about 48:34is I guess like open AI is going to go 48:36with like northeastern nuclear power you 48:39know company and Anthropic is going to 48:41go with you know southwestern you know, 48:43uh, wind power or something like that. I 48:45guess Gabe, like where does this all go 48:47in terms of market structure? Like, do 48:48we feel like we're going to almost have 48:49these like vertically integrated 48:51alliances start to emerge over time? 48:53Because I mean, traditionally, we've 48:55seen like Nvidia selling its chips to 48:57everybody. And I think this does really 48:59seem to put the the thumb on the scale a 49:00little bit, right? I think there's this 49:03is a big bet, frankly, uh, that bigger 49:06is still better, right? I think um we've 49:10hit a point now with some really really 49:12good frontier models trained on the 49:14infrastructure that already exists and 49:17um we're seeing that huge gains in 49:20capability are being achieved not 49:23necessarily just by training the newest 49:25models but by training by by building 49:27the systems around the models. And 2025 49:30the year of the agents and you know 49:32we've we've covered everything under the 49:34sun about and I keep coming back to you 49:36know models are just generating tokens. 49:37It's what you put around them that 49:39actually adds the value. And I think, 49:43you know, there's a real obvious loser 49:45in moving the innovation from the models 49:47to the systems and that's Nvidia, right? 49:50Because you don't need to buy more chips 49:52if you've already got enough, right? You 49:54don't need to boost your data center 49:56size if your models if you can keep 49:58refining the models with the same 50:00hardware you already have. And so I 50:02think this is a huge bet on Nvidia's 50:04part to say, "Hey, we believe that 50:08bigger is still better and we're going 50:10to make bigger happen and see what 50:12happens." Um, so they're putting their 50:14money where their mouth is very 50:16self-interestedly as you put, Tim, and 50:18it's going to come back to them in the 50:19chips anyway. But I think, um, 50:24does this where does this shake the 50:26industry out? It's really going to it's 50:28going to come down to the technology of 50:29whether big is in fact better. Um and 50:32Sandy to your point eventually these 50:34this power consumption the environmental 50:36impact um you know do we can can we 50:40source this power responsibly like we're 50:44going to hit some breaking points there 50:45when you're talking about orders of 50:47magnitude this big and I'd be curious to 50:51see I mean again it's really uncharted 50:54territory. It's a push to explicitly try 50:57to push into further uncharted territory 50:59so that um this critical resource, the 51:04next generation of GPUs, is still a 51:06commodity that is desperately 51:07desperately needed. We'll see where 51:09those uncharted territories lead. 51:10>> Gabe, to your point before we wrap up, 51:12I'm I'm curious whether the hundred 51:14billion will completely go to facility 51:17or they're maybe making that next 51:19investment in the next type of chip or 51:21the next architecture. like is that is 51:24this a play for what's whatever is next? 51:27>> Very possible. But I think you know the 51:29topline power numbers and the topline 51:32scale numbers still really hint at 51:34bigger is better. I think you know 51:36theoretically you could invest $und00 51:38billion in chips that are 10x more 51:40efficient rather than 10x more powerful. 51:43But I don't think that's what they're 51:44doing here. Um, and so it seems like 51:47this is clearly just trying to push the 51:50upper envelope of, you know, raw 51:52compute. Um, but I'm sure some of that 51:56will go into the ecosystem. And I don't 51:58know, I didn't get into the details of 51:59the article about how much is up to the 52:01discretion of OpenAI versus how much of 52:03this is just build the biggest data 52:05center that's ever been built and staff 52:07it up with our GPUs. 52:09>> Can you smell the AGI? 52:11>> Yes, I suppose we all can. Uh, it's a 52:13great note to end on. Um, that's all the 52:15time that we have for today. Uh, so, uh, 52:18Mihi, Sandy, Gabe, thank you for joining 52:20us as always. And, uh, thanks to all you 52:22listeners. Uh, if you enjoyed what you 52:24heard, you can get us on Apple Podcast, 52:25Spotify, and podcast platforms 52:27everywhere. And we'll see you next week 52:29on Mixture of Experts. 52:31[Music]