Learning Library

← Back to Library

Karpathy vs McKinsey: AI Design War

Key Points

  • A emerging conflict in AI pits business consultants, exemplified by McKinsey’s boardroom influence, against technical builders like Andrej Karpathy, highlighting divergent strategic visions.
  • Karpathy’s “Software 3.0” talk at Y Combinator frames large language models (LLMs) as computers, utilities, and operating systems, arguing that the next programming language will be English.
  • He introduces the term “people spirits” to describe LLMs as stochastic simulations of humans, emphasizing that designing software for these entities requires fundamentally new, human‑centric approaches.
  • Karpathy cautions against hype about autonomous AI agents, stressing that realistic progress over the next 1‑2 years will depend on substantial human supervision, a stance he presents as more realistic than many industry voices.

Full Transcript

# Karpathy vs McKinsey: AI Design War **Source:** [https://www.youtube.com/watch?v=xZX4KHrqwhM](https://www.youtube.com/watch?v=xZX4KHrqwhM) **Duration:** 00:11:48 ## Summary - A emerging conflict in AI pits business consultants, exemplified by McKinsey’s boardroom influence, against technical builders like Andrej Karpathy, highlighting divergent strategic visions. - Karpathy’s “Software 3.0” talk at Y Combinator frames large language models (LLMs) as computers, utilities, and operating systems, arguing that the next programming language will be English. - He introduces the term “people spirits” to describe LLMs as stochastic simulations of humans, emphasizing that designing software for these entities requires fundamentally new, human‑centric approaches. - Karpathy cautions against hype about autonomous AI agents, stressing that realistic progress over the next 1‑2 years will depend on substantial human supervision, a stance he presents as more realistic than many industry voices. ## Sections - [00:00:00](https://www.youtube.com/watch?v=xZX4KHrqwhM&t=0s) **AI Consultant vs Builder Clash** - The speaker contrasts business‑focused AI consultants like McKenzie with hands‑on builders such as Andre Karpathy, using Karpathy’s “Software 3.0” presentation to argue that his developer‑centric vision will likely dominate the AI landscape. - [00:03:32](https://www.youtube.com/watch?v=xZX4KHrqwhM&t=212s) **Constraining AI Output for Practical Validation** - The speaker debates Andre’s proposal to limit LLM‑generated content—favoring a streamlined validation loop over overwhelming reviewers—while also contesting the notion that English will replace all programming languages, emphasizing the continued need for skilled engineers to manage increasingly complex hybrid systems. - [00:06:38](https://www.youtube.com/watch?v=xZX4KHrqwhM&t=398s) **Skepticism Over Agentic Mesh Claims** - The speaker critiques the hype around a supposedly plug‑and‑play “agentic mesh,” noting its lack of empirical grounding, reliance on outdated models, and the broader disappointment with edge‑computing promises such as Apple’s recent bet. - [00:11:30](https://www.youtube.com/watch?v=xZX4KHrqwhM&t=690s) **Pressuring Firms on AI Accountability** - The speaker thanks Andre, urges organizations such as McKinsey to adopt a stronger stance on AI, acknowledges they may ignore the request, yet still hopes for a better response to AI challenges. ## Full Transcript
0:00There's a war at the heart of AI between 0:02the business consultants and the 0:04builders. And I want to outline how that 0:06popped out in sharp relief this week 0:08between Andre Karpathy and McKenzie. 0:12Both of them had major presentations uh 0:14this week, major papers this week. I 0:16want to talk about how stark a contrast 0:19they laid out and why Andre's vision is 0:22more likely to be correct. It's 0:25important to understand both though 0:26because McKenzie has tremendous 0:28influence in the boardroom. Okay, first 0:30understand the context for Carpathy's 0:33presentation. He's speaking to a bunch 0:34of entrepreneurs at Y Combinator Startup 0:37School. His presentation is titled 0:40software 3.0, which he is sort of 0:43uniquely qualified to talk about because 0:44he coined the term software 2.0 uh a few 0:47years ago, I believe, also at YC. So, 0:49he's coming back and he's basically 0:51saying there's a new paradigm. It's 0:52shaped obviously by AI. and he spends a 0:55lot of time in the presentation which 0:56I'm going to link encouraging you to 0:59think about AI as a design problem that 1:02is unique because of the qualities of 1:05the large language model. And so he 1:07talks about large language models as 1:09computers, large language models as 1:11utilities, large language models as 1:12operating systems. And he describes in 1:15detail how LLMs have qualities that 1:18match these. So as an example for 1:20utilities we meter their usage dollars 1:22per token the way we meter electricity 1:24right uh for oss we've already heard 1:27other major figures in AI talk about the 1:30fact that especially young people are 1:32using AI like an operating system and 1:34you have differences in preference for 1:36operating system the Windows versus Mac 1:38wars well similarly you have differences 1:41in preference for clawed versus open AAI 1:43so you have some of that same sort of 1:45dichotomy playing out but let's get to 1:47the heart of software 3.0 0 software 3.0 1:50is the idea that the next coding 1:52language is English and that we are not 1:55working with deterministic software. 1:57Instead, we are working with what 1:58Carpathy terms people spirits. So, 2:01stochastic simulations of people is the 2:03way he puts an LLM. I love that phrase. 2:06I'm going to keep and like share it a 2:08lot because it helps me explain why 2:11large language models feel so human but 2:14aren't. It explains why the intelligence 2:17of large language models feel so jagged. 2:20They are stochastic simulations of 2:22people. They're people spirits. And so 2:24if we're building software for this kind 2:26of interaction for people spirits, we 2:28have to think from the ground up how we 2:30design our software. And this is where 2:32Andre's caution comes in. And I think 2:34it's really needed in an age when we are 2:36hyping up agents so much. It is really 2:39really important to think about our 2:42building in the next six months, 18 2:45months, two years as building for people 2:47spirits that need a fair bit of human 2:50people supervision to go anywhere. And 2:53Andre is more honest about this than 2:55most of the other major figures in AI 2:57that I've seen. He is not overhyping and 2:59saying that AI agents will take over 3:01everything and be autonomous. And this 3:03is where you see an early conflict with 3:04McKenzie because what Andre is saying is 3:06essentially people's spirits or LLMs 3:09just don't have the uh reliable 3:13execution. They have too much jaggedness 3:16in their intelligence to be good at 3:18enough of everything to be trusted with 3:20highle tasks at this point. Instead, we 3:23should be building our software for the 3:25assumption that humans will need to be 3:26validators in the loop that AI can 3:28generate and human needs to validate. 3:30And we need to think about software as a 3:32design problem from that perspective. 3:34And he suggests there's two ways to make 3:36this easy. One is pretty obvious. Make 3:38the the checking responsible validation 3:41loop as easy as you possibly can. That's 3:43software 101. But the second is a little 3:45bit more controversial. Andre suggests 3:47putting the LLM on a short leash, 3:50deliberately constraining AI generation 3:53so that you don't have so much AI 3:54generation that you overwhelm 3:56evaluators. An example of this would be 3:58the AI generating hundreds of different 4:01ad variants, but the human only being 4:02able to validate 10 of them. Well, 4:04what's the point? You're just wasting 4:05energy at that point. And I appreciate 4:07his honesty on that front. Now, I do not 4:09think that Andre is entirely correct, or 4:12at least I disagree with him, that 4:14English will be the only programming 4:15language of the future effectively. I 4:18think in particular there will be a need 4:20for strong technical engineers who 4:22understand the construction of complex 4:25systems because systems are about to 4:26become more complex as we have 4:28traditional software interacting with 4:30this agentic augmented software that 4:32he's talking about. This is not going to 4:34be as clean as English driving code all 4:38the way through. But I do understand 4:40from his perspective as someone who's 4:42dipped in engineering from the beginning 4:44and like knows his code backward and 4:46forward the transition to English is a 4:48fundamental shift and he is to his 4:50credit honest about the limitations of 4:52the vibe coding revolution that he 4:54kicked off a few months ago. He was the 4:55one that said vibe coding and spawned a 4:58thousand startups. And so he talks about 5:00the fact really honestly that vibe 5:01coding right now is great for local 5:03environments, but there's a lot of other 5:05pieces in the deploy pipeline in CI/CD 5:08and integrations that don't work well 5:10with vibe coding right now. And I 5:11appreciated that honesty as well. So 5:13when you ladder all of that up, what he 5:15is basically saying, Andre is leaving us 5:17with this vision of software 3.0 is 5:20building like augmented iron man suits 5:22for ourselves where the agents expand 5:24our our span, our reach, our control. uh 5:27but we have to design our data systems 5:29to accommodate how they interact with 5:30data. We have to design our software so 5:32it's agent-friendly. We have to think 5:34about agent control systems so that you 5:36can have agents interacting with data 5:37and people validating it in a 5:39sustainable loop. It's a really 5:40interesting software design talk and 5:42it's scalable and it's empirical and 5:44because he is a builder you can feel the 5:46fingertip knowledge and that is the 5:48fundamental distinction between Andre's 5:50presentation of software 3.0 and 5:52McKenzie's presentation which is very 5:55very different. McKenzie is speaking to 5:57CEOs. McKenzie and look, I I get that 6:00Mistl blessed the McKenzie presentation. 6:03Uh it's all about agentic mesh. That's 6:05the theme. And like the CEO of Mistl has 6:07a nice introduction at the beginning. 6:08This is not an attack on Mistl. They do 6:10hard work. They produce great software. 6:12But McKenzie because of the way they 6:14speak to their audience is not able to 6:17successfully articulate anything that's 6:20buildable for tech teams. And that is 6:22the fundamental issue. I understand that 6:24they want to communicate to CEOs in 6:26their presentations, and I'll link to 6:28this as well, that it is important to 6:30think in terms of workflows. That's 6:32true. It is important not to just think 6:34in terms of LLM's automating tasks. 6:36That's true. If you think about agents, 6:38you have to think about autonomy. That's 6:39true. The problem lies when they go from 6:42general concepts to try and suggest a 6:45solution. The agentic mesh is a word 6:48salad that has no empirical grounding. 6:51it doesn't have the builder's touch. And 6:53that is what makes that presentation so 6:55concerning because I've seen over and 6:57over again as someone in sort of the 6:59product engineering side of things when 7:01you have a CEO come in fresh off a 7:03report like that and he's like this 7:05should just work. The McKenzie guys say 7:07that they can build an agentic mesh and 7:08you can plug any model in without 7:10additional work. Why don't we use uh you 7:13know Mistl small or why don't we use 7:15GPT3.5 turbo because McKenzie mentioned 7:18it. Both of those are in the 7:19presentation, by the way. And the tech 7:21teams roll their eyes because they're 7:22like, "These are ancient models. They're 7:23tiny." It relies on this assumption of 7:25edge computing that hasn't sustained 7:27very well because larger models just 7:28show sustained gains in intelligence 7:30that smaller models aren't matching. 7:32That's one of the big surprises of 2025 7:34is that edge computing for models is not 7:36working as well as people thought it 7:38would yet. Um, and to his credit, Andre 7:40still thinks there's room for edge 7:41computing. We will see. Apple made a big 7:43bet on it earlier last year, and it 7:46really hasn't paid off. 7:48uh it remains to be seen. I don't want 7:50to sort of rabbit hole us on edge 7:51computing. That's probably a different 7:52conversation. The point sort of for 7:55McKenzie is that they should be able to 7:58recommend something that is actually 7:59buildable. And if you recommend what is 8:02effectively a theoretical substrate for 8:04agents that allows them to plug in like 8:06USB ports and any agent can plug in and 8:09you can plug in any data, that is a 8:10fiction for a CEO. That makes a CEO 8:13sleep well at night. It is not true. It 8:16is not how you actually build things. I 8:18understand because I've had to work with 8:20boards that you do have to simplify 8:22technical concepts into a business 8:23narrative. I understand that. I 8:25understand that you have to have 8:26outcomes that you can talk about that 8:29are easy for non-technical people to 8:30understand. It is possible to take Andre 8:33Carpathy software 3.0 or a similarly 8:35clean technical vision and tell good 8:38business stories. you do not have to 8:40resort to the kind of um sophistry, the 8:44kind of word salad that McKenzie uses in 8:47order to communicate clear business 8:48narrative. And in fact, the fact that 8:51the fact that they're doing that, right, 8:52the fact that they are telling a story 8:54that isn't real at root because you 8:57can't just plug agents in like USB, like 8:59you can't just plug them in without 9:01modification from any from any source 9:03whatsoever and stick them into data and 9:05just expect it all to magically work. It 9:07does not work that way. And if you sell 9:09that vision, what you are selling is the 9:13reason why so many enterprise companies 9:15are walking away from AI after an 9:17investment and why so many enterprise AI 9:19projects don't launch. It's because of 9:21advice like this. And so part of why I 9:23am punching up on McKenzie a little bit 9:25is I I need people who have seuite and 9:28board ears to tell the truth about 9:31building AI to tell the truth about how 9:33complex AI systems are. That yes, there 9:36is a power law of payoffs. If you invest 9:38and you get true AI in Agentic systems 9:42and you can implement them at the 9:44enterprise level there there is big 9:46money on the table. It matters but it's 9:48hard to get there and if you are just 9:50starting out that may not be the place 9:51you want to start out. You don't want to 9:53necessarily start out with automating 9:55your entire customer success line or 9:58automating all of your 10:02oh what have you. Automating all of your 10:04uh retail uh orders and pickups. You get 10:07the idea. What you want to do is focus 10:10on a crawl, walk, run motion. Describe 10:12the culture change you want and start 10:14living into that. And that's the piece I 10:16want to leave you with today. What is 10:18the culture change that Andre is 10:20suggesting we need to create in our 10:23organizations that enables us to think 10:25in terms of software 3.0 that enables us 10:28to think and relate to LLMs not as 10:30people, not as programs, but as 10:32stochastic simulations of people in a 10:34probabilistic context. There's an 10:36emergent psychology to LLM that is 10:38relevant to talk about even if the 10:40psychology isn't quote real because 10:42these are quote simulations. We can 10:44still talk about it and understand it 10:46and that may be a window for us to 10:48understand how probabilistic agents 10:50interact with our software 10:51infrastructure. There's a lot to dig 10:53into, but I would much rather us dig 10:55into what is actually going on and tell 10:57business stories that actually matter 10:59than go to McKenzie's side of the fence 11:02to pretend everything is easy and get 11:05into a position where enterprise after 11:07enterprise starts on AI and walks away 11:10because they discover belatedly that 11:12it's much harder than the board deck 11:14says. It's just not true that you can 11:16plug in agents anytime. It's just not 11:18true that these tiny little edge models 11:20will do whatever you want and won't get 11:22eaten by the next large model that comes 11:24along. We need to do a better job 11:26telling truths up and down the stack. 11:28And I appreciate Andre for doing his 11:30best to lay that out. And I'm asking 11:32organizations like McKenzie to take a 11:35stronger stance there. And look, I who 11:37am I kidding myself? They're not going 11:38to hear me. They're not going to listen. 11:39That's okay. I can still ask. I can 11:42still expect a better response to the 11:45challenge of AI. Cheers.