Learning Library

← Back to Library

Building Kai: AI Personal Augmentation

Key Points

  • The session is split into a high‑level overview of how to augment yourself with AI, followed by tactical demos of Daniel’s recent AI projects and an open Q&A.
  • Daniel, a former security leader at Apple and Robinhood and current author of the “Unsupervised Learning” newsletter, has shifted his focus to applying AI for security consulting and human flourishing.
  • He’s constantly prototyping new AI‑powered tools—evidenced by nightly “sickest thing ever” screenshots he shares with the host.
  • Daniel will deep‑dive into his “Kai” system, explaining its cloud‑code‑based yet platform‑agnostic architecture, core design principles, and how others can replicate it through a quick demo and an FAQ.
  • The overarching goal is to give attendees concrete knowledge and inspiration to build their own AI‑augmented workflows that boost personal performance.

Sections

Full Transcript

# Building Kai: AI Personal Augmentation **Source:** [https://www.youtube.com/watch?v=Le0DLrn7ta0](https://www.youtube.com/watch?v=Le0DLrn7ta0) **Duration:** 00:48:39 ## Summary - The session is split into a high‑level overview of how to augment yourself with AI, followed by tactical demos of Daniel’s recent AI projects and an open Q&A. - Daniel, a former security leader at Apple and Robinhood and current author of the “Unsupervised Learning” newsletter, has shifted his focus to applying AI for security consulting and human flourishing. - He’s constantly prototyping new AI‑powered tools—evidenced by nightly “sickest thing ever” screenshots he shares with the host. - Daniel will deep‑dive into his “Kai” system, explaining its cloud‑code‑based yet platform‑agnostic architecture, core design principles, and how others can replicate it through a quick demo and an FAQ. - The overarching goal is to give attendees concrete knowledge and inspiration to build their own AI‑augmented workflows that boost personal performance. ## Sections - [00:00:00](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=0s) **AI Augmentation & Open Q&A** - The speaker introduces a two‑part session that explains high‑level approaches to augmenting oneself with AI, showcases Daniel’s recent AI‑driven tools and demos, and then opens the floor for audience questions. - [00:04:01](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=241s) **Getting Started in the Agentic Age** - The speaker cautions against haphazardly building AI “agentic” systems, emphasizes first clarifying personal goals and values, and explains how the emerging level‑two agentic ecosystem took off around 2024‑2025. - [00:08:50](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=530s) **Scaffolding Trumps New Model** - The speaker contends that strong scaffolding and orchestration around AI models outweigh the benefits of the latest model version, especially for vulnerability detection tasks, citing tools like OpenAI’s Arvar and successes in competitions such as AIXCC. - [00:12:27](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=747s) **Beyond Prompts to Deterministic Testing** - The speaker argues that relying solely on AI prompts is inefficient, highlights Anthropic’s recommendation to replace their MCP with TypeScript for more reliable execution, and stresses the importance of specifications, tests, and consistent evaluation to move away from “vibe‑based” development. - [00:15:36](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=936s) **Custom Slash Commands & AI Skills** - The speaker explains how natural‑language prompts and custom slash commands invoke various AI skills—like summarizing, multi‑level explanations, and automation—by chaining tools such as Fabric, Gina AI, and Kai’s terminal integration. - [00:20:31](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=1231s) **Self‑Updating Skills Architecture** - The speaker outlines Kai’s pipeline—from top‑level goals through prompts, command‑line tools, and modular skills—to implement self‑update, self‑healing, and self‑improvement, exemplified by an upgrade skill that automatically gathers, parses, and integrates the latest external resources into the system. - [00:23:41](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=1421s) **Custom Skill Routing & History System** - The speaker describes augmenting cloud code with an explicit routing layer for skills—using workflow tables and deterministic tools to boost accuracy—and implementing a custom history system that logs all agent actions, sessions, and decisions. - [00:27:28](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=1648s) **Exploring Kai’s Skill Directories** - The speaker walks through Kai’s skills folder, detailing workflow and tool categories, and then tests Kai by inputting a series of whimsical concepts to see how it generates visualizations. - [00:30:57](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=1857s) **Cost Analysis of AI Tool Subscriptions** - The speakers compare pricing, usage caps, and ROI implications of services such as Bun, 11 Labs, and subscription versus pay‑as‑you‑go models, arguing that higher monthly fees can be justified by increased productivity and revenue. - [00:34:03](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=2043s) **Choosing Cloud vs. Local AI Infrastructure** - The speaker compares self‑hosted GPU solutions with cloud‑based tools such as Cloud Code and Gemini CLI, arguing that Cloud Code’s scaffolding and command‑line integration with Google’s ecosystem make it a superior choice for building AI applications. - [00:37:20](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=2240s) **Layered AI Defense Strategies** - The speaker outlines a multi‑layered security framework—using prompt‑injection safeguards, Anthropic tool‑access controls, and sandboxed sub‑agents—to protect their AI system, while noting its approximate 85‑95% effectiveness and potential for future improvements. - [00:41:21](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=2481s) **Heavy Git Usage & Safety Practices** - The speaker outlines a Git‑centric workflow with robust safeguards against mixing multiple session tabs, leverages file‑system work trees and branch experiments, and notes plans for an upcoming deep‑dive training course. - [00:45:04](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=2704s) **Managing Private Kai and Public PI** - The speaker outlines how they keep personal AI work in a private system called Kai while releasing selected, high‑value components to a publicly accessible version called PI, describing their rapid update workflow and philosophy of democratizing content creation. - [00:48:10](https://www.youtube.com/watch?v=Le0DLrn7ta0&t=2890s) **Planning Follow-Up and Async Collaboration** - The participants agree to share resources, use asynchronous communication, and schedule another session in a few months as their workflow continually evolves. ## Full Transcript
0:02So today we're going to cover basically 0:04two things. Um so we're going to cover 0:07at a high level how to think about 0:08augmenting yourself with AI as well as 0:11some tactical examples and demos from 0:13some really cool stuff Daniel has built. 0:15And then the second part is basically 0:16open Q&A. So ask Daniel and I anything 0:19you want. So just sort of open 0:20discussion. Little bit about Daniel. 0:22He's a longtime great friend of mine. So 0:25I'm a huge fan of him as a person as 0:27well as his work. He runs unsupervised 0:29learning which is one of the biggest and 0:31best security newsletters in my opinion. 0:33He recently keynoted OASP global apps 0:36USA which is pretty cool. He was 0:37previously a security leader at Apple, 0:40Robin Hood and other places but recently 0:42the past couple of years he's been 0:44focusing on AI so both applying it to 0:46security where he does some consulting 0:48work with companies as well as using AI 0:50for human flourishing which is what 0:52we're going to be uh focusing on a bit 0:54today. And just a fun anecdote, 0:56basically every week sometime between 10 0:58PM and midnight, I get a text from 1:00Daniel that's basically like, "Dude, I 1:02just built the sickest thing ever." And 1:03then he sends me a screenshot of some uh 1:05cool new like dashboard or something 1:07he's built. So today, we're going to be 1:08covering a number of those things that 1:10he's been building cuz he's constantly 1:11tweaking and improving his stack. Yeah. 1:13With a focus on using AI to augment 1:15yourself. So that's a little bit of 1:16context, but yeah, let's jump right into 1:18it. Daniel, I pass it to you. 1:20>> Awesome. Thanks, Clint. Appreciate the 1:22intro and thanks for uh having me. I 1:24appreciate the time here. 1:24>> Appreciate it. 1:25>> So want to talk about why I built Kai. 1:27Some stuff you need to know before you 1:29can build a similar system if you want 1:31to do that. Core principles and 1:33engineering that I put like into it like 1:36the design concepts and everything. Many 1:38of which are actually different than uh 1:40cloud code natively. So the Kai system 1:42is built on cloud code, but it's kind of 1:44designed to be agnostic and not just 1:46have the base stuff that's in cloud 1:48code. I've extended it quite a bit. 1:50We're going to do a deep dive on an 1:52actual skill. We're going to do a quick 1:54demo and I'm going to show you how to 1:55get started on your own. And uh we got 1:57an FAQ which uh you could take a picture 1:59of. I don't think I'm going to read 2:00through that one. And then as Clint 2:02said, we have time for discussion and QA 2:04afterwards. So like Clint also said, I 2:06want to apologize beforehand because 2:08this could be like an hour for each 2:10section. And in fact, when I do it other 2:12places, it is actually much longer. So 2:14we will have to go pretty fast. But um 2:16I'm going to go in more detail about a 2:18lot of these things than I've done 2:20anywhere else. So that should be fun to 2:22uh get into. But the goal is ultimately 2:24to give you enough detail and direction 2:26to build your own system like this. So 2:29why I made it in the first place? The 2:31main reason is because I just wanted to 2:34get better overall at everything that I 2:36do. I don't like being surprised by 2:38things. I like understanding how things 2:40work. And when I'm surprised, that means 2:41I didn't know how it worked at whatever 2:43level. So that's basically like an AI 2:46augmentation system was the first thing 2:48I I thought about at the end of 22 when 2:50this all went crazy. I also think 2:52regular jobs are going away. I talk 2:55about this postc corporate world and I'm 2:57really worried about that for humans and 2:58I think the best way to get ready for 3:00that is to get really really good at 3:03being a human and that means 3:04understanding yourself and using AI to 3:06magnify yourself. So this is my model 3:08for thinking about how AI is going to 3:11impact like the job market and how much 3:13it's going to I guess get inside of the 3:15workflow of how we do work and 3:17eventually take over more and more of 3:19that. And I use this to sort of keep 3:21pressure on myself kind of thinking 3:23about where it's all going and basically 3:26where am I and how ready am I and how 3:28ready are other people as well. So it's 3:31five levels. Before 2022, we didn't have 3:34any AI, right? So we basically did all 3:36this work ourselves. 23 to 25 roughly. I 3:39mean, this is all general, right? This 3:41is like the first level of AI. It's chat 3:42bots. Everyone understands this. It's 3:44like you ask a question, you get a thing 3:46back, and then you could do manual work 3:48with that thing, which is still really 3:50valuable. And you notice at the bottom 3:52here, we're inside of the section for 3:54human- centered work. So the first three 3:56sections here are largely human focused. 3:59And what we're entering into now is like 4:01this whole agentic thing, which by the 4:03way, I hate the word agentic. I I think 4:05it's a good word. I think it's a cool 4:06word, but everything is agentic now. Got 4:09aentic toaster ovens or or whatever. 4:11It's uh it's being overused, but uh this 4:13has been going on for a couple of years. 4:15I I think it's starting now. Yeah, I 4:18think it sort of started at the end of 4:2024 a little bit and then they were like, 4:22"Oh, 2025 is going to be the year of the 4:24agent." Turns out that that was kind of 4:26true. And that's this layer here, right? 4:29That's level two. And the next two we 4:32could talk about later if you want to 4:33hit me up afterwards, but they're less 4:35human focused in terms of how much of 4:37the work is being done. And this system 4:40that I'm going to show you is firmly 4:42right here in the uh level two. So what 4:45can you do to get started? So the most 4:47important thing is understanding like 4:49you can't just start building, right? A 4:51lot of people try to just start building 4:53random stuff inside of a system like 4:54this and then they try a few things, 4:56they don't really work that well and 4:58they kind of abandon them. And I I 5:00taught a course on this four times so 5:02far and I always start with the same 5:04thing, which is who are you? What do you 5:06care about? What do you actually want to 5:08get good at? And how can technology save 5:11you time so you can actually do more of 5:12the stuff you care about and less of the 5:14stuff that's just like busy work. So for 5:16me, this is roughly what mine looks 5:18like. One through four is definitely 5:20like the most important to me. Reading, 5:22thinking, writing, and discussing 5:24things. Five is what I've been spending 5:26most of my time on, which is building. 5:28This used to be known as coding, by the 5:30way. I've stopped saying like when I 5:32talk to Clint or whatever, what have you 5:33been doing? Uh, coding, right? Cuz 5:35that's what I was doing before, and now 5:37it's like I'm actually thinking and 5:39writing, which is producing the coding. 5:41It's a weird abstraction. I also do a 5:43lot of consulting for customers, 5:44building a bunch of products. For 5:46physical stuff, I play table tennis, 5:48like the crazy kind where you're really 5:50far away from the table. I play drums 5:52and I'm getting into kickboxing. And 5:54perhaps most importantly, I orient a lot 5:56of my life around trying to help other 5:58people do the same thing that I'm I'm 6:01doing here, which is just self-discovery 6:03and like self-magnification. 6:05So, the system itself and its design. 6:08So, these are the principles that I 6:10built Kai on top of and what sets it 6:12apart from cloud code. Cloud code has 6:15some of this naturally built in, but 6:17I've kind of augmented it for the Kai 6:19system to be a lot more of of these 6:21things with a heavy focus on on these 6:23principles. So, what we're going to do 6:25is go one by one through these and talk 6:28about them as a concept and show you 6:29what it looks like inside the system. 6:31So, they go roughly in order of 6:33importance. And this is definitely the 6:35most important one. Even though it's 6:37kind of invisible, it it's become less 6:39talked about. So basically, if you 6:41remember back to early 24, I don't know, 6:43maybe in 23 as well, it's it's hard to 6:46actually remember because it flows 6:47together, but a lot of companies were 6:49talking about prompting. Oh, you got to 6:50learn how to prompt. Prompt engineering 6:52was the big word, right? Put out the 6:54fabric project, which is a whole bunch 6:55of crowdsourced prompts, and that was 6:58just kind of the thing that you did. AI 7:00was heavily associated with prompting. 7:02And in my opinion, prompting never 7:04became less important. In fact, I think 7:06it's more important than ever now. I 7:08think it's just more hidden because 7:10there's lots of other shiny things that 7:11people are talking about. The other 7:13reason is because AI is doing a lot of 7:15that prompting for us. The way I think 7:16about this is clear thinking is 7:19basically the center of everything and 7:21clear thinking becomes clear writing and 7:23then clear writing is essentially what 7:25prompting is and that that becomes 7:27really good AI. So a good heristic for 7:29this is like can you explain this to 7:31yourself especially to yourself like 6 7:33months later when you might not have 7:35remembered what you actually built. and 7:37you explain it to others and if you 7:38can't then AI really can't understand it 7:41either right when the AI is confused 7:43that's when everything goes sideways so 7:45I think prompting is still like the most 7:47important thing to AI cuz at the end of 7:49the day it's all language it's all 7:51instructions so it's all prompting so 7:53what this actually means in practice is 7:55I've spent many thousands of hours at 7:58this point working on my whole structure 7:59I don't know why this thing is 7 gigs I 8:01actually took it down from 10 gigs but 8:03that's a lot of text that's not actually 8:05the scaffolding cuz that would be a 8:07nightmare and nothing would be able to 8:09parse it. So there's a lot of outputs 8:11and other stuff, but the system is 8:12definitely growing. And the most 8:14important directories inside of here are 8:16things like skills and hooks and 8:17history, which we're going to talk 8:18about. So this leads really well into 8:21the next one, which is the scaffolding 8:23is more important in my opinion than the 8:25model. Now, there's some exceptions to 8:27this, and all the news, of course, is 8:29about the new models when they get 8:31released. you know, Gemini 3 recently, 8:33Opus 4.5 recently, but I've always been 8:36team scaffolding, and I I continue to be 8:39team scaffolding. Obviously, it's it's 8:41best to have both, right? If you have 8:42really good models, it magnifies the 8:44scaffolding, and if you have really good 8:46scaffolding, it magnifies the models. 8:48But if I had to choose between the 8:50latest model with not very good 8:52scaffolding or excellent scaffolding 8:54with a model from 6 months ago or even a 8:56year ago or even 18 months ago, 8:58honestly, I would definitely pick the 9:00ladder. And a quick quick note on this. 9:02Yeah. So I totally agree with Daniel 9:03about the scaffolding one thing. So I 9:05spent a lot of time reading and thinking 9:07about people who are using AI for 9:09vulnerability detection for example like 9:11analyzing source code or scanning 9:13running systems and I think it's 9:14difficult to say like a given model can 9:16or can't do this because again to 9:19Daniel's point like the scaffolding 9:20around it makes such a huge impact like 9:22if you see OpenAI's Arvar or deep sleep 9:25or codemen from Google basically they 9:27are giving all these sorts of tools and 9:29orchestration layers like around the 9:31models which allow them to perform 9:33orders of magnitude better on like the 9:35same task. So when someone says this 9:37current model can or can't do something, 9:39it's kind of actually hard to know that 9:40for sure given just better orchestration 9:43and context management and things like 9:45that can cause meaningfully different 9:47outcomes. So I think scaffolding is 9:49huge. So yeah, that I think that's a key 9:51point. So I just wanted to like 9:52emphasize that. Okay. Yeah, back to you. 9:54>> Yeah, I think those are all good points. 9:56Another one is the AIXCC competition. 9:59Trail of Bits really crushed it there. 10:00They they did a lot of scaffolding 10:02there. The Atlanta team as well. So good 10:04examples of that. So uh my skills 10:06directory is probably the most important 10:07center of the scaffolding. Currently 10:09have like 65 skills in here. A lot of 10:12these are just very uh pointed at my own 10:14stuff that not really useful to other 10:16people, but a number of these are just 10:18core and just essential to everything I 10:20do. Next one of is the last of my three 10:23central philosophies for the system. And 10:25this one is to be as deterministic as 10:28possible. Right? So what does that mean? 10:29It means in practice basically code 10:31before prompts is what that really turns 10:34into. If I have anything that I can do 10:35in code, I do it in code first. I don't 10:38even use AI at all. And if you think 10:40about it this way, Kai is more like a 10:43tech orchestration system. Not really an 10:45AI system cuz I had something kind of 10:48similar before before even AI obviously 10:51wasn't as good. But when you add AI on 10:53top of it, it really magnifies it. But 10:54it's more of an orchestration framework. 10:56This is my art skill for example which 10:58we're going to talk about more and the 11:00tools directory within that skill and 11:03this is just deterministic regular code 11:05right anything that my art skill does is 11:07actually running code at the end of it 11:09right at the end of the day it's 11:11actually just deterministic code and 11:12that provides as much consistency and 11:14control as possible and also has the 11:16upside of not involving AI at all for 11:18the step which saves a bunch of tokens 11:20and usage and everything. This one I've 11:22broken out just because it's so 11:24important, but really it's the same 11:25thing we just talked about. Code before 11:27prompts. I'm not really sure what my 11:29current percentage of this is, like 80 11:31versus 20. I think I'd like I don't 11:33know. Curious what you think, Clint, 11:35what this balance should be like the 11:37ideal balance, but I feel like it should 11:39be mostly deterministic with like AI 11:42wrapping it 8020. I don't know. I'll 11:44have to do more thinking on that and see 11:46how it plays out. Yeah, I was just going 11:47to say I think if there's something that 11:49can be done programmatically 11:50deterministically, I think code is the 11:52right solution for it because it's like 11:54cheaper and you know you're going to get 11:55the answer you expect. In the past 11:57creating that code maybe has been time 11:59or cost prohibitive but now with like 12:01cloud code and other similar coding 12:03agents like creating the deterministic 12:05code can be done in a fraction of the 12:06time. So yeah, I think it depends on the 12:09domain. When you need like a fuzzy 12:10answer that's sort of difficult to solve 12:12generically completely 12:13deterministically, then maybe sort of 12:16some sort of prompt and code system is 12:18better. But yeah, I think the core 12:19intuition here is like if you can solve 12:20it deterministically, like probably 12:22that's the better solution and maybe you 12:24vibe your way to like code that does it. 12:26But yeah, trying to solve everything 12:27with prompts, at least as of today, is 12:29going to be inefficient, costly, and 12:31like if you were like find all the 12:32routes in this repo, you're going to 12:34find a lot of them, but cloud code or 12:35whatever system is going to miss some of 12:36them. So, if you can do that with like 12:38another tool that you know is going to 12:40work, it's just better. But yeah, you 12:41were saying about Anthropic. 12:42>> Yeah, Anthropic came out with a thing 12:44about this actually kind of throwing 12:46shade at their own MCP. like they 12:48invented MCP and they're like, "Hey, you 12:51might want to actually just do this in 12:52Typescript and use the MCP to get the 12:55service that you want to use, turn that 12:57into TypeScript and actually run that 12:59instead cuz then you're not calling all 13:01these uh tokens before and after. You're 13:03just getting the results and then you 13:04could use the results to give to AI." 13:06So, I thought it was really cool and 13:08within like a little while of them 13:10releasing that, I upgraded a couple of 13:12my MCPs to not use MCP anymore. So this 13:15next one is specifications, tests and 13:17evals. And this is also playing at the 13:20whole concept of determinism and 13:22consistency. So there's a big tendency 13:24in AI to use vibes. This is just like a 13:27gentic everything is vibes. Vibe 13:28hacking. Now vibe marketing, of course, 13:30vibe coding. Big thing Clint and I 13:32actually talk about a lot is how do we 13:35know any of this is working, right? How 13:37do you actually test any of this stuff? 13:39How do we actually get consistency from 13:41what we build? This is the skill.mmd for 13:44my development skill. And you can see 13:46I'm starting with specd driven 13:47development. And this is roughly based 13:49off of GitHub's really excellent project 13:52called spec kit. And I basically 13:54simplified it because it was it was a 13:56little bit too involved. But um first 13:58you create specs, then you create plans, 14:00then you write tests, then you write 14:01code. And that's the flow that uh that 14:03system uses. And I'm always optimizing 14:06this, but this is the general flow that 14:07I follow. So, the next one I've been 14:09obsessed with since being in college in 14:11like the 1990s. And uh shout out to my 14:14buddy Kundi who uh showed me the command 14:16line for the first time, which honestly 14:18it might have been one of the best days 14:19of my life when I found out I could pipe 14:21the output of one command into the input 14:23of the next. I mean, it truly tripped me 14:25out. And I feel like I've been building 14:27systems based on that concept ever 14:29since. So the way it materializes inside 14:31of this system is I try to have each 14:34container do one thing well and I build 14:36different skills to call each other 14:37instead of replicating that 14:39functionality inside of each one. I've 14:41got a number of examples of this within 14:43Kai. But this is a red team skill and 14:45this red team can actually hit a network 14:47architecture, an application 14:49architecture, threat modeling. It could 14:51do like all sorts of different stuff. 14:53But I often use it to attack ideas that 14:55I have to see like blind spots if I'm 14:58missing something. But the red team 14:59skill calls a first principal skill and 15:02breaks that down further into other 15:04pieces, right? So it works in a flow to 15:06really break open ideas and attack the 15:08ideas. This is another really cool 15:10example which is called lifelog pulls 15:13off of my uh necklace pendant which I'm 15:15wearing right now. I I could show when I 15:17turn on the camera. It's basically a 15:18thing that I can turn on when I'm 15:20walking and I could say, "Hey, new idea 15:22or new blog or yeah, I should do a piece 15:24of content on this or whatever." And 15:26then I just talk and it captures it and 15:28then when I get back I could say okay 15:30that thing I just said when I was 15:31walking take this and do that with it. 15:33Right? So it goes and pulls the content 15:36from the transcript, pulls out the 15:37section, summarizes it from there. I 15:39could do research on it. I could blog on 15:41it. I could do whatever. I could red 15:42team the idea for example and it's all 15:44done using natural language prompt. Uh 15:47just me talking to Kai and it cross 15:49calls all those different ones. I also 15:51want to mention real quick custom slash 15:53commands. These are commands you could 15:54just type forward slash include code to 15:56run. And these are also calling one or 16:00multiple. The cse one is calling create 16:03story explanation which is a skill and 16:05it's just another way to call into 16:07skills as well worth mentioning. So if I 16:09take a piece of content I want to get 16:11information from like this is a great 16:12article on tlddrc and this was from 16:15Jason Chan who built this security 16:17program at Netflix. So you could take 16:19that and you could do like this for/cse5 16:22which will give me five levels of 16:24explanation of what this thing is about. 16:26And this is what it does when it's 16:28working. It actually uses fabric here 16:30which I forgot about that but it uses 16:32fabric switch u which goes actually uses 16:35Gina AI to do this and it pulls down the 16:37markdown for the thing and then it runs 16:39the cse skill and it returns the results 16:42in five levels. And cool little thing 16:44for Kai is Kai also updates the tab name 16:48inside of the Kitty terminal to be the 16:50result and reads it in Kai's voice which 16:53we'll see later. Next one is engineering 16:55or S sur principles which is also part 16:58of this determinism story which I feel 17:01like it's I might move that up to be 17:03like the most important one. But my 17:04background is hacking in like the 17:07security sense, but also in the more 17:09pure sense of like building and creating 17:12and breaking things, right? So, I've 17:14been a crappy programmer since the early 17:172000s, but I've never been an actual 17:19software engineer, right? And there's a 17:21huge difference, as anyone who's been 17:23both knows, there's a huge difference 17:25between an SWE and someone who programs. 17:29So what I'm doing now is I'm learning a 17:31lot of engineering stuff that I should 17:32have learned in college and trying to 17:34build that into the DNA of the system 17:36which usually manifests as tests and 17:39evals and stuff like that. The way it 17:41mostly manifests is through like the 17:43development skill where I'm going 17:45through you know true tested engineering 17:47practices of like building plans, 17:50test-driven development and all that 17:51sort of thing. This is a thing I talk to 17:53Clint a lot about. Most people don't 17:55know this because he never talks about 17:57it, but Clint Loki is a PhD in computer 18:00science, and it definitely shows when he 18:02starts talking about tests and evals. 18:05You will see him light up like a 18:06Christmas tree. So, that's always fun. 18:08>> For context, a lot of Daniel's skills 18:10and prompts and things like that. Some 18:12of them have like this detailed 18:13backstory about maybe the person's 18:15persona and their goals in life. And uh 18:17I'm like, Daniel, does that does that 18:19help? And he's like, you know, vibes, 18:20baby. But then also like test them 18:23rigorously in practice to make sure that 18:24they work consistently. 18:26>> Yeah. And I've got some examples of that 18:27when we talk about the voice system. I 18:29put all this effort into it and I'm 18:31always wondering myself and then 18:33especially when I talk to Clint, I'm 18:34like, "Okay, what if I had this same 18:37exact prompt without all this 18:38personality stuff?" Yeah. So, we're 18:40we're doing a bunch of eval stuff so we 18:42can test this stuff at scale. This one 18:44is super cool. This is a relatively new 18:46addition to the Kai system. So, not only 18:48am I trying to write code for as much as 18:50possible in the system, but I'm actually 18:52trying to have that be executed via CLI 18:55instead of just calling the code and 18:57having the model try to figure out how 18:59to actually run it. So, I love the 19:01command line so much. Terminal is my 19:03favorite place to live. And I love the 19:05fact that there's documentation, there's 19:07flags, there's switches, there's 19:09options, right? And it means you know 19:11how to use it. You know how to use a 19:13command line by running the help 19:14command. And you know who else loves 19:16that? AI loves that. AI absolutely loves 19:19when it doesn't have ambiguity in what 19:20it's supposed to do. So, going all the 19:22way back to the concept of clarity and 19:25AI not being confused, like there's 19:27nothing more clear than how to use a 19:28command line tool, assuming it's well 19:30documented. So, I've got a command line 19:32tool for launching Kai. Actually, when I 19:34type K, that used to just be a ZSH 19:37alias, but now I've got a actual command 19:39line tool for it. And the most useful 19:41switches I I think are actually the uh 19:43switch M switch to dynamically load 19:46MCPS. And uh shout out to Indie Dev Dan 19:49for this one. He's a bull developer 19:51who's doing AI stuff on YouTube. You 19:53should definitely check him out. He also 19:55did this and I was like, "Oh, that's a 19:57great idea." So I built the command line 19:59to be able to do that. And here's the 20:00actual command for generating images 20:03using my art skill. So I can pass in a 20:05model, but the default is actually nano 20:08banana pro for obvious reasons. It's 20:10just incredibly good. But I have all 20:12these different options. And this is 20:13what Kai actually uses to generate 20:15images. So this next one is a highle 20:17flow for a concept that solidifies a lot 20:20of what we've been talking about. It's 20:22just a way of thinking about how to 20:24organize the entire system. Basically, 20:26you figure out what you want to do. You 20:27figure out if you can do it in code. 20:29Then if I can, I build a command line 20:31tool around that. Then I use prompting 20:33to run the command line tool. And then I 20:35use skills or agents to call it or to 20:39run it in parallel. And that's that's 20:41kind of the flow and the structure. And 20:42this is basically how all these skills 20:44work is going from the top level goal 20:47all the way down to the codebased 20:49implementation. This next one is super 20:50fun. It is also super useful. Basically, 20:53I have a whole bunch of capabilities 20:55within Kai that are used to update Kai 20:58himself, right? So it's like 20:59self-update, self-healing, 21:01self-improvement, and not just like a 21:03little component or a module or 21:05something, but like the system overall. 21:07So you've seen all different components 21:09of the scaffolding. We have skills, we 21:11have workflows within the skills that 21:13execute things. We have code in the 21:15command line tools. Then we have the 21:17models, right? We also have different 21:19services that could be called via MCP or 21:22API or whatever. So the best example of 21:24this is I have an upgrade skill. I guess 21:27that's a good name for it if it's doing 21:28upgrades. So the upgrade skill, it's a 21:30universal skill that multiple sources, 21:33it hits multiple sources on the internet 21:35and I'm looking at those sources because 21:37they're constantly releasing stuff that 21:40I cannot keep up with manually and I 21:42don't want Kai to get behind. So when I 21:43say run the upgrade skill, it will go 21:45and find all these different sources. 21:47It'll parse the latest content. It will 21:49review all of Kai's documentation. 21:52There's a single file that documents all 21:54of Kai. So Kai will read that, 21:56understand how he works, and then 21:58understand all the updates that it just 22:00pulled from different sources, and then 22:02look for opportunities to improve. So 22:04one of the sources it looks at is the 22:06anthropic engineering blogs, all of 22:09their releases on GitHub. I mean, 22:10they're releasing stuff constantly, like 22:12every day, multiple times during the 22:14day. There's no way I could possibly 22:16follow it all. I also do this for 22:18YouTube channels. Somebody talks about a 22:19new technique, automatically parse it 22:21and bring it in. And security-wise, I do 22:23it for security research as well. So 22:25like all the talks that Clint puts out 22:27when he puts out like, oh, here's the 22:28latest videos from so and so conference 22:30or whatever, I parse those and update my 22:33testing methodology if there's like a 22:35new technique. So here's an actual 22:37example of me running this a little 22:39while ago. So Anthropic had a release 22:41saying how they could improve uh routing 22:43within skills using this keyword use 22:46when. and they basically emphasized, 22:48hey, look, you need to be using this 22:50because if you're not getting your 22:51skills to function the way that you 22:53want, they're not being triggered 22:54properly, you need to make sure you have 22:55this in the front matter. So Kai ran the 22:58upgrade skill, found this, and came back 23:00with this as the top recommendation. I 23:02said, "Okay, do it." And within like 5 23:04minutes, the entire Kai system was 23:06upgraded with this uh piece of 23:08functionality. And after that, skills 23:09worked way better. So, as I was making 23:11this like prepping for the conversation 23:13here, I said look for in our history for 23:16learnings because everything that I do 23:18for an upgrade actually across the 23:20system, it's all captured in the history 23:22system and it's broken down into 23:23structures which we'll actually see a 23:25little bit later. But this basically 23:27looked up our archive of things that 23:28we've done to learn over time. And 23:30that's exactly what it found. It found 23:32this use win thing. All right. So, this 23:34one is absolutely crucial. I should 23:36probably raise it in priority as well. 23:39custom skill management. So this is 23:41probably the most significant thing that 23:43I did on top of cloud code. It it's just 23:45a completely different structure. So 23:46skills are already really good inside of 23:49cloud code. It's pretty good at routing 23:50by itself, but what I did was add a 23:52supplemental system that's more explicit 23:54about the routing. And I'm getting like 23:56I don't know 95 98%. I don't know the 23:58actual number. Me and Clint will have to 24:00figure out the actual number. We need uh 24:02you know rigor around this thing. but 24:03it's basically routing in the system 24:05prompt which goes to a routing table of 24:08workflows that go to specific prompts 24:10and then within the directory there's 24:11also a tools directory that the 24:13workflows actually call and like we 24:15talked about those are ideally 24:17deterministic code as opposed to 24:18prompts. So, what this does is it 24:20produces way better results for being 24:22able to do multiple things inside of a 24:24category such as art. And I'm going to 24:27show the art skill later, but what it 24:29does is allows me to just speak in plain 24:31language and get exactly pretty much 24:33what I asked for. Last couple here, 24:35custom history system. We touched on 24:37this one a little bit, but basically I 24:39have sessions, learnings, research 24:41decisions, all sorts of different 24:43categories here. And when we get done 24:45doing anything, if any agent does 24:47anything, if I do it, if Kai does it, if 24:50any sub agent does it, Kai thinks about 24:52what we did, turns that into a summary 24:54and writes it into this history system. 24:56That's part of the reason I've got like 24:58six gigs of stuff grown over time here. 25:00But file system is cheap and file system 25:02is fast. So I I like this way better 25:06than rag for most things. This is what 25:08the directory actually looks like. And 25:10it basically allows me to understand 25:12where we've been, where we are, and 25:14where we're going, right? Cuz I hate 25:15making the same mistake over and over. I 25:17especially have a system like this for 25:19bugs, calling out bugs that keep coming 25:22up when I'm building web applications. 25:24So I could say, capture a learning on 25:26this, capture a learning on this, and 25:28have that be crystallized into a thing 25:30that Kai can then use to upgrade the 25:33development skill. So we don't do that 25:34anymore. And this last one is kind of 25:37flourish to be honest, but it can be 25:39useful as well. I basically have a fully 25:41customized voice system for Kai and all 25:44the different agents that we use within 25:45it. So we've got architects, engineers, 25:48researchers, QA testers, interns. They 25:50all have different personalities. This 25:52is the part that I'm not sure how much 25:53this is actually helping, but it's fun. 25:55And they all have different approaches 25:57to their work, too, right? Some are like 25:59library scientists, some are like super 26:01curious, and they have different voice 26:02characteristics based on their 26:04personalities. So, what this means is 26:06while the agents are talking to me or to 26:08each other, you can actually hear 26:10emotion in what they're talking about. 26:11If they come back with a finding and 26:13they're excited about it, like you can 26:15actually hear that in the voice 26:16rendition. I still need to do more evals 26:18on this uh as I talked about. So, the 26:21whole thing goes through a voice server, 26:22which we have, which is part of the PI 26:24system. It goes through 11 Labs API. 26:27then it gets read out on the local 26:29system in the particular voice of that 26:31particular agent. And uh it's mostly 26:33just pretty cool like I said, but I do 26:35get use out of it because if I'm 26:37building like 20 different things off on 26:39the side in their own sessions, they can 26:41be reporting back with what they did and 26:43I could actually tell by their voices 26:45who they are and plus they give me a 26:46summary of what they've done. And these 26:48are all the different agents that we 26:49have. Like I said, we got engineers, pen 26:51testers, all kinds of different folks in 26:53here. So that is the overview of the 26:56system and I just want to reiterate that 26:58the whole point of all of this is to go 27:00from the left to the right here. A human 27:02on the left who has human interests and 27:05human goals using a system that does all 27:07this different stuff with tech and AI or 27:10whatever with again the output to be 27:12human outcomes that help humans. So it's 27:15humans tech in the middle. It's not 27:17important and then humans on the right 27:18side as well. So I want to do a deep 27:20dive on a real skill or at least just 27:23show it live in a quick demo. All right. 27:26So first we're going to invoke Kai. 27:28>> Kai here ready to go. 27:29>> All right. So KS is to go to the skills 27:32directory. Kai skills. See the art. And 27:35inside of here we have that level. We 27:38have that structure. We have workflows 27:39and we have tools. So if we go inside of 27:42workflows, this is what that looks like. 27:44We've got annotated screenshots. We've 27:47got apherisms. I haven't even tried that 27:48one. Comics. This is like an XKCD type 27:51thing, although I didn't use the Randall 27:53Monroe style because I thought that was 27:55kind of rude. Comparisons essay. Essay I 27:58use all the time. It's actually what 28:00produces art for my site now. 28:02Frameworks.md is actually for talking 28:04about like architecture frameworks. 28:06Maps, mermaid diagrams. This one makes 28:08it look like mermaid. Stats, taxonomies, 28:11technical diagrams is another one that I 28:13use all the time. timelines and 28:15visualize is actually just if I say 28:17visualize so and so and I give Kai input 28:20he will decide which of these to use or 28:22which combination of these to use and 28:24then if we go into the tools directory 28:26we can see these here I think I showed 28:29one of these already so no need to show 28:31that but what I want to do 28:33>> hi here ready to go 28:34>> let's have somebody give a concept 28:36somebody from chat to give a concept 28:38>> okay we have a couple purple dogs skiing 28:40down the mountain ghost in the machine 28:42style timeline horizon for The AI 28:44revolution. Trees are younger than 28:46sharks. Adopting UTCP. Pig with a 28:50jetpack. Longing for my lost youth. Do 28:52any of those AI and humans living 28:54together? 28:54>> A human story arc. I like that one. And 28:57I love the uh the trees are younger than 29:00sharks. That blew me away. Sadly, I 29:03learned that only like two years ago. I 29:05had to Google multiple places cuz I 29:06thought it was fake news. Okay, I want 29:08you to visualize a human story arc. So 29:11basically all the way from I guess 29:13coming out of the ocean assuming you 29:15believe in that sort of thing and then 29:17moving through different stages of 29:18development agriculture I don't know we 29:21could talk about warfare we could talk 29:22about science we could talk about 29:24whatever you want to put in here but 29:26basically just show an arc and actually 29:28show where we're currently at and then 29:30move beyond that as well and show other 29:32parts like the AI transformation or what 29:35comes after that. All right. So, that 29:37was captured using dictation, which is 29:39how I do 99% of everything now. And 29:42we're going to see what Kai comes up 29:43with. 29:43>> Yeah. Shout out Whisper Flow. 29:45>> Absolutely. I think I'm at I'm about to 29:47cross 600,000 words for Whisper Flow. 29:50>> Does Whisper Flow have different tiers 29:52like you are a certified yapper or uh 29:55>> Yes, it does. You can actually set if 29:57you want to do formal. I always have 29:59mine set to formal. I like to capitalize 30:01the first word and use a period rather 30:03than a casual one is all lowercase all 30:05the time. While this is working, Daniel, 30:07a couple of questions that have popped 30:08up a couple times. Um, people curious 30:10about the um overall cost of this. So 30:13maybe the like cloud code costs for 30:15running Kai. Obviously, you're using 30:18Whisper Flow and various other services 30:20like Nano Banana from Google for 30:21generating images. Yeah, I'm curious 30:23maybe if you could concisely say like 30:25here's the main tech stack and like 30:26approximate maybe monthly costs. 30:28>> Yeah. Yeah, it's a great question. So 30:30first of all, Kai uses lots of different 30:32AIs inside. So he's calling often times 30:35Google stuff like this art stuff is 30:38going to use Google. I use openAI stuff 30:40sometimes but these are call outs from 30:42the base which is cloud code. On cloud 30:44code I'm using the maximum maximum which 30:47is $200 a month. So I would say that my 30:50usual cost for using Kai is less than 30:54$250 per month. 30:56>> Cool. That makes sense. Yeah. Thanks for 30:57sharing. 30:58>> All right. So look at that bun run. By 31:00the way, Anthropic just bought bun, 31:02which I'm super excited about. But you 31:03see, now he's using the art tool, the 31:06generate.ts, which is the command line 31:08tool to generate using Nano Banano Pro. 31:11Any other questions? Yeah, $250 31:13absolutely is cheap. Hey, what's up, 31:15Rowan? Oh, yeah. With 11 Labs, that's a 31:17point a good point. I didn't add that 31:18on. Maybe that's like another $20. I 31:21think that's also pretty cheap. I have 31:23only hit my claw limit once and it was 31:26actually a couple of days ago when I was 31:28going absolutely nuts for the entire 31:29weekend and then it switches over to 31:31using a key instead of using the 31:33subscription and that will hurt you very 31:35quickly if you're not inside the 31:37subscription. I think I worked for maybe 31:39an hour and it cost me I think like $70. 31:42Yeah, I think uh something Aaron 31:44mentioned in the chat, but I think 31:45thinking as like a person, you're like, 31:47"Yeah, 250 to 300 per month is like a 31:50lot, but if you're thinking of it from 31:52like a business cost point of view, and 31:53you're like, well, how much more output 31:55do I have? How many more things do I 31:56accomplish? Am I doing things that might 31:58lead to additional revenue or things 32:00like that? So, for example, you do a lot 32:02of security consulting. So, if you're 32:03like, yes, this costs me x amount of 32:05money, but then I deliver work, which I 32:08charge like this much for." So, it's 32:09like you're still making more money than 32:10it is costing. So I think that's also a 32:12useful frame. 32:13>> Yeah, completely agree. This is pretty 32:15interesting. It's kind of bunched up on 32:17the side over here. This is definitely 32:18not using my technical diagram. Are we 32:21making multiple? Cool. Oh, when I said 32:24visualize, I also have a visualization 32:26skill and that produces uh D3 graphs. So 32:29I think that's what he's trying to do 32:30there. Yeah, this one's pretty 32:31interesting. I'm actually curious. I 32:33want to I want to do a different one. 32:34>> Hi here. Ready to go. 32:36>> Use uh the technical diagram workflow to 32:38make this one inside of the art skill. I 32:40I think that would be a pretty cool 32:42visual. I put a lot of effort into that 32:44workflow recently. So I want to see how 32:46he does there. Oh, I forgot to show 32:47something. This is how I see what Kai is 32:50actually working on. So this Daniel 32:52here, you can see my prompt and you can 32:54see that Kai is working on a thing. We 32:56got the stages that it's doing AI 32:59transformation, current post AGI, cosmic 33:01unknown, future. So that's how it broke 33:03down the different pieces to visualize. 33:05And then yeah, this is my observability 33:08system for watching what the different 33:10agents are doing. This is a custom 33:11interface that I built. Obviously, I 33:13wouldn't release because it's too cool 33:15except for it's already released. It's 33:17already in the project. 33:18>> Yeah. Is that in the PI repo? 33:20>> It is. 33:20>> Oh, I actually didn't know that till 33:22just now. Awesome. 33:22>> Yeah. And I just updated it. So, it has 33:24all this new UI stuff. And so, check 33:26this out. Also, a UI workflow. It shows 33:28the activities of what it's working on. 33:30It shows the different skills, shows the 33:32different tools that are being used. It 33:34actually obuscates uh if there's keys 33:36being used, it won't show the keys in 33:38the flow as well, just in case I'm uh 33:41doing a YouTube video or a webinar. 33:43>> So, a bunch of people both in the normal 33:45chat as well as the Q&A are asking about 33:47like costs. You already talked about 33:49that a little bit, but I guess I just 33:51wanted to convey if you wanted to do a 33:52blog post on like a deeper dive into 33:55like example monthly costs for like 33:56specific services and workflow and stuff 33:58like that. Seems like there were five to 34:0010 questions about that. Maybe more. 34:02>> Yeah, that makes sense. I could 34:03definitely do that. 34:04>> And then there were broadly some 34:06questions about Emmy asked about sort of 34:09like the owning your own AI in terms of 34:11investing in GPUs to run local models. 34:14You know, should you buy your own 34:16hardware and run local models versus 34:17relying on third party providers? Other 34:20related questions I would say are like, 34:21well, could you do the same approach 34:23using like codecs or Gemini CLI? And 34:25what made you choose cloud code versus 34:27like open code? There's like a couple of 34:29questions around which models and like 34:32tooling to use and why. 34:33>> Yeah, great question. So, Cloud Code in 34:35my opinion has the best scaffolding. We 34:37already talked about the scaffolding and 34:38how important it is, but it's just to me 34:40vastly superior to any of the other 34:42options in terms of like an overall 34:44infrastructure to work from within. The 34:47way that I call cuz I saw a number of 34:48people ask that question. The way that I 34:51call is via command line. So, Gemini is 34:53a tool. Gemini is a command line tool. 34:55Codeex is a command line tool. For 34:57example, I have a Gemini researcher. 34:59Gemini researcher uses Gemini and then 35:02it uses my Google account to do a deep 35:04research using Google deep research 35:07because I want that to be a whole 35:08separate ecosystem of basically how 35:11Google thinks about search and how it 35:13has its own structure and biases and 35:14stuff like that. I have a Grock one. Oh, 35:17nano banana is down. That would explain 35:19it. That's why you pre-record demos 35:21>> while you're waiting. Question from John 35:24Roberts. So if you're not CLI savvy or a 35:27programmer, how far can you get? Clear 35:29thinking and writing doesn't require a 35:30CLI but the next steps do code before a 35:32prompt for instance. I guess broadly 35:34advice for people who are not CLI savvy 35:38or a programmer. 35:39>> Yeah, I would say I mean I'm not writing 35:41these CLI tools. Kai is writing most of 35:43this code. I have the ability to go into 35:46the code and fix it or change it or 35:48modify it. But more and more I am using 35:51my interaction with Kai to change the 35:54code and I'm using my scaffolding to 35:56control what code gets made in what way. 35:59But I'm not over here spending time 36:01writing code. That would massively slow 36:03me down. Honestly, it would be worse 36:04code compared to Kai writing code that 36:07is controlled by me and my scaffolding. 36:09So I would say you should not be 36:11intimidated. You should not be 36:12intimidated by I'm not a coder or 36:14whatever. The one thing I'm trying to do 36:16with PI is make it clear. You can be 36:18completely non-technical because again, 36:20I come in here and and I I could 36:22basically say anything that doesn't 36:24involve a model that's currently down. 36:27So, let's create a human arc story using 36:30one of the other models that you have 36:32access to using the technical diagram. I 36:35can't remember. I think I have Replet in 36:37there with a few models. So, look at 36:39this. Over here it shows that the art 36:41skill is running and over here it says 36:42the art skill is running. And then in 36:44the tab up here, it says human arc story 36:47creation, which is also in the visual. 36:49But just to complete that point, do not 36:51be intimidated. The whole point of PI is 36:53that you should be able to be 36:54non-technical and come in here and do 36:56all this stuff. Yeah. So you can see he 36:58can pivot to using a open AI model. I 37:01haven't used that one. I don't know if 37:02the key is good. 37:03>> One thing I was curious about though, 37:04Vashall asked, what controls or guard 37:07rails do you have in place to protect 37:09Kai from performing malicious 37:11activities? Especially because you're 37:12like, you know, scraping websites which 37:14could have arbitrary content. You have a 37:15lot of surface area to potential prompt 37:17injection, for example. 37:18>> Yep. So, that one I'm not going to talk 37:20too much about just because I I don't 37:22want to talk about the controls that 37:24much, but I've basically got four to 37:26five layers of different defenses uh put 37:28in there to to block that kind of stuff. 37:30A lot of it involves 37:31>> human story arc diagram generated with 37:33GPT image one open for review. 37:35>> Yeah, so this one is not bad. I mean, I 37:37like this. It's nowhere near as good as 37:39a Nano Banana Pro in my opinion, but 37:41still looks pretty cool. To finish 37:42answering that question, a lot of Kai 37:44understanding what our purpose is, what 37:46Kai's purpose is for us in the 37:48ecosystem, and then recognizing things 37:51that are attempting to hijack that 37:52purpose. So, a lot of things around 37:54prompt injection. I'm also using the 37:56anthropic controls for different types 37:58of control of tools, what it can and 38:00cannot call, especially in conjunction 38:02with a request, which is out of 38:05character. So, I've got a bunch of that 38:06stuff built in. I guarantee that I could 38:09probably break it and a lot of other AI 38:11oriented pen testers could probably 38:13break it as well. I would guess it's 38:14probably an 85 to 95% decent defense and 38:19then that's the type of thing I just 38:20keep improving as well. 38:21>> For some additional thoughts on that, I 38:23think anthropic shipped like a cloud 38:25code sandbox that tries to sandbox 38:28things by default and then to your 38:30point, you have different sub agents for 38:32different tasks. So one may be like a 38:33researcher or something that is talking 38:35to the internet whether that's like 38:36YouTube transcript or scraping from a 38:38web page or things like that and the 38:40capabilities that you give that agent 38:42could be mostly like readonlyish type 38:45things and then maybe you just write a 38:47summary of like the plan of what you're 38:49going to implement or a distillation of 38:51whatever you read and then what actually 38:53implements or makes file system changes 38:55or has access to execution or things 38:57like that on your computer could be like 38:59a separate agent that just like reads 39:01whatever was written by the previous 39:02step which you could as a human read 39:04first before just sort of blindly 39:07executing. So, so basically similar to 39:09um Simon Willis's blog post about the I 39:11forget what he exactly called it like 39:13the dangerous or the critical three. 39:15Basically having separation between 39:16external sources that you don't control 39:18and then like executing things locally. 39:20You could have like a human review uh in 39:22the middle. 39:22>> Yeah, I I think that's that's all very 39:25smart. 39:25>> Yeah, please leave a little trifecta. 39:27Thank you everyone in the chat. So, I'm 39:29sure everyone or mostly everyone 39:30probably already guessed this, but this 39:32was the actual prompt that was used to 39:34generate all the art for the entire 39:37presentation. And this is using the 39:39technical diagram workflow. Yeah. All 39:41that stuff that you've been seeing. Kai 39:42made all that stuff. So, getting started 39:44on your own. When I say that like the 39:47purpose of all this for me is to really 39:48enable people to be the best versions of 39:50themselves and like magnify themselves, 39:53like I'm actually quite serious. So, the 39:55whole thing it's all open sourced. 39:57There's a whole bunch of skills that 39:58aren't migrated over. Most of them are 40:01just complete garbage to anyone else. 40:03There's a few that are good for other 40:05people that I just have to be careful 40:07about how I move them over. Talk about 40:08stress. Every time I push from Kai to 40:11Pi, it's like one of the most stressful 40:13things because I've got, you know, 40:15sensitive stuff in here. Um, and stuff 40:17that's, you know, yeah, that should not 40:19be shared in a public repo. So, I've got 40:21like, yeah, pre-commit hooks. I've got 40:23all all sorts of defenses to make sure I 40:25don't share anything uh private 40:26hopefully, but I know for a fact that 40:28that's going to happen. So, I've also 40:30got a key rotation routine I could 40:32execute if necessary. So, the project is 40:34called Pi, which unfortunately rhymes 40:36with Kai. Kai was just my first role 40:38playing game, and it just happens to 40:39rhyme with Kai. It's unfortunate, but uh 40:41it is all out there for people to use 40:43and download and play with. So, these 40:45are some of the most common questions 40:47that I get, and I'm not going to go 40:48through these. I think we might have 40:50touched most of them in the content or 40:52in the questions, but if you want to 40:54just take a picture of this or 40:55screenshot or whatever, those might be 40:57helpful. And with that, I want to say 40:59thanks for your time. You can hit this 41:01QR code to connect with me and different 41:03projects I'm working on. Got a whole 41:05bunch of stuff coming out in January to 41:07get a lot more of this like on demand 41:09inside of uh videos. I also help 41:11companies like do this and train their 41:13uh people to do this at work. So, you 41:15can hit me up if you're interested in 41:16that. And with that, I'm happy to take 41:18questions. 41:19>> Maybe here's a quick one from Henry. Can 41:21you speak to the role of Git and give 41:23Git ops in your workflows? 41:25>> Yeah, I I would say everything is Git 41:27heavy heavy Git usage. I've got a bunch 41:30of security controls there as well. The 41:32context management also matters there a 41:34lot. Kind of the big thing you need to 41:35worry about is you got four tabs open, 41:38five tabs open, and they're all separate 41:39sessions. They're all doing really cool 41:41stuff. And maybe you go into the wrong 41:43one and you say, "Yeah, great work. Go 41:44ahead and push this." But maybe it got 41:47confused about something that happened 41:48earlier. It thinks this content should 41:50be pushed to that repo. So I got a whole 41:52bunch of defenses around that to make 41:54sure okay what is your current working 41:56directory? What is the history of our 41:57conversation? Therefore, what directory 41:59do I probably mean when I say that? So a 42:02lot of scaffolding around that. But in 42:04general, I'm huge file system. Get work 42:07trees are also really powerful. So you 42:08have separate branches on separate file 42:11systems doing different experiments and 42:13the one you like is the one that gets 42:14brought over and merged. I would say I 42:16use Git very very heavily. Good 42:18question. 42:18>> Uh there's a question. Are you going to 42:21do a deep dive course or training or 42:24something on this at some point? 42:25>> Yeah, absolutely. So I've been doing 42:27augmented since 2023. Early 2023. I did 42:30a AI course and I'm converting that over 42:33to be an online version instead of live. 42:35So the the previous four were all live 42:37and it just it doesn't scale and you 42:39have to be there and it's just like it's 42:40uh it's not great. So that's going to be 42:42what I release in January as part of 42:44this human 3.0 program that I'm doing. 42:47So it's going to be more modular too 42:49where you could just like watch the 42:50history system video or you could watch 42:52the skill building video and stuff like 42:54that. But also talking about the overall 42:56arc of like how to get into this if 42:58you're not technical, how to do this for 42:59work, lots of different stuff like that. 43:01>> Awesome. Yeah. Yeah. Uh question from 43:03Lad. How do you decide which workflows 43:06to automate that is add to Kai to get 43:08the highest ROI? So, you know, how do 43:10you avoid wasting time on some 43:12complicated skill or flow that's going 43:14to take too many iterations to make it 43:15work, right? And I think there was 43:17another question in the chat about how 43:19do you avoid spending all your time yak 43:21shaving versus actually building new 43:22things and making progress. 43:24>> Yeah, that that's a challenge. Um, it's 43:26a challenge because I love the tooling. 43:28Like I've also you're also talking to 43:30someone who spent hundreds of hours on 43:32their neov config. So I'm also a tech 43:34nerd, but I also think humans matter 43:36more than tech, right? So I'm 80% human, 43:40but also 20% tech nerd. So the answer is 43:43sometimes I overfocus on the tech and I 43:45have to remind myself, hey, what are you 43:47doing? That's why I have a bunch of 43:48sticky notes in front of me. I have a 43:49whole TLO system for staying on track 43:52and like goals and stuff like that. I 43:53think it's fine to nerd out and to go 43:55crazy with new tech and learning stuff 43:57as long as you could pull yourself back 43:59and get get back on track to the goal. 44:01>> Yeah. And and just to be open about it, 44:03I think like Daniel and I have this 44:04conversation often where we were 44:06chatting about evals recently and I was 44:07like, "Oh, here are some services that 44:08are like these built-out eval 44:10platforms." And he's like, "Well, what 44:11if I built my own eval platform from 44:13scratch?" And then like one day later 44:15sends me a screenshot of like a working 44:17version. So I think yeah, it's easy to 44:19rabbit hole with this stuff in terms of 44:20priorities. I think if you do like a 44:22time audit, like where am I spending my 44:24time today and what is low value versus 44:26high value? like the things that are 44:28taking a lot of time that are not maybe 44:30uniquely suited to you or your role. So, 44:32lots of time, low value are good places 44:34to try to automate and streamline or 44:36things that happen often that seem easy 44:38to streamline. It's also useful. Yeah, 44:39>> totally. It's a TLO te. It's also a 44:43GitHub repo. A lot of the stuff that 44:44I've talked about is is on GitHub. It 44:46it's basically a system for managing 44:49goals, priorities, challenges to your 44:51goals. It's like alignment towards a 44:53particular goal. So, you could use it at 44:54work. You could use it like to run a 44:56family. You could use it to run a United 44:58Federation of Planets. It really scales 45:00to any different size. 45:02>> Oh yeah. I think um some other broad 45:04questions are like they they see your 45:06setup and then they see the public pie. 45:08Let me describe it and you can feel free 45:10to correct me. My understanding is you 45:12have like Kai and this is your personal 45:13thing um like what you're working on and 45:16then you are taking the generalizable 45:18public safe to share pieces of it and 45:20putting it into pie which is kind of a 45:23subset of Kai or at least the public 45:24version of it that's not the rapidly 45:26changing stuff that you're working on 45:28dayto-day. 45:28>> Yeah, that's right. And and I'm even 45:30putting like the super high value stuff 45:33like the art skill for example is up 45:34there and a lot of people told me do not 45:36put that up there. That is like a 45:38massive advantage for you as a content 45:40creator because I could basically feed 45:42it a blog post and it makes the perfect 45:44art image. And if you modify your 45:46aesthetic file and you say what you want 45:48it to look like in the aesthetic file, 45:49your PI system will do that for you. But 45:51my answer to that person was, well, I 45:54want more people to be content creators. 45:55Like that's the whole point of this 45:56entire thing. So like I am essentially 45:59trying to get everything that I can into 46:01the public version as fast as possible. 46:03like I'm uh generally handling the 46:05issues and PRs inside of Pi within hours 46:08or a couple days. Try to do as fast as 46:10possible. 46:11>> Related question from Brandt. Do you 46:13have any good workflows for updating 46:15your personal PI? So basically you have 46:17like the public pie and then you have 46:19your local customizations and like how 46:21do you sort of manage those in terms of 46:23like maybe pulling the latest that you 46:25and other community members are sharing 46:26without maybe overriding or complicating 46:29your customizations people have made 46:30locally. Um, so I haven't had many 46:33situations where people have added too 46:35much that I've pulled back in just 46:37because it's so early. I think that'll 46:39start happening a lot more in like the 46:41next months and year, but right now it's 46:43mostly me pushing out. 46:45>> Yeah. But if you're like a user of Pi 46:47and you maybe are like trying to balance 46:49those changes with the ones you've made 46:51locally. 46:52>> Oh, right, right, right. Yeah, I haven't 46:54found a good solution for that yet. We 46:56are actively working on that inside of 46:58the Pi repo in the community and the 47:00discussions. We're trying to figure out 47:01if that looks like a minor fork 47:03situation, if that looks like a side by 47:06side type thing where their PI agent 47:09have it sitting on the side all the new 47:11stuff and then it will migrate over to 47:13the authoritative one. So that's that's 47:15one option for that. Yeah, I guess it 47:17could also be standard Git and GitHub 47:18workflows in terms of like you fork Pi, 47:20you have your private fork in a private 47:22repo which you then commit all of your 47:24personal or like stuff specific to you 47:27and then you're just periodically 47:28merging in from the remote that is like 47:29the official public pie merging it into 47:31your private one. 47:32>> That's right. I mean, and you can have 47:34your agent do that and decide what fits 47:36for you, right? 47:37>> Keith had a great point. Maybe you have 47:39like a local folder that gets get 47:40ignored where people can create 47:42customizations and then still get 47:43upstream stuff. 47:44>> Yeah, I think that's a good idea. Yeah. 47:46Also, just shout out to Keith. He and 47:47the Trail of Bits folks are doing 47:48awesome uh AI work. So, check out their 47:51blog. I enjoy reading it very much. 47:53>> Oh, yeah. Yeah. They're doing amazing 47:54stuff. Crushed it at the AI XTC as well. 47:57>> Absolutely. So, there are still a lot of 48:00questions, but I wonder perhaps for time 48:03sake if we might want to answer them 48:06like asynchronously after or how are you 48:08feeling? 48:09>> Yeah, I think we can uh get some of 48:10those together and yeah, maybe we do 48:12another session or something or send 48:13them out. Yeah, I think async is the way 48:15to go. 48:16>> Yeah, people seem to have a bunch of 48:17excellent questions and uh no shortage 48:19of them. So, also your workflow and how 48:22things work dramatically changes every 48:24few weeks in terms of it gets better and 48:25better. Happy to do this again uh maybe 48:27in a few months or so. Cool. Yeah, 48:29Daniel, thank you so much. This was 48:30awesome. Tons of stuff to dig into. So, 48:33really appreciate your time. Have an 48:35awesome rest of your day and talk