Learning Library

← Back to Library

Chatbots, Virtual Agents, and Humans

Key Points

  • The episode explores how virtual agents, chatbots, and human support differ in accuracy and usefulness, and how businesses can serve users who prefer either automated agents or human interaction.
  • Susan Emerson shares her career path of joining emerging tech companies, leading to her current role at Salesforce after her previous employer was acquired.
  • Nick Renotte recounts his early coding experiences with Excel VBA, a pivot to business studies, founding an ed‑tech startup, and now leading a global team of roughly 500 AI engineers.
  • The hosts set the stage to discuss the historical evolution of chatbots and virtual agents, emphasizing that these technologies have been around much longer than most people realize.

Sections

Full Transcript

# Chatbots, Virtual Agents, and Humans **Source:** [https://www.youtube.com/watch?v=348242wb1-E](https://www.youtube.com/watch?v=348242wb1-E) **Duration:** 00:32:42 ## Summary - The episode explores how virtual agents, chatbots, and human support differ in accuracy and usefulness, and how businesses can serve users who prefer either automated agents or human interaction. - Susan Emerson shares her career path of joining emerging tech companies, leading to her current role at Salesforce after her previous employer was acquired. - Nick Renotte recounts his early coding experiences with Excel VBA, a pivot to business studies, founding an ed‑tech startup, and now leading a global team of roughly 500 AI engineers. - The hosts set the stage to discuss the historical evolution of chatbots and virtual agents, emphasizing that these technologies have been around much longer than most people realize. ## Sections - [00:00:00](https://www.youtube.com/watch?v=348242wb1-E&t=0s) **Bots, Agents, and Humans** - The hosts introduce Salesforce’s Susan Emerson and IBM’s Nick Renotte to explore how accurate and helpful virtual agents are, clarify the distinctions between chatbots, virtual agents, and human support, and discuss strategies for serving users who prefer each option. - [00:03:08](https://www.youtube.com/watch?v=348242wb1-E&t=188s) **From Rule‑Based Bots to Generative AI** - The speaker outlines how customer‑service bots have progressed from simple keyword‑driven assistants in phone banking to advanced generative AI agents like IBM’s Watson Assistant, transforming interactions across industries. - [00:06:16](https://www.youtube.com/watch?v=348242wb1-E&t=376s) **Balancing AI Speed with Human Touch** - The speakers discuss when to rely on quick, generative AI answers versus connecting users to humans, weighing accuracy, speed, and personalization, and illustrate the shift from rule‑based bots using examples like banking queries and a florist’s daily requests. - [00:09:26](https://www.youtube.com/watch?v=348242wb1-E&t=566s) **Generative AI: Speed vs Accuracy** - The speakers debate using generative AI to tackle the 20 % of support interactions that lack first‑call resolution, weighing the need for rapid responses in some sectors against the requirement for highly accurate, compliant answers in regulated industries. - [00:12:32](https://www.youtube.com/watch?v=348242wb1-E&t=752s) **Balancing Virtual Agents with Human Oversight** - The discussion highlights the need to complement chatbot interactions with human support, structured SOPs, and escalation mechanisms to build trust and deliver a holistic customer experience rather than just a technical solution. - [00:15:35](https://www.youtube.com/watch?v=348242wb1-E&t=935s) **Balancing AI Agents with Human Support** - The speaker argues that generative AI virtual agents can handle routine queries but still require human fallback for edge cases, emphasizing that data quality is the foundation of their accuracy. - [00:18:50](https://www.youtube.com/watch?v=348242wb1-E&t=1130s) **Swiss Cheese Governance Model** - The speaker explains a multi‑layered safety framework for large language models—likened to stacked Swiss‑cheese slices—where input filtering, prompt engineering, and final guardrails work together to catch undesirable outputs. - [00:21:55](https://www.youtube.com/watch?v=348242wb1-E&t=1315s) **Specialized Banking Virtual Agents** - The speaker explains that a banking virtual agent comprises multiple specialized sub‑agents—such as trading, transaction handling, and debt recovery—using intent detection and task handoff to generate a unified customer response, illustrated through the “Swiss‑cheese” framework analogy. - [00:25:16](https://www.youtube.com/watch?v=348242wb1-E&t=1516s) **Pinpointing Friction for AI Task Agents** - The speaker discusses how to identify pain points and workarounds in customer‑service workflows, select tasks ripe for automation, and use generative AI‑powered virtual agents to provide sense‑and‑respond, proactive assistance instead of traditional decision‑tree systems. - [00:28:20](https://www.youtube.com/watch?v=348242wb1-E&t=1700s) **Proactive AI Coaching Framework** - The speaker outlines a proactive, persistent AI assistant that leverages IBM’s Granite LLM and the open‑source Instruct Labs framework to deliver faster, smarter, customized, and governed coaching. - [00:31:32](https://www.youtube.com/watch?v=348242wb1-E&t=1892s) **Value, Governance, and Virtual Agents** - The speakers emphasize quantifying business impact, maintaining long‑term governance of virtual agents, and express gratitude while inviting audience questions and feedback. ## Full Transcript
0:00Virtual agents have 0:01become a part of our daily lives, for better or worse. 0:04But I have to wonder. 0:05How accurate and helpful are they actually? 0:08What is the difference between a chat bot versus a virtual agent versus a human? 0:14Because, good grief, I've heard them all. 0:16So today I'm going to be joined by Susan Emerson from Salesforce 0:19and Nick Renotte from IBM. 0:21We're going to unpack the burning question: 0:23how do you serve 0:24the different people who want an agent and those who want the human? 0:28Because I think most of us want them both at different times. 0:31So Susan and Nick, welcome to AI in action. 0:34Well, thanks so much. 0:36Thanks for having us. For sure. 0:37Let's get on into this conversation. 0:39But, Susan, let's start with you first. 0:40You have such a storied career. 0:43It seems like you've been everywhere and worked with companies 0:45that are small and work with them until they've gotten large and beyond. 0:49But how did you end up at your current role at Salesforce? 0:52Well, I would say probably a long history of saying yes to young companies 0:57that are at the intersection of some new innovation and technology, 1:01which kind of always puts you unsteady because you're figuring things out. 1:04And that just leads always to be in a conversation of innovation. 1:09And that's how I got to Salesforce, the company I was working for 1:12was acquired. Excellent. 1:14And now, Nick, I know that you know a lot of people in this space as well too, 1:18because you've been, did you start like coding 1:21when you were like eight years old or something? 1:23Yeah, yeah, I did actually. 1:24So, funnily enough, I started at coding inside of Excel. 1:29So I was doing like a lot of VBA coding. 1:32My dad was doing like share trading and options pricing. 1:35So, I was like, hey, I can actually automate some of this stuff. 1:39It started off really basic. 1:40It was just like, record a macro. 1:42And then Jerry rig something to to get it to do something else. 1:45But funnily enough, like after that, like there was this huge period, right? 1:49I barely touched any tech. 1:51I'm like, 1:52I just really want to become a business owner, 1:54and I really want to get into business and everything's about business. 1:56So I went and studied business, not realizing that really, 1:59in order to build amazing businesses, you have to build amazing products. 2:02Which tech really helps. 2:04So, that kind of switched around, and I actually built an ed tech startup. 2:10and I currently lead, like, 500 AI engineers around the world. 2:15So, so they sort of take my lead in terms of, like, what we're doing. 2:18and, and YouTube's definitely played a huge part in that as well. Wow. 2:22I love again, we've got such a breadth of experience 2:25just represented right here on this specific episode. 2:27So I want to take full advantage of that during this conversation. 2:31Nick, I want to start off with you. 2:33I've been thinking a lot about chat bots and virtual agents. 2:36And I guess the thing is that if we really consider it, 2:39they both have been around a lot longer than people assume. 2:43Can you please briefly 2:44take us through exactly how we got to where we are with customer support. 2:49From a long time, a big part of interacting 2:52with an organization is making sure that you do have good customer support. 2:57We started off really simply like you would maybe go into a bank, 3:02you'd speak to a, customer support 3:05officer or a teller, and they'd be able to help you out. 3:08That eventually shifted to being able to perform things like phone banking. 3:13And then we had the rise of the internet, so we had chat bots coming into play. 3:17So that sort of shifted how we've been able to interact 3:20with different businesses. 3:21I'm using a bank as an example, but, this has been across 3:24a range of different organizations as well. 3:26Those chat bots in the past have kind of been reasonably simple. 3:30So, you might know them as like rules or heuristic based chat bots. 3:34So you type something in and maybe picks up a key word or two, 3:37and then it's able to go, hey, where you looking for? 3:40I don't know, support on tracking your particular product or did 3:44you want to go and transfer some funds? 3:47We've gone and shifted significantly away from that to to the ability 3:51to use generative AI to help us, speed things up quite a fair bit, 3:55but but again, they've been around for quite some time. 3:57So I've been at IBM for, I want to say like five years now. 4:01And since I started at IBM we've had virtual agents, through Watson assistant. 4:06That being said, they have evolved so rapidly to to the point 4:09that they're at now. Okay. You talk about that evolution. 4:12I remember some of those early experiences where I just felt like, 4:16just put me on the phone with the person, 4:19this is not this is not what I'm trying to do. 4:21Susan, can you tell me a little bit about your experience 4:25in terms of what you've seen happen with customer support over time? 4:29Well, I mean, one of the things that, 4:31kind of just listening to Nick talking about the evolution, I, I kind of think 4:35this is really completely not relevant to anything, but I, 4:38I remember a time when there were party lines on your phone 4:42and you would pick it up and you'd have to yell at the neighbor to get off. 4:45But anyway, so now thinking about like Gen AI answering the questions 4:48versus like, yelling at the Martha Becker, it's a whole new world. 4:52So the thing I would say around, like, why we're seeing such crazy transition 4:57and transaction with these things, it's, it's a number of things like generative 5:02AI, obviously, for those of us in tech, it's like the biggest transformation 5:06since the internet. 5:08And all of the companies that make investments in supporting 5:12customers are pressured by the executive staff, the board 5:16members of, how are you bringing Gen AI into the foreground? 5:20And it's one of these areas where you get that pressure. 5:23You get the fact that these call centers are literally the eyes and the ears, 5:28or maybe the ears and the eyes, depending on which order you want to go. 5:31And they represent everything from the question answering, 5:34the brand, the experience, the relationship 5:38and being able to serve people up on the channel that they choose. 5:42You talk about when you want a human and when you want the digital interaction. 5:47It really unlocks that full potential and what the technology has done 5:52is brought it forward in a way where the way you develop 5:55it is so much easier 5:57because you don't have to do that tree based stuff that Nick was talking about, 6:00where you would have to previously train and look for every kind of intent 6:04and keyword and have this stuff really, you know, rigorous down a decision tree. 6:09Now it can be generative, interactive, 6:12multi turn, but with all the guardrails that you need. 6:16So for the things where you don't want the human, you can get the quick answer. 6:20And the things that you do want the human, you get a quick connect. 6:24I think the quick connect is is very key. 6:28I keep thinking, all right, what's more important? 6:30Is it the accuracy that I want. 6:32Like is it the speed that I want. 6:34Is it the personalization. 6:36And I think the answer is yes. 6:38Like Nick, you talked about the adaptive element of this. 6:42So, I kind of want to get under the hood a little bit with you, Nick, 6:46as we're talking about the rule based chat bots moving into gen AI. 6:49It's such a radical shift. 6:51Nick, how do you even start thinking about making that? 6:54Let me give you an example of like how this actually comes into play 6:58and like where Gen AI is really suited 7:01for handling these because, I mean, you mentioned something, right? 7:05There's certain situations where you do want a human, and there's other situations 7:09where getting an answer from a virtual agent 7:11or just finding a response and getting a valid response is enough. 7:15Because let's say, for example, you wanted to check 7:17your bank account statement and it's more than enough 7:20for you to get that via virtual agent or just checking an app. 7:23You probably don't really want to be speaking to a human, because these 7:27are really quick interactions where you just want an answer really fast. 7:30So let me switch it up a little bit. Right. 7:33So let's let's take Bob. Right. 7:34So Bob owns a florist. 7:36So every single day he gets like a ton of requests for people 7:40that want to go ahead and buy buy flowers. 7:42So it's like, okay, all right, all right. Cool. 7:44This is perfect. I want a chat bot here. 7:46This chat bot’s going to help people just purchase flowers. 7:49They're going to be able to choose where they get them sent to. 7:51It's just a form, we’re able to fill that out. 7:53Now the problem is though like 80% 7:56of the interactions that Bob gets via his chat bot are to buy flowers. 8:00So so it handles that 80% absolutely perfectly. 8:04The problem is that there is like 20% of other questions. 8:08Where are the flowers sourced? 8:10What's your refund policy? 8:11Do you have hypoallergenic flowers? 8:13I don't know if that's actually a thing, but like, let's say it is. 8:17These are what we typically refer to as like long tail questions. 8:20Right. So questions 8:20that maybe don't come up so often, but you still want to be able to handle. 8:24Now like using rule a heuristic based chatbots, 8:27like in the past has been perfect for handling that 80%. 8:30But when it comes to the long tail questions, these are where you can 8:34very quickly start 8:35to spend like an absolute ton of time to get very minimal benefit. 8:40In the past, we've been able to do things like semantic search, 8:43and bring out an extract and then just dump an extract back to a user. 8:48But they kind of know that, hey, you're just dumping your response to me. 8:52It's not really all that personalized. 8:53I probably could have just searched 8:54for that via a search engine and probably got the response back. 8:59Gen AI is amazing for this, right? 9:01Because we can give that extract 9:03that we traditionally would have just dumped to the customer. 9:06We can actually pass it to an LLM, that LLM can personalize 9:09it based on the customer's interaction, because we have context 9:12and we can get a response. 9:13So like when somebody goes, hey, do you have hypoallergenic flowers? 9:17The virtual agent’s like, look, to be completely honest, I don't actually 9:21think hypoallergenic flowers exist, but you might try these as an alternative. 9:26So when it comes to handling that 20%, where we spend 9:29a ton of time where we don't typically get, first call resolution, 9:34this is perfectly suited for generative AI. 9:37Okay, Susan, I heard you throw in some mmhms and yeahs during that. 9:40What do you got? 9:42I was thinking about the hyper 9:43allergenic flowers, and I was thinking about the intersection of like, 9:47what do you want fast, and what do you need accurate? 9:50Because it's probably different for different industries. 9:52Like, 9:53you know, flowers are a great example, but maybe extracting ourselves from that, 9:57this idea of like the 80% of the questions are always sort of in the same domain. 10:01I would argue a little bit, maybe that heuristic 10:03has been fine for that, because there's a ton of set up 10:06in taking all those utterances and conversations and training them 10:08for intent. 10:09You can go a whole lot further with these generative experiences 10:14because they leverage the power of the pre-trained models. 10:16But then I would also say, like in a lot of other industries 10:19that might be regulated, you started off with an example of banking. 10:22It has to be right, and it has to be grounded, 10:25not just in a knowledge repository that is validated 10:28and known, but legal risk, compliance and things like that. 10:31Because of the impact of could be something as simple as getting 10:35a balance wrong or something like that. 10:37Some of these things have to be right. 10:39And so in some examples, you know, right is going to be much more appreciated 10:42than, than fast. 10:43But then there's the fast to deliver and fast to support and fast to innovate 10:47in the back of house and everything like that. 10:49So, you know, with traditional call center, 10:51a lot of these conversations start with like the two words 10:55that most people orient to things like call deflection, 10:58which I guess is a technical strategy as a thing. 11:01But as a consumer of any product or service, 11:03it sounds kind of like, oh, I hope, I hope I'm not really being deflected. 11:07I'm a customer, right. Like that kind of thing. 11:09But but there's this whole idea of like call deflection isn't always the goal 11:15because there's some categories of, of experience 11:19and being a customer where 11:20the human in the loop is the priority, not the speedy fast chat bot answer. 11:25And like doing the balance inquiry and like, 11:27did you send me my tax form and doing that all automated is great. 11:31But like you have like like a bad you know, I'm going with your banking 11:34example, Nick. 11:35You have like a bad fluctuation in the stock market. 11:38Like you want that human on the phone right then and there. 11:41And so this there's this whole mindset of like 11:43not one channel is better than the other, but have a point of view 11:47of what type of interactions you want to actually 11:50optimize your relationship with the customer 11:51on whether it's fast and speedy and automatic 11:54or human and personalized and have it all be part of the same platform. 11:58So you really always knowing the customer. 12:00It's funny, Susan, as you as you're talking about the banking and Nick 12:03because you brought this banking example up, I, this feels very personal right now. 12:08So I'm going to have a little confession right here. 12:09Maybe some of the listeners will vibe with me on this one. 12:12But every now and then, like perhaps I will forget 12:16to pay a credit card bill on time. 12:19And because I feel like I'm special, I always want the bank 12:23to drop that fee for me. 12:24I always want the credit card companies to drop that fee. 12:27And so that involves me often, you know, calling and trying to get that done. 12:32And sometimes the virtual agent doesn't care that I feel like I'm special. 12:37And sometimes I need the human in order to work with me through that. 12:41So yes. 12:42Yes and yes, to all of this, Nick, I know that you were about to share 12:47some more with us about that distinction between the virtual.... 12:50Nah, I'm with you, I’m with you on the... 12:52you need a little give and take. 12:53You don't always get that with the virtual agent or chatbot, right? 12:57But, yeah, to echo Susan’s point, like that, that is absolutely bang on. 13:00Right? 13:01I think it's important that we're creating an experience for our customer. 13:05It's not just a solution. 13:06So like from a technical perspective, the people who have to build 13:09and curate these things, it's a whole different world. 13:12And then sort of maybe, you know, adding on some of the more recent innovations. 13:17I know some of the stuff that we've been working 13:18on is like, you know, those SOPs, Nick, right, in terms of 13:22on is like, you know, those SOPs, Nick, right, in terms of 13:22these are the questions, 13:23these are the instructions 13:24that if we had a human, we'd take you step one, two, three, four. 13:28And these are the things we would permit you to do. 13:30And this is 13:30when you escalate to the supervisor or to the human or things like that. 13:34So having that kind of inbuilt infrastructure is one of the 13:38I think the big things that help customers make this transition because, 13:42you know, obviously many organizations still wrestle with what 13:45their point of view is around trust, fidelity, false certainty. 13:49You know, if this is customer facing, you know, to have that 13:53those those types of checks, balance controls and guardrails in place. 13:56Yeah, definitely. 13:56And one of the things that I've noticed, right, 13:58is that as we've gone and deployed generative AI solutions, 14:03let's say we generate an answer for a particular question. 14:06What we're actually doing now is we're taking those responses 14:09and we're caching them. 14:11So that means that it's not as costly. 14:13You don't need to send all your data 14:14to an LLM again, because if you know that you've got a standard question 14:18with the standard response, you can actually go 14:20and sort of save that off in turn that into a rule. 14:23Plus you can apply that additional overlay with, with your own corporate field. 14:27So if you have a specific way that you want to go and respond to it, 14:30you've already got a baseline. 14:31You just need to append or tweak the response that you've got. 14:35As you talk about this tree of responses, like thinking about the rule space. 14:40But now also again, of course, compared now to this virtual agent space that we're 14:44in, are you finding that people are trying 14:47to skip to the human portion less now? 14:51Like, because I can think about many times when folks would just try to always 14:54press zero. 14:55I think it really depends on on the industry. Right? 14:57So if you're like a high tech startup 14:59and you've gone and done this stuff before then, then it could be weeks. 15:03But if you're a bank or a government institution, then it's typically months 15:07because the process to govern, monitor and test is, 15:11significantly more stringent in terms of like people handing off 15:17or trying to get through to a human a lot faster. 15:20I think it's important to note that you're still going to want to go 15:24to a human for a lot of interactions, like you mentioned before, right? 15:27Like you're going to know that you can sort of you can pitch a human 15:31why you should get your your bank or credit card late fee remitted. 15:35Right? 15:35Like you're going to have a hard time doing that with like a gen AI solution. 15:38Like, I don't know how much give and take 15:39you're going to get when it comes to bargaining with that. 15:42I mean, you can give it a crack. 15:43It'll be very Australian of you to do so. 15:45But in terms of actually seeing people 15:49completely deviate away from a virtual agent 15:52all the way back to a human, I think the the attitude is still kind of similar. 15:58Right? 15:59So we've just added an additional layer of support to a virtual agent. 16:03That doesn't necessarily mean that you're going 16:04to want to deviate away from everything. 16:06That being said, I personally find myself using virtual agents 16:10a lot more because I'm like, I don't really want to speak to a human for this. 16:14I just want an answer a lot of the time. 16:16Where it's edge cases, 16:18then then I probably want to go and speak to someone 16:20because I'm probably going to get some guardrail 16:22because they probably aren't allowed to talk about that anyway. 16:25If it could answer it self-serve, first first pass. 16:28If you're asking for the human like, that's a whole different ballgame. Yep. 16:32And you know what? 16:33Like in terms of, regardless of what the situation is that we're dealing with, 16:37we know that data is still at the foundation of it. 16:40And the virtual agent 16:42is only as good as the data that is being placed inside of it. 16:46So how much can you program a virtual agent for accuracy with what you have? 16:52I mean, definitely the moniker of like data 16:55is the foundation of AI and AI needs data is is super true. 16:59A lot of organizations, 17:00you know, Nick was talking earlier about, standard operating practices. 17:04Most organizations have this stuff written down 17:06and they have a knowledge repository. 17:08And so many organizations are much further along to use these things, 17:11especially if they're this gen AI like chat 17:14bot experience can tap into that official knowledge repository, 17:18because then you get this power of like conversational turn 17:21and that great user experience, 17:23plus validated answers that are incorporated with the context 17:26setting of a of a large language model answering a question, 17:30but with all the valid responses that are in the validated, 17:33risk approved, marketing approved knowledge repository. 17:37Yeah, I think data is is absolutely critical and has been from day dot. 17:42So if we look at how we've got to where we are now, 17:45big data was a really big thing back in the day. 17:48Then it shifted over to data science and everyone wanted to do data science. 17:52And then with machine learning and now it’s generative AI. 17:55All of those principles or all of those spheres have been focused on data, right? 18:01Generative AI is no different. 18:02It's a derivative, that the, the underlying models which power them. 18:06So large language models originally trained as a deep learning model. 18:10It's based on a transformer architecture, 18:11which you need a ton of data to actually go and and train. 18:16The thing is, though, right, accuracy is always going 18:19to be like a fluid metric and a fluid dynamic. 18:23because it's only like one factor, right? 18:25So in terms of how we go and evaluate whether a generative 18:29AI solution is performing, it's a whole range of things, like how factual is it? 18:33How valid is the response? 18:34Is the response generating an answer from the context that we've given it? 18:38I think what's more important, rather than just focusing on accuracy 18:42and what we've been preaching, within our client engineering team is 18:47how do we monitor, test and govern these LLMs? 18:50Because you can go so far as to train the model 18:54to actually ensure that it's going to generate a correct response. 18:57But how do you know that the customer's going to ask a question 19:01which that underlying model has seen before or knows the answer to? 19:06So the governance framework around 19:08these solutions is, is probably just as critical. 19:11Right. 19:12And what we've actually started or I saw one of my engineers present to me 19:15the other day, is this Swiss cheese framework, right? 19:19I love cheese, and I was like, he recommended Swiss cheese to me. 19:21I'm like, perfect. Cool. Let's go down this route. 19:24So if you think about a piece of Swiss cheese, right, 19:26it's got a whole bunch of holes through it. 19:28Now imagine you're sending a question or a prompt through a slice of Swiss cheese. 19:33Some of this stuff is going to get through, right? 19:34Some of the bad stuff is going to get through. 19:36Now imagine you place another layer of Swiss cheese. 19:39Maybe it doesn't get through that layer. 19:40Then you're adding another layer of Swiss cheese. 19:42Maybe we finally catch it at that final layer. 19:44Each one of those layers of Swiss cheese is a different component 19:48in how we actually go and govern our virtual agents or large language models. 19:53We can perform input filtering, which would be that, that that first Swiss 19:57cheese layer. 19:58We can then go and perform 19:59a model prompt engineering and governance at the model layer. 20:02But we can also apply guardrails at the final layer as well to ensure that 20:06as we're going through, we're making sure that, 20:08hey, we're only taking in stuff that we want to be taking in, 20:11and we're only outputting stuff that we want to be outputting. 20:13On top of that, we can apply monitoring. 20:15So if we start to see user feedback that, hey, maybe we're not answering 20:20as well as we potentially could, we can actually give that feedback back 20:24to our engineering team to make sure that we're going, 20:27and potentially going and fine tuning a LLM. 20:29We're adding that data to the corpus so we can actually generate that response. 20:33So we think about accuracy a lot, but I think a lot of people 20:37maybe don't think about, hey, how do we fix it 20:40if we're getting the wrong response, how do we fix it? 20:42Because ultimately it comes back to bang for buck. 20:45If Gen AI is not performing well, then you're spending a lot of money 20:49for something that doesn't work. 20:50I can think of like all sorts of things that could define the Swiss cheese. 20:53Like it could be everything from your authorized use of generative AI. 20:57What is the framework of what you're comfortable with, 21:00and the types of use cases you put in front of your employees 21:03and your customers? 21:04And then it would be things like, what is the data 21:07that you have to ground this stuff, and what is the data safety 21:11that you have to maintain through this stuff 21:13in terms of ensuring that your data isn't inadvertently training someone else's LLM 21:17and it's not stored in the wrong place, and that you have like a full suite 21:22of capabilities to say what was asked, what was answered, 21:25what was the toxicity score, and did the human use it or change it like 21:28because you need to have all this governance 21:30and then you also have to have things like, 21:32as Nick said, prompt engineering because that gives 21:34the full set of instructions to the LLM 21:36and you want to ground it with the customer data. 21:38So when the thing goes off through the Swiss cheese filter, it's 21:41like giving instructions that are not naive and relevant enough. 21:44So like there's a whole like series of really like nice things 21:48in technology like Salesforce that give us those frameworks for interacting. 21:51Let's say, for example, you've got a virtual agent to to your point, right. 21:55Like you've got a banking virtual agent. 21:57Now, what is a banking a virtual agent? 22:00Within there there might be a, a 22:02specialized 22:03agent, which is really great at trading interaction. 22:06So like, hey, this is our banking agent, which actually is focused on 22:10stock trading, 22:11but we've also got one that's focused on transactional interactions. 22:15We've got ones, another one that's focused on debt recovery or payment plans. 22:20So having multi agents as, as I understand it, where in my context or in my world, 22:25is actually having specialized agents which are focused on different types 22:29of skills. 22:30Yeah. 22:30So, so in terms of how we roll that out, typically, 22:33you can have switching or task handoff. 22:35So let's say for example, you actually go and ground an agent in a specific SOPs 22:42when your initial agent detects that, hey, this intent is probably focused on that. 22:46We can hand off to that agent, generate that response, and then come 22:49in, present a unified response, or a unified answer back to the customer. 22:53Coming back 22:54to, like, have this with Swiss cheese framework fits around this, right. 22:58It wraps around the whole thing. 23:00If you think about, like, data as being, 23:04actually, I've got this, this great analogy. 23:06Right? 23:06So, like, I'm a big foodie. 23:08It's not going to be more cheese, is it? 23:10No, it's not, it's not cheese. It's not cheese. Right. 23:12I’m a massive foodie in case you haven’t noticed. Right. 23:14So I went to Japan, last year and I was actually presenting this to 23:18to my team there. 23:20And I had a good laugh, but have any of you had okonomiyaki before? No. 23:25No, okay, all right. 23:26So okonomiyaki is like a Japanese cabbage pancake, so it's like a savory pancake. 23:31Now, the basis of this pancake is cabbage, so it's like a ton of cabbage in there. 23:37And if you think about cabbage as being your data, right, 23:40it forms the foundation of that specific data. 23:45You can also add in other different toppings 23:46which sort of like spice it up and make it a little bit more interesting. 23:50And like let's say, for example, we go and chuck bacon on top of our okonomiyaki. 23:54The way that I think about that is the bacon is our large language model. 23:58It's what adds the spice to actually generating our responses. 24:01Because if we gave our customer raw data and they'd kind of be okay, 24:05but it wouldn't be as nice and bright and shining as it potentially could be. 24:09Governance or the Swiss cheese framework is the data that ties it all together. 24:13So we wrap up the cabbage and then we put the bacon on top, 24:17and then the batter ties it all together 24:18in terms of have the Swiss cheese framework works around that, right. 24:21It wraps around the entire multi-agent system. 24:25We have handoff in between, but we still apply these guardrails at each step. 24:29You have input guardrails which would wrap around all of your input. 24:33You'd have model guardrails 24:34which would apply at each stage, or each different agent that you'd have. 24:38And then you'd also have your final Swiss cheese layer, 24:40which would be, your output guardrails or your output filtering. 24:43So again, it's sort of if you just think about the batter holding together 24:47the pancake, that's the way a best way that I can describe, handling governance, 24:52which is what we actually use, Watsonx governance for at IBM. 24:56And I'm thinking more about as a vegan, I'm not thinking about 24:59baking or batter or like the, the cheese even too, so, 25:04but the, 25:05but like, 25:06kind of thinking 25:06of these virtual agents, like, I usually just like 25:08one of my favorite frameworks in the world is jobs to be done, 25:12you know, where you just have this really kind of programmatic thing of like, 25:15who is the user? 25:16Or in this case, the customer or the service agent 25:18that is experiencing this, this digital virtual task based agent? 25:23What are they trying to do? 25:25Where is the friction in the process? 25:27What are the workarounds that make it horrible? 25:29And if you could do it all better, why would it matter? 25:31And so thinking about these virtual agents 25:33like we're working with so many call centers 25:35where, you know, it starts with that hypothesis of where the friction is. 25:39And then is that the thing we should automate to the task based agent. 25:43Because not everyone is ready to take a call center and make it all digitized. 25:46They're still human. 25:47So what are the tasks that are really ripe for that? 25:50That you can just say, okay, agent, tell me when this is happening. 25:53And then, you know, maybe kind of taking the next step further in terms of 25:57like a more of a sense and respond culture where the agent is always on 26:03and they're identifying problems, like, you can use examples 26:05where you got real time signal in terms of a customer interaction. 26:08You know, maybe it's like, you know, watching things on Netflix or your mobile 26:12phone and consumption or like things that have like real time data 26:16like these task based agents can start to do the sense and respond 26:20and cue up the proactive outreach and things like that. 26:23So it really can be not just like a friction take out. 26:26And let's use gen AI as opposed traditional decision trees 26:30for a lot of organization, this is like the first step to new operating models, 26:34where they can really start to think about new ways that they connect 26:36with their customers. Okay, that sounds like a Salesforce commercial. 26:39Connect with your customers in a whole new way. 26:41I didn't mean to like do that. 26:43Like, but I guess I've been here too long. 26:45No, no, no. You're on brand. You're on brand. 26:49So I want to talk about the future now. 26:50I want to talk about what is going to change. 26:53What do we have to look forward to in this space? 26:57Well, I mean, we've been working on what we call like, 26:58these transformation frameworks because everyone, like, 27:01looks to partners like IBM and Salesforce in terms of what this could look like, 27:05where do we stand on the steps of looking better, better, best, 27:08and what should we be doing next if this is the step we're on? 27:11And so we put together this thing we called the maturity framework 27:14for gen AI, where, you know, most organizations start in some form 27:18with human in the loop when they're bringing gen AI 27:21to their employees and their customers. 27:23It's the things that gives them 27:24the confidence, the control and the learning around these things. 27:27And we kind of call that phase one. 27:30And phase one is usually like the start point is we have a hypothesis of value 27:35in terms of how we're using this to drive loyalty, 27:38customer experience, revenue, whatever it is. 27:41And it by and large will take friction out of the process. 27:45It will hit the top value of human capital and data in very nice ways together. 27:51And that's sort of phase one of this maturity model. 27:53But you can't get to fully autonomous until you start with the first one. 27:57We talked about like call it into the call center. 27:59And hey what's my balance? 28:01Why is Bob the banker first not calling Nick because Bob has the data. 28:05It's just not super relevant and available to him unless he's focusing on it. 28:10So what if we had these AI agents that are personalized and have this kind 28:14of persistent worldview in terms of how we achieve this business objectives? 28:18The AI agent tapping Bob on the shoulder. 28:20Hey Bob, 28:21like we noticed you haven't called Nick and we've seen all these things happening. 28:24We think you should talk to them. 28:25We think these are the topics you should talk to him about. 28:28And here's your summary 28:29of all the interactions you've had with them over the last couple of months. 28:33So it's proactive. It's persistent. 28:34It takes all the dirty work out of like that preparation 28:38and that identification of how to spend your time. 28:40So that's sort of I think like the next chapter of these autonomous agents 28:44where they're like personal, persistent, predictive and present and coaching 28:48and things like that. 28:49Susan just painted such an ideal version of the future. 28:54What does yours look like? 28:55Yeah, I think that like there's four words that I keep harping on to myself. 28:59Faster, smarter, customized and governed. 29:01And how we're doing that really is faster. 29:05Like, we've been training a ton of our LLMs 29:07and getting these trained on business based data so that, 29:11you're not likely to go off the rails. 29:13So, like, if you look at what IBM's released, we've got a range of LLMs. 29:17One of them called IBM Granite. 29:19Now that's great. 29:20You're probably thinking, hold on, Nick, there's like a ton of LLMs out there. 29:24Like, why am I going to use IBM's one? 29:26Or like, why am I going to go down this route? 29:28Well, this sort of brings me to like my second part, which is like smarter. 29:32So we actually released, in partnership with Red Hat, we released Instruct Labs. 29:37So like, it's a completely open source framework, which actually 29:40allows you to go and fine tune a large language model. 29:43The amazing thing about this, though, and I've been taking a look at it, 29:47is that we actually do something called synthetic data generation. 29:50So let's say for example, 29:52you've got a PDF and you want to go and train your large language model 29:56or your virtual agent to be able to go and respond based on those questions. 29:59So there’s 30:00currently two key ways that that that we can do that out there in the field. 30:03So we can go and build a pattern called retrieval augmented generation, 30:07where we think of it like just going and dynamically grabbing chunks 30:10out of that document that are relevant to the customer's question. 30:13We chuck that into our LLM prompt, and then the LLMs are able to answer 30:17based on that context. 30:18The other way that we can do that is using fine tuning or parameter 30:21efficient fine tuning, which basically means that we actually go 30:24and train the model to go in to answer better. 30:26But to do that, you need the data structured, prepared and formatted 30:30so that we can actually go and pass that to a, an LLM. 30:33The cool thing about Instruct lab is you literally just dump a PDF 30:37or a markdown document, and it's able to go and generate that data 30:41in that format to be able to go and fine tune that LLM. 30:44So I think I wrote a cheat sheet for for our team, and I think it's like 30:48eight commands and it literally goes and fine tunes your LLM and deploys it. 30:53That brings me to the next stage, right? Customize. 30:55So once you've actually gone and made your LLM smarter, 30:58I think it's really important that we start 31:00baking this stuff back into our workflow. 31:02Like AI is not a process in and of itself, it's part of your daily work. 31:07Like how do we just make you smarter by using these tools? 31:11That's something that that my team does. 31:12So like we just a hook to the next podcast that I'm going to be doing, 31:17we're going to be talking about how to ensure that you got successful 31:19pilots, but we actually bake this into into customer solution. 31:23So rather than going out to the AI system, 31:27it's inside of your CRM, could be inside of Salesforce, 31:30it's inside of something else. 31:32And quantifying the business value to making sure, to make sure that that's, 31:36relevant and going to deliver bang for your buck is absolutely critical. 31:39And then governance. 31:40Right. 31:40Swiss cheese framework, making sure that once you've gone and rolled something out 31:44that you've got the ability to make sure that it keeps performing out into 31:48the long run. 31:48So I think that that's it. 31:50Faster, smarter, customized, more governed. 31:52I just have to be 100% honest here. 31:54It's really neat to have this conversation and to see the connections 31:58between you and Nick and the customers that are being served. 32:03And then, like me, figuring out and just kind of like 32:05having all my synopses fire, as you say certain things. 32:08So Nick and Susan, thank you very much for being here. 32:12Thank you for opening up this world of virtual agents. 32:15And to everyone who's listening, everyone who's watching. 32:18thank you also for spending your time with us. 32:20It means a great deal. 32:22If you have any additional thoughts, any additional questions, please don't be shy. 32:26Just drop them in the comment section and we'll see. 32:28Hopefully we can get to some of those for you, but we'll see you again soon. 32:41Bye bye.