Learning Library

← Back to Library

Seven Essential AI Terms Explained

Key Points

  • AI is now ubiquitous—from everyday objects like toothbrushes receiving updates to rapid advancements that even tech professionals find hard to track.
  • “Agentic AI” refers to autonomous AI agents that perceive their environment, reason about next steps, act on plans, and observe outcomes, enabling roles such as travel booking, data analysis, or DevOps automation.
  • Large Reasoning Models are specially fine‑tuned LLMs that generate step‑by‑step “chain‑of‑thought” processes, allowing agents to tackle complex, multistep tasks with verifiable accuracy.
  • Vector databases store information not as raw files but as numerical embeddings (vectors) that capture semantic meaning, making similarity search and retrieval much more efficient.
  • Embedding models convert texts, images, and other data into these vectors, forming the backbone for AI systems that need to understand and retrieve contextual information quickly.

Full Transcript

# Seven Essential AI Terms Explained **Source:** [https://www.youtube.com/watch?v=VSFuqMh4hus](https://www.youtube.com/watch?v=VSFuqMh4hus) **Duration:** 00:11:03 ## Summary - AI is now ubiquitous—from everyday objects like toothbrushes receiving updates to rapid advancements that even tech professionals find hard to track. - “Agentic AI” refers to autonomous AI agents that perceive their environment, reason about next steps, act on plans, and observe outcomes, enabling roles such as travel booking, data analysis, or DevOps automation. - Large Reasoning Models are specially fine‑tuned LLMs that generate step‑by‑step “chain‑of‑thought” processes, allowing agents to tackle complex, multistep tasks with verifiable accuracy. - Vector databases store information not as raw files but as numerical embeddings (vectors) that capture semantic meaning, making similarity search and retrieval much more efficient. - Embedding models convert texts, images, and other data into these vectors, forming the backbone for AI systems that need to understand and retrieve contextual information quickly. ## Sections - [00:00:00](https://www.youtube.com/watch?v=VSFuqMh4hus&t=0s) **Agentic AI: Autonomous Reasoning Loop** - The speaker explains how agentic AI agents autonomously perceive, reason, act, and observe to accomplish tasks such as travel planning, data analysis, and DevOps engineering. - [00:03:07](https://www.youtube.com/watch?v=VSFuqMh4hus&t=187s) **Embedding Vectors and Retrieval Augmented Generation** - The speaker explains how an embedding model converts media such as images into numeric vectors, allowing similarity searches in a vector database, which are then leveraged by Retrieval Augmented Generation to enrich LLM prompts. - [00:06:17](https://www.youtube.com/watch?v=VSFuqMh4hus&t=377s) **Standardized MCP Connections & MoE Scaling** - The speaker explains that an MCP server offers a uniform interface for AI to access various tools, then outlines how mixture‑of‑experts models activate only the necessary specialized subnetworks to efficiently scale language models, citing IBM Granite 4.0 as an example. - [00:09:22](https://www.youtube.com/watch?v=VSFuqMh4hus&t=562s) **Promoting AI at IBM TechXchange** - The speakers discuss adding ASI to their AI term list, invite audience suggestions, and highlight upcoming AI workshops, demos, and live light‑board sessions at IBM’s TechXchange conference in Orlando. ## Full Transcript
0:00There are two things that hold true when it comes to artificial intelligence. One. 0:04It's everywhere. 0:06My toothbrush just got an AI update this week. 0:09And two. The field is changing rapidly, 0:11making it hard to keep up even for those of us who work in tech. 0:14So, I've put together 0:16my top seven AI terms 0:18that I think are important to be familiar with 0:21as AI continues to progress. How many do you already know? 0:25Well, let's find out. 0:26And I'm gonna to start at number one 0:30with something that I'm quite sure that you have heard of. 0:33And that's agentic AI. 0:36Everybody and their grandmother seems to be building 0:38the next generation of AI agents. 0:40But what exactly are they? 0:43Well, AI agents, they can reason 0:45and act autonomously to achieve goals. Something 0:48like a chatbot that only responds one prompt at a time, 0:51AI agents, they run autonomously. 0:54They go through a number of stages. 0:56So, first of all, they perceive their environment. 1:00Once they've done that, they move on to a reasoning stage, 1:05and that's where they look to see what the next best steps forwards are. 1:10Then they move on to act on the plan that it's built through the reasoning, 1:15and then observes the results of that action. 1:20And around and around we go. 1:23Now agents can work as well in all sorts of roles. 1:27They could be your travel agent to book a trip. 1:30They could be your data 1:32analyst to spot trends in quarterly reports. 1:35Or they could perform the role of a DevOps engineer, 1:39detecting anomalies in logs and spinning up containers 1:41to test fixes and rolling back faulty deployments. 1:44And AI agents are typically built 1:47using a particular form of large language models, 1:49and that is known as number 1:51two, large reasoning models. 1:54Now these are specialized LLMs that have undergone 1:57reasoning-focused fine tuning. So unlike regular 2:00LLMs that generate responses immediately, 2:03reasoning models, they're trained to work 2:06through problems step by step, 2:10which is exactly what agents need 2:12when planning complex, multistep tasks. 2:14Now, the reasoning model is trained on problems with verifiably correct answers. 2:19So math problems or code that can be tested by compilers 2:23and through reinforcement learning, the model learns to generate 2:27reasoning sequences that lead to correct final answers. 2:31So, every time you see a chatbot pause 2:34before it responds back to you by saying, thinking. Well, 2:38that's the reasoning model at work, 2:40generating an internal chain of thought to break down 2:43a problem step by step before generating a response. 2:46Now let's get a bit lower level 2:49and talk about number three, 2:52which is vector database. 2:55So, in a vector database, we don't store 2:59raw data like text files 3:02and like images just as blobs of data. 3:07We actually use something called an embedding model. 3:12And that embedding model is used to convert that data 3:16from these images here 3:18into actually a vector. 3:21Now, what is a vector? Well, 3:23a vector is essentially just kind of a 3:26a long list of numbers. 3:28And that long list of numbers captures 3:30the semantic meaning of the context. Now, 3:33what's the benefit of doing that? 3:36Well, in a vector database, 3:38we can perform searches as mathematical operations, 3:40looking for vector embeddings 3:43that are close to each other. 3:45And that translates to finding semantically similar content. 3:49So, we might start, with 3:51let's say a picture of a mountain vista. Something like this. 3:56And then that picture is broken down by the embedding model 4:01into vectors, 4:04a multidimensional numeric vector. And 4:06we can perform a similarity search 4:09to find items that are similar to that mountain picture 4:14by finding the closest vectors 4:16in the embedding space. 4:18Or it could be similar text articles, or it could be similar music files. 4:23Whatever you want. 4:24Now vector databases, they play a big role in implementing number four. 4:29And that is RAG or retrieval augmented 4:32generation. 4:34Now, RAG makes use of these vector databases. 4:40And it uses it to enrich prompts to an LLM. 4:44So, we start here with a RAG retriever component. 4:49Now that might take in an input prompt 4:52from a user. 4:54And it's going to turn it into a vector 4:58using an embedding model. 5:02That's the thing that ties it into that series of numbers. 5:05And then, once we've done that, we can perform 5:07a similarity search in the vector database. 5:11Now that vector database will return something, 5:14and we'll return that all the way back to the large language 5:18model prompt that we started with. 5:21And we'll embed into that prompt 5:22now the stuff that came out of that vector database. 5:26So, I can I can ask a question about let's say company policy. 5:29And then this RAG system is going to pull the relevant section from the employee 5:33handbook to include in the prompt. 5:35Now, number five, 5:38that Model Context Protocol or MCP. 5:43This is a really exciting one because for large language models to be truly useful, 5:47they need to interact with external data sources and services and tools. 5:51And MCP standardizes how applications provide context to LLMs. 5:56So, if you want your large language model 6:00here to be able to connect to stuff. 6:03Perhaps we want to connect to an external database, 6:08or maybe we want to go to some kind of code repository, 6:11or maybe even to an email server, 6:15or really any kind of external system. 6:17Well, MCP makes that connection standardized. So, 6:21instead of developers having to build one off 6:24connections for each new tool, 6:26MCP provides a standardized way for AI to access your systems. 6:30So basically we have here an MCP server. 6:35And that is how the AI knows exactly what to do 6:39to get through to any one of these tools. 6:41It connects through that MCP server connection. Okay. 6:44Now, for number six. 6:48That's a mixture of experts or MOE. 6:52And we've had the idea of MOE 6:54for a good while, actually, since the paper was published 6:56in a scientific journal in 1991. 6:58But, it's never been more relevant than it is today. 7:02You see, MoE divides a large language model 7:05into a series of experts. 7:08I'm just gonna to draw three, but there could be 100 plus of these. 7:13These are specialized neural subnetworks. 7:17And then it uses a routing mechanism to activate 7:22only the agents it needs for a particular task, 7:26or only the experts in this case that it needs for a task. 7:29And then, well, then it's going to perform a merge process. 7:34So, because we activated these two experts, 7:36we'll merge these two. 7:38And this performs mathematical operations 7:40to combine the output 7:42from these different experts into a single representation 7:45that continues through the rest of the model. 7:47And it's a really efficient way to scale up model size 7:52without proportional increases in compute costs. 7:54So, for example, MoE models, like IBM Granite's 4.0 series, 7:59that can have dozens of different experts here. But for any given token, 8:04it will only activate these specific experts it needs. 8:07And that means, though, while the whole model 8:09might have billions of total parameters, 8:12it only uses a fraction of those active parameters at inference time. 8:17And look, for number seven, 8:21I'm gonna throw in a big one, ASI, 8:24artificial superintelligence. 8:27It's the goal of all the frontier AI labs. 8:29And at this point, it is purely theoretical. 8:32It doesn't actually exist and we don't know if it ever will. 8:35Now, today's best models, 8:37they're slowly approaching a different standard, which is AGI. 8:42That's artificial general intelligence. 8:45Now that's also theoretical. But, 8:47if realized, AGI will be able to complete 8:50all cognitive tasks as well as any human expert. 8:54ASI is one step beyond that. 8:56So, ASI systems would have an intellectual scope 9:00beyond human level intelligence, 9:02potentially capable of recursive self-improvement. 9:05So, basically an ASI system could redesign and upgrade itself, 9:10becoming ever smarter in an endless cycle. 9:13It's the kind of development that would either solve humanity's biggest problems 9:18or create entirely new ones that we can't even imagine yet. 9:22And if that happens, well, 9:24I think it's probably a pretty good idea 9:26that we keep the term ASI on our radar. 9:30So, that's my seven. 9:32But I'm curious, what's the AI term 9:35you think that should have made it onto this list? 9:38Let me know in the comments. 9:40Hey, Martin. Hey Graeme. 9:42Hey, this is really cool stuff. 9:45This AI and these terms. Fascinating. 9:48Yeah, and I came up with seven, but I could have come up with 70. 9:50There's so much going on in this space. I I I bet you could. 9:53And you know what? There is so much going on. 9:55We are actually going to be talking about AI a lot at the IBM 9:59TechXchange conference in Orlando this October. 10:02And you know what? I'm gonna be there as well. 10:04I know it's gonna be so exciting! 10:05There's going to be so much going on. 10:07We are going to have. Let's see. 10:08We're gonna have boot camps. 10:10We're going to have workshops. There's going to be sessions. 10:13There's going to be live demos, certifications, all kinds of things going on. 10:18So much more when it comes to AI. 10:20But but Martin, what are you going to be doing there? 10:22Well, I'm going to be bringing my light board pens 10:25and this light board to the sessions as well. 10:27Oh my God! It's so exciting! I'm so excited have you there! 10:30Yes. So we're actually going to have a light board studio set up. 10:33And ah we're going to be performing light boards live. 10:35So if you always wondered how do I write backwards, 10:38you're going to find out in person at the event. 10:41And da also we'll be kind of teaching 10:43how to perform a light board 10:45video yourself. The sort of things that you need to know for that. 10:48Wow. So you get to meet a celebrity and maybe become one yourself. 10:51That sounds really exciting. 10:52I can't wait to welcome you down to Orlando. 10:55It's just going to be a blast. So looking forward to it. Can't wait. 10:58All right. Hope we see you there too. 10:59So, go to ibm.com/techXchange and we'll see you down there.