Learning Library

← Back to Library

Five Types of AI Agents

Key Points

  • In 2025 the AI community is saturated with “agentic” breakthroughs, but true progress requires understanding the different levels of agent intelligence rather than just hype.
  • AI agents are categorized by how they process information and act on their environment, with five main types ranging from simple reflex to advanced learning agents.
  • Simple reflex agents operate on fixed condition‑action rules (e.g., a thermostat) and work well in static, predictable settings but fail when rules are insufficient or past context is needed.
  • Model‑based reflex agents extend simple reflexes by maintaining an internal model of the world that updates with new perceptions, allowing them to remember past states and make more informed decisions in dynamic environments.

Full Transcript

# Five Types of AI Agents **Source:** [https://www.youtube.com/watch?v=fXizBc03D7E](https://www.youtube.com/watch?v=fXizBc03D7E) **Duration:** 00:10:09 ## Summary - In 2025 the AI community is saturated with “agentic” breakthroughs, but true progress requires understanding the different levels of agent intelligence rather than just hype. - AI agents are categorized by how they process information and act on their environment, with five main types ranging from simple reflex to advanced learning agents. - Simple reflex agents operate on fixed condition‑action rules (e.g., a thermostat) and work well in static, predictable settings but fail when rules are insufficient or past context is needed. - Model‑based reflex agents extend simple reflexes by maintaining an internal model of the world that updates with new perceptions, allowing them to remember past states and make more informed decisions in dynamic environments. ## Sections - [00:00:00](https://www.youtube.com/watch?v=fXizBc03D7E&t=0s) **Classifying AI Agents: Reflex Explained** - The passage outlines the hype around 2025 AI agents, then breaks down the simplest type—a reflex agent—showing how it uses sensor input and predefined condition‑action rules to act in its environment. - [00:03:22](https://www.youtube.com/watch?v=fXizBc03D7E&t=202s) **From Model-Based to Goal-Based AI** - The passage explains how agents use internal models of the environment to predict action effects and then extend this to goal-oriented decision making, illustrated with a robotic vacuum and a self‑driving car. - [00:06:29](https://www.youtube.com/watch?v=fXizBc03D7E&t=389s) **Utility and Learning AI Agents** - The passage contrasts a utility‑based drone that evaluates routes by scoring them for speed, safety, and energy efficiency with a learning agent that employs reinforcement‑learning components—critic, learner, problem generator, and performance module—to continuously improve its actions from environmental feedback, exemplified by an AI chess bot. - [00:09:34](https://www.youtube.com/watch?v=fXizBc03D7E&t=574s) **Cooperative AI Agents with Humans** - The passage explains how multiple AI agents collaborate in shared environments using generative AI, yet still rely on human oversight. ## Full Transcript
0:00In the world of AI, it seems that 2025 is the year of the AI Agent. 0:08New agentic workflows and models are released all the time, 0:12often accompanied by breathless declarations on social media that a task that previously required human expertise 0:20has now been entirely automated by the latest agentic breakthrough. 0:26But can you distinguish a simple reflex agent from an advanced learning agent? 0:33You see agents are classified based on their level of intelligence, 0:37based on their decision-making processes and how they interact with their surroundings to reach wanted outcomes. 0:43So let's explore the five main types of AI agents to understand what they can and cannot do. 0:51Now a simple reflex agent that is the most simple type of AI agent, 1:00the most basic type, and it follows predefined rules to make decisions like a thermostat. 1:07It turns on the heat when the temperature drops below a 1:09predefined threshold and then it turns it off again when a set temperature is reached. 1:14So let's break it down. 1:15We've got our agent here. 1:18Now the the environment over here that's the external world that the agent is embedded into and it next to. 1:28Then we've got precepts. 1:30These are the perceived input from the environment as measured through sensors. 1:34Then these sensors feed the precept into the internal logic of the agent 1:41which gives us a representation of what the world is like now, 1:46and knowing what the word is like now we can use condition action rules as the core logic of these simple reflex agent. 1:55Now, these are rules that follow an if condition. 1:58Then action structure. 2:00So if the temperature drops to 18 Celsius then turn on the heat. 2:06That's executed by actuators and that results in an action. 2:12The output behavior by the agent and that action affects the 2:18environment which in turn affects the next set of precepts and around and around we go. 2:26Simple reflex agents like this are effective in structured 2:30and predictable environments where the rules are well defined, but dynamic scenarios? 2:35They can trip these agents up, 2:37and because they don't store past information, they can repeatedly make the same mistakes 2:43if the predefined rules are insufficient for handling new situations. 2:50All right, well how about this one? 2:52This is called a model based reflex agent. 2:59So this is a more advanced version of the the simple reflex agent, 3:04and it uses condition action rules to make decisions as well but, 3:09it also incorporates an internal model of the world and that's stored in the state 3:15component and that state component is updated by observing how the world actually evolves. 3:23Essentially how the environment changes from one state to another. 3:29The agent also tracks how its own actions affect the environment. 3:34That's what my actions do. 3:36And all of this is used instead of just taking the raw precepts data for decision making. 3:42So take a robotic vacuum cleaner for example. 3:45The internal state that remembers where it's been and what areas are clean and where the obstacles are. 3:52It knows that if it moves forwards, it changes its location and that action has consequences. 3:58And it has condition-action rules, like if I think I'm in a dirty area and I haven't cleaned it yet, then vacuum it. 4:08It doesn't just react to what it immediately sees, it infers and it remembers parts of the environment it can't currently observe. 4:16That's model-based reasoning in action. 4:21now a goal-based AI model that is building on top of the model-based agent by adding decision-making that's based on goals. 4:35So we don't have any more condition action rules, we have goals, 4:39and they represent the desired output the agent is trying to achieve. 4:43So the agent uses its model, that's how the world evolves and what my actions do, 4:50to simulate future outcomes of possible actions, essentially predicting what will it be like if I do action A. 4:59Now that's a shift in decision making. 5:01The agent isn't just asking what action matches this condition, 5:05it's now asking what actually will help me achieve my goal based on the current state and predicted future. 5:12So consider a self-driving car. 5:14If the goal is to get to destination X, It'll consider its state, which is, I'm on Main Street. 5:22It will then generate a prediction. 5:23If I turn left, I'll head towards the highway, and it'll ask, will that help me reach destination X? 5:29And if the answer is yes, then the action will be to turn left. 5:35Goal-based agents are widely used in robotics and simulations where 5:39a clear objective is set and adaptation to the environment is required. 5:44Now a utility-based agent looks like this. 5:51And it considers not just if a goal is met, but how desirable different outcomes are. 5:57So utility here represents a happiness score or a preference value for a particular outcome. 6:03So for each possible future state, the agent asks how happy will I be 6:08in such a state or really the expected utility of the future state. 6:14And this lets it rank options, not just pick anything that meets the goal. 6:19So consider an autonomous drone delivery. 6:22The goal-based version might be to use a goal of 6:25deliver the package to address X, and it chooses an action that completes that goal. 6:30Doesn't matter if it gives you a bumpy energy-wasting route, 6:33but a utility-based person, 6:36that might instead be something like deliver the packages quickly and safely and with minimum energy usage, 6:42whereby now the drone simulates multiple paths, it estimates things like duration and battery level and weather, 6:48and it picks the route that maximizes its utility score. 6:52That's AI agent number four. 6:55Now, the fifth agent is the most adaptable and also the most powerful and it is the learning agent. 7:04So rather than being hard coded or being goal driven, it learns from experience. 7:09It improves its performance over time by updating its behavior based on feedback from the environment. 7:16So how does it work? 7:17There's a critic component and that observes the outcome of an agent's actions 7:22via the sensors and it compares them to a performance standard. 7:27Now that gives us a numerical feedback signal that's often called a reward in reinforcement learning 7:33and this reward is then passed to a learning element 7:37that updates the agent's knowledge using the feedback from the critic. 7:41Its job here is to improve the agents mapping from states all the way through to actions. 7:48Now the problem generator, that suggests new actions the 7:51agent hasn't tried yet, like try a different path, see if it's any faster. 7:56And then the performance element selects actions based on what the learning element has determined to be optimal. 8:03So think of an AI chess bot, the performance elements that plays the game using current learn strategies. 8:10The critic, you'll see that it lost the match. 8:13The learning element adjusts its strategy based on the outcomes of 8:17thousands of games and the problem generator suggests new moves that it hasn't explored yet. 8:23So a simple reflex agent reacts. 8:29It's fast to execute but it has no memory and it has no understanding of history. 8:35A model-based reflex agent, we can really think that the difference there is that that remembers. 8:43It does that by tracking state over time. 8:45It doesn't plan, it's still reactive. 8:48Now a goal-based model, that aims. 8:52It aims by using goal-directed behavior but any way of meeting that goal... 8:58Will do. 8:59Whereas, an utility-based agent that takes a different path, it evaluates. 9:06It does that by choosing the best outcome, but requires an accurate utility function to do so. 9:13And then a learning agent that improves by learning from experience, 9:20but this can be the slowest and most data intensive process. 9:24Now in many cases, we will want to use multiple agents together. 9:30That is called a multi-agent system. 9:35And that's where multiple agents operate in a shared environment, 9:39working in a cooperative way, working towards a common goal. 9:43And as agentic AI continues to evolve, particularly with learning agents that are making uses of advances in generative AI, 9:51AI agents are becoming increasingly adept at handling complex use cases, 9:58but it's not really all over for us just yet. 10:02AI agents typically work best with a good old human in the loop. 10:07At least for the time being.