Learning Library

← Back to Library

Machine Learning: AI Hierarchy and Types

Key Points

  • Machine learning (ML) is a subset of artificial intelligence (AI) that uses algorithms to learn patterns from training data and make predictions on new, unseen data, while deep learning (DL) is a further subset of ML that employs multi‑layered neural networks.
  • The core process of ML involves training a model on a representative dataset so it can perform accurate inference—running the trained model on fresh inputs to generate predictions.
  • ML learning paradigms fall into three main categories: supervised learning (using labeled data), unsupervised learning (discovering structure in unlabeled data), and reinforcement learning (optimizing behavior through trial‑and‑error rewards).
  • Supervised learning includes regression models that predict continuous values (e.g., linear and polynomial regression) and classification models that predict discrete categories (e.g., binary fraud‑vs‑legit classification).
  • Unsupervised techniques such as clustering, dimensionality reduction, and anomaly detection enable models to uncover hidden patterns without human‑provided labels.

Full Transcript

# Machine Learning: AI Hierarchy and Types **Source:** [https://www.youtube.com/watch?v=znF2U_3Z210](https://www.youtube.com/watch?v=znF2U_3Z210) **Duration:** 00:10:27 ## Summary - Machine learning (ML) is a subset of artificial intelligence (AI) that uses algorithms to learn patterns from training data and make predictions on new, unseen data, while deep learning (DL) is a further subset of ML that employs multi‑layered neural networks. - The core process of ML involves training a model on a representative dataset so it can perform accurate inference—running the trained model on fresh inputs to generate predictions. - ML learning paradigms fall into three main categories: supervised learning (using labeled data), unsupervised learning (discovering structure in unlabeled data), and reinforcement learning (optimizing behavior through trial‑and‑error rewards). - Supervised learning includes regression models that predict continuous values (e.g., linear and polynomial regression) and classification models that predict discrete categories (e.g., binary fraud‑vs‑legit classification). - Unsupervised techniques such as clustering, dimensionality reduction, and anomaly detection enable models to uncover hidden patterns without human‑provided labels. ## Sections - [00:00:00](https://www.youtube.com/watch?v=znF2U_3Z210&t=0s) **Machine Learning, AI, and Deep Learning Explained** - The speaker outlines how AI encompasses machine learning, which in turn includes deep learning, and describes model training and inference. - [00:05:23](https://www.youtube.com/watch?v=znF2U_3Z210&t=323s) **Clustering Techniques and Dimensionality Reduction** - The speaker explains k‑means and hierarchical clustering for segmenting data and introduces dimensionality‑reduction methods that compress feature spaces while preserving essential information. ## Full Transcript
0:00What is machine learning? Well, I think you have a basic idea. Right? It's the tech that 0:07recommends which video to watch next on YouTube. The machine is learning your preferences 0:14and its bot powers chatbots. But is ML or machine 0:20learning synonymous with AI? Or what about deep 0:27learning? DL. Are deep learning and machine learning the same thing? Well, no, 0:33it's a hierarchy. So, machine learning is a subset of 0:40AI that focuses on algorithms that can learn the patterns of training data and then make 0:47accurate inferences about new, unseen data. Essentially, the machine is learning pattern 0:53recognition that lets models make decisions or predictions without explicit hard-coded 0:58instructions. And then deep learning is a subset of machine learning that uses neural 1:05networks with many layers to learn hierarchical representations. So, ML sits inside 1:12AI and then DL sits inside ML. Now the central premise of 1:18machine learning is that if you optimize the machine's performance on a dataset of tasks that 1:25adequately resemble the real world, that's through a process called model training. 1:32Well, if you train the model in such a way, that model can make accurate predictions on new 1:38data. In essence, a trained model is applying patterns. It's learned from training data to infer 1:44the correct output for a real-world task. And then the deployment of this fully trained 1:50model, that is called AI inference. this is when we actually run the model. 1:57That's when we feed new data to this trained model. And it makes predictions based on the 2:03patterns it's learned. Now most machine learning is grouped into three learning 2:09paradigms. So there is supervised learning, that trains a model to predict the 2:15correct output for an input using labeled examples or ground truth. For instance, we might 2:21have a bunch of emails and they've been labeled spam or not spam. It's supervised because it 2:27generally requires a human in the loop to provide the ground truth data. Then there is 2:34unsupervised learning, and that uses unlabeled data to 2:40discover structure on its own. So that's the tasks like clustering and dimensionality reduction and 2:47anomaly detection. And then there is also reinforcement learning, which optimizes a 2:54policy through trial and error with rewards and penalties. Now there are several types of 3:00supervised machine learning models. Let's take a look at some of them. So one type of 3:06model is called a regression model. And that predicts continuous 3:13numerical values so think like price prediction or think temperature forecasting. Now there 3:20are different types of regression. So one type of regression is called 3:27linear regression that finds the best fit line through data points. Another 3:34type is called polynomial regression. And that captures 3:40nonlinear relationships. So that's regression. There's also 3:46classification. That's another type of supervised learning. And that predicts 3:52discrete classes. So one type of classification might be binary 3:59classification. So fraud or legit it's gonna be one or the other. It might be 4:06multi-class classification. That's where we have one 4:12of many categories. Or it might be multi-label 4:18classification as well. That's where we have multiple tags all at once. And modern 4:25supervised learning. It actually frequently uses a combination of this called ensemble 4:31methods which means combining models for better accuracy. Oh, and before we get to 4:38unsupervised learning, I should mention there's something that kinda sits between here called 4:44semi-supervised learning that sits between these two because you train with a small 4:50labeled data set plus a large unlabeled pool to improve the supervised model. it 4:57uses the labeled examples to generalize over the unlabeled data. So you need far fewer costly 5:03labels. All right. Now unsupervised learning that works with unlabeled data to uncover structure on 5:10its own. So let's look at two common families. One of those families is called 5:16clustering. Now clustering groups similar items so things that behave alike end up together. 5:23A well-known clustering method is called k-means clustering, so you 5:30choose k-groups. Repeatedly assign each item to the nearest group average. Then you recompute the 5:36averages until they're stable. So for example, you could split customers into k equals four segments. 5:42And those four segments might be ah bargain hunters, loyal repeaters, big ticket buyers and just 5:47browsers. And then you could target offers to each of those groups. Now another clustering method is 5:54called hierarchical. And hierarchical is where we start with every item by 6:01itself. Then keep merging the most similar groups to build a tree, and you cut the tree later for 6:07let's say, three or 5 or 12 clusters. So for example, you could cluster IT ticket text. So 6:13that's the subject and the description of the ticket. You could cluster those into themes like 6:18like password reset and laptop won't boot. And you could get broad buckets, or you can have 6:25fine-grained sub-buckets, depending on where you cut the tree. And that's useful for autorouting 6:30and spotting new service issue types. Now another type of unsupervised learning is 6:37called dimensionality reduction. Now dimensionality 6:43reduction are algorithms that reduce the complexity of data points by representing them 6:49with the smaller number of features that is fewer dimensions while still kinda keeping all of the 6:55meaningful characteristics. And they're often used for preprocessing data, as well as for tasks such 7:00as data compression or data visualization. Now, some examples of dimensionality reduction are 7:07PCA. That's principal component analysis. And we also have 7:13encoders as another example. Then there's reinforcement learning. So an 7:20agent interacts with an environment. And it does that by observing its state. 7:27That's what it sees now, and then by choosing an action 7:34to take. So what to do next. And then the environment either will 7:40reward that action, which is to say it moves correctly, ah or 7:47it will punish that action with a penalty for the incorrect 7:54action. And over time, the interactions with the agent teaches it a policy that 8:01maximizes long-term rewards because rewards can be delayed. An agent might also have to balance 8:06exploration of trying new actions with exploitation, which is repeating what works. So as 8:12an example, consider a self-driving car. So the state that comes from the from the 8:19GPS and the cameras, and well, for most cars the lighter as well. And the actions 8:26are to steer and to break and to accelerate and the model rewards the safe and 8:33move ah the smooth progress like, well, staying in its lane and obeying traffic signals. it will 8:40penalize things like hard braking, and it will heavily penalize the really bad stuff like 8:46collisions. The the rewards and the penalties help the model learn. Now, most of what 8:53we've covered here regression, classification, clustering and reinforcement learning that stuff 8:57all kind of falls under what you might call classic machine learning. These techniques have 9:02been around for years and are still everywhere in business today, from predicting prices to 9:08segmenting customers. But today, we're also seeing these old ideas applied in new 9:15ways. And of course, the most famous example of that is LLMs or large 9:21language models. And those are built on top of something called a 9:27transformer. And these transformer architectures are a newer neural network design, but they still 9:33rely on the same ML principles of pattern recognition and data and model training and 9:39inference. And in fact, even reinforcement learning has made a bit of a comeback with 9:45RLHF. That's reinforcement learning with human feedback. So instead of teaching an 9:52agent to say, play a game, RLHF trains LLMs to better align with human 9:59preferences. So human annotators, they provide rewards and penalties on model outputs, which fine 10:05tunes the system's behavior. So, while the buzz today is all around LLMs and generative AI and 10:11agentic AI, the foundations are still classic ML concepts. just scaled up 10:18and combined in new ways, which just goes to show that human learning keeps finding new ways to 10:25apply machine learning.