Accéder au contenu principal

Types of AI: A Comprehensive Guide

Explore the spectrum of artificial intelligence, from discriminative AI and predictive models to autonomous agents and the agentic framework.
29 déc. 2025  · 15 min lire

If it feels like Artificial Intelligence (AI) is suddenly everywhere, that’s because it is. According to recent market research by McKinsey, the global AI adoption has grown from around 50–60% of organizations in 2020–2021 to a staggering 88% in 2025.

However, this explosive growth has created a linguistic problem. "AI" has become a catch-all buzzword used to describe vastly different technologies. It is used interchangeably for simple rule-based chatbots, image generators, and futuristic sci-fi robots.

For data practitioners, this ambiguity is a roadblock. When a stakeholder asks for an "AI solution," you should know: do they need a simple regression model, a generative agent, or a computer vision system?

In this comprehensive guide, we’ll look at this terminology. We won't just list definitions; instead, we’ll look at it through four distinct lenses: how it is built, what it can do, how it decides, and how it fits into real-world workflows.

If you’re new to the topic, I recommend taking the Understanding Artificial Intelligence course.

What is AI?

Artificial Intelligence (AI) is a subfield of computer science focused on creating intelligent agents capable of performing tasks that typically require human intelligence, such as problem-solving, speech recognition, and decision-making.

AI is an interdisciplinary science with many approaches. It can be rule-based (operating under predefined conditions) or use machine learning algorithms to learn from data, allowing systems to adapt to unforeseen scenarios.

However, defining AI precisely proves surprisingly difficult due to what researchers call the "AI Effect." This phenomenon describes how society's perception of AI constantly shifts as technologies mature. When a capability becomes commonplace, people stop calling it "AI" and start viewing it as ordinary software. 

Because of this shifting definition and broad scope, a single classification isn’t enough. We need multiple lenses to get the full picture:

  • Technology: The underlying algorithms and architectures. 
  • Capability: Measuring intelligence level from narrow to superintelligent.
  • Functionality: How do systems process information and maintain memory?

Types of AI, divided by capabilities, functionalities and technologies.

For example, a self-driving car can be classified in three ways simultaneously:

  • Narrow AI (based on its specific capability).
  • Deep Learning (based on its underlying technology).
  • Limited Memory (based on its functionality).

In the following sections, we’ll explore these specific types.

For a strictly non-technical primer on the basics, I recommend you read our AI Guide for Beginners.

Technology-Based Types of AI

The most rigorous way to classify AI is by the underlying technology, the "engine" that drives the system. This classification separates systems based on how they process data and learn. Let’s see those different types.

Machine learning

Machine Learning (ML) is a subset of AI focusing on systems that learn from data and improve without being explicitly programmed for every single rule. Instead of writing code that says "if this, then that," you feed the algorithm data and let it find the patterns.

Machine learning divides into three primary learning paradigms, each suited to different problem types:

  • Supervised Learning: The most common form in business. It uses labeled data (input-output pairs) to train algorithms. For example, you might show the model 1,000 emails labeled "Spam" or "Not Spam," and it learns to classify the next email. Common algorithms include Linear Regression, K-Nearest Neighbors (KNN), and Decision Trees.
  • Unsupervised Learning: The algorithm is given unlabeled data and must find hidden structures or patterns on its own. This is often used for customer segmentation (clustering). K-Means Clustering and Principal Component Analysis (PCA) are among the most commonly used algorithms.
  • Reinforcement Learning (RL): Taking a different path entirely, RL teaches agents through trial and error using a reward-and-penalty system. An RL agent receives feedback on its interaction with the environment, and gradually learns which behaviors maximize cumulative rewards over time. This is the method used to train robots and advanced game-playing AI, such as AlphaGo.

Deep Learning is a subset of Machine Learning, which is a subset of Artificial Intelligence.

To get a good foundation in machine learning, I recommend you work through the Machine Learning Fundamentals with Python skill track.

Deep learning

Deep learning (DL) is a specialized subset of ML inspired by the human brain's structure. It uses artificial neural networks with many layers (hence "deep") to model complex, non-linear relationships in vast datasets. While traditional ML might plateau with more data, DL often continues to improve.

Key architectures include:

  • Convolutional Neural Networks (CNNs): The gold standard for image processing. They use "filters" to automatically scan images and detect spatial hierarchies like edges, shapes, and textures.
  • Recurrent Neural Networks (RNNs): Designed for sequential data like time-series or speech. They maintain a "memory" of previous inputs to influence the current output.
  • Transformers: The architecture behind modern Large Language Models (LLMs), such as GPT 5.2 and Gemini 3. It utilizes an attention mechanism to process entire sequences of data simultaneously, weighing the importance of different words relative to each other rather than processing them one by one.
  • Generative Adversarial Networks (GANs): A framework where two networks compete, a "generator" creating fake data, and a "discriminator" trying to detect the fakes as such. This competition results in highly realistic synthetic outputs.
  • Diffusion Models: The engine behind tools like Midjourney. They generate high-quality images by learning to systematically reverse-engineer random noise into clear, visually distinct images.

If you are looking to build these models yourself, I recommend you start with the Introduction to Deep Learning with PyTorch course.

Natural language processing

Natural language processing (NLP) focuses on enabling computers to understand, interpret, and generate human language in meaningful ways. NLP bridges the gap between human communication and machine comprehension by handling the ambiguity, context-dependence, and complexity inherent in natural language.

Techniques include the following: 

Let's see an example of a sentiment analysis using transformers in Python:

# Example: Sentiment analysis using transformers
from transformers import pipeline
# Load pre-trained sentiment analysis model
sentiment_analyzer = pipeline("sentiment-analysis")
# Analyze multiple texts
texts = [
    "This product exceeded my expectations!",
    "Terrible customer service, very disappointed.",
    "The quality is okay, nothing special."
]
results = sentiment_analyzer(texts)
for text, result in zip(texts, results):
    print(f"Text: {text}")
    print(f"Sentiment: {result['label']}, Confidence: {result['score']:.2f}\n")
Text: This product exceeded my expectations!
Sentiment: POSITIVE, Confidence: 1.00

Text: Terrible customer service, very disappointed.
Sentiment: NEGATIVE, Confidence: 1.00

Text: The quality is okay, nothing special.
Sentiment: NEGATIVE, Confidence: 0.98

Computer vision

Computer vision enables machines to derive meaningful information from digital images and videos, allowing computers to perceive and understand visual data in a way similar to human perception. This field deals with identifying objects, measuring distances, or even recognizing patterns invisible to human observers.

Computer vision systems combine image processing techniques with machine learning, particularly deep learning, to extract increasingly abstract representations from raw visual data. Key tasks include:

Medical imaging represents one of computer vision's most impactful applications, with AI systems detecting tumors in X-rays, CT scans, and MRI scans with accuracy levels that match or exceed those of human radiologists. These systems don't replace medical professionals but support them, for instance, by flagging potential abnormalities for expert review.

Audio processing

Audio processing concentrates on recognizing, interpreting, and synthesizing auditory signals, including both human speech and environmental sounds. The goal is to enable machines to listen, understand, and generate audio content, bridging the gap between acoustic waves and semantic meaning. 

Automatic Speech Recognition (ASR) converts spoken language into written text, and powers voice assistants, transcription services, and accessibility tools. Modern ASR systems use transformer architectures to directly map audio waveforms to text sequences, achieving near-human accuracy on clean speech and steadily improving on challenging conditions like background noise or accented speech. 

Text-to-Speech (TTS) engines like Amazon Polly perform the inverse operation, synthesizing natural-sounding speech from written text using neural vocoders that generate realistic prosody, intonation, and emotional expression. Contemporary TTS systems can clone voices from just a few minutes of sample audio, raising both creative possibilities and ethical concerns.

Applications extend beyond speech processing into creative and practical domains. Voice cloning technology enables personalized virtual assistants, audiobook narration with consistent voices, and controversial deepfake audio.

Robotics and embodied AI

This represents the intersection of AI and engineering. Embodied AI involves putting advanced "brains" (like vision-language models) into physical robot bodies. While much of AI exists purely in digital form, embodied AI must deal with gravity, friction, and the unpredictable complexity of physical environments. 

This grounding in reality introduces challenges that are absent from purely computational AI, but also enables direct interaction with the physical world. Traditional robots required explicit programming for each task, with engineers manually coding precise movements and decision trees to achieve specific outcomes. 

Modern embodied AI systems understand natural language commands, such as "pick up the apple" or "open that door," using visual perception to identify objects and plan appropriate actions without requiring task-specific programming. 

This advancement addresses Moravec's Paradox, the observation that tasks trivial for humans, like folding laundry, opening doors, and navigating cluttered rooms, prove extraordinarily difficult for robots, while tasks humans find challenging, like chess or calculus, are comparatively easy for AI. 

Embodied AI, using foundation models, is finally bridging this gap by providing robots with the perceptual understanding and adaptive planning that humans take for granted. Applications include consumer devices like robotic vacuums, industrial collaborative robots (cobots), and advanced humanoids such as Boston Dynamics' Atlas.

Capability-Based Types of AI

While technology defines how an AI is built, capability defines what it can do relative to human intelligence. This classification serves as a timeline of AI evolution, moving from current limitations to theoretical futures.

Artificial narrow intelligence

Artificial Narrow Intelligence (ANI), also known as Weak AI, represents the artificial intelligence that exists today. It means systems that are designed to excel at specific, predefined tasks but fail completely if tasked with anything outside their narrow training parameters. A chess bot cannot drive a car, and a medical diagnostic tool cannot write poetry.

However, the "narrow" definition is expanding. The current state-of-the-art involves advanced multimodal Large Language Models (LLMs). We are seeing the release of models like GPT-5.2 and Gemini 3, which can process text, audio, and visual inputs simultaneously. 

While they mimic reasoning, they still lack a genuine understanding of the world. Developers interact with these ANI systems via APIs to integrate them into applications.

The next step within ANI is Agentic AI. These are systems that move beyond passive chat; they can autonomously perform multi-step actions like browsing the web to book a flight, acting as a bridge toward broader intelligence.

Artificial general intelligence

Artificial General Intelligence (AGI) is the holy grail of AI research. It refers to a theoretical stage where a machine possesses the ability to learn, reason, and solve problems across completely unrelated domains, matching human cognitive flexibility.

The key differentiator is generalization. An ANI system trained to speak Portuguese cannot use that knowledge to learn Spanish faster. An AGI system could apply logic from one field to another without manual retraining.

The timeline for AGI is a subject of intense debate. Optimists in the industry predict its arrival between 2026 and 2029. Conversely, Skeptics argue we lack the fundamental architecture (specifically, accurate "world models") to achieve this, placing the timeline at 2040 or beyond.

Artificial superintelligence

Artificial superintelligence (ASI) describes a hypothetical stage where AI surpasses the smartest human minds in practically every field: scientific creativity, general wisdom, social skills, and problem-solving. ASI wouldn't just match Einstein or Da Vinci; it would exceed them as dramatically as they exceeded average human intelligence.

The most significant concern surrounding ASI involves recursive self-improvement, where an ASI redesigns its own architecture to become smarter, then uses that enhanced intelligence to make further improvements, creating a runaway feedback loop of exponentially accelerating intelligence. 

This intelligence explosion could rapidly produce systems so far beyond human comprehension that predicting their behavior becomes impossible, raising profound safety questions about alignment and control. If an ASI's goals misalign even slightly with human values, its superior intelligence could pursue those goals in ways harmful to humanity.

Artificial Narrow Intelligence vs Artificial General Intelligence vs Artificial Superintelligence

No credible roadmap exists for building ASI, and many researchers question whether superintelligence is even achievable or coherent as a concept. Nevertheless, ASI's theoretical possibility motivates serious academic work on AI alignment to ensure that increasingly capable systems remain beneficial and controllable. 

Functionality-Based Types of AI

While capability measures power, functionality measures how the system interacts with the world and processes data. This classification allows us to distinguish between a chess computer from the 1990s and the social robots of the future. Let’s look at each type.

Reactive machines

Reactive Machines represent the most basic and oldest form of AI. These systems have no concept of past or future, but operate strictly on immediate inputs based on pre-programmed rules.

The defining characteristic of reactive AI is that it is stateless and deterministic. If you feed the exact same input into a reactive machine ten times, it will produce the exact same output ten times. It cannot learn from experience because it doesn’t store memory.

One good example of this type is IBM’s Deep Blue, the chess computer that defeated Garry Kasparov. It didn't learn Kasparov's psychology. It simply calculated the best move based on the current board state. Standard spam filters that flag keywords without analyzing your email history also fall into this category.

Limited memory AI

Limited memory AI dominates modern applications. By temporarily storing recent data or context to make informed decisions, limited memory AI systems enable dynamic behavior. Unlike reactive machines, they reference historical information within a limited context window.

Almost all Predictive AI and Generative AI tools used today belong to this category. LLMs like Claude Opus 4.5 maintain conversation history, remembering what you said minutes ago to provide coherent responses that build on previous exchanges. The context window determines how much historical information the system can reference. 

Self-driving cars track the speed and trajectory of nearby vehicles over several seconds, predicting their future positions rather than reacting only to their current location.

Theory of Mind AI

Theory of Mind AI occupies a contested space between research prototype and practical capability. Systems falling into this category would model human mental states, beliefs, intentions, emotions, and desires to predict behavior accurately. 

This capability is essential for natural social interaction, allowing AI to interpret implicit cues, detect deception, or adapt responses to a user’s mood. 

Earlier frontier models like OpenAI's GPT-4 already achieved human-level results on false belief tasks, such as predicting actions based on incorrect beliefs, as shown in Stanford researcher Michal Kosinski's 2024 PNAS study

Yet, experts debate whether this means that modern LLMs can already be classified as genuine theory of mind systems or just apply pattern matching from language training.​ Prototype efforts include DeepMind's ToMnet, designed to infer other agents' beliefs in simulated environments.

Self-aware AI

Self-aware AI describes a concept of machines possessing consciousness, sentience, and a clear understanding of their own existence separate from the world around them. 

This represents not just intelligence but subjective experience. A machine like that wouldn't just process information about itself but would have a felt sense of "I exist" with associated feelings, needs, and desires.

This level transcends computational capability, entering philosophical territory about the nature of consciousness itself. Would a self-aware AI experience pain, pleasure, or fear of deactivation? If so, does it deserve moral consideration or rights? These questions shift the conversation from engineering to ethics.

No existing systems approach this level. Fictional examples, like HAL 9000 from 2001: A Space Odyssey or Data from Star Trek, illustrate the concept, but it remains speculative if it will ever be reached.

Key Types of AI Models

Beyond the high-level theory, data practitioners must understand the specific architectures they will encounter in production. Broadly speaking, the types of AI models used in the industry today can be classified either as creators or analyzers. You can read more on this distinction in this piece on generative versus discriminative models.

Generative AI: The creators

Generative AI models are designed to generate new data instances that resemble your training data. These models learn the underlying probability distribution of a dataset to produce original text, images, code, or audio. This shift from analysis to creation offers many possibilities for scalable content generation, creative augmentation, and simulation. 

Popular examples include LLMs like Claude for writing and coding, diffusion models like Midjourney or Nano Banana Pro for image synthesis, and video generation models like Sora that can simulate complex scenes. Notably, not all image generators rely on diffusion; models like GPT Image 1.5 utilize an autoregressive architecture instead.

Generative AI represents a major advancement in agentic AI workflows, where models not only respond but actively create assets or plans. We will look into AI agents later in more detail.

Discriminative AI: The analysts

While Generative AI captures the headlines, discriminative AI, also known as predictive AI, remains the backbone of enterprise operations. These models do not create new data; instead, they learn the boundary between classes in a dataset to classify inputs or predict future values.

Generative AI vs Discriminative AI

The business value of discriminative AI lies in its ability to inform decision-making and assess risk. Organizations use these models to automate judgments requiring consistent application of learned criteria.

Unlike generative models, discriminative models are not focused on creativity but prioritize accuracy and reliability, often operating in high-stakes scenarios where errors carry significant consequences.

Examples include:

  • Credit scoring models predicting loan default risk.
  • Spam filters classifying emails as junk or legitimate.
  • Medical diagnostic tools identifying anomalies in X-rays.
  • Financial monitoring systems detecting fraudulent transactions. 

Foundation models

Foundation models train on vast, diverse datasets including text, images, code, and structured data, learning general patterns applicable across domains. A foundation model doesn't just learn to classify images or generate text; it develops a broad understanding of language, reasoning, and world knowledge that transfers to new tasks with minimal additional training.

Foundation Models

This approach dramatically reduces the time and data required to deploy AI for new applications. Rather than building a separate model for each specific problem, organizations fine-tune a single foundation model for diverse tasks ranging from legal document analysis to code generation to language translation.

We are currently seeing a rise in Small Language Models (SLMs). These are efficient, distilled versions of foundation models designed to run locally on devices (like laptops or phones), offering privacy and speed without the massive computational cost of cloud-based giants.

Types of AI Agents

AI is evolving from passive tools to autonomous agents capable of executing complex workflows. While a standard model waits for an input to produce an output, an AI Agent can perceive its environment, reason about how to achieve a goal, and take action to execute it. To understand the architecture behind agents and how to build them, I recommend starting with our Introduction to AI Agents course.

Types of AI agents are generally categorized by their level of complexity and autonomy. Let’s take a look at them.

Simple reflex agents

Reflex agents are the simplest form of an agent. They operate on a strict "Condition-Action" rule, reacting only to their current perception, and ignoring history. A smart thermostat exemplifies this with rules like these: If the temperature drops below 68°F, turn on the heat. If it exceeds 72°F, activate cooling. 

Assembly line robots are another example; they perform actions when sensors detect conditions that indicate parts have reached designated positions. Reflex agents excel at repetitive tasks in stable environments, but struggle when conditions change unexpectedly, as they lack contextual understanding to adapt.

Model-based reflex agents

Model-based reflex agents maintain an internal world model, tracking state over time to handle partially observable environments. They not only react, but remember key information necessary to operate correctly.

Autonomous emergency braking systems exemplify this by fusing data from radar and cameras to track vehicle trajectories over time, rather than merely reacting to immediate threats. They maintain accurate state estimates even with incomplete information, such as remembering a car's estimated position after it disappears behind a truck.

Goal-based agents

Goal-based agents act to achieve specific objectives rather than merely reacting. Given a goal like "plan a vacation to Italy," they consider multiple action sequences and evaluate which leads to goal satisfaction. This represents the cutting edge of agentic AI.

Tools like Roo Code for software engineering or autonomous booking assistants execute multi-step workflows like searching, comparing, checking constraints, and making decisions to complete entire jobs. Workflow platforms like n8n enable building complex processes where AI agents handle decision points and adaptive logic.

If you are looking to build a strong foundation in AI Agents, I recommend taking our AI Agent Fundamentals track.

Utility-based agents 

Utility-based agents take it even one step further: they optimize for the best path to goals rather than any working path. They do this by employing utility functions that assign numerical scores to outcomes, enabling quantitative comparison. 

While a goal-based agent books any flight to Rome, a utility-based agent considers factors such as cost, duration, layovers, and departure times to maximize overall utility.

Examples include algorithmic trading systems that balance profit against risk, and navigation apps like Waze, which optimize routes based on user preferences such as fastest time, shortest distance, or avoiding tolls.

Learning agents

Learning agents operate in unknown or changing environments, continuously improving through experience without human intervention. The architecture includes a critic providing performance feedback and a learning element that modifies the agent's policy. 

Mars rovers learn which terrain features indicate safe paths as they encounter novel geological formations. Netflix's recommendation algorithm is another example: it continually refines its understanding of user preferences based on viewing history, thereby improving suggestions over time.

Agentic frameworks: Building the agents

Agentic frameworks serve as the operating system for complex agents, connecting the reasoning core (typically an LLM) to memory, tools, and multi-agent coordination. They manage:

  • Orchestration: Deciding which agent handles each subtask and managing workflow between components
  • Planning: Breaking complex goals into executable steps and adapting when circumstances change
  • Tool Use: Providing agents access to search engines, databases, APIs, or code execution
  • Memory: Maintaining conversation history and long-term context beyond immediate interactions

Popular frameworks include:

  • LangChain/LangGraph: Industry standard for connecting LLMs to data sources with extensive integrations.
  • Google ADK: Open-source framework for building sophisticated agents on Vertex AI.
  • CrewAI: Orchestrates multiple role-playing agents collaborating on complex tasks.

For a quick reference on these architectures, I recommend you go through the AI Agents cheat sheet. 

Conclusion

We’ve covered a lot of ground here; let’s wrap up. 

Rather than viewing AI as a monolithic technology, we should recognize it as a diverse ecosystem of specialized approaches, each suited to particular problems. Understanding AI fully requires examining it through the three distinct lenses of technology, capability, and functionality.

We currently operate in the transition from narrow AI focused on specific tasks to early forms of agentic AI that autonomously execute complex workflows, making this a critical inflection point for adoption across industries.

Take the first step towards becoming an AI expert by enrolling in the AI Fundamentals skill track.

Types of AI FAQs

What are the main types of AI?

AI is classified through three distinct lenses: technology-based (machine learning, deep learning, NLP), capability-based (ANI, AGI, ASI), and functionality-based (reactive machines, limited memory, theory of mind, self-aware AI).

What is the difference between generative AI and predictive AI?

Generative AI creates new content like text, images, or code, while predictive AI (discriminative AI) classifies data and predicts outcomes based on historical patterns.

What are AI agents, and how do they work?

AI agents are autonomous systems that plan and execute multi-step workflows using a language model for reasoning, tools for external interaction, and an orchestration layer to manage tasks.

What is Agentic AI?

Agentic AI refers to autonomous systems that independently execute complex, multi-step workflows to achieve goals, moving beyond simple question-answering to active task completion.

What is the difference between ANI and AGI?

Artificial Narrow Intelligence (ANI) excels at specific tasks but struggles outside its training domain. At the same time, Artificial General Intelligence(AGI) would match human cognitive flexibility across all domains, a capability that doesn’t yet exist.


Author
Rajesh Kumar
LinkedIn

I am a data science content writer. I love creating content around AI/ML/DS topics. I also explore new AI tools and write about them.

Sujets

AI Courses

Cursus

Principes fondamentaux de l'IA

0 min
Découvrez les principes fondamentaux de l'IA, apprenez à l'utiliser efficacement dans votre travail et explorez des modèles tels que chatGPT pour vous orienter dans le paysage dynamique de l'IA.
Afficher les détailsRight Arrow
Commencer le cours
Voir plusRight Arrow
Contenus associés

blog

Types of AI Agents: Understanding Their Roles, Structures, and Applications

Learn about the main types of AI agents, how they interact with environments, and how they are used across industries. Understand simple reflex, model-based, goal-based, utility-based, learning agents, and more.
Vinod Chugani's photo

Vinod Chugani

14 min

blog

AI Security: A Comprehensive Guide With Examples

Learn about the importance of AI security, the various threats AI systems face, effective defense mechanisms, and emerging trends in AI security.
Dr Ana Rojo-Echeburúa's photo

Dr Ana Rojo-Echeburúa

8 min

blog

Artificial Intelligence (AI) vs Machine Learning (ML): A Comparative Guide

Check out the similarities, differences, uses and benefits of machine learning and artificial intelligence.
Matt Crabtree's photo

Matt Crabtree

10 min

blog

Understanding AI Agents: The Future of Autonomous Systems

Discover the transformative potential of AI agents. Explore their applications, benefits, and challenges. Learn how to leverage AI agents for innovation and efficiency in your projects.
Vinod Chugani's photo

Vinod Chugani

9 min

blog

AI in Marketing: A Complete Guide With Examples

Discover how AI in marketing can empower your business with this complete guide.
Austin Chia's photo

Austin Chia

12 min

blog

What is AI Literacy? A Comprehensive Guide for Beginners

Explore the importance of AI literacy in our AI-driven world. Understand its components, its role in education and business, and how to develop it within organizations.
Matt Crabtree's photo

Matt Crabtree

15 min

Voir plusVoir plus