Skip to main content
Menu
📚 Concepts

AI Security Concepts

Foundational definitions and theoretical frameworks for understanding adversarial AI, LLM security, and machine learning vulnerabilities.

Understanding the Foundations

AI security concepts differ fundamentally from traditional cybersecurity terminology. In conventional security, we discuss vulnerabilities as discrete flaws—a buffer overflow exists or it doesn't, a misconfiguration is present or absent. AI security operates in a more probabilistic space where vulnerabilities emerge from learned behaviors, statistical patterns, and architectural decisions that don't map cleanly to binary categories.

This section establishes precise definitions for the field's core terminology. These aren't just academic distinctions—they're operational requirements. When a security team assesses an AI system, when a red team scopes an engagement, when a vendor communicates risk to customers, shared vocabulary prevents costly misunderstandings.

Core Concepts Index

Foundational

Concept Definition Relevance
Adversarial AI The study and practice of attacking and defending AI systems Defines the entire field
Prompt Injection Manipulating LLM behavior through crafted inputs Primary LLM vulnerability class
AI Red Teaming Adversarial testing methodologies for AI systems Practical application of concepts

The AI Attack Surface

Understanding AI security concepts requires a mental model of where attacks can occur:

Training Time Attacks

Attacks during model creation—poisoning the well.

Inference Time Attacks

Attacks against deployed models through user interaction.

Extraction Attacks

Stealing information from the model or its training data.

System-Level Attacks

Targeting infrastructure and integrations around the model.

Start Learning

New to AI security? Begin with these foundational entries in order:

  1. 1 Adversarial AI — The field overview
  2. 2 Prompt Injection — The defining vulnerability
  3. 3 AI Red Teaming — Putting concepts into practice