The AI Glossary: Setting the Vocabulary Straight for Beginners

People have been calling everything “AI” lately, but the truth is that not all AI is the same. In fact, some of the tools people call “AI” aren’t even in the same category.

Language has a funny way of shaping how we think about technology. When all innovation gets flattened under the same three letters, we lose nuance, and nuance is where understanding actually begins.

What Artificial Intelligence Actually Means

“AI” stands for Artificial Intelligence, but that’s only the headline.

AI isn’t one product, platform, or model. It’s an entire field of computer science focused on getting machines to mimic some aspect of human reasoning, learning, or problem-solving.

That means AI covers everything from a rule-based chatbot that follows a flowchart to an image generator that creates an original digital painting from text. One isn’t “smarter” than the other. They’re all just different branches of the same family tree.

If AI were a family, rule-based systems would be the meticulous older sibling who color-codes everything, predictive models would be the planner who always knows what’s coming next, and generative AI would be the creative cousin who makes wild art at 3 a.m.

The Foundational Types of AI

A single blog post also cannot capture all the nuance and vocabulary surrounding AI, but let’s start here:

Rule-Based AI (Symbolic AI)

Rule-based AI, also called symbolic AI, operates on a fixed set of pre-programmed rules or logic statements, often expressed as “if–then” conditions. It doesn’t learn from data, it simply executes explicit instructions. Every possible scenario must be manually coded into the system.

Think of this as the grandparent of AI. If the condition matches, it acts; if not, it waits. This type of AI still runs in medical diagnostic tools, finance fraud detection, and customer service chatbots that follow decision trees.

Predictive AI

Predictive AI uses statistical and machine learning models to analyze historical data and estimate the probability of future outcomes. It often employs techniques such as regression analysis, time-series forecasting, and gradient boosting to model trends or behavior patterns.

Basically, predictive AI is the planner friend who can look at your past five Mondays and say, “Yeah, you’re going to need extra coffee tomorrow.” It doesn’t make new data, it simply uses what it already knows to anticipate what’s coming next. Weather forecasts, stock market models, and spam filters all live in this predictive space.

Discriminative AI

Discriminative AI models learn to classify or distinguish between categories of data. They focus on modeling the boundary between classes rather than generating new examples. In mathematical terms, they estimate the conditional probability P(y|x) or the likelihood that input x belongs to category y.

Discriminative AI is your sharp-eyed friend who can instantly spot the difference between two shades of beige. It separates, sorts, and labels. When you upload a photo and your phone says, “Dog detected,” that’s discriminative AI at work. It powers facial recognition, sentiment analysis, spam detection, and quality control, or anything that needs a firm “yes, that’s it” or “nope, try again.”

Analytical AI

Analytical AI integrates data mining, statistical analysis, and pattern recognition to interpret complex datasets and identify relationships, correlations, or anomalies. It’s designed to interpret and explain patterns in the data.

Analytical AI is the data detective of the group. It crunches numbers then tells stories through them. It looks for the “why” behind the “what.” This is the kind of AI behind your streaming recommendations, sales dashboards, and healthcare risk models.

Generative AI

Generative AI is the creative one in the family. Technically speaking, Generative AI refers to models that learn the underlying distribution of training data, often represented as P(x) in statistics. Instead of just recognizing or predicting patterns, these systems can generate new data that fits the same patterns. This includes models such as Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and transformer-based architectures like Large Language Models (LLMs) and diffusion models.

These are the engines behind today’s most headline-grabbing tools like ChatGPT, Midjourney, RunwayML, DALL·E, and Adobe Firefly. They turn text prompts into essays, images, code, or videos.

In practice, Generative AI falls into two main categories:

  • Text-to-X Models: systems that turn language into another medium like text, image, video, or audio.

  • Data-Driven Synthesis Models: systems that learn directly from visual, auditory, or numerical data to generate new versions that statistically align with what they’ve studied (for example, GANs producing new faces or diffusion models rendering realistic textures).

Generative AI doesn’t “think” creatively. It doesn’t have curiosity, motivation, or inspiration. What it does have is mathematical rhythm. It predicts what should come next in a sequence, either word by word, pixel by pixel, or frame by frame based on probabilities learned from its training data. The result feels almost human.

That sentence might make you tense, especially if your work depends on human creativity. I get it. I’m saying this as a professional writer who has spent years learning syntax, metaphor, and rhythm the old-fashioned way. But data, research, and thousands of peer-reviewed studies show it’s true: machines can now reproduce certain outcomes of creativity, even if they don’t experience the feeling of it.

Generative AI has earned both fascination and scrutiny because it CAN paint, write, compose, or design, but only by remixing vast amounts of prior human work into new forms. Any debate about this capability is already a lost game.

We’ll unpack those ethical debates in a later post, don’t worry.

The Everyday AI Vocabulary People Mix Up

If you’ve ever heard someone say “machine learning” and “AI” as if they were twins, you’re not alone. The language of AI has turned into alphabet soup, and it’s hard to tell which letters actually belong in the bowl.

Here’s a quick glossary that clears up the most commonly mixed-up terms.

Machine Learning (ML)

ML is a subset of artificial intelligence focused on algorithms that learn patterns from data and improve performance over time without being explicitly programmed for every task.

Machine learning is the engine that powers most modern AI. Instead of following a fixed list of instructions, it studies examples until it can make decisions on its own. It’s how email filters learn to spot spam or how Netflix guesses what you want to watch next. Think of it as experience-based learning for computers.

Deep Learning

Deep learning is a specialized branch of machine learning that uses artificial neural networks with multiple layers (hence “deep”) to model complex relationships in data.

Deep learning is what happens when machine learning gets really good at pattern recognition. It’s the brain behind voice assistants, self-driving cars, and facial recognition. The “deep” part refers to all the hidden layers of logic that help it understand messy, real-world inputs like images, sound, and speech.

Neural Networks

A neural network is a form of computational architecture made up of layers of interconnected “nodes” (or neurons) that process input data through weighted connections to produce an output.

Neural networks are math inspired by biology. Picture a simplified version of the human brain where each “neuron” is just a calculator making tiny decisions. One layer notices shapes, another sees patterns, and together they reach conclusions like recognizing a friend’s face in a photo.

Large Language Model (LLM)

LLM refers to a deep learning model trained on massive text datasets to predict the probability of word sequences. It uses techniques such as transformer architecture and attention mechanisms to generate human-like language.

These models are behind ChatGPT, Gemini, and Claude. They’ve read an enormous portion of the internet, learned the rhythm of human speech, and now predict the most likely next word in a sentence until it sounds natural. They don’t understand meaning the way humans do, but they’ve gotten scarily good at sounding like they do.

Training Data

Training data refers to any dataset used to teach a model patterns, relationships, or behaviors. The model learns by adjusting internal parameters until its outputs align with the examples in this data.

Training data is the model’s childhood. It’s everything the system “grew up” reading, seeing, or hearing, including articles, code, art, audio, and more. The quality of that data shapes its worldview. Biased data creates biased models; clean, diverse data builds better behavior.

Parameters

Parameters are numerical values that a model adjusts during training to minimize prediction errors. In large-scale models, these parameters can number in the billions.

Parameters are basically the adjustable knobs and dials of an AI’s brain. The more of them it has, the more detail it can capture (but also the more complex it becomes to control). Think of them as the “experience settings” that tell the system how to respond, shaped by every lesson it’s ever learned.

Prompt

A prompt is the input text or instruction provided to a generative model to guide its output.

In other words, a prompt is how you talk to AI. It’s your half of the conversation, including the question, command, or idea you feed it. We’ll have to revisit prompting in another blog post, because it runs DEEP.

Token

The smallest unit of text that a model processes is a token. This can be a word, subword, or even a single character. Tokens are used to calculate input and output length in AI systems.

Basically, tokens are the puzzle pieces of language for AI. They’re how the model measures text by chunks of meaning. That’s why AI tools charge by “tokens,” not “words.” A sentence like “AI is fascinating” might be five tokens, depending on how the model slices it.

Training, Fine-Tuning, and Guardrails

Fresh new AI models are a lot like interns. They are eager to learn and curious, but they need direction.

The process of turning an AI intern into something useful happens through training, fine-tuning, and guardrails.

If this anthropomorphized analogy makes you uncomfortable, I get it. But I can’t hammer home the point enough that by ignoring the very real research and capabilities that already exist, any meaningful conversation surrounding AI’s future is lost. I may not always choose the right words either, but I am part of the crowd that is trying to make sure we all DO collaborate on the language we’re using together as though humanity depends on it (it very well might).

Training: How Models Learn from Large Datasets

Training is the initial learning phase where an AI model processes enormous datasets to identify patterns, relationships, and probabilities. During this process, the model adjusts billions of internal parameters to reduce prediction errors and generalize from its examples.

Think of training as the AI’s early education, except the equivalent for a human is reading every book in the library at lightning speed, then trying to summarize what language means. A large language model like ChatGPT trains on text from the internet, books, research papers, and code repositories, learning how words tend to appear together and what structures make sentences flow.

By the end, it doesn’t “remember” individual pages; it understands statistical patterns and the grammar of thought rather than the thoughts themselves. If you’ve ever wondered why an AI can nail an output one moment and totally flub it the next, this is why. It’s still learning how to balance the textbook knowledge it was raised on with the tone of an actual human conversation.

Training is massive, expensive, and energy-hungry. It’s also where models gain their base personality, which then defines what they know, how they respond, and the limits of their worldview.

We’ll revisit this energy conversation in another blog post, too. There is simply WAY too much misinformation out there about that.

Fine-Tuning: How Models Adapt to Specific Tasks or Audiences

Fine-tuning is a secondary training phase in which a pre-trained model is exposed to smaller, domain-specific datasets. This refines its behavior and adapts it for specialized purposes without starting from scratch.

Fine-tuning is a lot like graduate school. The model already knows the basics of language, math, and logic, but now it’s specializing by learning legal phrasing, medical terminology, or the polite tone of customer service chatbots.

It’s also where a general-purpose model learns manners. If the base model is an overexcited intern blurting out everything it knows, fine-tuning turns it into a professional who knows which details matter in which context. That’s why a medical AI and a creative writing AI can come from the same foundation but behave entirely differently in conversation.

Fine-tuning is also how companies personalize AI for brand voice or compliance standards like teaching the model that your company says “clients,” not “customers.”

Reinforcement Learning from Human Feedback (RLHF): How Models Learn What People Prefer

RLHF is a post-training process where human evaluators review and rank model outputs. The rankings are fed back into the system to help it adjust its responses, rewarding answers that align with human values and penalizing those that do not.

RLHF is basically the etiquette coach of AI training. It’s what teaches models to be more socially aware and less awkward at dinner parties. After the initial training, real humans rate thousands of AI responses on things like helpfulness, tone, accuracy, and bias. The system then learns what we like (e.g., less lecturing, fewer conspiracy theories, and no unsolicited poetry).

If you’ve noticed that AI tools have become less blunt or more polite over time, you’re witnessing RLHF in action. It’s what takes the raw data-driven brain of a model and adds a moral compass (albeit one built from consensus rather than conscience).

Guardrails and Filters: The Ethical Layers That Keep AI in Check

Guardrails are predefined constraints, filters, and moderation systems designed to prevent harmful, illegal, or inappropriate content from being generated. These mechanisms are integrated into the model’s architecture and its deployment interface.

Us millennials love boundaries, right? Guardrails are the boundaries that keep the system from wandering into chaos. They’re like parental controls for a supercomputer and set the limits for what it can say, share, or imagine. Without them, models might accidentally repeat misinformation, generate biased outputs, or produce content that crosses ethical lines.

These filters are built from safety datasets, community guidelines, and sometimes legal requirements. They stop bad behavior and at the same time influence tone, enforce empathy, and encourage accuracy.

If you’ve ever asked an AI a spicy question and got a gentle “I can’t help with that,” you’ve met the guardrails.

They exist for a reason: to keep creative power from turning reckless.

How It All Comes Together

Training gives AI its foundation.
Fine-tuning gives it focus.
RLHF gives it feedback.
Guardrails give it boundaries.

None of these steps is perfect yet (and might not ever be), but without any of this knowledge or vocabulary, it is very difficult to have grounded debates and grounded action.

Common AI Misconceptions to Retire

  • “AI is self-aware.” → It’s not. It’s predictive math.

  • “AI replaces human creativity.” → It amplifies or imitates it.

  • “All AI tools are trained the same way.” → Far from it; data sets, model sizes, and architectures differ dramatically.

  • “AI is new.” → Nope. The field dates back to the 1950s; the hype is what’s new.

Language Is How We Think About Technology

Every generation redefines its relationship with technology through the words it chooses. In the early internet era, we talked about “surfing the web.” Then we started “scrolling.” Now we talk about “training,” “prompting,” and “fine-tuning.” Each shift in language rewires how we think about innovation and about ourselves.

When we flatten all AI into one bucket, we lose the vocabulary that helps us ask better questions. We start blaming the wrong systems, fearing the wrong outcomes, and missing the nuance that shapes real progress.

The words we use determine how we treat creativity, ethics, and authorship. They influence policy debates, classroom discussions, and even how we credit human labor in the age of automation. Vocabulary is POWER, people!

So learn the language. Use it freely, ask boldly, and stay curious.

Next
Next

The Power of Pacing: How to Manage Workload and Chronic Illness Symptoms