This article provides a clear, easy-to-understand overview of foundational terminology used in Large Language Models (LLMs) and modern AI systems. Each concept includes simple explanations and relatable examples to help non–data scientists quickly grasp how these technologies work.
-
Artificial Intelligence (AI)
Software designed to perform tasks that normally require human intelligence.
Example: Siri answering your questions. -
Deep Learning
A way for computers to learn from many examples using neural networks.
Example: Recognizing cats in pictures. -
Neural Network
A layered system of connected nodes that mimic how the brain processes information.
Example: A system that learns to play video games. -
Transformer
A neural network architecture designed to understand and generate language efficiently.
Example: A chatbot holding a natural conversation. -
Large Language Model (LLM)
AI trained on enormous amounts of text to understand and generate human language.
Example: Writing stories or answering questions. -
Parameter
Internal values a neural network adjusts while learning.
Example: Tuning a guitar to get the right sound. -
Positional Encoding
Helps models understand word order.
Example: Knowing “dog chases cat” differs from “cat chases dog.” -
Self-Attention
Allows models to focus on the most relevant words in a sentence.
Example: Identifying “cake” as key in “I want to eat cake.” -
Encoder
The part of a Transformer that reads and understands input.
Example: Understanding the question “What’s the weather?” -
Decoder
The part that generates output or responses.
Example: “Today is sunny and warm.” -
BERT
A Transformer for understanding language (NLU).
Example: Detecting whether a review is positive or negative. -
GPT-3 / GPT-4
Transformer models optimized for generating text.
Example: Writing summaries or essays. -
T5
A model designed to both understand and generate text.
Example: Translating English to Spanish. -
Unsupervised Learning
Learning patterns without labeled answers.
Example: Grouping similar photos. -
Foundation Model
A large, general-purpose model that can be adapted to many tasks.
Example: One model that writes emails, translates languages, and answers questions. -
Zero-Shot Learning
Performing new tasks without training examples.
Example: Playing a new game with no practice. -
Few-Shot Learning
Learning from only a few examples.
Example: Learning a song after hearing it twice. -
Fine-Tuning
Adjusting a pre-trained model for a focused task.
Example: Teaching a model to answer dinosaur questions. -
Prompt Tuning
Improving results by phrasing prompts more effectively.
Example: Asking “What’s the capital of France?” instead of “Where is Paris?” -
Adapters
Small add-on modules that give a model new capabilities without retraining the entire model.
Example: Adding a new skill to a game character. -
Natural Language Processing (NLP)
The AI field focused on human language.
Example: Chatbots or grammar checkers. -
Natural Language Understanding (NLU)
Understanding meaning in human language.
Example: Knowing “I don’t like cats” is negative. -
Natural Language Generation (NLG)
Creating human-like text.
Example: Writing stories. -
Tokenization
Splitting text into words or pieces (“tokens”).
Example: “I have a dog” → “I”, “have”, “a”, “dog”. -
Vocabulary
The set of tokens a model knows.
Example: Knowing “apple” and “banana” but not “kiwifruit.” -
Pre-Training
The initial learning phase using large text datasets.
Example: Reading millions of documents to learn grammar and facts. -
Transfer Learning
Applying knowledge from one task to another.
Example: Learning cats helps with learning dogs. -
Sequence-to-Sequence (Seq2Seq)
Models that turn one sequence into another.
Example: Translating English to French. -
Attention Mechanism
The model highlights important inputs when generating outputs.
Example: Prioritizing the word “pizza” in “I want to eat pizza.” -
Beam Search
A method to choose the most likely sequence of generated words.
Example: Selecting the best next word in a sentence. -
Perplexity
Measures how well a model predicts text. Lower = better.
Example: Fewer surprises in predicting next words. -
In-Context Learning
Models adapting behavior based on provided examples in a prompt.
Example: Answering sports questions correctly after reading sports examples. -
Data Augmentation
Expanding training data by modifying existing samples.
Example: Rephrasing sentences. -
Bias
Systematic errors caused by skewed training data.
Example: Assuming all doctors are male. -
Explainable AI (XAI)
Making AI decisions understandable to humans.
Example: Explaining why a model classifies a movie as a comedy.