Skip to main content
← The Builder’s Log/100 Top AI Vocabulary Terms Every Developer Should Know
100 Top AI Vocabulary Terms Every Developer Should Know
aimachine-learningllmglossaryeducationdeep-learning

100 Top AI Vocabulary Terms Every Developer Should Know

B

Bunlong Heng

February 20, 2026

100 essential AI terms. One page. No fluff.

🧱 Foundations

TermDefinition
AISystems performing tasks that typically require human intelligence
Machine LearningSystems that learn patterns from data instead of explicit rules
Deep LearningML using multi-layer neural networks to learn hierarchical representations
Neural NetworkLayers of connected nodes that transform inputs into outputs
ParameterA learnable value (weight/bias) adjusted during training
WeightScalar value scaling the connection between two neurons
BiasParameter that shifts a neuron's activation independent of input
Activation FunctionNon-linear function applied after each layer — ReLU, GELU, sigmoid
TrainingAdjusting model parameters on a dataset to minimize prediction error
InferenceRunning a trained model on new inputs — no learning happens

📦 Data

TermDefinition
DatasetStructured collection of examples for training, validation, or testing
Training SetData used to adjust model parameters
Validation SetHeld-out data for tuning hyperparameters and monitoring overfitting
Test SetUnseen data used only to evaluate final model performance
LabelGround-truth output for a training example
FeatureAn input variable used by the model
TokenizationBreaking text into tokens — roughly 4 chars or 0.75 words each
TokenBasic unit of text an LLM processes
CorpusLarge text collection used to train language models
Data AugmentationCreating modified training copies to expand the dataset

⚙️ Core ML

TermDefinition
Loss FunctionMeasures how wrong predictions are — training minimizes this
GradientDirection and magnitude of steepest increase in the loss
BackpropagationComputes gradients by propagating error backward through the network
Gradient DescentIteratively moves parameters in the direction that reduces loss
Learning RateControls step size during optimization — too high = unstable
BatchSubset of training examples processed before each parameter update
EpochOne complete pass through the entire training dataset
OverfittingModel memorizes training data and performs poorly on new data
UnderfittingModel too simple to capture patterns — poor on both train and test
RegularizationTechniques to reduce overfitting — L1/L2, dropout, early stopping

🏗️ Architecture

TermDefinition
LayerGroup of neurons performing the same type of transformation
Dense LayerEvery neuron connects to every neuron in the next layer
Conv LayerApplies learned filters to local patches — great for images
RNNNetwork with loops allowing information to persist across steps
LSTMRNN variant that learns what to remember or forget long-term
AttentionMechanism to weigh importance of different input parts
TransformerDominant architecture based entirely on self-attention
EncoderTransformer part that reads input into contextual representations
DecoderTransformer part that generates output tokens autoregressively
Residual ConnectionShortcut adding a layer's input to its output — aids deep training

🧠 LLMs

TermDefinition
LLMLarge Transformer trained on massive text — GPT, Claude, Gemini
Context WindowMax tokens an LLM processes at once — its working memory
PromptInput text guiding an LLM's response
CompletionText generated by an LLM in response to a prompt
TemperatureSampling randomness — low = deterministic, high = creative
Top-pSamples from tokens whose cumulative probability exceeds p
Top-kRestricts sampling to the k most probable next tokens
Greedy DecodingAlways picks the single most probable next token
Beam SearchKeeps multiple candidate sequences; picks the globally best one
PerplexityHow well an LLM predicts text — lower is better

🏋️ Training Techniques

TermDefinition
Pre-trainingLarge-scale training on broad data — builds general capabilities
Fine-tuningFurther training on a smaller task-specific dataset
Transfer LearningUsing a pre-trained model as a starting point for a new task
RLHFHumans rank outputs → reward model → LLM optimized against it
DPODirectly trains on human preference pairs — simpler than RLHF
LoRAFine-tunes small adapter matrices added to frozen weights
QLoRALoRA on a quantized model — fine-tunes on consumer GPUs
Instruction TuningFine-tuning on (instruction, response) pairs to follow directions
PEFTParameter-Efficient Fine-Tuning — LoRA, adapters, prefix tuning
DistillationTraining a small student model to mimic a large teacher model

🔍 Retrieval & Memory

TermDefinition
RAGAugments LLM with retrieved documents at inference time
Vector DatabaseStores embeddings for fast nearest-neighbor search
EmbeddingDense numerical representation where similarity = proximity
Semantic SearchSearch by meaning rather than keyword overlap
ChunkingSplitting documents into pieces for embedding and retrieval
Context StuffingPutting all relevant info directly into the prompt
Long-Context ModelLLM with 100k+ token window for entire documents
Memory (agents)Mechanisms to persist info across turns — context or external storage

🤖 Agents

TermDefinition
AI AgentAutonomously plans, uses tools, and pursues goals over multiple steps
Tool UseLLM calling external functions — search, code exec, APIs, files
Function CallingLLM outputs structured JSON to invoke predefined functions
ReActPrompting pattern: alternate reasoning steps with action calls
Chain-of-ThoughtModel outputs intermediate reasoning before a final answer
Agentic LoopObserve → think → act → observe. Repeats until goal is achieved
Multi-AgentMultiple agents collaborating — orchestrator + worker agents
MCPOpen standard connecting AI models to tools and data sources
SubagentAgent spawned by an orchestrator to handle a specific subtask
ScaffoldingInfrastructure managing an agent's loop, tools, context, memory

📊 Evaluation

TermDefinition
BenchmarkStandardized test suite to compare model capabilities
MMLU57-subject knowledge benchmark — law, medicine, math, etc.
HumanEvalCoding benchmark — models write Python functions, pass unit tests
HallucinationLLM confidently generates false info not grounded in data
GroundingConnecting outputs to verifiable facts to reduce hallucination
PrecisionOf all positive predictions, what fraction are actually correct
RecallOf all actual positives, what fraction did the model find
F1 ScoreHarmonic mean of precision and recall
BLEUTranslation metric — measures n-gram overlap with reference text
ROUGESummarization metric — measures n-gram recall vs reference text

🛡️ Safety & Compute

TermDefinition
AlignmentEnsuring AI pursues goals consistent with human values
Constitutional AIAnthropic's method — model critiques itself using a set of principles
Red TeamingAdversarially probing a model to find failure modes before deployment
JailbreakPrompt that bypasses a model's safety guardrails
Prompt InjectionMalicious content in retrieved data hijacks agent instructions
BiasSystematic prediction errors correlated with sensitive attributes
InterpretabilityUnderstanding what's happening inside a model's computations
GPUParallel hardware essential for deep learning training
TPUGoogle's custom AI accelerator chip for large-scale Transformer training
QuantizationReducing weight precision (float32 → int8) to shrink model size
Inference ServingInfrastructure hosting a model for prediction requests at scale
Scaling LawModel performance improves predictably with more params, data, compute
ShareX (Twitter)LinkedIn

Comments

Be the first to leave a comment.

Leave a comment

Related Posts