Skip to main content

Glossary

Plain-English definitions
for AI & dev tooling

Terms used across Codegen Blog, explained clearly for developers and non-technical readers alike.

20 Terms
4 Categories

Showing 20 terms

Agentic Coding
A development approach where AI agents autonomously plan, write, test, and iterate on code with minimal human intervention.
API Orchestration
The coordination layer that manages how multiple APIs communicate, sequence calls, and handle failures in complex workflows.
AST (Abstract Syntax Tree)
A tree representation of source code structure used by compilers and analysis tools to understand program logic.
Autonomous Agent
An AI system capable of independently completing multi-step tasks by planning, executing, and self-correcting without human prompting.
CI/CD
Continuous Integration / Continuous Delivery: automated practices for building, testing, and deploying code changes.
Code Review Agent
An AI agent that automatically reviews pull requests, checking for bugs, security issues, and style violations before human review.
Context Window
The maximum amount of text (measured in tokens) that an LLM can process in a single request, including both input and output.
Fine-tuning
The process of training a pre-trained model on a smaller, task-specific dataset to adapt its behavior for a particular use case.
LLM (Large Language Model)
A neural network trained on massive text datasets to predict and generate human-like text, powering all modern AI coding tools.
MCP (Model Context Protocol)
An open protocol by Anthropic that standardizes how AI clients communicate with external services, tools, and data sources.
Multi-file Editing
The ability of a coding agent to make coordinated changes across multiple files in a repository in a single operation.
Prompt Engineering
The practice of designing and refining instructions given to an AI model to produce accurate, relevant, and useful outputs.
Pull Request
A method for submitting code changes for review before merging into a shared codebase, serving as the AI-to-human handoff point.
RAG (Retrieval-Augmented Generation)
A technique that enhances LLM responses by retrieving relevant documents from an external knowledge base before generating output.
Sandboxing
Running AI agent code in an isolated environment that cannot affect live systems, enabling safe parallel execution.
SWE-bench
A benchmark for evaluating AI coding agents against real GitHub issues, measuring whether agents can produce correct fixes.
Technical Debt
The accumulated cost of shortcuts and deferred maintenance in a codebase that slows future development.
Telemetry
The collection and analysis of performance data from AI agent runs, including cost, execution time, and success rate metrics.
Tokenization
The process of converting text into smaller units called tokens that a language model can process and reason over.
Vibe Coding
A development style where developers defer to AI, accepting generated code without detailed review, coined by Andrej Karpathy.

Build faster with AI-powered agents

See how Codegen automates the full development workflow — from ticket to pull request.

Try Codegen free →