Distinctionary

A Distinctionary combines a searchable glossary with a hierarchical taxonomy visualization, allowing you to explore AI agent terminology both through direct lookup and conceptual relationships.

Term
Definition
A/B testingComparing variants under controlled conditions to choose the best performing approach.
Access controlMechanisms like RBAC that restrict who can run models, view data, or invoke tools.
ActionAn operation taken by the agent, such as calling a tool, retrieving data, or generating content.
AdapterA small parameter module inserted into a model to enable efficient task adaptation.
Agent architectureA category for concepts that define how agents plan, act, observe, and manage state over time.
Agent typesA category grouping classical and modern agent forms by complexity and capabilities to guide architecture selection.
Agentic AIAn approach where autonomous or semi‑autonomous agents carry out end‑to‑end tasks using planning, tooling, memory, and verification loops.
Agentic workflowA coordinated process where one or more agents plan, act, and verify to complete a task or case.
AGIArtificial general intelligence with human‑level performance across most cognitive tasks, a debated future goal.
AI agentA software entity that senses inputs, selects actions, and optimizes behavior to achieve objectives in an environment.
AI AgentsAn umbrella concept for software systems that perceive, decide, and act toward goals with varying autonomy, encompassing simple agents through multi‑agent, tool‑using, and memory‑equipped systems.
AI risk managementCoordinated activities to direct and control risks associated with AI systems and their use.
AlignmentEnsuring model goals and behaviors are consistent with human values and policies.
AnnotationHuman or tool labeling that creates training or evaluation data for models.
AnonymizationIrreversibly removing personal identifiers so individuals cannot be reidentified.
API integrationConnecting agents to external services via authenticated, schema‑defined endpoints.
ASIArtificial superintelligence that would surpass human capabilities across domains, a speculative concept.
Assistant messageModel or agent responses that continue the conversation and may include tool calls.
AttentionA mechanism that lets the model focus on the most relevant parts of the input when producing outputs.
Audio embeddingA vector representation of audio useful for search and classification.
Audit trailTamper‑evident logs that record actions for compliance and forensics.
BackoffIncreasing wait times between retries to reduce contention and protect services.
Barge-inA feature that lets a caller interrupt prompt playback with speech or DTMF to speed interactions.
Barge-in handlingPolicies and logic to safely process interruptions without losing dialogue state.
BenchmarkA standard task suite for comparing systems under consistent conditions.
BiasSystematic error introduced by data, modeling choices, or deployment context that skews outputs.
BM25A classic probabilistic ranking function for lexical retrieval commonly used as a baseline.
CachingReusing prior results to reduce latency, cost, and tool pressure for repeated queries.
Call orchestrationCoordinating ASR, NLU, business tools, and TTS across the call lifecycle for quality outcomes.
Canary promptA sensitive prompt used to detect regressions in safety or reasoning after changes.
Canary releaseDeploying a change to a small slice of traffic to validate safety and quality before full rollout.
Chain-of-thought promptingEliciting intermediate reasoning steps to improve solution quality on complex tasks.
ChunkingSplitting documents into appropriately sized passages to improve retrieval precision and context fit.
CitationA reference to a source that supports an output statement for verification.
ComplianceMeeting legal, regulatory, and policy requirements in AI development and operations.
Confidence scoreA numeric or categorical estimate of output reliability for ranking or escalation decisions.
ConsentExplicit user permission for data collection or processing in specific contexts.
Content filteringBlocking or transforming unsafe or off‑policy content before it reaches users.
Context awarenessUsing history and situational data to tailor responses to the user's current environment and needs.
Context lengthThe token capacity of the model's input window that bounds how much can be considered.
Context stuffingOverfilling the prompt with too much or irrelevant context that can degrade output quality.
Context windowThe maximum token span a model can consider at once, functioning like short‑term memory.
Conversation historyThe prior messages that inform current context and continuity.
Cost budgetThe maximum allowed spend in tokens or currency for a task or session.
Cost per taskTotal spend in tokens and tools required to complete a task successfully.
Cost trackingMeasuring token, tool, and infrastructure spend at task and system levels.
Data governancePolicies and controls that ensure responsible data use across the AI lifecycle.
Data minimizationCollecting and retaining only data necessary for a specified purpose.
Data retentionPolicies that define how long data is kept and when it is deleted.
DeduplicationRemoving duplicate or near‑duplicate items to reduce noise and bias in retrieval or training data.
DelegationAssigning a subtask to another specialized agent or a human to improve efficiency and outcomes.
Differential privacyAdding noise to data or queries to protect individual privacy while enabling aggregate analysis.
Document storeA repository for unstructured or semi‑structured documents used by retrieval systems.
DPIAData Protection Impact Assessment to evaluate and mitigate risks to individuals' rights from processing.
DriftDegradation due to changes in data, user behavior, or external tools over time.
DTMFDual‑Tone Multi‑Frequency signals from keypad input used for IVR navigation and data entry.
Dynamic resource allocationAdjusting model size, tools, or retrieval depth at runtime to meet task complexity and constraints.
EmbeddingsNumerical vector representations of text or other data that preserve semantic similarity for search and clustering.
Emergent behaviorCapabilities that arise in large models without explicit training for those skills.
EncryptionEncoding data at rest and in transit so only authorized parties can read it.
EvaluationSystematic measurement of agent quality, robustness, cost, and safety using tasks and metrics.
Evaluation and metricsA category for measuring agent quality, robustness, cost, and safety.
Evaluation harnessReusable tasks, datasets, and metrics for consistent measurement across iterations.
EventA notable change such as a tool result or user input that can trigger agent logic.
EvidenceSpecific retrieved passages or data points used to substantiate an output or decision.
F1 scoreThe harmonic mean of precision and recall providing a single effectiveness measure.
Failure modes taxonomyA categorization of common agent failure patterns to inform controls and testing.
Feedback loopCollecting and incorporating outcomes to improve future model or agent performance.
Few-shot promptingSupplying a few labeled examples in the prompt to guide the model to the desired task behavior.
Fine-tuningUpdating a pretrained model's weights on domain data to specialize behavior.
FoundationsA category for core model, prompting, and training concepts that underpin agent capabilities.
Function callingStructured tool invocation where the model outputs a function name and arguments for execution.
Functional agentA deterministic integration‑oriented agent that bridges AI prompts to APIs or devices with minimal autonomy or learning.
GDPREU regulation governing personal data processing, rights, and cross‑border transfers.
GlossaryA controlled vocabulary that standardizes definitions across teams and systems.
Goal-based agentAn agent that selects actions by evaluating how choices advance explicitly defined goals.
Governance frameworksA category for structured practices to manage AI risk, accountability, and security across agents.
GradientThe derivative of the loss with respect to parameters used to update the model during training.
Ground truthAuthoritative labels or facts used to train or evaluate models and agents.
GroundingLinking outputs to cited sources or retrieved evidence to ensure factuality and traceability.
GuardrailsControls that prevent unsafe, off‑policy, or low‑quality outputs during agent execution.
Handoff latencyThe time required to transfer a session from an agent to a human while preserving context.
Hierarchical agentAn agent organized in tiers where higher levels decompose tasks and supervise lower‑level execution.
HIPAAA U.S. healthcare regulation protecting medical information and governing its use and disclosure.
Human-in-the-loopHumans review or intervene in critical steps to ensure safety, accuracy, and accountability.
Hybrid searchCombining lexical (e.g., BM25) and semantic retrieval to improve recall and precision.
IdempotencyDesign that ensures repeated requests have the same effect as one, preventing duplicates.
In-context learningAdapting behavior from examples and instructions in the prompt without weight updates.
Incident responseCoordinated steps to triage, mitigate, and learn from production failures or breaches.
IndexA data structure that accelerates search over documents or vectors for fast retrieval.
InstructionsRules and objectives given to a model or agent that shape behavior and output quality.
IVRInteractive Voice Response systems that interact with callers via menus, speech, or DTMF.
JailbreakA technique to circumvent safety controls and elicit restricted outputs from a model.
Key managementSecure generation, storage, rotation, and use of cryptographic keys.
Knowledge and retrievalA category for grounding methods that connect agents to external knowledge.
Knowledge baseA curated repository of documents and answers used to support retrieval and support agents.
Knowledge graphA graph of entities and relationships that supports reasoning and semantic retrieval for agents.
Large language modelA transformer‑based model trained on large corpora to understand and generate language and code.
LatencyThe time from request to response, often tracked at p50, p95, and p99 for reliability.
Latency targetA specified real‑time responsiveness goal for voice interactions (e.g., <300 ms turn‑start).
Latent spaceThe continuous vector space where models represent features and relationships of data.
Learning agentAn agent that adapts behavior from feedback or data to improve performance over time.
LoggingRecording inputs, actions, tool calls, and outputs for debugging, audits, and evaluation.
Long-term memoryPersistent knowledge stored outside the context window and retrieved when relevant to the task.
LoRAA lightweight fine‑tuning technique that learns low‑rank updates to model weights.
Loss functionA scalar objective that measures prediction error during training and guides learning.
Max tokensThe maximum number of tokens the model can generate in the response.
MemoryStored information from prior interactions or facts that an agent can recall for continuity and accuracy.
ModelA parameterized system that maps inputs to outputs based on learned patterns from data.
Model versioningTracking and controlling model variants and their metadata across environments.
Model-based reflex agentA reflex agent that maintains an internal model of the world to infer hidden state while still lacking planning.
MonitoringContinuous measurement of agent behavior, latency, cost, and quality during operation.
Multi-agent systemA system of multiple interacting agents that coordinate or compete to achieve complex objectives.
MultimodalModels or agents that process and generate across text, images, audio, and other modalities.
NIST AI RMFA framework with core functions Govern, Map, Measure, Manage to manage AI risks and improve trustworthiness.
ObservabilityUsing logs, metrics, and traces to understand internal system behavior from external outputs.
ObservationFeedback received after an action, such as tool results or user responses, used to update state.
Operations and SREA category for running agents in production with reliability, cost, and change management controls.
OrchestrationCoordinating models, prompts, tools, data, and control flow to execute AI tasks reliably and repeatably.
Orchestration layerThe component that sequences prompts, tools, memory, and control logic for agents.
OverfittingWhen a model memorizes training data patterns and fails to generalize to new data.
OWASP agentic taxonomySecurity‑oriented taxonomy and patterns for building and testing secure autonomous agents.
PersonaA defined role and tone that constrain how an agent communicates and prioritizes actions.
PIIPersonally identifiable information that requires heightened protection and access control.
PlanA structured, ordered set of steps that guides the agent through execution toward a goal.
PlanningCreating a sequence of steps or subgoals that an agent follows to accomplish a task.
PlaybookGuided patterns and templates for implementing repeatable AI workflows.
PolicyA set of rules and preferences that constrain agent behavior for safety, compliance, and quality.
Policy complianceAdhering to organizational, legal, and platform rules in prompts, tools, and outputs.
Positional encodingSignals added to tokens so a transformer can represent order in sequences.
PostmortemA blameless analysis after incidents to identify root causes and corrective actions.
PrecisionThe fraction of returned items that are relevant, used to assess retrieval or classification.
PrivacyProtecting individuals' data and limiting use to authorized, disclosed purposes.
Privacy and complianceA category for safeguarding data and meeting regulatory obligations across the lifecycle.
ProfileStructured attributes about a user or entity that personalize an agent's decisions and responses.
PromptInput text (and other signals) that instructs a model on task, style, and constraints.
Prompt injectionA malicious input that manipulates the model to ignore instructions or perform disallowed actions.
Prompting and samplingA category for techniques that steer generation and control randomness and diversity.
ProvenanceMetadata describing where data and outputs came from and how they were produced.
PseudonymizationReplacing identifiers with tokens while retaining linkability under controlled conditions.
Query expansionImproving retrieval by adding related terms or reformulations to the original query.
QueueA buffer that evens out bursty workloads and supports retries and ordering.
Rate limit budgetThe allotted quota of calls or tokens the agent can consume over time.
Rate limitingRestricting request frequency to control load, cost, and fair use of APIs.
Re-rankingReordering retrieved results with a secondary model to improve relevance at the top of the list.
Real-time monitoringObserving latency, errors, and quality during active sessions for rapid intervention.
ReasoningThe process of transforming inputs and knowledge into intermediate thoughts that justify and guide actions.
RecallThe fraction of relevant items that are returned, complementing precision.
Red teamingAdversarial testing of AI systems to find failures, safety gaps, and abuse pathways.
ReflectionHaving the agent review its own steps or outputs to correct errors and improve future actions.
RegularizationTechniques that reduce overfitting by discouraging overly complex or brittle models.
RetriesRe‑issuing failed calls with safeguards to improve reliability without runaway costs.
RetrievalSelecting relevant items from a corpus based on a query to inform downstream generation.
Retrieval-Augmented Generation (RAG)Generating answers grounded in external knowledge fetched at query time to reduce hallucinations.
Retry policyA strategy for repeating failed calls with limits and delays to improve robustness.
RunbookStep‑by‑step procedures for handling common operational tasks and issues.
SafetyPractices and controls that reduce harm, misuse, and unintended behaviors in AI systems.
Safety and governanceA category for practices and controls that reduce harm and ensure responsible use.
Safety policyDocumented rules describing allowed and disallowed content and behaviors.
Safety violation rateThe frequency of outputs that breach safety policies over a defined sample.
SchedulerA component that runs tasks or checks triggers at specified times or intervals.
Secret managementManaging credentials and tokens required for tool and API access.
Self-attentionComputing attention within a sequence so each token can use information from other tokens.
Semantic searchFinding relevant content by meaning rather than exact keyword matches using embeddings.
SessionA bounded interaction period that groups turns, state, and budgets for an experience.
Short-term memoryThe active context available inside a model's context window during the current turn or task.
Simple reflex agentA rule‑based agent that reacts to current percepts without internal state or planning.
SLAA contractual commitment on availability and performance targets for a service.
Sliding windowA strategy that feeds overlapping context segments to a model to fit long inputs within limits.
SLOAn internal target for key reliability metrics such as latency and uptime.
Smoke testA lightweight check that validates basic functionality and safety before full runs.
SOC 2A service organization control standard for security, availability, processing integrity, confidentiality, and privacy.
Source citationAttaching references to outputs so users can verify origin and relevance of information.
Source groundingConfining answers to retrieved or known sources to improve factuality and trust.
Speech-to-textAutomatic conversion of spoken language into written text.
StateThe internal information an agent maintains across steps, such as goals, plans, and intermediate results.
System promptThe highest‑priority instructions that define role, goals, and boundaries for the assistant or agent.
Task decompositionBreaking a complex goal into simpler, solvable subtasks to improve reliability and control.
Task success rateShare of tasks completed to specification without human escalation.
TelemetryCollected metrics and traces that describe agent performance and errors in production.
TemperatureA randomness parameter controlling output diversity, with higher values yielding more varied text.
Termination criteriaConditions that signal an agent to stop, such as achieving goals, hitting limits, or detecting failures.
Terms of useConditions governing how users and organizations may access and use an AI system.
Text-to-speechSynthesis of natural‑sounding audio from text.
ThroughputThe number of tasks or tokens processed per unit time under given resources.
TimeoutThe maximum wait time before abandoning a tool call or model request.
TokenThe smallest unit of text the model processes, such as words or subword pieces.
Token usageThe number of tokens consumed across prompts, contexts, and generations.
TokenizationConverting text into tokens according to a model's vocabulary rules.
Tool catalogThe set of available tools with schemas and permissions that an agent can invoke.
Tool invocationExecuting an external function with structured inputs produced by the model.
Tool outputThe result returned by an invoked function that the agent uses as new context.
Tool schemaA formal specification of function name, parameters, and types for safe calling.
Tool selectionChoosing which available tools to call and in what order based on the task and context.
Tool useSelecting and invoking specific tools during a task to extend an agent's capabilities beyond text generation.
Tool-augmented retrievalUsing tools or APIs during retrieval to transform queries or fetch structured data beyond static corpora.
ToolsExternal functions or APIs an agent can call to obtain information or take real‑world actions.
Top-kSampling that restricts token choices to the top k most probable candidates.
Top-pNucleus sampling that limits choices to the smallest set whose cumulative probability exceeds p.
Training and optimizationA category for methods that adapt or improve models for tasks and domains.
TranscriptionConverting speech audio into text with ASR models and pipelines.
TransformerA neural architecture built on self‑attention that powers modern LLMs.
TriggerA condition or signal that initiates a workflow step or agent action.
TruncationDropping least‑relevant context when inputs exceed the model's token limit.
Trustworthy AI glossaryA NIST resource that standardizes terminology to align teams on responsible AI concepts.
Turn countThe number of conversational exchanges needed to resolve a user's request.
User promptUser‑provided instructions or questions that the model responds to within constraints.
Utility-based agentAn agent that optimizes a numeric utility function to balance trade‑offs among competing outcomes.
VarianceThe sensitivity of a model's predictions to changes in training data, often trading off with bias.
Vector databaseA database optimized for storing and searching embeddings via nearest‑neighbor similarity.
Voice activity detectionDetection of speech segments in audio to control turn‑taking and barge‑in.
Voice agentAn agent that communicates over telephony or voice channels with ASR and TTS in real time.
Voice and multimodalA category for real‑time speech, telephony, and non‑text modalities in agent interactions.
WebhookAn HTTP callback that delivers event notifications to trigger workflows.
WindowingTechniques that slide or segment long inputs to fit within the context window.
WorkflowAn ordered set of steps and decision points that structure how tasks are completed.
Zero-shot promptingPerforming a task from instructions alone without any examples in the prompt.