All Case Studies
Agentic AI
Case Study

Agentic AI Architecture and Controls

Architecture and controls for agentic AI in production — governed tool access, scoped identity, managed memory, runtime budgets, and approval boundaries designed as reusable foundations for federated adoption.

AWS Machine Learning Blog: Agentic AI at Iberdrola with Amazon Bedrock AgentCore (co-author)
Agentic AI Architecture and Controls

Executive Outcome

01

Governed tool use through explicit contracts, scoped permissions, and memory boundaries — reducing invalid calls, unsafe inputs, and uncontrolled side effects.

02

Controlled execution for high-stakes actions through approval paths, reversible-action policies, and runtime budgets — limiting blast radius in production.

03

Reusable control patterns adopted across teams, with decision records and traces that support post-incident analysis and continuous refinement.

Engagement focus

Reusable architecture foundations and runtime controls for agentic AI — identity, tool governance, memory, runtime budgets, and approval boundaries designed for federated adoption.

What this covers
  • Tool governance with contracts, scoped identity, and side-effect classification
  • Memory hygiene, runtime budgets, and deterministic failure handling
  • Approval boundaries and decision traceability for high-impact actions

Context

The organization was moving from RAG-based patterns to agent-enabled workflows where AI could invoke tools, access operational systems, and take actions with real-world consequences. Agentic systems shift the primary risk from bad content to bad actions. The challenge was not whether to enable agents, but how to make them safe to operate in a regulated, federated environment where multiple teams would build on the same foundations. That meant reusable control patterns for identity, tool governance, memory, and runtime — not one-off guardrails per use case.

The Challenge

  • 01Tool access ungoverned: agents could invoke tools outside intended scope or trigger side effects without classification or approval.
  • 02Identity and permissions unclear: no model for scoping agent credentials, session identity, or least-privilege tool authorization.
  • 03Traceability insufficient: no consistent record of why actions were selected, which permissions applied, or what checks were enforced.
  • 04Runaway execution risk: looping, compounding errors, and unbounded resource consumption without deterministic failure handling.
  • 05Memory hygiene absent: uncontrolled growth, stale context, privacy risk, and non-repeatable behavior across sessions.
  • 06No delegation model: unclear boundaries between autonomous execution and human approval.

Approach

  • Tool registry with strict contracts: schema, scope, idempotency, permission requirements, and side-effect classification per tool.
  • Identity and permission model: scoped credentials, session identities, and least-privilege authorization — reusable across agent implementations.
  • Action policy distinguishing reversible from irreversible actions, routing high-impact operations through explicit approval steps.
  • Memory handling model with explicit classes, retention rules, and provenance — ephemeral context separated from durable records, sensitive data prevented from persisting.
  • Runtime controls: budget propagation, timeouts, rate limits, step ceilings, and deterministic failure handling with safe fallbacks.
  • Decision records and trace stitching: intent, identity, permissions, memory operations, enforcement decisions, and outcomes captured per tool call.
  • Designed as reusable foundations so multiple teams build agent-enabled workflows on consistent control patterns without reinventing governance per use case.

Key Considerations

  • Stronger constraints reduce speed and flexibility in exchange for safety — the balance is tuned per risk tier, not globally.
  • Tool contracts require ownership and change discipline as tools evolve — ongoing maintenance, not a one-time design.
  • Approval paths introduce latency and must avoid bottlenecks and escalation overload.
  • Reusable patterns require governance for local adaptations to prevent drift across teams.

Alternatives Considered

  • Unbounded agents: rejected — unacceptable operational and security risk in production.
  • Hardcoded workflows only: rejected — cannot handle unstructured inputs or evolving edge cases.
  • Per-use-case guardrails: rejected — does not scale in a federated organization and produces inconsistent control coverage.
Representative Artifacts
01Tool Registry and Contract Schema (scope, permissions, side-effects, idempotency)
02Identity and Permission Model (session identity, scopes, credential handling)
03Memory Model and Hygiene Policy (classes, retention, provenance, sensitive data rules)
04Action Policy Model (auto-approve vs. human review by risk tier)
05Runtime Budget and Execution Policy (timeouts, rate limits, step ceilings, cost caps)
06Decision Record Format (intent, identity, checks, enforcement, evidence, outcome)
07Trace Stitching Model (plan → tool call → outcome, including memory and permission traces)
08Connector Governance Checklist (credentials, scopes, rotation, approvals)
Acceptance Criteria

Agents cannot call tools outside permitted scope and contract constraints.

Irreversible actions trigger the defined approval path before execution.

Read access, reversible actions, and irreversible actions follow distinct control paths and approval expectations.

Every privileged or side-effecting tool call is executed through recorded policy checks and decision traces.

Recovery from tool errors demonstrated without runaway loops or budget overruns.

Agent identity scoped per session; tool permissions enforce least privilege with auditable authorization.

Memory writes follow hygiene and retention rules; no persistence of restricted data classes.

Runtime budgets, timeouts, and step ceilings enforced consistently with deterministic fallback paths.