Skip to content

Deep Dive

The five technical layers, in depth

Each layer builds on the one below. Skip one, and the system breaks. This is why most AI deployments fail. They cherry-pick layers instead of building the stack.

01

CONTEXT

The system that understands

Your company's knowledge, engineered for AI. Not a static document, but a living, multi-level architecture that evolves with your business.

Context engineering describes the core skill better: the art of providing all the context for the task to be plausibly solvable by the LLM.

Tobi Lutke, CEO Shopify

Identity

Who you are, what you do, why, for whom. Your company's constitution. Changes rarely.

e.g. Vision, mission, values, positioning, ICP definition, brand voice

Operations

How you function. Processes, rules, decision frameworks. Changes regularly.

e.g. Sales process, pricing rules, onboarding SOPs, compliance requirements

Situation

Where you are now. Current priorities, active projects, real-time focus. Changes continuously.

e.g. Q2 goals, active deals, hiring priorities, current initiatives

Key insight: Cognizant announced plans to deploy 1,000 context engineers in 2025 to industrialize agentic AI. Context engineering is becoming a recognized discipline, not an afterthought.

02

DATA

The system that sees

Live operational data connected through a semantic layer. Not a data dump, but structured entities, relationships, and permissions with just-in-time retrieval.

Find the smallest set of high-signal tokens that maximize the likelihood of your desired outcome.

Anthropic, Context Engineering Guide

Semantic Layer

Entities (clients, deals, projects), relationships (client X → deal Y), temporality (this week vs last month), permissions (who sees what).

e.g. CRM contacts, calendar events, document metadata, pipeline stages

Connectors

MCP (Model Context Protocol) servers that bridge your tools to the AI. The emerging standard with 10,000+ public servers, adopted by ChatGPT, Cursor, Gemini, VS Code.

e.g. HubSpot MCP, Google Calendar MCP, Slack MCP, custom internal tools

Just-in-Time Retrieval

The AI doesn't load everything into memory. It maintains lightweight pointers and pulls data on demand, like a human who knows where to look without memorizing every file.

e.g. Querying CRM on mention of a client, pulling meeting notes when preparing a brief

Key insight: The industry is moving from raw data connections to semantic layers. Microsoft's GraphRAG combines knowledge graphs with retrieval for better reasoning. The semantic layer is becoming as critical as the database was to analytics.

03

SKILLS

The system that acts

Two modes: Intelligence (perceive, analyze, alert, safe by default) and Execution (send, create, update, governed). Skills are composable and chain into workflows.

The real bottleneck isn't what the model can do. It's whether your business processes are encoded in a way the model can act on.

Adapted from Anthropic, Context Engineering Guide

Intelligence Mode

Read-only operations. The AI observes, analyzes, and surfaces insights. No risk, nothing changes.

e.g. Daily brief, pipeline analysis, meeting insights, anomaly detection, market intelligence

Execution Mode

The AI acts on your behalf. Requires governance (Layer 4). Each action has an autonomy level.

e.g. Send email, update CRM, create report, schedule meeting, onboard employee

Composition

Skills combine. summarize_meeting + update_crm + send_followup = automated post-meeting workflow. Build once, run on every meeting.

e.g. Post-meeting flow, weekly reporting pipeline, prospect qualification chain

Key insight: This separation matters: start with Intelligence (zero risk), then progressively enable Execution as trust grows. Most AI deployments fail because they try to automate before they observe.

04

GOVERNANCE

The system that controls

Autonomy is a slider, not a switch. Four trust levels, earned through demonstrated safe behavior, like a new employee earning responsibility over time.

Autonomy and authority must be deliberate design variables, not accidents.

World Economic Forum, 2026

Level 1: Foundation

Week 1–2. Direct supervision. Every action requires human approval. The AI proposes, the CEO decides.

e.g. AI drafts an email → CEO reviews and sends. AI suggests a pipeline action → CEO approves.

Level 2: Supervised Autonomy

Month 1. Simple predefined tasks run automatically. Everything else requires review.

e.g. AI sends meeting summaries automatically. AI flags stale deals. AI still asks before emailing clients.

Level 3: Conditional Autonomy

Month 2–3. Autonomous within a defined scope. Escalates only for decisions outside its boundaries.

e.g. AI handles routine client communication. AI manages calendar scheduling. AI escalates pricing questions.

Level 4: Trust-Based Autonomy

Month 6+. Expanded authority earned through consistent performance. Hard guardrails remain active forever.

e.g. AI manages the full prospect qualification pipeline. AI never sends contracts without approval (guardrail).

Key insight: 80% of organizations have already encountered risky behavior from AI agents, including improper data exposure and unauthorized system access (McKinsey, 2025). The companies that build governance from day 1, not as an afterthought, are the ones that scale safely.

05

MEMORY

The system that learns

Four types of memory that make the system exponential. Your AIOS at month 12 is unrecognizable from day 1, and the accumulated memory is a competitive advantage no one can copy.

Context engineering is what we do instead of fine-tuning.

Simon Willison

Working Memory

The current conversation. Ephemeral, limited by the context window. Optimized by context engineering.

e.g. The CEO asks a question → AI has the current conversation in memory.

Declarative Memory

Structured facts about the company. Persistent, stored in the Context Layer. The 'knowledge' of the organization.

e.g. "Our ICP is 50–200 employee service companies in Switzerland." "Our pricing requires approval above 50K."

Episodic Memory

Timestamped history of interactions, decisions, and outcomes. Searchable, auditable.

e.g. "On March 15, deal X moved to phase 3 after the call with Y." "Last week's pipeline review flagged 3 stale deals."

Procedural Memory

Patterns learned from accumulated feedback. The system's 'know-how' that improves over time.

e.g. "Emails with a case study get 3x more responses." "Pipeline reviews on Monday produce better weekly outcomes."

Key insight: Memory is cumulative and proprietary. A competitor can copy your tools, but never your 12 months of accumulated organizational learning. The system at month 12 is unrecognizable from day 1.