Walid Boulanouar on Building an AI Operating System
A practical take on Walid Boulanouar's AI OS idea: roles, permissions, memory, and agent-ready workflows for scale.
Walid Boulanouar recently shared something that caught my attention: "creating an ai os for your company is one of the smartest moves you can make today." He followed it with the reality many teams hit after early growth: "tasks everywhere" and the constant effort of "tracking work in notion/clickup" while trying to understand what each engineer was doing.
That combination - rapid client growth plus rising operational chaos - is exactly where an "AI operating system" stops being a buzzword and starts becoming a competitive advantage.
In his post, Walid described a very specific turning point: they grew to 14 clients, added more, and then the complexity of running the business slowed them down. Their response was not to keep stacking tools. They built their own internal AI OS (they call it "aytalent"), with clear roles, strict permissions, and a principle that every new feature should be usable by an AI agent, not only by a human.
This is my attempt to expand on what Walid is pointing to, and translate it into a practical blueprint you can apply even if you are not building a full internal platform.
The real problem is not productivity - it is coordination
When Walid writes "too many moving parts" and "we slowed down," he is naming a classic scaling failure mode: coordination costs grow faster than headcount.
At 3-5 people, you can manage work with informal updates, a shared Notion, and a weekly call. At 10-20 people across multiple clients, you start paying a tax in:
- Context switching (what is the latest status, where is the source of truth?)
- Permission ambiguity (who can see what, who owns what?)
- Accountability drift (who changed this, why, and with what context?)
- Tool sprawl (each new workflow gets a new SaaS)
An AI OS, as Walid frames it, is less about "adding AI" and more about redesigning operations so the company can reliably execute with traceability.
"each person only sees what they are allowed to see: clear permissions, clear responsibilities."
That sentence is easy to skim past, but it is foundational. AI systems amplify whatever structure you already have. If your permissions, roles, and ownership are fuzzy, your AI layer becomes another source of confusion.
What an "AI OS" actually means in practice
Most teams hear "operating system" and imagine a huge platform rebuild. Walid's post suggests something more concrete and incremental.
An AI OS is a set of company primitives that define how work moves:
- Identity and roles (admins, clients, engineers, managers)
- Permission boundaries (what can be viewed, edited, exported)
- A workflow engine (tasks, approvals, handoffs)
- An execution layer (humans, automations, and agents)
- An audit trail (the fingerprint of who did what, and why)
- Memory and retrieval (context that can be recalled instantly)
Walid highlights several of these directly.
1) Roles and permissions are not "enterprise overhead"
Walid explicitly built around roles: "admins, clients, engineers, managers." That is not bureaucracy. That is what makes scaling possible without turning every request into a DM thread.
If you want AI agents to participate in work, you need to decide:
- Which agent acts on behalf of which role?
- What data is the agent allowed to access?
- What actions can it take without approval?
Think of it like least-privilege access for humans, extended to software workers.
2) Build features so an agent can execute them
Walid wrote: "we decided that every new feature must be built in a way that an ai agent can execute it or interact with it."
This is the core shift from "automation" to "agent-ready operations."
Automation usually means: trigger - do step - output result.
Agent-ready design means:
- The workflow is expressed in a way a machine can reason about (inputs, constraints, expected outputs)
- The system exposes actions via APIs or tools (create task, update status, request approval)
- The system captures decisions (why something was done, not only that it was done)
If you adopt this rule, your internal tooling evolves into something AI can safely operate.
3) The "fingerprint" is your audit trail and your training data
Walid describes going beyond automation to record:
- "who did what"
- "what context was used"
- "what decisions were made"
This is the missing piece in many AI implementations. Teams plug in a chatbot, get outputs, and then struggle with reliability because nothing is traceable.
A fingerprint gives you:
- Accountability (humans can review and correct)
- Debuggability (you can see which context led to which action)
- Improvement loops (you can refine prompts, tools, or policies)
- Compliance readiness (especially when client data is involved)
It also turns everyday execution into structured data you can later use to standardize best practices.
Why the MCP layer matters: connecting execution with memory
Walid says, "then we added an mcp layer. that changed everything." He also mentions they are heavy users of Claude Code and that everyone uses it daily.
Even if you are not deep in the jargon, the idea is straightforward: your AI tools need a consistent way to access the right context and take the right actions.
In other words, an MCP-style approach helps you:
- Log context properly (store the "why" with the work)
- Retrieve it instantly (no scavenger hunt through docs)
- Connect execution with memory (the system gets smarter)
This matters because most teams fail at AI for a simple reason: the model is not the bottleneck. The bottleneck is context.
If an engineer asks, "What is the latest decision on this client integration?" and the answer is scattered across Slack, Notion, and a ticketing tool, the AI cannot reliably help without a memory layer.
A practical blueprint: build an AI OS MVP this weekend
Walid's call to action is refreshingly direct: "stop adding tools. build your ai os, start small. build an mvp this weekend."
Here is what a realistic MVP can look like in 2 days.
Step 1: Pick one workflow that is currently painful
Choose something high-frequency and high-friction, like:
- Client onboarding
- Bug triage and assignment
- Weekly status reporting
- Change request intake
The goal is not breadth. The goal is a tight loop you can measure.
Step 2: Define roles and permissions for that workflow
Keep it simple:
- Client: can submit requests and view status
- Engineer: can see assigned tasks, add notes, push updates
- Manager: can reprioritize and approve changes
- Admin: can configure and audit everything
Write down the rules explicitly. Those rules become your AI constraints.
Step 3: Instrument a fingerprint
For every state change, store:
- Actor (human or agent)
- Timestamp
- Inputs/context used (links, files, prior decisions)
- Output (decision, created ticket, updated status)
- Confidence or rationale (even a short note helps)
This is where your system starts to feel like an OS instead of a task list.
Step 4: Add one agent that can safely operate
Start with a low-risk agent role, for example:
- Intake agent: turns client messages into structured tickets
- Summarizer agent: produces weekly updates from the audit trail
- Routing agent: assigns issues based on ownership rules
Give it limited permissions. Require approval for anything destructive.
Step 5: Build retrieval that matches how people ask questions
Do not overcomplicate embeddings on day one. You can get far with:
- A searchable timeline per project
- A decision log
- A "latest context" snapshot per task
Then integrate with the coding assistant your team already uses, so context is one command away.
Data-first, then AI-first
Walid closes with a point that more teams should adopt: "be data-first and then make the move to ai-first company."
If your data is inconsistent, your AI outputs will be inconsistent. So the sequence matters:
- Standardize how work is represented (statuses, owners, definitions)
- Capture context and decisions as data
- Only then scale agent capabilities
Once you do this, AI stops being a novelty and becomes a reliable part of operations. The payoff is not just faster execution. It is calmer execution.
The takeaway I am borrowing from Walid
What resonates most in Walid Boulanouar's post is the restraint: when things got messy, they did not keep buying tools. They built a system where roles are clear, permissions are explicit, and every feature is designed so an AI agent can participate with traceability.
If you are feeling the same "tasks everywhere" chaos, consider adopting just one principle this week: every workflow change should produce a clean fingerprint. That alone will make your company easier to run, and it sets you up for the AI-first leap when you are ready.
This blog post expands on a viral LinkedIn post by Walid Boulanouar, get one engineer with swarm of agents | aiCTO ay automate & humanoidz | building with n8n, a2a, cursor & ☕ | advisor | first ai agents talent recruiter. View the original LinkedIn post →