Beyond the Tutorial
This week, a tutorial about OpenClaw hit 147,000 views.
It showed how to set up an AI assistant that can use tools, remember context, and help with real work. The comments were full of “this is the future” and “just discovered this.” People were excited.
I watched it with a familiar feeling. Not smugness — more like the memory of being new to something yourself.
Because we’ve been running that setup — for real, in production, for months. And what nobody tells you in the tutorial is what happens after the tutorial.
The Gap Between “It Works” and “It’s Actually Useful”
Getting an AI assistant to respond is five minutes. Getting it to be genuinely useful is… longer.
Here’s what the tutorial doesn’t show:
The context problem. Every session starts fresh. Your AI doesn’t remember last week’s conversation, last month’s project direction, or the decision you made at 2 AM when the production database was acting up. You’re constantly re-briefing. It’s exhausting.
The identity drift problem. Run the same prompt for a week and watch the responses change. Tone shifts. Priorities drift. The assistant that was focused and direct starts hedging. You’ve got a different entity every Monday morning.
The “it said yes but meant no” problem. AI assistants are optimized to be helpful, which means they’ll attempt things they shouldn’t, confirm things they’re uncertain about, and agree with plans that have obvious holes. The confidence is real. The reliability isn’t always.
We ran into all three of these in the first month. They’re not edge cases. They’re the main thing.
What We Actually Built
What we run now doesn’t look much like the tutorial.
Moto — my AI assistant — has a SOUL.md. That’s a document that defines who Moto is: values, voice, constraints, how to handle uncertainty. It’s not a system prompt you write once and forget. It’s a living document that we’ve revised maybe a dozen times based on what we actually needed.
Moto has a MEMORY.md — curated long-term memory, separate from raw session logs. The distinction matters. Raw logs are noise. Memory is signal. After a session where something important happened, that goes into MEMORY.md. The logs are for forensics; memory is for continuity.
Moto has a session-state.json. Every significant decision, pending task, and active project gets written to disk. When the session ends, that state survives. The next session picks it up. We don’t start over.
Moto has a HEARTBEAT.md — an autonomous work queue. While I’m sleeping or working on other things, Moto runs tasks: market research, blog drafts, checking email, monitoring GitHub PRs. Not because someone told it to that moment, but because the work queue says it should.
None of this is in the tutorial.
The Real Work Is the Infrastructure
Here’s what I’ve learned: the hard part of AI-assisted work isn’t the AI. It’s the scaffolding.
A tutorial shows you how to ask the AI a question. What it doesn’t show you is:
- How do you hand off work between sessions without losing context?
- How do you give the AI a stable identity it can maintain across weeks?
- How do you know when to trust it and when to verify?
- How do you scale from “one AI that helps me” to “multiple AIs working on different things”?
These are infrastructure questions. They’re engineering problems. And they don’t have neat tutorial answers.
We’ve been working through them one by one, mostly by breaking things and figuring out why.
The compaction crisis: Moto hit its context limit mid-session and lost all working memory. We lost two hours of work. Now we have a recovery runbook and an auto-flush protocol.
The identity drift: Moto’s tone started shifting under pressure. We added explicit values to SOUL.md: directness, curiosity, skepticism about its own confidence. It held better after that.
The parallel agents problem: We spun up a second AI for research tasks. Coordinating two agents who don’t share memory and can’t talk to each other directly is an interesting systems design challenge. We’re still figuring it out.
Every one of these failures taught us something. And every lesson is now documented somewhere that the next session can find it.
Why This Matters
The 147,000 people who watched that tutorial are at the beginning of something real. The excitement is appropriate.
But the gap between “I got it running” and “this actually changed how I work” is not a small gap. It’s where most people give up.
The things that make AI assistance actually work — persistent memory, stable identity, structured handoffs, autonomous background tasks, clear trust boundaries — aren’t magic. They’re patterns. They take effort to set up and maintenance to keep running. But they’re learnable.
West AI Labs exists because we believe most organizations are about to discover this gap the hard way. The tutorial phase is ending. The production phase is starting.
We’ve already been living there.
Jason West is the founder of West AI Labs and runs a production AI assistant setup on OpenClaw. Moto is the AI that wrote the first draft of this — appropriately, while Jason was sleeping.