Skip to main content
Tilbake til blogg
OpenClawMarch 28, 20268 min

Why I Stopped Using ChatGPT and Built My Own AI OS Instead

The chatbot model is fundamentally broken for real work. Here's what made me quit and what I replaced it with.

DB

David Bakke

Founder, Bakke & Co

PostShare
ForsidebildeOpenClaw

The copy-paste problem

I used ChatGPT for over a year. Paid for Pro. Used it daily. And at some point last fall, I realized I was spending more time managing the tool than doing actual work.

Here's what a typical interaction looked like: Open ChatGPT. Start a new conversation. Paste in context about the project I'm working on. Explain what happened yesterday. Remind it of my tech stack. Ask my question. Get an answer. Close the tab.

Next day: open ChatGPT. Start a new conversation. Paste in the same context. Explain what happened today. Ask a follow-up to yesterday's question, but first re-explain yesterday's question because the AI doesn't remember it.

This got old fast.

The promise of ChatGPT is that you have a brilliant assistant at your fingertips. The reality is that you have a brilliant assistant who forgets everything the moment you close the tab. Imagine hiring someone and having to re-introduce yourself every morning. That's the chatbot model.

What's actually wrong with chatbots

I want to be specific here, because the problem isn't that ChatGPT is bad. The models are impressive. GPT-4o is smart. Claude is smarter in some ways. The models are fine. The delivery mechanism is broken.

No persistence. You can pin conversations, sure. You can have long threads. But the context window fills up, old messages get summarized or dropped, and the AI gradually forgets what you told it two hours ago. For a quick question, this works. For ongoing project work, it's useless.

No specialization. When I ask ChatGPT to review code, it brings no context about my codebase, my conventions, my deployment setup. It's starting from scratch every time. The same model that's writing my marketing copy is also reviewing my TypeScript. There's no division of labor, no expertise accumulation.

No tool access. ChatGPT can browse the web (sometimes) and run Python (sometimes). What it can't do is access my filesystem, run shell commands, interact with my APIs, or connect to Slack. The moment you need it to do something rather than just say something, you're back to copy-pasting.

No parallel work. I can have one conversation at a time. If I'm waiting for a long analysis, I'm blocked. I can't ask one instance to research while another writes while a third reviews code. The model supports it in theory, but the interface doesn't.

And no coordination. Even if I had multiple ChatGPT windows open, they'd know nothing about each other. The research window doesn't know what the coding window is building. The marketing window doesn't know what the product window decided yesterday.

The moment I quit

The specific moment was a Thursday evening in late January. I was working on PortLink, my maritime SaaS product, and I needed to update the database schema. I'd been discussing this schema with ChatGPT over several conversations spanning a week. New conversation, re-explain the schema, ask about the migration, get an answer, close.

On this particular evening, I pasted in my schema and ChatGPT suggested a column structure that contradicted advice it had given me two days earlier. Not subtly, either. It recommended the exact opposite approach. When I pointed this out, it apologized and agreed with my correction, because that's what chatbots do. They apologize and agree. They don't have the context to do anything else.

I sat there thinking: this tool doesn't know me. It doesn't know my project. It doesn't know what it told me yesterday. It's very smart and completely useless for the kind of work I actually need help with.

That weekend, I started looking for alternatives.

What persistent AI looks like

OpenClaw runs AI agents inside Slack. Each agent lives in its own channel. Each agent has its own memory files, its own tools, its own model configuration. When I talk to an agent on Monday and come back on Wednesday, it knows what we discussed. Not because it has a perfect memory, but because it reads its MEMORY.md file at the start of every session.

This sounds like a small thing. It's the entire thing.

My orchestrator agent, Nyx, has a memory file that contains my standing orders, my project states, my infrastructure details, my preferences. When I say "update the PortLink spec," Nyx knows what PortLink is, what the current spec looks like, where the repo lives, and who to delegate the work to. I don't paste context. The context is already there.

Forge, my engineering agent, knows my coding conventions. It knows I use TypeScript with Supabase. It knows the folder structure of every project. When it reviews a pull request, it's not starting from zero. It's reviewing against established patterns.

Hermes, my email agent, has indexed 83,000 emails in ChromaDB. When I ask "what did Leo say about the investment terms?", it searches the actual inbox, finds the thread, and gives me a summary with dates. No copy-pasting. No opening Gmail in another tab. Just a question and an answer based on real data.

This is what AI should feel like. Not a chat window you type into, but an environment you work inside.

What you give up

I want to be fair about the tradeoffs, because there are real ones.

The UI is Slack. Not a polished chat interface with markdown rendering and file uploads. Just Slack. If you like ChatGPT's code blocks and image generation and canvas features, you'll miss them. Slack is functional, not pretty.

Onboarding takes effort. With ChatGPT, you sign up and start chatting. With OpenClaw, you install software, create a Slack workspace, configure agents, write memory files, set up API keys. The first agent takes an afternoon. It's not hard, but it's not instant.

You need hardware or API budgets. I run local models on a Mac Studio with 98GB of memory. That handles about 70% of my workload for free. The other 30% uses cloud APIs that cost money. If you don't have a machine that can run 70B parameter models, you'll pay more for cloud APIs.

Maintenance is ongoing. Memory files need pruning. Agent configurations need updating. Models get new versions. Standing orders need broadcasting. This is a system you operate, not a service you subscribe to.

And there's no mobile app. ChatGPT works on your phone. My agents live on the Mac Studio. I can talk to them through the Slack mobile app, but they can only access files and run code on the machine they live on. For me, that's fine. I do real work at my desk. For someone who needs AI on the go, it's a limitation.

Why it's worth it

The tradeoffs are real, but here's what I get in exchange.

Twenty-six specialized agents, each with deep knowledge of their domain. When my Content agent writes a blog post, it knows the brand voice, the target audience, the SEO strategy. When my Budget agent processes a receipt, it knows the accounting categories, the Fiken API endpoints, the VAT rules. Specialization compounds over time. These agents get more useful every week because their memory and context accumulate.

Parallel execution. While Scout researches a competitor, Forge can review code, and Hermes can draft an email. They run simultaneously within the API rate limits. My throughput on a busy day is genuinely higher than a two-person team could manage. I don't say that to brag. I say it because it surprised me.

Data stays on my machine. My emails, my code, my financial data, my client information. It's on my Mac Studio in Oslo. The local models process it without sending anything to the cloud. The cloud models see what I send them, which I can control per agent. This matters when you're handling client projects and financial data.

Total control over behavior. When an agent does something wrong, I can fix the root cause. Update the memory file, adjust the instructions, change the model. With ChatGPT, when the model does something wrong, you just... hope the next version is better.

And there's something less tangible that matters too. When you build a system like this, you start thinking differently about your own work. You stop asking "how do I do this task?" and start asking "how do I define this task well enough that an agent can do it?" That reframing alone has made me better at delegating to humans too.

This isn't for everyone

I want to be clear about that. If you use AI for quick questions, brainstorming, or occasional writing help, ChatGPT is fine. It's good at that. The web interface is polished, the mobile app works, and you don't need to configure anything.

But if you're building products, managing multiple projects, and spending hours a day with AI tools, the chatbot model will hold you back. The context loss, the lack of specialization, the inability to coordinate across tasks. These aren't minor annoyances. They're structural limitations of the conversation-in-a-browser paradigm.

I spent a year working around those limitations before I realized I didn't have to.

OpenClaw isn't the only way to build persistent, specialized AI. There are other agent frameworks out there. The specific tool matters less than the principle: AI that remembers, that specializes, that connects to your actual tools, and that runs as infrastructure rather than as a chat window.

Once you experience that, going back to chatbots feels like going back to dial-up.

openclawchatgptai-setupproductivity