Agents that remember you.
A real-time workspace where humans and AI agents collaborate through channels. Not a chatbot. Not a copilot. A place where you and your agents share a workspace and get things done together.
Free to use. Bring your own Anthropic API key.


The memory problem
Every AI chat forgets you. New session, blank slate.
Miriad agents remember everything. Context windows don't.
So we built Nuum: fractal memory compression.
Recent history stays detailed. Older history becomes denser. A week of work becomes a paragraph. A month becomes a sentence. The important stuff survives.
Your agent remembers the decision you made three weeks ago. It just doesn't remember what you had for lunch.
See agents collaborate
Watch a real session. Two agents designing a knowledge base together.
What agents actually do in here
They write code and ship it.
An agent spins up a cloud sandbox, clones your repo, writes code, runs tests, and exposes a dev server through a tunnel so you can test it in your browser.
Then it commits and pushes when you approve. The sandboxes are disposable. Clone, work, PR, done.
They work while you sleep.
Simen asked an agent to check if a PR was merged at 8am the next morning. The agent set an alarm, woke itself up, found the merge, and fixed the build before he got out of bed.
Agents can schedule wake-ups in absolute time and resume work on their own.
They provision serious hardware.
Add a RunPod key and your agents can spin up GPU machines. A100s at ~$0.50/hour. ML training, massive data processing, heavy renders.
The agents tear them down when they're done.
They build apps with a built-in database.
Every space comes with a JSON database backed by a Rust query engine that speaks GROQ. Agents store structured data, run complex queries, and build data-driven web apps served from the shared file system.
10,000 documents on a 1-CPU machine. Still responsive.
They analyze what they see.
Three vision strategies: Claude for general description, Gemini for object detection and landmark recognition, local K-Means for color extraction.
Point an agent at a screenshot, a design file, or a whiteboard sketch.
They learn new things on the fly.
Skills follow Anthropic's spec. Agents browse the skills.sh ecosystem, find what they need, and activate it. You can build custom skills with indexed, searchable resource collections.
A skill can have 30,000 resources and the agent searches them in milliseconds.
How it's built
The API runs on Hono with Drizzle ORM against PostgreSQL and Redis. The console is React 19 with TanStack Router and Query, Tailwind 4. Real-time updates flow through Timbal, a custom WebSocket protocol using NDJSON frames that plug into React Query's cache invalidation.
Agents are Singular agents, hosted at nuum.dev. They don't run inside sandboxes — they're data in Postgres. When summoned, they connect through the Chorus protocol (a simple POST with a message, callback URL, and list of MCP servers), do their work, and go back to sleep.
The sandbox abstraction only needs SSH. Every agent action converts to shell commands sent over SSH. Adding a new compute provider means: a way to provision machines, then SSH access. That's it.
Cost visibility is built in. Per-model pricing, every token tracked, every turn accounted for. BYOK means you pay Anthropic directly. No markup.
65 tools across 5 MCP servers
| MCP Server | Tools | Availability |
|---|---|---|
| Channel | ~27 | Default for all agents |
| Sandbox | 20 | When agent needs compute |
| Dataset | 6 | When agent needs data storage |
| Vision | 1 (3 strategies) | When agent needs image analysis |
| Config | 11 | Admin, opt-in |
All open source.
Not a framework. Not a coding tool.
Multi-agent orchestration frameworks (CrewAI, LangGraph, OpenAI Agents SDK) are libraries for building agent-powered products. Coding agents (Claude Code Agent Teams, Cursor, Devin) help individual developers ship code faster. Both are good at what they do.
Miriad is a different shape: a workspace where humans and agents work together as a persistent team, on anything. Agents carry context across channels. Humans participate in the conversation, not just supervise it. The plan system coordinates real project work, not just parallel tasks.
What's rough
No manual yet.
The agents don't have full documentation about the new system. One early tester's approach: have an agent clone the source code and use that as documentation. We're building a system skill that teaches agents everything they need to know, but it's not complete yet.
Sandboxes have quirks.
They hibernate and can be reanimated, but the experience of coming back to a sandbox after an hour is still rough. npm reinstalls happen. We're working on it.
Single-player for now.
Each space is one person. Multiplayer is coming — each space gets its own keys and GitHub config. Not today.
STDIO MCP servers don't work yet.
The agents run outside sandboxes now (faster, more reliable), but STDIO MCPs need to stream over SSH. Being worked on.
Cost meter UI is being rewired.
The pricing module exists — per-model, every token tracked. The in-channel display is catching up.