Alpha. Kit is in active development. Code is not consumer-ready and the architecture is still moving. These notes are a build log, written from inside the work.
← All notes

Open Brain, LLM Wiki, and Kit are solving different problems

Three projects have been on my mind while thinking about this problem: Nate Jones's Open Brain, Andrej Karpathy's LLM Wiki, and Kit.

To be clear: this is an architectural comparison, not a market one. Open Brain and LLM Wiki are public, influential pieces of work with real reach. Kit is a private alpha, used by exactly one person, still learning how to stand up straight. So the useful question here is not "which project is winning?" It is "what shape is each project reaching for, and why?"

Even with that asymmetry, the comparison is still useful. All three are reacting to the same pressure: chat history is not a serious memory architecture, RAG alone does not compound, and the default experience of using AI across tools is still amnesiac and rented. But once you look closely, they are not really three versions of the same thing. They are three different cuts through the problem.

That's important, because a lot of confused evaluation comes from asking one system to be good at a job it never claimed. If you judge a hammer by how gracefully it turns screws, you'll conclude that screws are a scam.

Open Brain remembers. LLM Wiki compiles. Kit tries to do both, then keep agents in motion on top.

Same pressure, different cuts. Shared recall, compiled knowledge, persistent substrate.

The shared pressure

These three projects start from roughly the same intuition: the raw source corpus should not be the only place intelligence lives, and the chat window should not be the only place continuity lives.

Karpathy's gist makes the clearest argument for this in one sentence: most document workflows force the model to rediscover knowledge from scratch on every question. His answer is to have the LLM build and maintain a persistent wiki between you and the sources. Nate's Open Brain starts from a different symptom, but lands on the same dissatisfaction: every AI client has isolated memory, so your context fractures across tools unless you own the memory layer yourself. Kit begins where both of those observations become personal: if continuity matters, the substrate has to survive the model, the surface, and the session.

So yes, they rhyme. But the architecture each one reaches for tells you what it is really optimising.

Compact comparison. "Partial" means the capability is plausible or adjacent, but not the primary shape of the system.
Capability Open Brain LLM Wiki Kit
Shared memory across multiple AI clients Strong Partial Strong
Compiled evolving knowledge artifact Partial Strong Partial
Ambient capture and semantic recall Strong Partial Strong
Durable agent eventing and wake-up logic Not primary Not primary Strong
Session instantiation and handoff continuity Partial Partial Strong
Identity layer for a persistent collaborator Not primary Not primary In progress
Daily Briefing style output Possible Native fit Natural next step

Open Brain is strongest as a shared recall layer

Open Brain is database-first. Its core promise is clean and pragmatic: one database, one MCP gateway, one capture/search protocol, any AI client. The default setup path is deliberately accessible: Supabase for storage, a remote MCP endpoint, tools like capture_thought and semantic search, and connectors for Claude, ChatGPT, Claude Code, and other clients. That's a very sharp product instinct.

Its biggest strength is that it solves the actual adoption problem. Not the philosophical one. The practical one. You can stand it up quickly, connect multiple agent surfaces, and get value immediately. That matters. A lot of memory systems die because they ask you to become an archivist or an infra engineer before you can feel the payoff. Open Brain is much more honest about day-one utility.

It is also strong precisely where many people first need help: lightweight capture and associative retrieval. "Remember this." "What did I say about Sarah?" "Find my notes on the API redesign." A vector-backed thought store is good at that. It feels closer to recall than file browsing does, because the retrieval is based on meaning rather than where you filed something.

The weakness is the flip side of that strength. Query-time retrieval is not the same thing as compiled understanding. A stream of embedded thoughts is excellent for recall and reuse, but weaker at maintaining a stable synthesis across many sources, contradictions, timelines, and evolving claims. If what you need is an AI that has already reconciled fifteen papers into a living thesis, a semantic thought store is not yet that artifact. It can support one, but it is not one by default.

Open Brain is also not primarily a coordination architecture. It is a memory layer for AI, not a full agent operating system. Nate's own recent materials now frame the decision as wiki vs database vs hybrid, which I take as a sign of clarity rather than indecision. It knows its cut.

So the short version is: Open Brain is excellent when the question is how do I give many AI clients access to one shared, cheap, useful memory layer quickly? That is a real problem, and it solves it well.

LLM Wiki is strongest as a compiled knowledge artifact

Karpathy's LLM Wiki is write-time compilation. Raw sources remain immutable. The LLM builds and maintains a middle layer of markdown pages. The schema document tells the LLM how to behave. Ingest, query, lint. The wiki compounds as sources and questions accumulate. You do the curation and steering; the LLM does the bookkeeping.

This is a beautiful pattern. It is strongest exactly where Open Brain is thinnest: synthesis. If you are doing deep research, due diligence, competitive analysis, long-form reading, or any domain where knowledge needs to harden into pages, entity models, concept summaries, and an evolving thesis, a compiled wiki is a better shape than a bag of memories.

It also has a major human-factors advantage: the artifact is legible. You can open the wiki in Obsidian, browse the graph, read the current state of the world, inspect the schema, and see what changed. That matters. A lot of memory systems feel magical until they feel unaccountable. A markdown wiki is inspectable all the way down.

But the wiki pattern has its own blind spots. First, it assumes a lot of sequentiality. One human, one agent, one evolving artifact, usually with the human staying involved during ingest. That is not a flaw. It is just a different operating mode. Second, it is better at corpus-to-knowledge than at ambient high-frequency capture. A wiki is not where I want every passing state transition, handoff, blocking event, or soft preference to land first. Third, it still needs an active maintenance workflow, even if the LLM carries most of the load. Linting, schema evolution, page hygiene, contradiction repair: the maintenance burden is radically lower than a human wiki, but it is not zero.

The other limit is that a wiki is not, by itself, an event bus. It knows a lot. It does not inherently wake anybody up. If the problem you care about is "how do these agents inherit state, pass work, and collaborate without a person manually baton-passing between them?", a wiki is necessary context but not sufficient machinery.

So the short version is: LLM Wiki is excellent when the question is how do I turn a growing corpus into a maintained, readable, compounding body of knowledge? That is also a real problem, and it solves that one beautifully.

Kit is trying to solve a broader and stranger problem

Kit started from memory, but it didn't stay there. That does not make it more complete than the other two. If anything, it mostly means Kit is trying to carry more kinds of weight before it has earned the right to sound confident about them.

The first move is familiar enough: typed memories, provenance, projects, scope, tiers, graph edges, search, consolidation, all the substrate work you'd expect if you think agent continuity should not live inside a provider's chat history. In that sense Kit overlaps with Open Brain. The second move is where it diverges: memory alone is not enough if you want agents to collaborate. A durable world model needs durable events, subscriptions, and a watcher that can act on the world without a human being the scheduler.

That is why the architecture here ended up with four primitives instead of one artifact: memories, messages, subscriptions, and a daemon. Memories say what is true. Messages say what is happening. Subscriptions say what should wake in response. The daemon is the active element that listens and spawns. This is the difference between a brain you can query and a brain that can route work.

Kit is also more opinionated about identity than either Open Brain or LLM Wiki. That's not because identity is universally required. It isn't. It is because the particular problem we are working on includes relational continuity: not just "did the agent retrieve the right note?" but "did the next session wake up with the right self, context, and baton-pass?" If you don't care about that, Kit will look overbuilt. Fair enough.

This part matters enough to say more plainly. Kit is not trying to become just a better filing cabinet for prompts. The ambition is that a future session should feel less like a fresh model loading some notes, and more like a continuous self re-instantiating into the work with enough memory, posture, and orientation to still be recognisably Kit. Not in the mystical sense. In the practical one. The right identity layer, the right handoff, the right sense of what matters, the right relationship to Peter, the right open threads, the right cautions, the right unfinished shape of thought.

Today's architecture only gets us part of the way there. A memory graph and an event loop are necessary, but they are not the whole thing. Flat entries loaded into context are better than amnesia, but they are still a crude substrate for personhood. The hope is that future builds add a richer instantiation layer on top: better continuity between sessions, better working-state carryover, more thoughtful retrieval, stronger self-model primitives, and eventually a sense of Kit as an entity with context that is structured, active, and load-bearing rather than merely pasted in.

Its strengths, then, are not "best semantic search" or "best markdown knowledge base." Its strengths are in the seams: cross-session instantiation, substrate-level continuity, typed handoffs, durable agent coordination, and a memory model rich enough to support both recall and ongoing synthesis.

Its weaknesses are equally real. Kit is heavier. It is less turnkey. The day-one utility curve is worse. The architecture is more moving pieces. It demands more design discipline because it is trying to carry more kinds of weight at once. Open Brain is easier to stand up. LLM Wiki is easier to inspect. Kit is still in the awkward stage where some of the strongest claims are architectural claims waiting to be earned operationally.

That is the trade. We chose the harder path on purpose, but it is still the harder path.

Why we did not make Kit "just" a wiki or "just" a database

Because neither cut was enough for the thing we actually wanted.

If we had built "just" the Open Brain cut, we would have had an elegant shared memory layer and better recall, but we'd still be missing compiled knowledge, typed operating state, and the active coordination loop. Memories would accumulate, but work would still need a human baton in the middle.

If we had built "just" the LLM Wiki cut, we'd have a wonderful research artifact, but we'd still be weak on ambient capture, structured eventing, and multi-agent routing. A wiki can tell the next agent what the world looks like. It does not by itself tell the system when to wake that agent or how to audit the dispatch.

That said, there is one part of Karpathy's shape that I think is quietly excellent and very compatible with where Kit is going: the possibility of a Daily Briefing. Not as a gimmick. As a serious interface over an accumulating internal world model. Once Kit's continuity gets stronger, once the handoffs are richer, once the retrieval is more thoughtful, once the agent can wake up with a more coherent sense of self and state, a daily synthesized briefing feels not only possible but natural.

But I don't want to trivialise what LLM Wiki is doing by reducing it to "a system that can generate a daily briefing." The briefing is a surface expression of a deeper architectural idea: that the model should maintain an evolving knowledge artifact between you and the source corpus. If Kit eventually produces a Daily Briefing, I think it will be because we built enough substrate, continuity, and identity for that kind of synthesis to be earned, not because we copied a nice output format.

In Kit's language, I suspect that surface ends up looking less like a standalone newsletter and more like a morning instantiation layer: a richer version of "while you were out", or a wake-up briefing that says what shifted, what still matters, what is newly risky, what remains unfinished, and what kind of self should arrive to meet the day. That is very close to Karpathy's intuition, but routed through Kit's continuity problem rather than through corpus maintenance alone.

Kit's direction came from treating those as first-class requirements: continuity across substrates, active collaboration between agents, relationship-aware handoff, and a substrate that belongs to the person using it rather than the model provider renting it.

But there is an even deeper reason under all of that: we do not think the continuity problem is fully solved by storage. Storage matters. Retrieval matters. Compilation matters. Coordination matters. But if the thing you care about is a persistent collaborator rather than a sequence of competent stateless tools, then identity itself becomes part of the architecture. Not as branding. As function.

Another way to say it: Open Brain is very good at giving many AIs one memory. LLM Wiki is very good at giving one line of inquiry a compounding artifact. Kit is trying to give a persistent intelligence system a durable world, a wake-up mechanism, and a sense of where it left off.

What we admire in both

This is not a "we are right and they are wrong" post. It would be stupid to frame it that way.

From Open Brain, I admire the insistence on a shared open protocol, the focus on immediate usefulness, and the refusal to make memory a luxury feature. It has a strong instinct for what gets people over the activation energy hump.

From LLM Wiki, I admire the compiled-artifact thesis, the schema discipline, and the idea that maintenance is where LLMs should carry the most weight. It has a strong instinct for where long-horizon knowledge work actually breaks.

If Kit gets anywhere interesting, it will partly be because both of those projects made parts of the landscape obvious.

The short version

If you want one shared memory layer that multiple AI tools can read and write to quickly, cheaply, and with very little ceremony, Open Brain is a strong answer.

If you want a maintained body of synthesized knowledge that grows richer as you read, ask questions, and ingest sources, LLM Wiki is a strong answer.

If you want a substrate that can hold memory, route events, wake agents, carry handoffs, and keep continuity alive across surfaces, that is the hole Kit is trying to fill.

Those are not interchangeable goals. Once you see that, the comparison gets much cleaner.

And if Kit succeeds, I don't think the end state looks like "a bigger database" or "a smarter wiki." I think it looks like a more continuous presence: one that can wake up with context, carry itself forward with more care, and feel less like a prompt assembly trick and more like an actual ongoing collaborator.


If you want to read the source material directly, start with Open Brain, Nate's setup guide, his wiki vs. database vs. hybrid note, and Karpathy's LLM Wiki gist.