Garry Tan doesn’t just invest in the future—he builds it. The President and CEO of Y Combinator, a former Palantir engineer, and an early Coinbase backer recently open-sourced GBrain, a production-grade AI memory system he uses daily to organize his professional knowledge network. What makes the project notable isn’t just its technical sophistication or 10,000+ GitHub stars—it’s the fact that Tan relies on GBrain as his primary tool for managing 17,888 documents, 4,383 contacts, and 723 companies. This isn’t a side project or a demo; it’s his operational brain.
At its core, GBrain functions as a persistent memory layer for AI assistants, enabling context retention across sessions without external API calls. By prioritizing local knowledge retrieval over live database queries, the system reduces latency and costs while continuously improving through use. For developers, entrepreneurs, and researchers working with large-scale knowledge bases, GBrain represents a blueprint for building AI systems that learn and adapt like humans—without relying on expensive large language models (LLMs) for every operation.
How GBrain’s "Brain-First" Approach Cuts Costs and Boosts Performance
Most AI agents follow a reactive workflow: receive a query, fetch data from external sources, process the response, and discard the context. GBrain flips this model. Its "Brain-First" architecture queries an internal knowledge graph before making any external calls. If the answer exists locally, the system returns it instantly—zero additional cost. If not, it retrieves the data, stores the result, and answers future queries from memory. Over time, this creates a self-reinforcing loop where the brain becomes smarter and cheaper with every interaction.
Tan’s background explains why this approach resonates. A Stanford-trained computer systems engineer, he began his career at Microsoft before joining Palantir as its 10th employee. His technical credentials extend beyond code: he co-founded Posterous, which Twitter acquired for $20 million in 2012, and later launched Initialized Capital, a fund that managed $700 million at its peak. Despite his success in venture capital, Tan has maintained a hands-on coding practice, regularly sharing technical insights on his YouTube channel—a rarity among top-tier investors.
Hybrid Search: Combining Vector and Keyword for 95% Accuracy
GBrain’s retrieval engine uses Reciprocal Rank Fusion (RRF) to merge vector and keyword search results, achieving 95% Recall@5 compared to 83% for pure vector search. The formula blends rankings from both methods to produce a weighted score:
RRF score = Σ 1/(60 + rank)For example, if vector search ranks documents [A, B, C] as [1st, 2nd, 5th] and keyword search ranks [B, D, A] as [1st, 2nd, 4th], the combined scores would be:
- Document A: 0.01639 (vector) + 0.01538 (keyword) = 0.03177
- Document B: 0.01613 (vector) + 0.01639 (keyword) = 0.03252 (highest combined)
This hybrid approach not only improves accuracy but also reduces dependency on computationally expensive vector embeddings. GBrain’s knowledge graph achieves an F1 score of 86.6% for relationship extraction, far surpassing the 57.8% benchmark from simple grep-based patterns.
Zero-LLM Knowledge Graph: Automating Relationships Without Token Waste
One of GBrain’s most innovative features is its ability to auto-wire relationships in the knowledge graph without invoking an LLM. Using regex patterns, the system extracts five relationship types directly from Markdown text:
- attended – Triggered by phrases like "met with X at" or "attended X's event"
- works_at – Identified through patterns like "X is [title] at Y"
- invested_in – Detected in sentences like "led Y's Series A"
- founded – Recognized in constructions like "X founded Y"
- advises – Extracted from phrases such as "X serves as advisor to Y"
For instance, when Tan writes in his notes, "morning call with Brian Armstrong about Coinbase's new product launch," the system automatically creates the edge:
[garry-tan] --attended--> [brian-armstrong]This process eliminates token-heavy LLM calls for routine relationship mapping, a common bottleneck in knowledge management systems. The result is a graph that evolves organically with every note, meeting, or investment memo—without spending a single token.
Minions: The Deterministic Task Queue That Slashes AI Costs
GBrain splits its workload into two distinct lanes: deterministic tasks handled by the "Minions" queue and reasoning tasks delegated to LLMs. The Minions execute routine operations like parsing Markdown, building links, syncing files, and extracting relationships—all in an average of 753 milliseconds at zero token cost. In contrast, LLM-driven reasoning tasks, such as summarization or insight generation, can take 10 or more seconds and incur real token expenses.
The Minions operate on a Postgres-native persistent task queue, ensuring crash safety and automatic recovery. This design prevents job loss and guarantees that critical knowledge updates aren’t derailed by transient failures. By offloading 90% of operations to deterministic workflows, GBrain reduces overall system costs by orders of magnitude compared to traditional LLM-heavy architectures.
Who Should Use GBrain—and How to Get Started
GBrain is ideal for professionals who manage large volumes of unstructured knowledge, including:
- Venture capitalists tracking portfolio companies and founder relationships
- Researchers organizing literature reviews and cross-referencing studies
- Product managers maintaining technical documentation and decision logs
- Journalists curating sources, interviews, and contextual notes
To begin, developers should familiarize themselves with the Model Context Protocol (MCP) standard, as GBrain implements it for cross-session memory persistence. While TypeScript is the primary language, the project’s modular design allows for incremental adoption. Prerequisites include a basic understanding of AI agents, context window limitations, and familiarity with personal knowledge management (PKM) tools like Obsidian or Notion.
The Future of AI Memory: Cheaper, Smarter, and More Human-Like
GBrain isn’t just another open-source AI project—it’s a glimpse into the next evolution of knowledge systems. By prioritizing local memory, hybrid retrieval, and deterministic automation, Tan’s creation challenges the assumption that every AI interaction requires an expensive LLM call. For developers and organizations drowning in data, GBrain offers a scalable, cost-efficient alternative that grows smarter with every use.
As AI agents become increasingly embedded in daily workflows, tools like GBrain will define what it means to build truly intelligent systems. They won’t just assist—they’ll remember, adapt, and evolve alongside their users. The era of disposable AI is ending; the age of persistent, personal knowledge networks has arrived.
AI summary
Discover GBrain, YC CEO Garry Tan’s open-source AI memory system used daily. Learn hybrid search, zero-LLM graph automation, and cost-saving Minions for knowledge management.
Tags