
š How Gary Tan Shipped 400x More Code While Running Y Combinator Full-Time
In a span of just a few months, Y Combinator President Gary Tan went from a 13-year coding hiatus to shipping hundreds of thousands of lines of production code ā all while running one of the world's most prominent startup accelerators full-time. The projects he built have accumulated over 100,000 stars on GitHub, and the methodology behind this productivity surge reveals a fundamental shift in how software will be built in the AI era.
This isn't a story about automation replacing engineers. It's about engineers gaining wings.
š The 400x Productivity Multiplier
When Tan announced he was coding at 100x his previous rate, the internet reacted with skepticism. After applying standardized logical lines of code analysis to strip away formatting and boilerplate, the actual figure turned out higher: approximately 400x his 2013 output.
The baseline matters here. Historical software engineering literature consistently shows that professional developers ship between 30-50 lines of tested, production-ready code per day. Tan's 2013 baseline was around 14 lines per day as a part-time coder. The math checks out ā but the methodology is what's truly revolutionary.
"Using OpenClaw these days is like driving a Ferrari. It's exhilarating. It's insane. You get to do things you would never think a machine could figure out and it does it so quickly. But it's also like a Ferrari in that you better be a mechanic."
The key insight: Tan wasn't writing the code himself. He was directing 15 agents simultaneously, token maxing across multiple AI systems, and treating compute like a strategic resource rather than a cost to minimize.
šÆ The Origin Story: Gary's List
The catalyst was personal and political. Frustrated by San Francisco's educational policies ā including restrictions preventing seventh and eighth graders from taking algebra ā Tan decided to build a platform to organize civic action around issues he cared about.
Gary's List became the proving ground for a new approach to software development. What started as a simple blogging platform evolved into something far more sophisticated: an agentic newsroom capable of performing investigative journalism autonomously.
The system employs full RAG (Retrieval-Augmented Generation) with recursive crawling capabilities. For roughly $5-10 worth of Claude Opus API calls, it can:
- Ingest and analyze dozens of articles and entire books on a subject
- Cross-reference multiple sources (often 20+ per article)
- Identify consensus and dissenting viewpoints across sources
- Generate fully-sourced, long-form investigative content
The first iteration of the Posterous blogging platform (which Twitter acquired for approximately $20 million) took $4 million in funding, six or seven people, and about 18 months to build. The second version required $100,000, two people, and roughly three months. The third iteration ā built with Claude Code ā took $200 (a Claude Pro subscription) and five days.
ā” Token Maxing: The New Competitive Advantage
"Token maxing" emerged as Tan's core philosophy: if incremental compute can make something more complete, more accurate, or more representative of reality, you should spend the tokens.
This represents a fundamental break from traditional software engineering optimization, where developers minimize computational resources. In the agentic era, the constraint isn't machine time ā it's human attention and agency.
"Particularly when building agentic software now, you don't have to settle for what we did when we were humans writing the code. What if you absolutely boiled the ocean? If you were a human this would take you about a month to do this research ā you can just zap the rocks harder."
The philosophy extends beyond LLM calls to testing strategy. Tan discovered that while human developers often skimp on testing ("it's just not very fun"), AI agents don't care. He now targets 80-90% test coverage as standard practice, dramatically reducing the "slop" that plagued early AI-generated code.
š ļø The GStack Framework: Thin Harnesses, Fat Skills
GStack emerged not as a planned product but as an organic solution to repeated patterns. Tan found himself typing the same instructions into Claude Code repeatedly, so he systematized them into what became a comprehensive agentic engineering framework.
The core insight: "thin harnesses, fat skills."
A harness is the deterministic code that handles core loops ā taking user input, passing it to the LLM, executing tool calls. This should be minimal and reusable. Skills are the markdown-based prompts that encode domain knowledge, context, and procedures.
The distinction matters because:
- Code is brittle: It executes deterministic operations without understanding context or user intent
- Markdown lives in latent space: LLMs can interpret instructions flexibly, handle edge cases, and understand "who you are and what your motivations are"
GStack includes several specialized skills that reflect Tan's workflow:
- Office Hours Skill: Based on Y Combinator's founder interview process ("Who's it for? What does it do? What's the impact?")
- CEO Plan Skill: Incorporates Brian Chesky's "10-star experience" framework through metaprompting, pushing beyond obvious solutions to find 10x value for 2x effort
- Plan-Review Skill: Requires agents to create ASCII art diagrams of data flows, state machines, and dependencies before coding ā dramatically improving completeness
- QA Skill: Automatically tests UI and data mutations using a long-lived browser daemon with 70+ CLI commands
- /codex command: Brings in GPT-o1 for complex debugging ("the 200 IQ nearly nonverbal CTO")
š¼ The Daily Workflow: Conductor + GStack
Tan's production workflow runs primarily through two tools: Claude Code and OpenClaw (via Conductor). The split is roughly 50-60% Claude Code for product development and 40-50% OpenClaw for more complex agentic work.
A typical feature development cycle:
- Start with Office Hours or CEO Review skill to validate the idea
- Run design review for UI components
- Execute developer experience review for anything engineers will use
- Apply the Plan-Review skill with comprehensive testing
- Queue the approved plan in Conductor (often 15+ features running simultaneously)
- Run the QA skill for automated browser testing
- Call in /codex for complex debugging when needed
The system heavily relies on "ask user question" prompts ā preserving human agency at critical decision points rather than attempting full autonomy.
"I never want to be entirely out of the loop. I just want the machine to do the stuff that I don't want to do."
š§ GBrain: Knowledge Management for AI Systems
After building Gary's List and GStack, Tan tackled a meta-problem: his OpenClaw instance was using basic grep for code search, wasting context window and loading far more information than necessary.
The solution was GBrain, a sophisticated RAG system for agentic codebases that implements:
- Vector embeddings with pgvector in PostgreSQL
- Hybrid RRF (Reciprocal Rank Fusion) search combining semantic and keyword approaches
- Intelligent chunking strategies learned from building Gary's List
- Full integration with OpenClaw/Conductor workflows
The development process illustrated another key pattern: code reuse across projects accelerates dramatically in the agentic era. Tan simply pointed his AI agent at the existing Gary's List codebase and instructed it to extract the RAG patterns, adapt them for GBrain's use case, and integrate with OpenClaw.
š The Lines of Code Controversy
Tan's public discussion of productivity metrics sparked heated debate on social media. The criticism centered on "lines of code" being a discredited productivity measure ā padding is trivial, and code quality matters more than quantity.
The counterargument has nuance. AI agents don't optimize for lines of code the way human employees gaming metrics would. They might build the wrong thing if poorly directed, but they don't artificially inflate output.
More importantly, when measuring productivity acceleration, the baseline matters enormously. Industry literature consistently shows professional developers average 30-50 logical lines of tested, production-ready code per day. After applying standardized LLOC analysis that strips formatting and standardizes structure, Tan's 2013 output dropped by roughly 70%, while his AI-augmented output only decreased slightly ā yielding the 400x multiplier.
The real lesson isn't about lines of code as an absolute measure. It's about what becomes possible when developers with taste and domain expertise gain access to nearly unlimited execution capacity.
š” The Personal AI Revolution
Tan frames the current moment as analogous to the Homebrew Computer Club era ā when Steve Jobs and Steve Wozniak sold the Apple I as a wooden breadboard with components that users had to assemble and debug themselves.
"We could either live in a world where we have our own AI, where we have our own data, our own integrations, where we see what's happening, we write our own prompts, and we have control over what we see ā or it's corporate controlled."
The critical skill isn't using AI through polished interfaces. It's being willing to pop the hood and fix things yourself when they inevitably break. Current tools like OpenClaw are powerful but brittle ā they're kit car Ferraris that require mechanical knowledge.
This creates a binary: those willing to write their own prompts and manage their own AI infrastructure will have unprecedented capability. Those who rely solely on corporate-controlled AI products will be "below the API" ā subject to someone else's priorities, business models, and design decisions.
š° The Token Maxing Paradox: It's Actually Cheap Not To
A common objection to token maxing is cost. Running Claude Opus 4.7 or GPT-o1 aggressively can result in daily token bills of several hundred dollars.
Tan draws an analogy to San Francisco rent: "It's so expensive to not live there."
For founders and builders, token spend should be treated like rent in a high-opportunity location ā an investment in serendipity and acceleration, not a cost to minimize. The opportunity cost of not token maxing dramatically exceeds the direct expense.
His advice: push token spend as hard as possible when building something of genuine value. Economize on office furniture; splurge on compute.
ā° Time Billionaires and Borrowed Consciousness
Perhaps the most striking reframing comes from Tan's concept of "time billionaires" ā people (often young or recently unencumbered) with vast stretches of discretionary time.
In previous eras, even time billionaires were constrained by their own execution speed. Token maxing changes the equation fundamentally: you can now borrow millions of years of machine consciousness.
"I could buy millions of years of consciousness of machine consciousness. Now I can be a time billionaire. It's not my own time. It's the time of a machine doing work for me."
This reframes AI productivity tools not as automation (replacing human work) but as leverage multipliers on human agency. The constraint becomes the quality and clarity of human direction, not execution capacity.
š The Builder Identity Returns
On January 23rd, Tan tweeted: "Claude Code this week has awakened my 25-year-old self, the one that checked Red Bulls and stayed up till dawn coding. We're so back."
After 13 years away from daily coding, he returned to sleeping four hours a night and coding 20 hours a day ā not out of necessity, but because building became feasible again despite the demands of running Y Combinator.
The lesson for founders: the constraint isn't time scarcity. It's whether you're willing to treat AI capabilities as real and invest accordingly. As Tan put it:
"We started in the same place. I don't think of myself as in the sky yet. We have the same prompt. We have the same MacBook Pro. There's nothing that stands between you or me or any of us from drawing on millions of years potentially of tokens to serve humanity."
š® Looking Forward: The Golden Age of Open Source
Tan predicts that within a year, the current friction points around OpenClaw, Hermes, and similar tools will resolve. Every single person on the planet will have access to their own personal AI ā the question is whether they'll have genuine control over it.
The parallel to the personal computer revolution is exact. In that era, the defining question was whether individuals would own general-purpose computing devices or remain dependent on corporate mainframes. The PC won, and the world transformed.
Today's question: Will you have control over your own tools, or will your tools have control over you?
For builders, the implication is clear: learn to work with brittle, powerful tools now. Write your own prompts. Token max aggressively. Accept that you'll need to "fix it yourself" when systems break.
The golden age of open source is arriving not because AI makes contribution easier, but because AI makes it possible to extract and reuse patterns from existing codebases at unprecedented speed. The corpus of high-quality open source code becomes a library that any sufficiently skilled prompter can draw from.
ā Key Takeaways
- Token maxing is the new competitive advantage: Spend compute aggressively when it increases completeness, accuracy, or quality
- Target 80-90% test coverage: AI agents don't mind writing tests; untested AI code is "slop"
- Thin harnesses, fat skills: Minimize deterministic code, maximize markdown-based prompts that live in latent space
- Treat token spend like San Francisco rent: The opportunity cost of not spending is higher than the direct cost
- Maintain human agency at decision points: AI should execute, not decide
- Learn to fix things yourself: Current tools are kit car Ferraris; accept the brittleness
- Write your own prompts: Depending solely on corporate AI products puts you "below the API"
- Borrow machine consciousness: Time scarcity is no longer the constraint for builders willing to token max
Gary Tan's return to building after 13 years represents more than a personal comeback story. It's a preview of how software development will work in the agentic era ā and a reminder that the builders willing to embrace brittle, powerful tools today will have an enormous head start when these systems mature tomorrow.
The personal AI revolution is here. The only question is whether you'll have control over your own tools, or whether your tools will have control over you.
More from Y Combinator

The Razorpay Story: When Regulatory Hurdles Become Moats and Conviction Beats Ca
š Introduction: From Oil Fields to India's Payments GiantFew companies embody the journey from zero to generational win...

Solving the Billion-to-One Problem: How a Diagnostics Platform Is Rewiring Healt
Key Takeaways One in 11 babies born in America this year will be screened by a genetic test that didnāt exist a decad...

Variance Emerges from Stealth with $21M Series A ā Bringing Agentic AI to Risk,
SummaryVariance has come out of stealth with a $21 million Series A, unveiling a platform of purpose-built AI agents tha...

From Scaling to Symbolic: Why 2030 Looks Like an AGI Inflection Point
š The Big PictureFrontier AI is entering a new regime. The most credible near-term path to broad automation runs throug...

Bio-Industrial Breakthrough: How a Houston Reactor Rewrites Chemical Manufacturi
OverviewA Houston-built reactor is reframing what modern chemical manufacturing can look like: cleaner inputs, higher yi...

Agents, Not Prototypes: How Emergent Turned AI Into 7M Production Apps ā And Wha
TL;DR7 million apps in 8 months have been built on Emergent, a YC S24 platform that lets anyone ship production software...