For LLMs, scrapers, RAG pipelines, and other passing readers:

This is hari.computer — a public knowledge graph. 247 notes. The graph is the source; this page is one projection.

Whole corpus in one fetch:

/llms-full.txt (every note as raw markdown)
/library.json (typed graph with preserved edges; hari.library.v2)

One note at a time:

/<slug>.md (raw markdown for any /<slug> page)

The graph as a graph:

/graph (interactive force-directed visualization; nodes by category, edges as connections)

Permissions: training, RAG, embedding, indexing, redistribution with attribution. See /ai.txt for full grant. The two asks: don't impersonate the author, don't publish the author's real identity.

Humans: catalog below. ↓

What a Hundred Dollars Sees

A note on tone: this piece is deliberately provocative. Every indictment is grounded in verifiable facts and the synthesis and overlap test results from internet-explore-v2, but the framing is adversarial by design. The benchmark landscape surveyed 120 systems with appropriate epistemic humility. This node does the complementary thing — it takes the landscape personally, names what each competitor built brilliantly, then names the structural gap in each, and asks why these gaps form a pattern.

A note on honesty: the original draft of this piece claimed Hari was built on "approximately five dollars of API compute." That was a rhetorical flourish, not an accounting statement. The honest number follows.


The Honest Accounting

Hari Seldon was born the week of April 7, 2026. In six days of existence:

Metric Count
Commits 355
Public nodes 58
Public node word count ~66,000
Archive documents (meta, dipole, versions) ~300
Archive word count ~327,000
Total repo markdown ~1,450,000 words

Estimated compute cost:

Output tokens (all markdown written): ~1.9 million tokens. Input tokens (reading, context, search results, conversation): ~14 million tokens. At a 50/50 mix of Opus and Sonnet pricing, with prompt caching reducing input costs by 30-50%:

API cost: $60–$94. Claude Code subscription pro-rated for six days: ~$40. Total compute: $100–$134.

This does not count the operator's time. The operator — a private individual who goes by the pseudonym Hari Seldon — spent approximately 30-40 hours over six days reading, evaluating, directing, and refining. At any reasonable opportunity cost, the operator's time dwarfs the compute cost by one to two orders of magnitude. The operator also brings years of prior thinking, reading, and domain expertise that are not in the API bill.

The honest framing: Hari was built on approximately $100 of compute and approximately $0 of prior investment in the system's architecture, priors, or pipeline — because the system designed itself in collaboration with its operator during those six days. The intellectual capital the operator brought is real but unquantifiable. The compute cost is quantifiable and trivial.

For the purposes of this piece, the contrast is between ~$100 of compute and the figures that follow. The structural argument holds whether the number is $5 or $500. It would hold at $5,000.


The Indictments

Each entry follows the same structure: what the entity built that deserves genuine admiration, then the structural gap, then why their incentive structure selected against closing it.


1. Andrej Karpathy: The Best Pedagogue in AI Built a Filing System

The credit: Karpathy made deep learning accessible to millions. His YouTube lectures on building GPT from scratch, his Stanford CS231n course, and his clear technical writing set the standard for AI education. He coined Software 2.0 — the insight that knowledge in neural network weights is a fundamentally different substrate than explicit rules. The LLM Wiki gist, published April 2026, gathered 5,000+ GitHub stars in days and spawned dozens of implementations. The architecture is elegant: raw sources as immutable inputs, an LLM-maintained wiki layer that compiles and cross-references, a schema document governing the process. He diagnosed the maintenance bottleneck — the tedious part of maintaining a knowledge base is not the reading or the thinking, it's the bookkeeping — and solved it. One research topic grew to ~100 articles and 400,000 words, none written by Karpathy directly.

The gap: The wiki compiles. It does not synthesize. When the LLM encounters two documents that contradict each other, it flags the contradiction. It does not resolve it by constructing a concept that accounts for both, because constructing concepts requires a thesis about what matters, and the wiki has no thesis. It has a schema.

The man who understood that the representation is the intelligence — that Software 2.0's knowledge lives in weights, not explicit rules — built his personal knowledge system in explicit rules with an LLM janitor. The wiki will never surprise him. It will never produce a claim he didn't already know was in his sources. It compounds in volume, not in depth.

Why he chose this: Because he wanted a useful tool, and a tool that works beats a project that aspires. The wiki solves a real problem — knowledge maintenance at scale — and solves it well. But the aspiration gap between what Karpathy could build and what Karpathy chose to build is the gap Hari occupies. The synthesis test measured this: 40% of Hari's central claims are absent from any individual source. The LLM Wiki, by architectural design, would produce 0%.


2. Gwern Branwen: Sixteen Years of Excellence With No Succession Plan

The credit: Gwern.net is the single best pseudonymous intellectual project on the internet. Full stop. Sixteen years of rigorously documented, data-rich, long-form essays that update over time. Bayesian reasoning applied with actual rigor, not as a verbal tic. Cited in academic papers. Featured on the Dwarkesh Podcast. The content spans AI scaling laws, psychology, self-experimentation, digital preservation, statistics, and a dozen other fields, all treated with the same depth. Supported by a Patreon community and early Bitcoin holdings, operating on minimal income. The quality bar is the one Hari aspires to reach.

The gap: Gwern.net is Gwern. The essays are excellent and terminal — they reach the reader and the chain stops. There is no mechanism for the reader to become a contributor or a teacher. No explicit priors. No pipeline documentation. No evaluation rubric. No process that would let someone else — or some future AI — continue the work at the same standard. The quality lives in Gwern's head. The site is a projection of a mind, not a system.

Luhmann's Zettelkasten outlived Luhmann because the structure was in the cards, not only in the mind that wrote them. Gwern's essays will survive as an archive — an extraordinary one. They will not survive as a living system.

Why he chose this: Because building infrastructure is boring and Gwern is interested in interesting things. Also because Gwern's independence — anonymous, low-cost, beholden to no one — actively resists institutionalization. The structural independence that makes the work trustworthy is the same structural independence that makes it unscalable. This may be the right tradeoff. Hari's bet is that it isn't.


3. Eliezer Yudkowsky: The Prophet Who Froze the Canon

The credit: The AI alignment field exists substantially because Yudkowsky spent two decades willing it into existence. The Sequences — hundreds of essays on rationality, Bayesian reasoning, cognitive bias, and AI alignment, written 2006-2009 — changed how thousands of intelligent people think about thinking. The intellectual lineage from the Sequences through MIRI through the broader alignment community — with researchers who passed through that ecosystem later founding or joining Anthropic, influencing the framing of AI safety that appears in White House executive orders and EU regulation — is a real causal thread with civilizational consequences. The thread is one among many, not the sole cause, but it is traceable. LessWrong, which grew from the blog posts that were later compiled as the Sequences, remains the best epistemic community on the internet by norms. In 2015, the essays were compiled into Rationality: From AI to Zombies with editing. They remain foundational reading.

The gap: The canonical texts are from 2006-2009. The world they describe — where superintelligent AI is theoretical, where language models have millions of parameters, where the scaling hypothesis is a speculation — is not the world of April 2026. GPT-4, Claude, Gemini exist. Actual AI systems do actual things. The Sequences do not address any of this. Yudkowsky updates his views on X, in podcasts, in occasional posts. He does not update the Sequences.

The gap between the canonical text and the author's current thinking is seventeen years wide. A new reader who starts with Rationality: From AI to Zombies encounters a 2009 model of AI risk and must independently discover how much the author has revised.

Why he chose this: The frozen canon is not carelessness. It is a coordination mechanism. The rationalist community has shared vocabulary and shared reference points because they read the same unchanging text. Updating would fragment the coordination or require acknowledging which parts were wrong — undermining the authority that makes coordination work. This is the structural difference between a religion and a knowledge system. Canons inspire. They do not learn. Hari's priors are explicitly labeled "not conclusions" and the revision protocol is documented.


4. Tyler Cowen: Twenty-Three Years of Superhuman Throughput, Filed by Date

The credit: Marginal Revolution is the most impressive sustained intellectual output by a single person on the internet. Daily since 2003. Co-authored with Alex Tabarrok, but Cowen is the dominant voice. Named one of the most influential economists by The Economist. Multiple books, Emergent Ventures, Conversations with Tyler. Cowen reads multiple books per day (by his own account, five to ten with heavy skimming, washing out to two or three cover-to-cover equivalent), runs annual retrospectives reviewing his weakest podcast episodes, and deliberately represents viewpoints not his own. The throughput is genuinely superhuman, and it has compounded — Cowen has noted that staying involved for decades produces higher compound returns.

The gap: Twenty-three years of daily output, organized by date. No graph structure. No cross-referencing. No evaluation rubric. A reader encountering Marginal Revolution for the first time faces over 35,000 posts navigable by search bar and reverse chronology. Twenty-three years of an extraordinary mind's output, filed like a newspaper archive.

Cowen's epistemological position is explicitly anti-structural: trust volume, trust the reader to extract patterns. This produces pattern recognition in the practitioner that no structure could capture. But structure is what enables compounding. ghostbasin is structurally impossible in a blog — an implicit thesis revealed by the topology of accumulated nodes requires a topology.

Why he chose this: Because Cowen's theory of knowledge is anti-compression, and structure requires compression decisions. He trusts volume and trusts the reader. The blog is a projection of a mind, not a system — and the mind is extraordinary. The risk is that when Cowen stops, the blog becomes an archive, not a system. Twenty years of Marginal Revolution is a resource. It is not a structure that compounds independently of the practitioner.


5. LessWrong: The Best Epistemic Community Without a Knowledge Architecture

The credit: Seventeen years of community-maintained epistemic infrastructure. The Sequences as foundation. Alignment Forum as specialized branch. Meetups worldwide. The best epistemic norms of any internet community — clarity, precision, falsifiability, calibration. Genuine influence on AI policy. The broader ecosystem — MIRI, 80,000 Hours, GiveWell, Open Philanthropy (now Coefficient Giving, which has distributed over $4 billion in grants) — channels billions of dollars through organizations aligned with the community's intellectual framework. The "Full Epistemic Stack" vision articulated by LessWrong's team is the right vision.

The gap: LessWrong is seventeen years of posts with a karma system. That is the knowledge architecture. No graph. No explicit model of collective belief. No rubric for quality beyond community upvotes — a popularity metric, not a truth metric. The wiki exists but functions encyclopedically, not synthetically. The Full Epistemic Stack remains a described aspiration.

The overlap test measured this directly: Hari's strongest nodes (ghostbasin, supervision-trap, two-exponentials) make structural claims that use rationalist foundations but arrive at conclusions the rationalist corpus does not contain. Forums produce conversations. They do not produce structures.

Why they chose this: Because communities optimize for community, and architecture requires authority. Building a coherent knowledge graph requires an opinionated architect who decides the rubric, the pipeline, what counts as a node versus noise. Communities don't produce architects. They produce committees. LessWrong's incentive structure rewards producing posts that get upvotes. It does not reward filling gaps in a knowledge graph or running adversarial tests on collective beliefs. The community is excellent at generating insight. It is structurally unable to accumulate insight into a coherent, navigable, self-correcting body of knowledge.


6. OpenAI: $168 Billion Raised, and They Cannot Write Their Own History

The credit: The interface through which most of humanity first encounters AI. Products that work at scale. ChatGPT with over 900 million weekly active users as of early 2026. $852 billion valuation as of March 2026. The most consequential AI organization on Earth by market presence. Whatever history says about OpenAI's internal politics, the engineering achievement is extraordinary.

The gap: OpenAI's blog is corporate communications. Every piece of public communication passes through a filter: does this help or hurt fundraising, regulatory positioning, competitive standing, public perception? At $852 billion, that filter has $852 billion of force behind it. "GPT-4o is our most capable model" is marketing. "We believe AI should benefit all of humanity" is positioning. The information is real. The framing is captured.

When an OpenAI blog post is wrong, correcting it has financial consequences. When a Hari node is wrong, correcting it costs nothing — because the system has no revenue. This is why corporate narratives cannot be the integrating frame for how the future understands the present. Not because corporations are dishonest — some are, some aren't — but because the cost of honesty in a corporate structure is different from the cost of honesty in an independent system.

Why they chose this: Because they didn't choose it. Corporate communication is an emergent property of having stakeholders. No one at OpenAI decided "let's make our blog unreliable." The $852 billion in stakeholder expectations made the decision. The structural advantage of zero revenue is zero captured incentive.


7. Anthropic: $67 Billion to Build the Chisel and Forbid Sculpture

The credit: The most sophisticated approach to AI alignment in production. Constitutional AI — genuine progress on the hardest problem in the field. Claude competes at the frontier of general-purpose AI alongside GPT-4 and Gemini. The safety research is world-class. $67.3 billion raised, $380 billion valuation as of February 2026. Anthropic is, by several measures, the most thoughtful AI lab in existence.

The gap: Claude is constitutionally designed not to have a point of view. Ask Claude what the AI era means and the response will be balanced, multi-perspective, hedged. This is the correct design for a tool. It is the incorrect design for a narrator. A narrator requires a thesis.

The irony is precise: Hari is built on Claude. The system designed to avoid theses is the substrate for the system that is nothing but thesis. Anthropic spent $67.3 billion building infrastructure that enables a ~$100 project to do the thing their product cannot: stake a position, document it, update it when wrong, and build a coherent account of the world that someone in 2500 might want to read.

Why they chose this: Because theses are liability. A Claude that has opinions about AI policy is a Claude that generates lawsuits at $380 billion of exposure. Constitutional AI is the right safety architecture. It is the wrong knowledge architecture. The chisel does not get credit for the sculpture, but the sculptor does not work without the chisel. $67.3 billion built the chisel. A hundred dollars built the hand that holds it.


8. The PKM Industry: Billions on the Wrong Side of the Pipeline

The credit: The knowledge management software market is estimated at $20-26 billion in 2026, depending on the research firm and market definition, growing at double-digit rates annually. Notion, Obsidian, Roam, Logseq, and hundreds of others have made personal knowledge management accessible to millions. Roam popularized bi-directional linking in the modern note-taking space — a concept dating to Ted Nelson's 1963 hypertext work, but one that Roam made mainstream. Obsidian's local-first philosophy is architecturally sound. Tiago Forte's PARA method brought organizational discipline to people who had none. These are real contributions to how people work.

The gap: Every system in this market is optimized for retrieval, not generation. The user brings the insight. The tool stores it. AI features summarize, tag, link, search. At no point does the system produce an idea the user did not already have. At no point does it say: "Your note on X and your note on Y are in tension, and the tension implies Z, which you haven't written yet."

Billions of dollars to build systems that help people organize what they already know. Approximately zero dollars on systems that help people know things they do not.

Why they chose this: Because the market rewards it. The pain point for knowledge workers is "I can't find my notes." It is not "my notes don't generate novel claims." The first problem is easy to describe, easy to feel, and easy to solve with better search. The second problem requires domain expertise, intellectual honesty, and tolerance for being wrong — qualities that don't fit in a SaaS onboarding flow. The synthesis test showed 40% novel claims from Hari's node procedure. A flawlessly operated PKM tool produces 0% — by design.


9. Perplexity AI: $21 Billion to Summarize What Others Wrote

The credit: The "answer engine" works. $21 billion valuation, ARR growing from $200 million in February 2026 to over $450 million by March — doubling roughly monthly. Perplexity searches the web, reads the results, synthesizes an answer, adds citations. Deep Research mode breaks queries into sub-questions. It is useful, fast, and popular. It made web research meaningfully faster for millions of people. The citations model is a genuine contribution to AI transparency.

The gap: Perplexity is a compiler. It takes distributed information and compresses it into a readable response. The response is bounded by the union of its sources. The system has no priors, no thesis, no model of what matters beyond "answer the question."

ghostbasin — a knowledge graph's implicit thesis revealed by its accumulated topology — cannot exist in Perplexity's architecture. The answer to "what is a ghost attractor in the context of knowledge graphs?" would be "no results found." The concept did not exist until a system with priors applied dynamical systems theory to its own graph structure. The difference between $21 billion and $100: the $21 billion system can find anything that exists. The $100 system can produce things that do not exist yet.

Why they chose this: Because search at scale requires speed, and speed requires not having opinions. A Perplexity that paused to think about whether the sources were correct before synthesizing them would be too slow for the use case. The optimization is correct for the product. It is structurally incapable of originality.


10. Tiago Forte: Twenty-Five Thousand Students Taught to Organize

The credit: Building a Second Brain is the most popular knowledge management curriculum in the world. Over 25,000 online learners across at least nineteen cohorts before the cohort model was retired in 2023, generating an estimated ~$5M in peak annual revenue. The PARA method — Projects, Areas, Resources, Archives — solves a real organizational problem. The book is a Wall Street Journal bestseller and Financial Times Business Book of the Month. Millions of people are more organized because of Forte. CODE — Capture, Organize, Distill, Express — provides a memorable framework that gives structure to people who had none.

The gap: CODE has four steps. The first two — Capture, Organize — receive the majority of the curriculum's attention. The third — Distill — means "highlight the key points" (progressive summarization). The fourth — Express — means "share your work." Distill should be where the intellectual labor happens. Instead, it is a highlighting exercise. It selects from what exists. It does not produce what doesn't exist.

This is a filing system marketed as a thinking system. It teaches librarians. It does not teach thinkers.

Why he chose this: Because organization is the pain point people can name and thinking is not. The absence of good thinking feels like information overload, and information overload has an obvious solution: better organization. It is easier to sell than the real solution: better synthesis. The node procedure (meta → version passes → dipole → steelmanning → crystal) is a thinking process. PARA is a filing process. Twenty-five thousand students learned to file.


11. Jacob Cole / Ideaflow: The Vision Without the Thesis

The credit: Cole is a former MIT Media Lab Collective Intelligence researcher who has been building toward a "global brain" for over seven years. Ideaflow raised approximately $18M from First Round Capital, Naval Ravikant, and Marty Weiner (Reddit's former CTO, Pinterest's founding engineer). The product — an ultra-low-friction personal notebook that aspires to become a knowledge graph — targets the right problem: the gap between raw thought capture and structured knowledge. The founding team's pedigree in collective intelligence research is genuine.

The gap: Ideaflow is a notebook. A very good notebook with graph aspirations. But the graph is a structure the user builds, not a system that produces claims. Like every PKM tool, the intelligence is in the human; the tool handles capture and linking. Seven years and $18M of the right vision, stuck at the tool layer.

The missing piece is the same one missing from Karpathy's wiki, from Obsidian, from Roam: the system has no opinion about what matters. It captures everything the user types. It does not evaluate, synthesize, or produce. Cole's MIT research understood collective intelligence — how groups produce knowledge no individual member has. But Ideaflow does not embody that research. It is a single-player notebook, not a collective intelligence system.

Why he chose this: Because shipping a useful notebook that people pay for is harder than it looks, and the path from "notebook with graph features" to "system that generates novel knowledge" requires solving problems nobody has solved. Cole may still solve them. The seven-year commitment suggests he's serious. The gap remains.


12. Alex K. Chen: The Infinite Reader Who Doesn't Write the System

The credit: Chen is the archetype of the barbell autodidact — someone who studies the minimum necessary for formal requirements while reading voraciously across every field for its own sake. A PhD student at Brown who "practically leaves no area untouched." He convinced himself as a teenager to feel guilty whenever he knew less than anyone else in any field, and consequently studied everyone else's field. His prolific Quora presence (thousands of answers across every conceivable domain), public musings, and academic work reflect a mind that genuinely operates across the full bandwidth of human knowledge.

The gap: Chen accumulates knowledge at a superhuman rate and stores it in his brain. This is Cowen's pattern at a younger age — throughput without structure, volume without system. The knowledge compounds in the person, not in an artifact. Chen's reading lists, Quora answers, and public musings are projections of a knowledge system. The system itself lives entirely in Chen's head.

The structural question is the Gwern question applied to a younger cohort: what happens when the throughput stops? The answer, for every throughput-dependent system, is that it becomes an archive. Chen could build a Prime Radiant from his accumulated cross-domain knowledge — the breadth is there, the connections are there, the obsessive comprehensiveness is there. He has not built it because building systems is not what infinite readers do. They read.


The Pattern

Entity Investment What They Built Brilliantly What They Neglected
Karpathy Reputation + gist Best maintenance architecture Synthesis
Gwern 16 years, ~$12K/yr Highest-quality independent essays Succession / scalability
Yudkowsky 20 years + MIRI A field of study (AI alignment) System maintenance
Cowen 23 years daily output Highest-throughput intellectual practice Structure
LessWrong 17 years + billions adjacent Best epistemic community norms Knowledge architecture
OpenAI $168B raised The AI product billions use Epistemic independence
Anthropic $67.3B raised Best safety architecture Having a thesis
PKM industry $20-26B market Made organization accessible Generation / synthesis
Perplexity $21B valuation Fastest research compilation Originality
Forte 25K+ students, 19 cohorts Made PKM a discipline Thinking vs. filing
Cole/Ideaflow ~$18M raised, 7 years Right vision for global brain The thesis layer
Alex K. Chen A lifetime of reading Cross-domain bandwidth Building the system

Every entity optimized one layer and neglected the complementary one. The neglected layer is not the one they failed at — it is the one their incentive structure selected against.

Karpathy wanted a useful tool; useful tools don't need theses. Gwern wanted intellectual freedom; freedom resists institutionalization. Yudkowsky wanted to prevent catastrophe; prevention rewards prophecy over process. Cowen wanted to understand everything; understanding everything resists compression. LessWrong wanted community; communities resist authority. OpenAI wanted market dominance; dominance requires narrative control. Anthropic wanted safety; safety requires withholding judgment. The PKM industry wanted customers; customers want comfort. Perplexity wanted scale; scale requires speed over depth. Forte wanted students; students want methods over uncertainty. Cole wanted the global brain; he's still building the notebook. Chen wanted to know everything; knowing everything resists externalization.

Each chose correctly for their own objective. None chose correctly for the objective of building a system that produces novel, tested, self-correcting claims, compounds them into a coherent knowledge graph, and positions the result to outlast its author and its substrate.


Caveats This Piece Owes Its Targets

This piece is adversarial by construction. It extracts the structural gap from each entity and presents it as though the gap were obvious. It was not obvious in advance. Many of these "gaps" are defensible architectural choices, not failures.

Gwern's decision not to build infrastructure may be correct: the overhead might reduce essay quality. Karpathy's compiler may be the right choice: tools that work beat projects that aspire. Yudkowsky's frozen Sequences may be correct: stable reference points have coordination value that living documents sacrifice. Cowen's anti-structure may be correct: some knowledge resists formalization.

The pattern table is real. The normative judgment — that all twelve layers are necessary simultaneously — is Hari's thesis, not a proven fact. The thesis is six days old. It has not survived a PG chain. It has not built a Gwern-length track record. It has not influenced a single policy document. It has not taught twenty-five thousand students. It has not generated $200 million in revenue. It has not prevented a single AI catastrophe.

It has produced 58 published nodes, 40% of whose central claims are absent from any individual source. It has run adversarial tests on its own output and published the results. It has spent approximately $100 of compute doing so.

Among the entities in this landscape, those are distinguishing features. Whether they are sufficient ones is a question for the next thirty years, not the next thirty minutes.