MasteryMade · Product Specification · April 2026
A new product category: software that gets smarter with every user.
What it is, how it works, why it's defensible, and what we're building first.
Every SaaS product in the world sits on a pile of user experiments it throws away.
When 500 businesses use the same CRM, each one independently discovers what email subject lines work, what follow-up cadence converts, what lead scoring thresholds matter. They each pay the same tuition — weeks of trial and error — and the platform learns nothing. User #500 starts from the same blank slate as User #1.
The entire AI industry has converged on RAG (Retrieval-Augmented Generation) as the solution: chunk documents, embed them as vectors, retrieve fragments when asked. It works for search. But it doesn't learn. Ask a subtle question requiring synthesis across five sources, and the system re-derives the answer from scratch every time. Nothing accumulates. Day 100 is exactly as smart as Day 1.
Karpathy's insight: knowledge management is a compiler problem, not a search problem. Don't retrieve and re-derive. Compile — synthesize knowledge once, keep it current, and make it available as a persistent, interlinked, self-correcting artifact.
We took that insight and asked a different question: what happens when you compile knowledge not just for one user, but across an entire community of users — weighted by how relevant each person's experience is to your specific situation?
The Knowledge Flywheel is a system where every user's experience is compiled into the collective knowledge base in real time, weighted by contextual relevance, and returned to all users as statistical distributions — not expert opinions. The more users contribute, the better everyone's outcomes.
It operates at three layers:
Your product learns from YOUR usage. Every interaction — decisions, outcomes, configurations, mistakes — is compiled into persistent, structured knowledge. Your experience on Day 30 is informed by everything that happened on Days 1-29. Not as chat history. As compiled, interlinked articles that your AI agent reads before every action.
personal switching cost proven mechanism
Everyone in your community's experiments compound for everyone. An agent compiles all interactions, all shared outcomes, all implementation attempts into community knowledge — not just what the instructor taught, but what 50 people discovered when they actually did it. Bridge-weighted to your specific context: industry, business model, experience level, goals.
network effect bridge-weighted needs anti-fragility mechanisms
Patterns that emerge across communities — insights nobody in any single group can see. The SEO community and the e-commerce community independently discover the same lead gen pattern. Neither knows the other exists. The platform bridges them.
cross-network effect speculative privacy constraints apply
Not all community knowledge is equally relevant to you. Bridge-weighting scores each compiled insight by its contextual relevance to YOUR specific situation — your industry, your business model, your team size, your goals. The result isn't "here are 50 implementations, good luck reading them." It's:
"Student B's approach is 87% relevant to your context (B2B, similar model). Key insight: use case study links, not product shots. But Student A's DM automation technique works cross-context — adapt the Instagram flow to LinkedIn. 3 other solo founders in B2B averaged 18 leads/week with this hybrid approach."
This is the difference between a library (here are all the books) and a research assistant who read every book and tells you which three pages answer YOUR question.
| Claim | Evidence |
|---|---|
| Compiled knowledge > RAG | Proven by Karpathy + 8 independent implementations. Synthesize once > re-derive per query. |
| Real-time statistical inference beats expert bottleneck | This is what makes Netflix, Spotify, and Google work. n=many user outcomes > n=1 expert opinion. |
| Network effects in knowledge sharing compound value | Stack Overflow, Wikipedia — more contributors = more valuable. Proven platform economics. |
| Personal learning compounds (Layer 1) | Every memory/context system validates this. Your agent getting smarter from your usage works. |
| Claim | Risk | Mitigation |
|---|---|---|
| Community knowledge compounds for users, not just the platform | Most valuable knowledge won't be shared (competitive tension) | Works best for non-competing users. Pool by industry × geography. |
| Bridge-weighting produces relevant results at scale | False bridges = noise that looks like signal | Confidence scoring. Show distributions, not conclusions. Let users judge. |
| Cross-community patterns are useful (Layer 3) | Privacy concerns. False cross-domain connections. | Park for later. Don't promise until Layers 1-2 are proven. |
| Failure Mode | What Goes Wrong | Built-In Protection |
|---|---|---|
| Convergence | Everyone follows compiled consensus, kills innovation | Outlier preservation: surface non-consensus successes alongside patterns |
| Survivorship bias | Only successes get compiled, failures invisible | Failure signal: track abandonment as data point |
| Gaming | Users inflate results knowing they'll be compiled | Outcome verification: tracked metrics weight higher than self-reported |
| False confidence | "Community says X" replaces critical thinking | Distribution view: show spread + confidence intervals, not conclusions |
| Staleness | Compiled knowledge outdated in fast-moving domains | Confidence decay: recency-weighted scoring, not just count-weighted |
| Expert conflict | Compiled community wisdom contradicts instructor | Expert override with transparency: show both views, suppress neither |
These are not optional nice-to-haves. They are structural requirements. Without them, the Flywheel degrades into a confidence machine that scales misinformation.
The first application of the Knowledge Flywheel is an AI agent setup service — the same surface the market recognizes (5 agents, 4 pricing tiers), but with fundamentally different economics underneath.
Template agents. Pre-built configurations. Copy. Paste. Running by tonight. Client #1 and Client #50 get the exact same thing. Templates are commodity. Race to the bottom on price.
Agents that learn. Same 5-agent surface. But underneath: a Knowledge Flywheel that compiles every interaction, every outcome, every pattern. Client #50's agents are 49x smarter than Client #1's day-one agents.
| Agent | What the Client Sees | What the Flywheel Does Underneath |
|---|---|---|
| Content | Social posts, newsletters, blogs from brand voice | Compiles what content performs for businesses like theirs. "Plumbers get 3x engagement with before/after posts vs tips." Client doesn't research this — the Flywheel knows from 20 prior deployments. |
| Lead Gen | Scores leads, enriches data, routes to CRM | Compiles lead quality patterns by industry. Distribution view: "For B2B SaaS under $50K ACV, LinkedIn outperforms Google Ads 2.3:1." Confidence-scored, not opinion. |
| Outreach | Personalized cold email/DM sequences | Compiles response rates by approach × industry × size. Bridge-weighted to YOUR niche. "Solo consultants: 'mutual connection' template → 4.7% vs 'cold intro' → 1.2%." |
| Research | Competitor monitoring, trend alerts | Compiles competitive intelligence across clients in adjacent (non-competing) verticals. Anonymized patterns, not raw data. |
| Admin | Support FAQs, ticket triage, scheduling | Compiles common support patterns. "Businesses like yours: 70% of tickets are about X. Auto-response resolves 85%." Improves with every ticket across every client. |
| Tier | Price | Scope | Flywheel Access |
|---|---|---|---|
| Quickstart | $1,500 one-time | 1-2 agents, basic config, 30-day support | Personal only (Layer 1) |
| Growth | $4,500 + $500/mo | 5 agents, brand voice, CRM, quarterly optimization | Personal + Community (Layers 1-2), bridge-weighted |
| Scale | $12,000 + $1,500/mo | 5 agents + custom workflows + weekly calls | Full Flywheel with distribution views, confidence scoring, outlier alerts |
| Partner | $5,000/mo retainer | Everything + ongoing agent development + strategic advisory | Full Flywheel + new agents built from compiled intelligence. A fractional AI team with institutional memory. |
| Metric | Template Agency | Knowledge Flywheel Agency |
|---|---|---|
| Client #1 delivery | 4-8 hours | 4-8 hours (same) |
| Client #10 delivery | 4-8 hours (same) | 2-3 hours (compiled patterns from #1-9) |
| Client #50 delivery | 4-8 hours (still same) | 30-60 min (template + Flywheel context) |
| Value to client over time | Flat | Increasing |
| Churn driver | "I can do this myself" | "I'd lose my compiled intelligence" |
| Moat after 50 clients | None | 50 verticals of compiled deployment data |
The Knowledge Flywheel runs on infrastructure we call The Loom — a compilation pipeline built on top of the Knowledge Fabric Service (KFS), our entity graph + temporal reasoning engine.
| Component | What It Does | Status |
|---|---|---|
| KFS (Knowledge Fabric Service) | Entity graph + semantic search + temporal reasoning + multi-tenant API. Where compiled knowledge lives. | 80% built, production |
| The Loom (Compilation Pipeline) | Capture → Compile → Query (with file-back) → Lint. How raw input becomes compiled knowledge. | Building now (PRD complete, Phase 1 in progress) |
| Bridge-Weighting | Graphiti graph intelligence — detects contextual relevance, finds bridges between concepts, identifies knowledge gaps. | Built (Graphiti + FalkorDB + NetworkX) |
| Anti-Fragility Layer | Distribution views, confidence decay, outlier preservation, failure tracking, expert override. | Designed, builds with Phases 3-5 |
Raw sources (sessions, agent outputs, outcomes, decisions)
→ Normalized by source adapters (automatic, zero manual effort)
→ Compiled by LLM into structured articles (concepts, connections, Q&A)
→ Dual-written: filesystem (human-browsable) + KFS (API-queryable)
→ Bridge-weighted by context (industry, model, size, goals)
→ Served to users as distributions with confidence scores
→ Good answers filed back as new articles (compounding loop)
→ Daily lint checks for quality (self-correcting)
Cost per client: ~$25-50/month (Supabase share + LLM calls). At $500-5,000/month revenue, margin is 90-99%.
The Loom and KFS are designed as independent services, not features embedded in a single product. This means:
| Consumer | How They Use It | What They Get |
|---|---|---|
| Forge Builds (agent setups) | Per-client KFS tenant + Flywheel compilation | Agents that learn. Compiled intelligence across deployments. |
| MasteryOS (expert platforms) | Per-expert KFS vault + subscriber interaction compilation | Expert knowledge GROWS with every subscriber conversation. Network effect. |
| Process Factory (clone builds) | Cross-build pattern compilation | Each build is informed by prior builds. Quality compounds. |
| NowPage / HC Protocol | Content gap detection via KFS | "What do we know but haven't published?" Instant content calendar. |
| Any external SaaS | KFS API + multi-tenant auth | Add a Knowledge Flywheel to ANY product. Spreadsheets, CRM, project management — the pattern is universal. |
Phase 1 serves Forge internally. Phase 2 extracts KFS as a standalone repo with multi-tenant API keys. Phase 3 adds MCP (Model Context Protocol) so any LLM session on any platform can read from and write to KFS automatically. Phase 4 opens KFS to external products.
The endgame: Knowledge Infrastructure as a Service (KIaaS). Not SaaS (software that does things). Not AI-as-a-Service (models that generate things). KIaaS: infrastructure that learns from everything everyone does, and makes everyone better at doing things.
| Layer | Moat Type | What a Competitor Would Need to Replicate |
|---|---|---|
| Code + Architecture | None (open patterns) | Days — anyone can build a compilation pipeline |
| Personal compiled knowledge (Layer 1) | Switching cost | Can't replicate — it's the user's history |
| Community compiled knowledge (Layer 2) | Network effect | Can't replicate — it's months of community experiments |
| 50+ client deployment patterns | Data moat | Months to years of real deployments |
| Bridge-weighting precision | Tuning moat | Requires the data to tune against — chicken-and-egg |
A competitor can copy the code in a week. They cannot copy the compiled knowledge of 50 clients across 20 verticals, bridge-weighted and confidence-scored over 12 months of real outcomes. That's the moat.
| Channel | Approach | Why It's Different |
|---|---|---|
| Content | Publish the distributions. "Here's what actually works for [vertical] — compiled from real deployments." | Nobody else has this data. It's compiled knowledge used as lead magnet. |
| Outreach | "We've deployed agents for 12 businesses in your vertical. Here's the #1 pattern and the #1 mistake." | Specific, data-backed insight — not a generic pitch. |
| Referrals | "Refer a non-competing business in your industry → both of you get richer data." | Structural incentive, not monetary. The Flywheel gets better for both. |
| Will's cohorts | 14-day sprint cohorts → each cohort IS a tribal learning group. | GTM = product development. Cohort implementations become seed data. |
| Communities | Free "Flywheel diagnostic" — show what compiled data says about businesses like theirs. | Value-first. The diagnostic itself demonstrates the product. |
Template agencies sell hammers.
We sell hammers that learn.
Same price. Different trajectory.
| Knowledge Flywheel | The full concept: compile → distribute → improve → compound |
| The Loom | The infrastructure: KFS + compilation pipeline + bridge-weighting |
| Bridge-weighting | Contextual relevance scoring — personalizing community knowledge to YOUR situation |
| KIaaS | Knowledge Infrastructure as a Service — the product category |
| Distribution view | Showing the spread of outcomes, not just "the right answer" |
| Confidence decay | Recency-weighted scoring — recent validations count more |
| Outlier preservation | Deliberately surfacing non-consensus successes |
| Compiled apprenticeship | What the Flywheel produces — structured knowledge transfer at scale |