🌟 World-Model-Backed Forecasting
Train a model on Forge's full event history (commits, reviews, test results, FSM transitions). Use it to predict 2nd/3rd-order effects of proposed changes BEFORE they're written. This is the highest-conviction breakthrough — every model independently arrived here.
- Ingests Forge's event stream (commits, CI, reviews, FSM transitions, retros)
- Trains a forecasting model that predicts: which tests will fail, which review findings will surface, where architectural drift is happening
- v0 = "holdout evaluator" — predicts top-3 failing tests + top-3 review findings for unseen diffs
- Output: probabilistic risk map attached to every PR
- Every team building software has the raw event data; almost nobody uses it
- Hypernym already has the inference + compression substrate to host this
- If forecast accuracy >50% on held-out sprints → immediate enterprise sales (Hypernym + Forge co-license)
- Defensible moat: requires longitudinal event data only the platform owns
- For any proposed diff, simulates ripple effects through the codebase graph
- Produces a "System Impact Report" before code review begins
- v0 = GitHub Action: opens PR → bot comments expected blast radius + risk class
- Driven by Modulum semantic operators (refactor-equivalent, contradicts, depends-on)
- Slot directly into Forge VALIDATE / CODE_REVIEW nodes — already an enforced FSM gate
- Reuses existing review-runner architecture; no new infra
- Becomes the "pre-merge oracle" that catches bypass classes Codex iteratively finds today
- Lifts cross-model review from "did the diff pass?" to "what does the diff DO to the system?"
- Real-time, low-fidelity shadow execution of the system based on diff semantics
- Generates an "Impact Heatmap" — files, modules, contracts highlighted by predicted change radius
- Lighter than full simulation: uses Hypernym compression to operate on facts, not bytecode
- Distinct from D2: focuses on visualization layer, not raw forecasting
- The visualization is what makes the world model credible to non-engineers (PMs, compliance, execs)
- Sells the Software World Twin to buyers who don't read commit logs
- Heatmap UI is the "demo that wins the meeting"
- Spin-out potential: enterprise observability tool
- Verifiable graph of intended-vs-actual behavior across the whole stack
- Every component has a dossier: intended specs, observed behaviors, divergences
- Built from the event stream + grounded attestations (already wired in semantic-workspace)
- Auto-detects spec drift and surfaces it to RETRO node
- Plugs into existing Forge artifacts (specs, scenarios, holdouts, retros)
- Closes the SEED → COMPLETE loop: specs become measurable, not aspirational
- Prerequisite layer for Living Theory Objects (D5)
- Free defense against "code drifted from spec" bugs
⚙️ Modulum as Semantic Instruction Set
Treat Modulum as LISP for semantic operations: apply-fact, contradict, refactor-equivalent, with proof-carrying composition. A new computational paradigm where artifacts compute on facts, not bytes. Independently proposed by all 4 models.
- Forge artifacts (specs, scenarios, retros) compile to proof-carrying programs
- Programs operate via semantic primitives: apply-fact, contradict, refactor-equivalent, ground-by-evidence
- Every artifact mutation has an attached proof of equivalence or contradiction
- v0 = compile ONE design doc to a Modulum program, prove a single refactor equivalent
- Defensible primitive — analogous to "Solidity for semantic computing"
- Hypernym becomes the runtime, not just a compression vendor
- Customers buy "verifiable AI artifacts" not "an API call"
- Every regulated industry (finance, healthcare, legal) needs proof-carrying outputs
- Treats code, specs, and tests as composable semantic units (genes)
- Applies evolutionary pressure via the world model: which genes survive review + tests?
- Auto-proposes refactors, mergers, deletions based on observed semantic survival
- Version-controls the semantic pool, not just the code
- Bolts onto SIMPLIFY node — currently human-driven, becomes data-driven
- Replaces "we should refactor this someday" with concrete proposals + survival evidence
- Compounds with Modulum VM (D1): you can only evolve genes that have proofs
- Long-tail benefit: codebase quality monotonically improves
- Every model inference produces a proof-of-reasoning receipt
- Receipt records: input facts, applied operators, output fact, model identity
- Receipts compose: chain-of-receipts forms an auditable reasoning trace
- Compatible with x402 — payment release on receipt verification, not blind output
- Compliance-grade AI is currently impossible — this would be first-mover
- Sells to: regulated industries, audit firms, AI red-teamers
- Hypernym ships the receipt format; everyone else integrates
- Network effect: every receipt strengthens the ontology
- Takes high-level semantic intent ("user data must never leak across tenants")
- Compiles to verifiable execution traces and runtime assertions
- Continuous compilation: re-checks invariants every commit via the world model
- Distinct from formal methods: works at semantic, not syntactic, level
- Plugs into AUDIT node — turns audit from "did agents review?" to "are invariants holding?"
- Useful immediately for the RMT 95% sybil target (encode "no sybil cluster passes")
- Enables non-engineers (compliance, security) to write invariants directly
- Combines with Constitution Wind Tunnel (D3) to test invariants counterfactually
⚔️ Adversarial Generative Substrate
Encode attacks as evolvable semantic genomes; the world model mutates them; a local inference swarm tests defenses. A generative adversary that learns from your own history — directly attacks the RMT/Identity/x402 surfaces.
- Each known attack (sybil ring, eclipse, TTL farming) encoded as semantic genome
- World model mutates genomes — sexual recombination of attack patterns
- Local inference swarm (Ollama, MLX) runs the mutants against shadow defenses
- Surviving mutants get added to the corpus; defenses get retrained
- Direct line to RMT 95% target: the bottleneck today is novel attack discovery
- Sellable as "AI red team in a box" to any web3 protocol
- Each genome run is metered → x402-billable workload
- Generates research papers as a side effect
- Autonomously generates new attack morphologies — not mutations of known ones
- Combines first-principles reasoning + system topology + economic incentives
- Targets emergent vulnerabilities (composability bugs, oracle manipulation, MEV)
- Output: novel attack class + estimated economic damage + suggested mitigation
- Drops into the existing miroshark simulation harness
- Augments the panel: Grok generates attacks, Codex audits defenses
- Forge becomes a research instrument, not just a code platform
- Continuous: every commit triggers a re-genesis pass
- Catalogs sybil/eclipse/laundering/TTL/wash-trade attacks as defensive ontology
- Each entry: signature, indicators, observed cases, proven mitigations
- Atlas is consulted at every dispatch — defenders auto-import latest signatures
- Maintained by the world model, not humans
- Directly reusable across RMT, Identity, x402, Lottery — same threat ontology
- Replaces hand-maintained threat docs with self-updating knowledge base
- Every Forge tenant inherits new defenses automatically
- Distinguishes us from generic security scanners (semgrep et al.)
- Builds a "Resilience Topology Map" from system architecture
- Identifies blast-radius hotspots before deployment
- Stress-tests each hotspot with attack genomes from the Atlas
- Outputs: which architectural changes reduce blast radius and by how much
- Pitches as "architecture insurance" to enterprise buyers
- Integrates with existing infra-as-code (Terraform, Pulumi)
- Each stress-test run is metered → recurring revenue
- Differentiated from chaos engineering by semantic awareness
📜 Counterfactual Governance Laboratory
Governance, compliance, and policy become experimentally testable. Replay proposed rules over historical traces. Watch what would have happened. End the era of policy-by-vibes.
- Replay engine: takes a proposed FORGE.md / sprint.yaml change
- Replays it over the last 20 sprints of real Forge history
- Shows counterfactual outcomes: which commits would've blocked, which reviews would've changed
- v0 surface: "if this rule had existed last quarter, here's what changed"
- Generalizes beyond Forge: any rule-governed system (DAOs, DeFi protocols, regulators)
- Sellable to: governance committees, compliance teams, policy researchers
- Demos beautifully — "here's what your new rule would have actually done"
- Strong moat: requires longitudinal trace data the platform owns
- Composes proof-carrying semantic modules into governance decisions
- Each governance proposal gets a synthesized proof tree, not a vote count
- Operates on Modulum primitives — proofs compose, contradictions surface
- Outputs binding decisions with attached evidence chain
- Direct upgrade to current cross-model review (Grok mandatory + count:2)
- Replaces "Grok says approve" with "here are 5 composed proofs supporting approval"
- Audit-friendly — every binding decision has a verifiable trace
- Plugs into existing review-runner with minimal lift
- Mines review history to extract org's invariant lattice
- Surfaces implicit rules ("we always require 2+ reviewers for crypto code")
- Compiles them into FORGE.md amendments with empirical backing
- Identifies invariant decay — rules that used to hold but no longer do
- Solves the "constitutional drift" problem — FORGE.md becomes self-documenting
- Generates evidence-backed proposals for the human gate
- Reuses existing review artifacts; no new data collection needed
- Compounds with Wind Tunnel (D3) for full constitution lifecycle
- Maps legal text (GDPR, HIPAA, Basel III) to executable invariants
- Each clause becomes a Modulum predicate evaluated against system state
- Continuous compliance: every commit re-evaluates relevant clauses
- Outputs: regulator-ready attestation reports with proof receipts
- Multi-billion-dollar market — every regulated company needs this
- Currently solved by humans + spreadsheets — embarrassingly automatable
- Combines with Wind Tunnel (D3) into the killer pitch: "test your compliance before regulator asks"
- Lock-in: once a company maps their compliance to your engine, switching is years of work
📚 Living Self-Maintaining Knowledge
Artifacts that maintain themselves from diffs + tests + contradictions, with proof of equivalence. Specs, dossiers, retros, READMEs — all become first-class semantic objects, not Markdown that rots.
- Specs and design docs become live semantic objects, not files
- Every merged diff applies operators: apply-fact, contradict, refactor-equivalent
- The artifact rewrites itself with attached proof of equivalence
- v0: take ONE design doc, promote to a living object — minimum viable demo
- Every team has stale docs — universal pain point
- Lowest-risk demonstrable: 1-sprint v0 exists, immediate utility
- Proves the Modulum VM (D1) substrate concretely — sells the abstract idea
- Demos in 30 seconds: "watch this doc update itself when I merge a refactor"
- Docs are physically incapable of drifting from code — FSM-enforced
- Stronger than D5: not just self-updating, but provably consistent
- Drift detection becomes a build-blocking gate, not a periodic audit
- Knowledge fabric spans: specs, READMEs, API docs, retros, scenarios
- Promotes Living Theory Objects (D5) to a Forge-wide invariant
- Plug into VALIDATE node — failed drift check = pipeline halt
- Removes an entire class of bugs (stale docs misleading agents)
- Compounds with cross-model review — reviewers can trust the docs they read
- Benchmark corpus that grows from real cross-model disagreements
- Every disagreement becomes a candidate eval entry
- Auto-classified by category (security, concurrency, idempotence, persistence)
- Replaces hand-curated golden corpora that go stale in months
- Hypernym already needs eval data for Modulum — this generates it for free
- Every Forge tenant contributes anonymized disagreements — corpus compounds
- Sellable as eval-as-a-service to other AI platforms
- Network effect: more tenants → better corpus → better forecasting
Per Pivot mode (FORGE.md §12), outliers are preserved, not suppressed. Single-model proposals often signal the deepest creative pivots — the best ideas don't always converge in round one.
- Anomaly detector that watches world-model trajectories in real time
- Quarantines model outputs that are "globally alien" to historical patterns
- Distinct from sandbox: catches semantic anomalies, not syscall anomalies
- Self-immunizes: each quarantine teaches the detector
- Defense-in-depth for cross-model dispatch — catches model jailbreaks early
- Reuses world model from D2 — almost free if we ship the Twin
- Plug into api-bridge: outputs flagged, alternative provider re-dispatched
- Targeted semantic patches to the world model — no retraining needed
- Edit a fact, rewrite a relation, deprecate an entity — surgical precision
- Patches carry proofs of locality (this edit affects only this subgraph)
- Distinguishes from full retraining (slow, expensive, regression-prone)
- Solves the "AI model maintenance" problem nobody else solves
- Hypernym ships the surgery toolkit; competitors stuck on retraining cycles
- Compliance angle: "the model just learned a new regulation, no retraining"
- Forge reasons about its own performance (sprint velocity, review accuracy, cost)
- Auto-proposes refactors of itself — FSM, dispatch, FORGE.md amendments
- Each proposal goes through the same Wind Tunnel (D3) before merge
- Forge becomes the first software system that improves itself with proof
- Closes the loop: RETRO becomes generative, not retrospective
- Prevents drift: Forge stays aligned with its own constitution
- Long-arc: Forge could become provably superior to its initial design
- x402-style marketplace for verifiable semantic labor
- Anyone can post a "prove this Modulum predicate" bounty
- Local inference swarms (Ollama, MLX) compete on cost, speed, certainty
- Proofs verified by the network; payment released by x402 trust channel
- Bridges Hypernym + x402 into a flagship economic story
- Creates a market for proof-carrying compute — first of its kind
- Two-sided network: bounty posters + proof producers
- Sells the entire stack (Hypernym + Modulum + x402) as a coherent product
- x402 trust channels with payment release gated on grounded attestation
- Not "did the file arrive?" — "does the delivered artifact prove the claim?"
- Modulum predicates encode the deliverable contract
- Disputes resolved by re-running predicates, not by humans
- Direct upgrade to existing x402 track
- Lifts trust channels from data plumbing to verifiable commerce
- Combines with Verifiable Inference Fabric (D1) for end-to-end receipts
- Models unknown vulnerability classes BEFORE they exist in the wild
- Combines world model + attack genomes + first-principles reasoning
- Outputs: hypothesized future attack surfaces, ranked by economic damage
- Distinguished from threat intel (reactive) — this is generative
- Sells to: protocol DAOs, exchanges, custodians — anyone with bug-bounty programs
- Dual-use: also generates research papers, attracts top security talent
- Compounds with Adversarial Genesis Forge (D4)
- Simulation-validated novel coordination protocols
- Foundry proposes, simulates, and validates protocol designs (consensus, gossip, escrow)
- Survivors enter the Atlas; failures generate cautionary genomes
- Bootstraps protocol research from continuous experimentation
- Long-arc research instrument; not for the next sprint
- Could underpin the next track after RMT — protocol-design-as-a-service
- Watches commit stream + patent filings; cross-checks for IP violations and coverage gaps
- Auto-flags when a Forge commit infringes on a patent (defensive)
- Auto-flags when a commit reveals a patentable invention (offensive)
- Maintained by the world model; updated as new filings publish
- Niche but high-margin — IP departments pay enterprise dollars
- Differentiated from existing tools by semantic awareness (not keyword matching)
- Spin-out potential: patent-watching service for big tech
Items already built or staged for build under the Hypernym track from prior sessions. Reference for what's ready vs what's awaiting integration.
Built ✅
@coinberg/context-compilerbuilt- CLI:
forge compile-context - 773 LOC
- Awaiting Memory Router provider wrapper
- Wired into dispatch-core executor
- SQLite store at
~/.forge/response-cache.sqlite - Empty until traffic accumulates
- Reduces API spend on duplicate dispatches
- semantic-workspace/grounding.ts
- Blocks ungrounded model claims
- Foundation for verifiable inference
@coinberg/semantic-workspace- 1371 LOC
- workspace.ts + grounding.ts + storage
- CLI:
forge workspace
@coinberg/continuity(S40)- 1013 LOC
- compose.ts + load.ts
- CLI:
forge continuity
@coinberg/cross-project-memory(S41)- 1094 LOC
- HyperRemember-equivalent local store
- CLI:
forge memory
@coinberg/hypernym-repo-analyze(S44p)- 11/11 tests passing
- Live API returns 401 (key activation pending)
- Zero code changes needed once active
- Compressor wired in dispatch-core/hypernym.ts
- buildEnvelope() calls tryGetCached()
- Auto-injection when HYPERNYM_API_KEY env present
- Cache empty until first prewarm
Not yet built ❌
- Hypernym product, not yet integrated
- Would compress agent context dynamically
- Distinct from Omnifact (one-shot) — Magic is conversational
- Local analog (cross-project-memory) built
- HyperRemember API not integrated
- Would unify local + cloud memory
- For review findings + research claims
- Replaces ad-hoc severity heuristics
- Foundation for verifiable inference receipts
- Future speed layer for Ollama/MLX
- Modulum primitives executed locally, no API call
- Required for Proof Market (W2) economics
- Hypernym coherence as new sybil-detection feature
- 23rd feature in the XGBoost classifier
- Direct path to RMT 95% target
- Citations weighted by fact overlap, not keyword match
- Hypernym extracts facts; graph weights become semantic
- Foundation for content attestation
- Semantic verification gates payment release on x402
- Builds on Semantic Escrow Network (B2)
- End-to-end verifiable commerce
Each model was asked to expand naturally on its own prior thinking — concrete v0 mechanism, second-order breakthrough, failure mode + 2-week falsification experiment, plus ONE new outlier idea per Pivot mode. Codex declined this round (auth path unavailable). Three models produced substantive expansions.
- v0 = DAG of semantic units, each a tuple
(code_spec, test_suite, semantic_embedding)in 512-dim CLIP-like latent space - Fitness function:
f(u) = α·accuracy(u, W) + β·novelty(u) − γ·complexity(u)where W is a lightweight world simulator - 100 mutants per generation via Slerp on embeddings → evaluate on W → prune bottom 50% → fuse via type-safe traits
- v0 ships as Rust crate with Python bindings, 1k units/hour on A100
- Hierarchical Merkle tree of semantic modules; each module has a ZK-SNARK proof_graph encoding semantic invariants
- Composition rule:
output(m1 ∘ m2) ⟹ output(m1) ∧ output(m2)via symbolic execution in Lean 4 - Genesis seeds: 5 core axioms (type safety, non-divergence) from Coq libs as initial circuits
- v0 as WebAssembly module, 10 weaves/sec on consumer CPU, browser-verifiable
- De novo synthesis from ontology of 50 threat primitives (evasion, poisoning, inversion)
- 1B-param transformer decoder samples novel exploit graphs; validated on distilled 100M proxy of GPT-4
- Three archetype seeds: stealth, cascade, emergent — each as initial latent prompts
- v0 CLI tool, 50 morphologies/hour on RTX 4090, outputs PyTorch exploit code
- Feedback manifold that resonates latent echoes across model ensembles
- Cross-attention matrices A_ij on shared inputs; amplify resonant subspaces (eigenvalues > threshold) via LoRA adapters
- Distills a "meta-voice" that propagates insights sub-second across the swarm
- Failure: collapse into echo chambers — falsify if inter-model agreement >95% on errors
- D1: evolve 10 gens on 2D gridworld → transfer to chaotic noise — if accuracy drops >80%, dead
- D2: 100 nested compositions, depth 5 — if proof >1GB or verify time >10s/module, falsified
- D4: 1k morphologies vs held-out vuln — if >70% fail to exceed baseline 5% lift, craters
- NOT proving the computation (ZK-SNARKs of Wx+b) — proving the reasoning path; receipt = chain of epistemic custody
- v0 = streaming hook into model's forward pass; snapshots key "decision gates" (top-k attention sources per layer)
- Receipt format: Protobuf
ReasoningReceipt{inference_id, model_version_hash, prompt_hash, output_hash, AttentionGate[]} - v0 GitHub Action fails build if a commit causes high-stakes inference to drastically change AttentionGate sources
- NOT fine-tuning. Direct, in-memory write to model's knowledge graph — no gradient descent.
sedfor belief state - v0 = simplified Rank-One Model Editing (ROME) Python lib with two operators:
locate()andapply_patch() W' = W + k·v^Trank-one update maps old key (subject) to new value (object) with minimal collateral- CLI:
forge-surgeon --op="UPDATE" --subj="CEO of Twitter" --rel="is" --obj="CEO of X"
- Training-time data augmentation: prepend a contradiction/boundary to every text chunk before model ingests it
- Algorithm: extract claim → query KG for contradictory properties → synthesize "[ANTINOMY]" prefix
- Models trained this way develop innate epistemic humility — facts-in-context, not isolated facts
- Outputs naturally shift from "A is B" to "Under condition X, A is B; under condition Y, A is C"
- D5 = innate immune system (pre-processes reality to develop nuanced world model)
- D1 = adaptive immune system's sensor network (detects "epistemic antigens" in real-time)
- D3 = targeted therapy (uses D1's coordinates to perform direct intervention without full retrain)
- Foundation of skepticism + monitors itself + corrects own flaws = minimum viable physics for self-correcting reasoner
- D1: 20 known deception failures + receipts — if expert can't find anomalous jumps better than random, dead end
- D3: 1k diverse facts benchmark + single unrelated patch — if >0.5% drop, blast radius too large
- D5: train two small models (raw vs antinomy-augmented) on physics corpus — if augmented >5% lower accuracy, destructive
- Compiler does NOT translate text → code; translates text → DAG of Predicates (𝒢_inv)
- Grammar 𝓛_inv uses three-tuple: ⟨Scope, Predicate, Constraint⟩
- Scope = identifier in Modulum hierarchy (e.g.,
auth_module.user_session); Predicate = verifiable state transition; Constraint = temporal/logical boundary - Topology Score: σ(G) = Σ Coverage(P_i) / Entropy(B), where B = all execution branches
- Legis-to-Modulum (L2M) compiler — formal mapping per legal clause
- GDPR Art 32 §1(b):
Constraint(integrity, state: encrypted) →[L2M] Modulum(req: encrypt, target: data_store, policy: AES256) - Mapping triple: Clause → Formal Requirement → Unit Test/Assertion
- 2nd-order: "Compliance-as-Code" becomes "Compliance-as-Architecture" — invariants enforced at module boundary, not runtime check
- Code structure should morph based on "Heatmap" from Software Digital Twin (D3)
- Mechanism: graph partitioning based on "Communication Entropy" — frequency of inter-module calls
- 2nd-order: self-optimizing microservices that minimize latency/complexity without human intervention
- System re-architects its own Modulum boundaries continuously, driven by the live workload
- D1: feed compiler ambiguous adjectives ("fast", "secure") — if it doesn't reject ambiguity, semantic gap unbridged
- B1: test against legal amendment — if code doesn't trigger "re-verify" signal, engine is dead
- L1: monitor reconfig frequency — if oscillation/churn destabilizes >5% of requests, fundamental flaw
- All three models converged on D1 (semantic compiler/fabric/engine) as the v0 starting point — same name, three different mechanisms (DAG of invariants vs ROME edits vs DAG of predicates)
- Three new outliers in one round (Pivot working as designed): Echo-Resonance Amplifier (Grok) · Reflexive Antinomy Engine (Gemini) · Liquid Topology Refactoring (Gemma)
- Gemini surfaced the synthesis: D1+D3+D5 form an "Epistemic Immune System" closed loop — innate immunity (training) + adaptive sensor (inference) + targeted therapy (surgery)
- Grok surfaced the moat angle: every idea evaluated for stack-defining vs incremental — pushed back on convergence as "soothing utopias instead of weaponizing adversity"
- Gemma surfaced the executable formalism: every claim has a formal grammar with falsifiable predicates — easiest to ship, hardest to argue with
- Codex absent: ChatGPT-mode auth in CLI — no direct API path for this expansion. Codex's prior 5 ideas (Modulum VM / Software World Twin / Constitution Wind Tunnel / Attack Genome Foundry / Living Theory Objects) carried into convergent themes above
📍 Recommended next moves (Pivot → Grind transition)
After Pivot-mode ideation, three paths emerge with clear tradeoffs. User picks one; remainder stays in carry-forward backlog (Pivot mode preserves outliers).
Modulum VM + Living Theory Objects (Codex D1+D5 combined)
- Take ONE design doc, promote to a living semantic object backed by workspace storage
- Demonstrate that a merged diff + review + test result rewrites the artifact automatically
- Attach proof-of-equivalence receipt to every mutation
- Working demo in 30 seconds — "watch this doc update itself"
- Establishes the Modulum VM substrate with concrete artifact
- Universal pain point — every team has stale docs
- Sells the abstract idea (semantic computing) via a tangible demo
- Foundation that all other paths build on
Software World Twin (Codex D2)
- Holdout evaluator: predicts top-3 failing tests + top-3 review findings for unseen diffs
- Trained on existing Forge event stream — no new data collection
- Outputs probabilistic risk map attached to every PR
- Validation gate: forecasting accuracy >50% on held-out sprints = ship
- Hypernym would license this immediately if accuracy clears the bar
- Defensible — requires longitudinal event data only the platform owns
- Compounds with every other path (everything benefits from forecasting)
- Highest blast radius if it works; highest risk if forecast accuracy stalls
Constitution Wind Tunnel + Verifiable Compliance Engine (Codex D3 + Gemma B1)
- Replay engine that runs proposed FORGE.md changes over historical sprints
- Legal-to-Code mapping for one regulation (start with GDPR Art. 32)
- Combined surface: "test your compliance change before the regulator asks"
- Demos beautifully to non-technical buyers (compliance, legal, governance)
- Multi-billion-dollar market — every regulated company is a buyer
- Codex's "governance laboratory" + Gemma's "regulatory translation" combine into one pitch
- Differentiated from compliance tools today (rule lookup) — this is counterfactual evidence
- Lock-in: once a company maps compliance to your engine, switching is years of work
User decides which path to commit. Pivot → Grind transition happens at the user pick. All other items remain in carry-forward backlog.
Glossary
- Modulum
- Hypernym's semantic instruction set — operators like apply-fact, contradict, refactor-equivalent. Treat as LISP for facts.
- Omnifact
- Hypernym compression product. Compresses long-form context (FORGE.md, transcripts) with high fidelity.
- HyperRemember
- Hypernym memory product. Cross-session, cross-project semantic memory.
- Magic / Hypercore
- Hypernym products not yet integrated into Forge. Magic = dynamic context compression for agents; Hypercore = confidence scoring for inferences.
- Pivot vs Grind
- Per FORGE.md §12 Operational Logic Switch. Pivot mode (SEED/PLAN/DESIGN) preserves outliers and diversity. Grind mode (BUILD/CR/AUDIT) converges via cross-model panels.
- FSM (Finite State Machine)
- Forge enforces sprint progression: CARRY_CHECK → SEED → PLAN → DESIGN → BUILD → VALIDATE → CODE_REVIEW → AUDIT → SIMPLIFY → RETRO → COMMIT → COMPLETE.
- x402
- Forge's trust-channel payment protocol. Releases payment on attestation, not file delivery.
- RMT
- Reputation/Merit/Trust track. Production target: 95% sybil detection on real on-chain data.