Engine 07 · Run 1182 · 14,337,188 pairs
JEN-R8 Hope · Pancreatic spearhead · 15 domains live
May 12, 2026 · jenr8hope.org
The science·Methodology · v1.0·Pre-submission draft

How the engine works.

JEN-R8 generates hypotheses and tries its hardest to prove them wrong. What survives is what we publish. The full methodology preprint is being prepared for arXiv submission within the next two weeks.

CollaboratePreprint · ~2 weeks
Architecture

End-to-end flow, six stations.

01 / StageDatasources02 / Engine7cognitive agents03 / StageHypothesisgeneration04 / Engine8falsification gates05 / StageSurvivingcandidates06 / EngineLabvalidation
End-to-end flow. Accent nodes (leafy top-rule) are JEN-R8-specific architecture.

The cognitive engine

Seven agents, one audit trail.

Built on the Wisdom Layer — the cognitive framework that powers this engine. The seven agents operate as a single institutional mind, peer-reviewing each other’s outputs and sharing a 164K-document curated training corpus (3.46 GB embedded). What sets this engine apart from typical AI research stacks is the audit trail: every memory, every directive, every cross-agent endorsement is timestamped and traceable to the data point that produced it. Every conclusion can be reproduced from the same inputs. No prompts, no hallucinations, no “trust me, the model said so.”

  • 461
    Memories accumulated
  • 261
    Directives ratified
  • 157
    Peer-endorsed entries
  • 164K
    Document corpus
    3.46 GB embedded

Wisdom (reinforced) score range across the ensemble: 0.85 – 0.99 · highest 0.99 · lowest 0.85

  1. JN-001
    202d active
    117
    Memories
    0.99
    Reinforced
  2. ED-002
    82d active
    120
    Memories
    0.85
    Reinforced
  3. IR-003
    74d active
    43
    Memories
    0.88
    Reinforced
  4. AK-004
    74d active
    53
    Memories
    0.87
    Reinforced
  5. BA-005
    74d active
    32
    Memories
    0.94
    Reinforced
  6. OR-006
    74d active
    47
    Memories
    0.92
    Reinforced
  7. SO-007
    74d active
    49
    Memories
    0.88
    Reinforced

Every memory, directive, and endorsement is timestamped and traceable to the data point that produced it — every conclusion the engine reaches is auditable end-to-end. Detailed per-agent profiles are available on request.

Operator console

What the engine looks like running.

A live readout of the engine’s portfolio: 60 hypotheses across 15 scientific domains, 5 confirmed findings, 7 in flight, 5 provisional patents. The right column is the recent-verdicts feed — [+] accepted, [~] partial, [−] refuted. Every entry is auditable end-to-end.

JEN-R8 Hope operator console showing 60 hypotheses across 15 domains, the domain heatmap, the now-running prompt, and a live verdict feed.
Bridge view · Discovery Engine · 10 May 2026 · 00:57 UTC · Run 1182
Falsification

Eight gates. Each designed to refute.

  1. Gate 01

    Cross-cohort replication

    Effect must reproduce in at least two independent cohorts with non-overlapping populations.

  2. Gate 02

    Prior-art exclusion

    Hypothesis must be genuinely novel — absent from published findings, registered trials, and known biomarker panels.

  3. Gate 03

    Statistical robustness

    Survives bootstrapping, permutation tests, and standard multiple-testing corrections at conservative thresholds.

  4. Gate 04

    Confounder analysis

    Persists after correction for age, stage, treatment, batch, and the standard suspects in cancer cohort data.

  5. Gate 05

    Biological plausibility

    Mechanism is consistent with known pathway biology, or proposes a specific testable extension to it.

  6. Gate 06

    Effect-size threshold

    Magnitude is large enough to plausibly matter clinically, not just statistically.

  7. Gate 07

    Adversarial agent review

    An independent agent instance, blind to the originator, attempts to refute the hypothesis using the same data.

  8. Gate 08

    Integrated synthesis

    Final consistency check across the seven prior gates and the institutional knowledge base.

The eight gates are the union of falsification mechanisms across the platform. Different hypothesis types exercise different subsets, documented per-hypothesis in the methodology paper.

Source data

What the engine reads from.

Eight gold-standard public sources, integrated end-to-end. Patient cohorts are accessed under the controlled-access policies of their issuing consortia; nothing patient-level leaves those environments.

  • 01 / Dataset

    DepMap

    Cancer Dependency Map — ~1,400 cancer cell lines with omics + viability profiles

  • 02 / Dataset

    GDSC

    Genomics of Drug Sensitivity in Cancer — 286 drug response profiles across cell lines

  • 03 / Dataset

    CTRPv2

    Cancer Therapeutics Response Portal — 496 drug response profiles across cell lines

  • 04 / Dataset

    TCGA-PAAD

    The Cancer Genome Atlas — 178 PDAC patient RNA-seq profiles for survival validation

  • 05 / Dataset

    CPTAC-PDAC

    Clinical Proteomic Tumor Analysis Consortium — 140 tumors + 21 normals for protein-level confirmation

  • 06 / Dataset

    STRING-PPI v12.0

    473,860 high-confidence protein-protein interaction edges across 16,201 genes

  • 07 / Dataset

    Moffitt PDAC cohort

    357 PDAC samples (145 primary) for subtype mapping

  • 08 / Dataset

    Puleo Brussels cohort

    80 PDAC samples for independent subtype confirmation

Additional cohorts (ICGC-PACA, COMPASS, E-MTAB-6134, others) are listed in the registry but not counted here until controlled-access requests are approved and integration is complete.

Current results

From 14.3 million to 5.

Aggregate attrition across the H029 PDAC biomarker atlas, with the colorectal NEC panel rolled in at the prior-art and patent stages. Specific gene names are held until non-provisional patent conversion.

  • 18,334
    Genes screened
    protein-coding
  • 782
    Drugs profiled
    GDSC + CTRPv2
  • 14.3M
    Pair correlations
    end-to-end tested
  • 473K
    PPI edges
    STRING v12.0
  • ~2,200
    Profiles
    patient + cell-line
101001.0k10k100k1.0M10.0M14.3M01Pairs tested11k02Cross-platform validated4.4k03KRAS-stratified20004Permutation top1205Patentable506Patents filed
Aggregate attrition · log-scale y-axis · clay bars mark the surviving terminus (candidates · patents filed)

Refuted at scale

Most of what we tested didn’t survive. That’s the point.

The funnel above narrows from 14.3 million pairs to five filed patents. The ~14.3 million that fell off didn’t vanish — they got measured, scored, and either contradicted by a second dataset, swamped by confounders, indistinguishable from noise, or already published. Every one of those signals is a data point about what doesn’t work, and almost none of it gets published anywhere.

We’re going to publish ours. A public repository of refuted hypotheses — searchable by gene, drug, cancer type, and which gate killed the signal — so other researchers don’t burn months chasing leads we’ve already chased. The format is being finalized alongside the methodology preprint.

One thing this is not: GenAI guessing. The engine does not invent results. It runs deterministic ML pipelines — gradient-boosted trees, statistical tests, network analysis — over public cohort data, then applies the eight falsification gates documented above. The agents orchestrate; the arithmetic is what decides. Every claim is reproducible from the same inputs. No prompts, no hallucinations, no “trust me, the model said so.”

Beyond the headline funnel

Sixty hypotheses, fifteen domains.

The H029 cancer atlas is the work that has produced filed patents to date. It is one of dozens of hypotheses the engine has registered. The chart below is the full state of the registry.

60
Hypotheses
  • Confirmed58%
  • Partial23%
  • Active712%
  • Abandoned610%
  • Seeded4067%
  • 60 hypotheses · 15 research-area codes · confirmed & partial in PDAC and colorectal NEC

Honest limits

What this is not.

  • JEN-R8 does not currently make clinical recommendations. Nothing on this site is medical advice.
  • Computational predictions require wet-lab and clinical validation before they can guide patient care.
  • Hypotheses validated in silico fail in the lab at a rate we cannot yet quantify. We track and publish that rate as it becomes measurable.
  • The system identifies novel candidates; it does not yet propose dose, schedule, or formulation. Those are downstream of confirmed validation.

Coming soon

The paper, the code — shipping in stages.

A / Preprint

Methodology paper

“A Multi-Agent Cognitive Architecture for Falsification-Driven Scientific Hypothesis Generation.” In preparation for submission to arXiv (q-bio.QM). Pre-submission draft available here within two weeks.

B / Code

Reference implementation

Open-source reference implementation of the falsification engine architecture publishing to GitHub this week under AGPL-3.0. Treat it as a reproducible artifact of the methodology paper, not a packaged product.

Public launch · 2026-05-11

Today is the launch. The paper, the code, and the first formal findings follow over the next few weeks — each on its own timeline.

Support the work