Competing-Risk Cumulative Incidence Functions as a Unified Protein Therapeutic Lifetime Predictor
A survival statistics framework borrowed from actuaries could predict exactly how—and when—engineered protein drugs will break down in the body.
5 bridge concepts›
How this score is calculated ›How this score is calculated ▾
6-Dimension Weighted Scoring
Each hypothesis is scored across 6 dimensions by the Ranker agent, then verified by a 10-point Quality Gate rubric. A +0.5 bonus applies for hypotheses crossing 2+ disciplinary boundaries.
Is the connection unexplored in existing literature?
How concrete and detailed is the proposed mechanism?
How far apart are the connected disciplines?
Can this be verified with existing methods and data?
If true, how much would this change our understanding?
Are claims supported by retrievable published evidence?
Composite = weighted average of all 6 dimensions. Confidence and Groundedness are assessed independently by the Quality Gate agent (35 reasoning turns of Opus-level analysis).
RQuality Gate Rubric
0/10 PASS
| Criterion | Result |
|---|---|
| ABC Structure | true |
| Test Protocol | true |
| Counter-Evidence | true |
| Novelty | true |
| Precision | true |
| Groundedness Adequate | true |
| Mechanism | true |
| Confidence | true |
| Falsifiable | true |
| Claim Verification | true |
Claim Verification
Empirical Evidence
How EES is calculated ›How EES is calculated ▾
The Empirical Evidence Score measures independent real-world signals that converge with a hypothesis — not cited by the pipeline, but discovered through separate search.
Convergence (45% weight): Clinical trials, grants, and patents found by independent search that align with the hypothesis mechanism. Strong = direct mechanism match.
Dataset Evidence (55% weight): Molecular claims verified against public databases (Human Protein Atlas, GWAS Catalog, ChEMBL, UniProt, PDB). Confirmed = data matches the claim.
Two very different fields are colliding here in an unexpectedly elegant way. The first is survival analysis — a branch of statistics originally developed by actuaries to model when and why people (or machines, or financial instruments) fail. It's been refined for over 200 years and is routinely used in clinical trials to track patient outcomes. The second field is brand new: AI-designed protein therapeutics. Scientists can now use tools like RFdiffusion to essentially 'dream up' entirely new protein drugs from scratch — proteins that don't exist in nature but are engineered to treat diseases. The problem? These custom proteins are fragile. Once injected into the human bloodstream, they face a gauntlet of ways to fail: they can clump together (aggregation), get chewed up by enzymes (proteolysis), unfold in the heat of the body, oxidize, or trigger an immune attack. Right now, researchers test for these failure modes somewhat separately, without a unified way to think about how they compete and interact. This hypothesis proposes borrowing a specific statistical tool — the Competing Risks Cumulative Incidence Function — to treat each of those five failure modes as 'competing causes of death' for the protein. The elegant insight is mathematical: because the probabilities of all failure modes must add up to no more than 100%, there's a built-in conservation law. If you engineer the protein to resist aggregation, you're implicitly asking: does that trade off against, say, increased oxidation risk? The framework makes those tradeoffs visible and quantifiable, rather than leaving them as educated guesses. Preliminary computational work suggests all five failure modes can happen on overlapping timescales — anywhere from 30 minutes to two weeks — making the competition between them real and practically important. The reason this matters is that designing a protein drug is currently a bit like building a car without understanding how different parts wear out together. You might fix the brakes only to discover the engine fails faster. This framework would give protein drug designers a single dashboard — a unified lifetime predictor — that shows not just when a protein is likely to fail, but which failure mode is likely to win, and what the engineering tradeoffs look like before expensive lab work begins.
This is an AI-generated summary. Read the full mechanism below for technical detail.
Why This Matters
If confirmed, this framework could fundamentally change how AI-designed protein therapeutics are evaluated and optimized during drug development, potentially cutting the time and cost of identifying stable candidates before clinical trials. It could provide pharmaceutical companies with a principled way to compare protein designs — not just 'does it work?' but 'how long will it last and what kills it first?' — enabling smarter engineering decisions earlier in the pipeline. The mathematical conservation law built into the framework also means it could flag hidden tradeoffs that current ad hoc testing misses entirely, reducing the chance of late-stage failures. Given how rapidly AI protein design is accelerating, building robust lifetime-prediction tools now is urgent — and this is a testable, concrete framework ready for experimental validation.
Mechanism
Each designed therapeutic protein entering the bloodstream faces K=5 competing failure modes with cause-specific hazard functions: aggregation h_agg(t), proteolysis h_prot(t), thermal unfolding h_unfold(t), oxidative degradation h_ox(t), and immunogenicity h_immune(t). The cumulative incidence function CIF_k(t) gives the probability of failing from cause k by time t. The CIF constraint (sum_k CIF_k(infinity) <= 1) forces a conservation law on failure probability, making tradeoffs between failure modes mathematically explicit.
Computational validation confirmed all 5 failure modes operate on overlapping timescales (30 min - 14 days) for designed miniproteins at physiological conditions. The proteostasis network is tightly interconnected (STRING scores 0.809-0.999), but Fine-Gray subdistribution hazard correctly handles correlated competing risks.
Supporting Evidence
Key strength: Complete framework with detailed assay panel; all other hypotheses depend on this foundation. Groundedness: 8/10. Claims verified: 7, failed: 0.. Application pathway: enabling_technology (Biopharmaceutical development)
How to Test
Cause-specific longitudinal assay panel in mouse serum at timepoints 0, 1h, 4h, 24h, 72h, 168h: (1) SEC-MALS for aggregation fraction, (2) LC-MS/MS intact mass for proteolytic fragments, (3) Met sulfoxide quantification for oxidation, (4) DSF for unfolded fraction, (5) ADA ELISA at days 7, 14, 21. Each protein molecule is assigned a failure time T and failure cause K based on the first assay detecting degradation above threshold.
Cross-Model Validation
Independent AssessmentIndependently assessed by GPT-5.4 Pro and Gemini 3.1 Pro for triangulation. Assessed independently by two external models for triangulation.
Other hypotheses in this cluster
The Dominant Competing Risk Theorem -- Optimizing One Failure Mode Provably Accelerates Another
PASSFix one way a protein drug breaks, and you mathematically guarantee another weakness gets worse.
Competing Risks Censoring Correction for Immunogenicity -- Anti-Drug Antibodies as Interval-Censored Competing Risk
CONDITIONALFixing a hidden flaw in drug safety testing: fast-failing proteins mask their immune risks until it's too late.
Nelson-Aalen Cumulative Hazard Decomposition Reveals Hidden Failure Modes in Accelerated Stability Studies
CONDITIONALSplitting protein drug degradation into its hidden failure modes could make shelf-life predictions far more accurate.
Related hypotheses
Gaussian Mixture Model Analysis of Cryo-EM OMV Populations Distinguishes Biogenesis Pathways in P. aeruginosa
PASSAI-powered microscopy could reveal how bacteria decide what to pack into their tiny 'mail packages'.
Abiotic vs Enzymatic PLOOH Regioselectivity as Chemical Fossil of Antioxidant Evolution
PASSThe chaotic chemistry of ancient iron reactions may have driven evolution of the precise enzymes that now control cell death.
Pyocyanin-GPX4-Ferroptosis Bidirectional Axis
PASSBacteria may hack their own iron supply by triggering a specific type of cell death in human lung cells.
Can you test this?
This hypothesis needs real scientists to validate or invalidate it. Both outcomes advance science.