Date: 2026-04-03 |
Target: PD-L1 (02_PDL1) |
Stage: Generate only |
DEMO RUN
Key Findings
1. Pipeline works end-to-end.
Generated 2 binder candidates for PD-L1 in 82.92 seconds on 1x H100.
AF2 reward scoring completed successfully.
2. Both samples fail production thresholds — expected at 100 steps with no refinement.
i_PAE × 31 = 25.76 and 27.32 (need ≤ 7.0).
pLDDT = 0.212 and 0.239 (need ≥ 0.9).
3. Sample 0 (n=262) is the better candidate:
lower i_PAE (0.831 vs 0.881),
higher i_pTM (0.147 vs 0.068),
much lower RMSD (11.25Å vs 45.21Å).
4. Reward = -i_PAE only.
All other reward weights (con, plddt, dgram_cce, etc.) are 0.0.
Consider multi-objective reward for production runs.
Metrics Comparison
Key AF2 metrics for both generated binder candidates
Quality Assessment
Demo values (blue/orange) vs production pass-rate thresholds (red dashed)
Criterion
Sample 0 (n=262)
Sample 1 (n=234)
Threshold
Status
i_PAE × 31 ≤ 7.0 (binding)
25.76
27.32
≤ 7.0
FAIL
pLDDT ≥ 0.9 (structure)
0.212
0.239
≥ 0.9
FAIL
scRMSD < 1.5Å (designability)
N/A (evaluate stage not run)
< 1.5Å
—
Sample Profile
Normalized radar chart comparing structural quality metrics across both samples
Timing & Configuration
Left: generation timing. Right: demo config vs recommended production settings
Parameter
Demo
Production
Impact
Diffusion steps
100
400
4× better convergence
Samples
2
32–64
Cover binder length space
Replicas
1
4–16
More candidates per length
Search
best-of-n
beam-search
Smarter exploration
Refinement
None
sequence_hallucination
Key to SOTA quality
GPUs
1× H100
4–8× H100
Parallel generation
Next Steps
Production run: 400 steps, 32 samples, beam-search, sequence_hallucination refinement on PD-L1
Full pipeline: Add filter → evaluate → analyze for designability and pass-rate metrics
Multi-target: Run on PD-1, IFNAR2, CD45 to compare difficulty across targets
Search algorithm sweep: Compare best-of-n vs beam-search vs MCTS