Ranomics
Scientific research and computational biology
binder discoveryacademic researchprotein designde novo designRFdiffusionBindCraftyeast displayCROgrant budget

Binder Design on a Grant Budget: Scoping a Single-Target Campaign

Most published benchmarks for AI protein binder design assume a well-resourced industrial pipeline: GPU clusters, in-house yeast display, FACS, NGS, and enough budget to run several hundred to several thousand designs through the full experimental funnel. Most academic labs running their first de novo binder campaign do not have that setup. This post is a practical scoping guide for running a single-target binder campaign on a defined grant budget — what to prioritize, what to cut, and what actually determines cost.

The Cost Drivers in a Binder Design Campaign

Before deciding what to cut, it helps to understand where the money actually goes in a binder campaign. In rough order of contribution:

  1. Gene synthesis of the filtered candidate pool. This scales linearly with the number of designs you carry into the wet lab. A 1,500-design pool is a different cost category from a 300-design pool.
  2. FACS or MACS selection rounds. Each sorting round has instrument time, reagents, and bench-scientist overhead. Multi-round campaigns compound.
  3. NGS sequencing of sorted populations. One MiSeq or NextSeq run per sorted population is the standard readout.
  4. Display platform prep. Yeast display construction and transformation is a relatively fixed cost per library.
  5. GPU compute for generative design. RFdiffusion and BindCraft campaigns are real GPU hours, but they are usually a smaller line item than gene synthesis for a moderate-sized pool.
  6. Bioinformatics and reporting. Hit calling, enrichment analysis, and writeup — bounded by design pool size.

Computational design is often the smallest line item. The big cost drivers are the synthesis and experimental follow-through.

What to Cut: Scope, Not Rigor

The single most effective way to bring a binder campaign inside a grant budget is to cut pool size, not methodology rigor. A single-round campaign on a 200-500 design pool with a clean NGS readout is defensible, publishable, and answers a specific scientific question. A 2,000-design multi-round campaign is a production pipeline, not a focused experiment.

Specifically, at grant scale you can typically run:

  • One algorithm, not three. Pick RFdiffusion plus ProteinMPNN sequence design, or pick BindCraft. Running RFdiffusion, BindCraft, and Boltzgen in parallel is a flagship-scale choice.
  • One selection round, not two or three. A single FACS round on a well-designed library gives you enough enrichment signal to rank candidates. Multi-round affinity maturation is a separate experimental question and a separate campaign.
  • Yeast display, not mammalian display. Yeast display is cheaper, faster, and well-validated for extracellular target binding. Mammalian display is worth the cost only when post-translational modifications on the displayed protein are the scientific question.
  • NGS hit calling, not characterization. SPR, BLI, and ITC affinity measurements are expensive. A ranked NGS hit list plus one or two purified-protein validations of top hits is usually enough for a first paper.

This is how the Binder Pilot program is scoped. It is not a diminished version of the flagship; it is a deliberately focused campaign for teams where the output is one solid ranked hit list plus the NGS and sequence data to back it up.

What Not to Cut: Computational Quality Gates

The computational side of a binder campaign is the cheapest part. Do not cut it.

Specifically, keep:

  • Self-consistency filtering. Before you synthesize any designs, fold each candidate with an independent predictor (ESMFold, ColabFold, or Boltz-2) and compare to the intended backbone. Designs with RMSD > 2 Å from the intended fold go in the bin. This single filter typically removes 50-80% of raw generative output and drastically improves wet-lab hit rates.
  • Hotspot selection. Where you direct the design matters more than how many designs you generate. Spend an hour on Epitope Scout or manual PyMOL inspection identifying surface patches that are hydrophobic, rigid, accessible, and ideally contain known hotspot residues (Trp, Tyr, Arg, Phe). Bad hotspot selection makes a 2,000-design pool perform worse than a 300-design pool on the right epitope.
  • Structure quality triage on the target. If your starting structure is an AlphaFold model, check pLDDT locally around the intended binding region. Disordered or low-confidence regions are a warning sign. Crystal structures are preferred; high-confidence models are usable; low-confidence models are a coin flip.

Choosing Your Target

The most common failure mode for academic binder campaigns is not computational — it is target choice. Some targets are hard even for well-resourced pipelines. At grant scale, pick a target where the biology is on your side:

  • Soluble, extracellular, folded. Yeast display works best for targets that can be surface-exposed and bound by designed proteins in a largely native conformation.
  • Ideally with a known binding partner structure. Even if you are not engineering an antibody fragment, a co-crystal structure of any protein bound to your target tells you which surfaces are bindable.
  • Rigid and structurally well-defined. Flexible loops, disordered termini, and conformational ensembles are where computational design struggles and where hit rates drop.
  • Not a PTM-dependent epitope. If the native binding partner only recognizes a glycosylated, phosphorylated, or acetylated version of your target, a yeast-display de novo binder campaign is the wrong tool.

Timeline Expectations

A single-round, single-algorithm, grant-scale binder campaign typically runs 4-6 weeks end-to-end from kickoff to delivered hit list, depending on synthesis turnaround. That is roughly half the timeline of a flagship multi-algorithm program, and it is compatible with most grant or dissertation cycles.

The slowest steps are almost always the physical ones: gene synthesis (1-2 weeks) and NGS run scheduling. The computational side runs in days.

What You Should Publish

A focused, grant-scale binder campaign is not a diminished result. Published work in the RFdiffusion and BindCraft literature routinely reports single-round campaigns against single targets with 2-12 confirmed binders per 1,000 screened candidates. A clean single-round Pilot-scale campaign hitting that range, with the NGS enrichment traces to back it up, is a publishable dataset in its own right.

The things to include in a paper or preprint:

  • The generative algorithm, version, and seed parameters.
  • The hotspot selection procedure and justification.
  • The filtering pipeline (self-consistency thresholds, expression filters, any physics-based triage).
  • The library size at synthesis and at sort.
  • The NGS enrichment distribution for the top hits, not just the headline hit count.
  • At least one orthogonal validation of the top-ranked hit (purified protein binding assay on a handful of candidates).

If You Are Deciding Between Running It In-House or Outsourcing

This is a judgment call, not a formula. Run it in-house when you have a yeast display pipeline, a FACS core, and a scientist or postdoc with bandwidth. Outsource when any of those is missing and when the project timeline cannot absorb a six-month ramp to set up display from scratch.

A well-scoped outsourced Pilot at grant scale typically comes in at roughly the cost of gene synthesis plus one FACS sort plus one NGS run, plus the bioinformatics overhead. There is no line item for GPU compute that you would not already be paying in cloud or shared-cluster time if you ran it yourself. The main thing you are buying is experimental throughput and reproducible methodology, not access to the algorithms themselves — RFdiffusion, BindCraft, and ProteinMPNN are all open-source.

Starting Points

If you have a target structure, an idea of the binding region, and a grant or core-facility budget, the fastest path to scoping a campaign is:

  1. Run Epitope Scout on your target (free) to get a ranked list of candidate epitope patches.
  2. Pick one patch. Send us the structure and the selection. We will assess feasibility and scope a Binder Pilot around it.
  3. Expect a kickoff call to define deliverables, then 4-6 weeks to a ranked hit list.

For multi-target pipelines, multi-round optimization, or teams that need the 100% binder guarantee that the flagship AI Binder Sprint includes, the Pilot is not the right scope. For a first single-target campaign on a grant, it usually is.

  • Binder Pilot: Grant-scale starter program — single-round, smaller design pool.
  • Epitope Scout: Free self-serve epitope identification to de-risk the target before you spend.
Share

Ready to start a project?

Tell us about your protein engineering challenge. We will scope a program and get back to you within 24 hours.

Start a project →