For CFOs & Finance Leaders4 min read

Why Drug Discovery Still Burns Billions on Guesswork

Chemical space holds up to 10^100 possible molecules — your lab screens about a million, and most of that spend is wasted.

The Problem

Nikola Tesla once said that "a little theory and calculation" could have saved Thomas Edison 90% of his labor. More than a century later, the pharmaceutical and materials science industries are still running Edison's playbook — synthesizing and physically testing massive libraries of compounds, hoping to stumble on a winner. The numbers reveal exactly how badly this approach has broken down.

The space of possible drug-like molecules is estimated between 10^60 and 10^100. Your best high-throughput screening campaign tests roughly one million compounds. That means you are covering 0.000000000000000000000000000000000000000000000000001% of the available options. The whitepaper puts it bluntly: trying to map the Pacific Ocean by dipping a teaspoon into the water at random intervals. If you are testing new materials in a wet lab without simulating them first, you are lighting money on fire.

This is not a technology curiosity. It is a structural crisis in R&D economics. The "Edisonian" method — brute-force trial and error — substitutes labor for intelligence. That trade-off becomes exponentially worse as the complexity of your targets increases. The low-hanging fruit of simple small molecules has been picked. What remains requires a fundamentally different approach: simulate first, then synthesize only the highest-probability candidates.

Why This Matters to Your Business

The financial damage is already visible in your industry's books. Here are the numbers that should concern your board:

  • $2.23 billion per new drug — that is the average cost to develop a single pharmaceutical asset as of 2024. This figure absorbs the cost of thousands of failures for every success.
  • 1.2% internal rate of return — pharmaceutical R&D hit a 12-year low in 2022 before recovering to 5.9% in 2024. That rebound was driven largely by outliers like GLP-1 agonists, not by systemic improvement.
  • ~90% failure rate — the vast majority of R&D candidates never reach market. Every dollar spent physically testing a compound that simulation could have eliminated is a dollar diverted from a viable candidate.
  • Cost-informed AI can cut reagent costs by up to 90% — algorithms that factor in the price of experiments alongside their informational value have demonstrated dramatic savings.

This declining R&D productivity has a name: Eroom's Law — Moore's Law spelled backwards. Each generation of drugs gets harder and more expensive to discover. Your competitors face the same wall, but the ones who move to simulation-first discovery will cross that wall years ahead of you.

For your P&L, the implication is direct. High-throughput screening is OpEx-heavy: consumables, reagents, personnel time, and equipment hours. Every experiment you avoid through better prediction drops straight to the bottom line. Meanwhile, patent clocks keep ticking. Every day saved in R&D is an extra day of market exclusivity. AI-designed small molecules from companies like Exscientia reached Phase I trials in roughly 12 months, compared to the industry average of four to five years.

What's Actually Happening Under the Hood

The core failure is that traditional discovery screens compounds instead of designing them. Screening assumes your answer already exists inside a pre-built library. But as you move into complex biologics, multi-element alloys, and nanostructured materials, the probability of the optimal solution sitting in any physically manageable library drops to near zero.

Think of it this way. You have a combination lock with 100 digits. Traditional screening tries random combinations one by one. A simulation-first approach studies the lock's mechanics, narrows the possibilities to a handful, and then turns the dial.

The market is flooded with black-box AI models that learn only from data correlations. These models break down in novel domains because experimental data is sparse, noisy, and expensive to collect. Standard machine learning performs well within its training data but fails when asked to predict properties of entirely new material classes. Worse, generative AI tools built on large language models can "hallucinate" — generating molecule descriptions that look plausible but violate basic chemistry. They might propose a structure that breaks valency rules or ignores conservation of mass. Molecules are three-dimensional graphs of atoms and bonds, not text strings. Treating them as sentences leads to physically impossible outputs.

The alternative is Physics-Informed Machine Learning, or PIML — models that embed the actual laws of physics (conservation of energy, thermodynamics, quantum mechanics) directly into the AI's architecture. These models need far less training data because the rules of chemistry are baked in. They extrapolate better into unknown territory. And they cannot propose molecules that violate fundamental physical constraints. One system, called FlowER, explicitly tracks how electrons redistribute during chemical reactions, ensuring mass and charge are always conserved. No standard language model can do that.

What Works (And What Doesn't)

Three approaches that look promising but fall short:

  • Bigger screening libraries: Scaling from one million to one billion compounds still covers a vanishingly small fraction of chemical space. You cannot brute-force your way through 10^60 possibilities.
  • Generic AI wrappers on public language models: These tools help with literature review and protocol drafting, but they lack the domain-specific physics constraints needed for molecular design. A wrapper cannot enforce conservation of mass in a chemical reaction.
  • "Fail fast" without simulation: Failing in a wet lab is expensive — reagents, synthesis time, and equipment all burn capital. Fast physical failure is still slow and costly compared to virtual failure that takes milliseconds of compute time.

What actually works is a three-step closed-loop architecture:

  1. Predict and select (input): A Bayesian Optimization engine — a mathematical framework that balances exploring unknown chemical territory with refining known promising areas — identifies the highest-value experiment to run next. It does not guess randomly. It calculates where the biggest information gain lies relative to cost. Multi-fidelity versions of this algorithm blend cheap computer simulations with expensive lab results, triggering physical experiments only when the simulation's confidence is too low.

  2. Automate and synthesize (processing): Robotic platforms receive instructions from the AI, synthesize the predicted material, and run characterization tests — spectroscopy, X-ray diffraction, microscopy — without human intervention. The A-Lab at Lawrence Berkeley National Laboratory used this approach to synthesize 41 novel inorganic compounds in 17 days. When a synthesis failed, the AI analyzed the results, adjusted the recipe, and retried autonomously.

  3. Learn and update (output): Every result — success or failure — feeds back into the model. This is where negative data becomes gold. Failed experiments sharpen the AI's understanding of what does not work, mapping dead-end regions of chemistry permanently. Your organization never wastes resources on those paths again. This accumulated knowledge of failure landscapes is durable intellectual property.

For your compliance and audit teams, this loop creates something traditional R&D lacks: a complete, traceable record of every prediction, every decision rationale, and every experimental outcome. The AI does not just say "try this compound." It shows you why it chose that compound over every alternative, what it expected to learn, and how the result changed its model. That decision trail is auditable, repeatable, and defensible.

Your digital twin — a virtual replica of your physical lab — adds another layer of protection. Before any robot moves, the experiment runs in simulation. This catches protocol errors, timing conflicts, and equipment collisions before they destroy samples or equipment. Real-time comparison between the twin's predictions and actual sensor data flags anomalies like clogged pipettes or drifting temperatures before they ruin a batch.

The economic case is straightforward. Autonomous equipment runs 24/7 with near-100% uptime, compared to the 30-40% typical of human-staffed labs. Insilico Medicine moved an AI-discovered fibrosis candidate from target discovery to preclinical stage in under 18 months at a fraction of the usual cost. These are not projections. They are results already delivered in production environments.

Key Takeaways

  • Chemical space holds up to 10^100 possible molecules — even screening a billion compounds covers a negligible fraction, making brute-force discovery statistically doomed.
  • Developing a single new drug now costs $2.23 billion on average, and pharmaceutical R&D returns hit a 12-year low of 1.2% in 2022.
  • Physics-informed AI models embed the laws of chemistry directly, preventing the hallucinated molecules that standard language models produce.
  • Closed-loop autonomous labs like the A-Lab synthesized 41 novel materials in 17 days — a task that would take human researchers months or years.
  • Cost-informed Bayesian Optimization can reduce experimental reagent costs by up to 90% while reaching the same results.

The Bottom Line

The math is clear: physical trial-and-error cannot search a space of 10^60 molecules, and every unsimulated wet-lab experiment risks burning capital on compounds that violate basic physics. Closed-loop AI that embeds physical laws, learns from every failure, and creates an auditable decision trail is how you restore R&D economics. Ask your AI vendor: when your model proposes a new molecule, can it prove that the prediction obeys conservation of mass and thermodynamic stability — and show you the full decision trail for why it chose that candidate over every alternative?

FAQ

Frequently Asked Questions

How much does it cost to develop a new drug in 2024?

The average cost to develop a single new drug is approximately $2.23 billion as of 2024. This figure includes the cost of thousands of failed candidates that precede each success. Pharmaceutical R&D returns hit a 12-year low of 1.2% in 2022 before recovering to 5.9% in 2024, largely driven by outlier drug classes like GLP-1 agonists.

Can AI really speed up drug discovery?

Yes, with documented results. AI-designed small molecules from Exscientia entered Phase I clinical trials in roughly 12 months, compared to the industry average of 4-5 years. Insilico Medicine moved an AI-discovered fibrosis candidate from target discovery to preclinical stage in under 18 months. The A-Lab at Berkeley synthesized 41 novel inorganic materials in just 17 days using autonomous AI.

Why can't regular AI tools like ChatGPT design new drugs?

Large language models treat molecules as text strings, which can lead to hallucinations — generating descriptions of molecules that violate basic chemistry rules like valency or conservation of mass. Molecules are 3D structures defined by atoms, bonds, and geometric constraints. Physics-informed machine learning models embed physical laws directly into the AI, ensuring predictions remain chemically valid even in unexplored regions of chemical space.

Build Your AI with Confidence.

Partner with a team that has deep experience in building the next generation of enterprise AI. Let us help you design, build, and deploy an AI strategy you can trust.

Veriprajna Deep Tech Consultancy specializes in building safety-critical AI systems for healthcare, finance, and regulatory domains. Our architectures are validated against established protocols with comprehensive compliance documentation.