flowchart LR
Start["Phase II<br/>Results"] --> Q1{"Efficacy?"}
Q1 -->|No| Stop1["Stop"]
Q1 -->|Yes| Q2{"Safety?"}
Q2 -->|No| Stop2["Stop/<br/>Redesign"]
Q2 -->|Yes| Q3{"Dose<br/>identified?"}
Q3 -->|No| More["More<br/>dose-finding"]
Q3 -->|Yes| Q4{"Commercial<br/>viable?"}
Q4 -->|No| Reassess["Reassess"]
Q4 -->|Yes| Go["Phase III"]
classDef stopNode fill:#ffebee,stroke:#c62828,stroke-width:2px,color:#000
classDef goNode fill:#e8f5e9,stroke:#2e7d32,stroke-width:2px,color:#000
classDef decisionNode fill:#fff3e0,stroke:#f57c00,stroke-width:2px,color:#000
classDef actionNode fill:#e3f2fd,stroke:#1976d2,stroke-width:2px,color:#000
class Stop1,Stop2 stopNode
class Go goNode
class Q1,Q2,Q3,Q4 decisionNode
class More,Reassess actionNode
8 Phase II: Therapeutic Exploratory
Phase II is where clinical research shifts from the question “is this drug safe?” to the more consequential question “does this drug actually work?” These studies represent the first controlled exploration of whether an investigational drug can produce the therapeutic effects that preclinical research suggested it might.
The transition from Phase I to Phase II marks a fundamental change in emphasis. Phase I studies established that the drug can be given to humans with acceptable safety. Phase II studies must now establish that giving the drug to humans accomplishes something worthwhile.
This exploration occurs in two stages (see Table 8.1). Early Phase II studies—sometimes called Phase IIa—seek proof of concept: evidence that the drug produces some measurable therapeutic effect in patients with the target condition. Later Phase II studies—Phase IIb—take that signal and try to understand it better (International Council for Harmonisation 2021):
| Aspect | Phase IIa (Proof of Concept) | Phase IIb (Dose Finding) |
|---|---|---|
| Primary Objective | Detect therapeutic signal; establish proof of concept | Characterize dose-response; select Phase III dose |
| Sample Size | 20-100 patients | 100-300 patients |
| Design | May be open-label or informally controlled | Randomized, controlled, often with multiple dose arms |
| Duration | Shorter (weeks to months) | Longer (months) |
| Endpoints | Exploratory; biomarkers and early efficacy signals | Primary efficacy endpoints; comprehensive safety |
| Key Question | “Does this drug do anything beneficial?” | “What dose balances efficacy and safety?” |
| Go/No-Go Decision | Proceed to dose-finding or terminate | Proceed to Phase III or terminate |
8.1 The Dose-Response Relationship
Understanding how response changes with dose is one of the most critical—and challenging—objectives of Phase II (International Council for Harmonisation 2021). The relationship is rarely simple.
At very low doses, there may be no detectable effect; the drug is present but at concentrations too low to produce meaningful target engagement. As doses increase, response emerges and grows. Eventually, a plateau is reached where higher doses produce little additional benefit. At still higher doses, toxicity may begin to outweigh efficacy.
The goal of dose-finding studies is to characterize this curve well enough to select the optimal dose for Phase III—high enough to produce meaningful efficacy, low enough to minimize toxicity, positioned in the portion of the curve where small dose differences do not produce large response differences. Common dose-finding approaches are summarized in Table 8.2.
| Approach | Description | Advantages | Disadvantages |
|---|---|---|---|
| Parallel Dose Groups | Fixed doses compared in separate arms | Straightforward analysis; clear comparisons | Requires larger sample sizes; less efficient |
| Dose Escalation Within Subjects | Individuals receive increasing doses sequentially | More efficient use of subjects | Effects may reflect cumulative exposure; carryover effects |
| Adaptive Designs | Allocation modified based on interim data | Concentrates patients in informative dose ranges | More complex; requires careful pre-specification |
| MCP-Mod | Multiple comparison procedure with modeling | Extracts maximum information about dose-response curve | Requires sophisticated statistical expertise |
| Bayesian Adaptive | Uses Bayesian updating to allocate to optimal doses | Efficient; incorporates prior knowledge | Complex implementation; regulatory acceptance varies |
Whatever the approach, certain principles apply (U.S. Food and Drug Administration 2019): Placebo should almost always be included—it is required for understanding the magnitude of any effect and for accounting for the substantial placebo response seen in many indications. The doses studied should bracket the expected optimal dose, including doses both above and below the target. And the analysis should incorporate pharmacokinetic data, since exposure-response relationships are often more interpretable than dose-response relationships.
8.2 Endpoint Selection
The choice of endpoint—what is measured to determine whether the drug works—is critical in Phase II. The ideal endpoint would be the clinical outcome we ultimately care about: survival, resolution of disease, or clinically meaningful symptom improvement. But these outcomes often require large samples or long follow-up to observe.
Phase II studies therefore often use surrogate endpoints: measurements that are expected to predict clinical outcome but can be observed more quickly and in smaller populations (see examples in Table 8.3) (International Council for Harmonisation 1998). When selecting surrogate endpoints for Phase II, the ICH E9(R1) estimands framework provides a structured approach to defining precisely what treatment effect is being estimated (International Council for Harmonisation 2019). This includes specifying how intercurrent events—such as treatment discontinuation, use of rescue medication, or switching to alternative therapy—will be handled in the analysis. For example, a Phase II diabetes study using HbA1c as a surrogate must define whether the estimand reflects treatment effect while patients remain on therapy (a “while on treatment” strategy) or the effect regardless of treatment adherence (a “treatment policy” strategy).
FDA’s Patient-Focused Drug Development guidance series emphasizes incorporating patient perspectives into endpoint selection, particularly for clinical outcome assessments (COAs) that capture treatment benefit from the patient’s perspective (U.S. Food and Drug Administration 2023). This patient-centric approach ensures that endpoints reflect outcomes that matter to patients, not just biomarkers convenient for measurement.
| Therapeutic Area | Surrogate Endpoint | Clinical Outcome | Validation Status |
|---|---|---|---|
| Cardiovascular | Blood pressure reduction | Myocardial infarction, stroke, cardiovascular death | Well-validated |
| Oncology | Tumor shrinkage (objective response rate) | Overall survival | Partially validated; varies by cancer type |
| HIV/AIDS | Viral load (HIV RNA) | AIDS progression, death | Well-validated |
| Diabetes | HbA1c reduction | Diabetic complications (retinopathy, nephropathy) | Well-validated |
| Alzheimer’s | Amyloid plaque reduction (PET imaging) | Cognitive decline | Controversial; under evaluation |
| Osteoporosis | Bone mineral density | Fracture risk | Reasonably well-validated |
Surrogates offer efficiency, but they carry risk. Not every change in a surrogate translates to clinical benefit. Drug A might lower blood pressure more than Drug B yet produce no better cardiovascular outcomes. Using surrogates in Phase II is generally accepted—the goal is to detect signals and select doses—but the relationship between surrogate and clinical benefit must be established if the surrogate is to support regulatory approval.
8.3 Adaptive Designs
The traditional approach to Phase II—design a study, run it to completion, analyze the results—has given way to more flexible adaptive designs that allow modifications based on accumulating data.
Modern Phase II studies frequently employ adaptive designs that allow for data-driven modifications during the trial. These may include sample size re-estimation to maintain statistical power in the face of unexpected variability, or response-adaptive randomization to prioritize better-performing arms. Furthermore, researchers may drop ineffective doses or add new ones to better characterize the dose-response relationship, and in some cases, utilize seamless Phase II/III designs that transition directly into confirmatory stages based on interim evidence.
The FDA has issued guidance supporting the use of adaptive designs when appropriately planned. Key requirements include pre-specifying the adaptation rules, controlling Type I error across the adaptation, and maintaining the integrity of the trial despite the modifications. For Phase II, advanced Bayesian frameworks are increasingly used to handle these adaptations—including stopping for futility, dropping arms, or response-adaptive randomization—supported by rigorous simulation workflows to maintain statistical integrity (Granholm et al. 2025).
8.4 The Phase II Decision
The transition from Phase II to Phase III is often called the Phase II decision, and it represents one of the most consequential moments in drug development (see Figure 8.1). By this point, a sponsor has typically invested $100–200 million and several years (DiMasi, Grabowski, and Hansen 2016). Proceeding to Phase III will require several hundred million dollars more. And the probability of success in Phase III, even with positive Phase II data, is far from certain.
The Phase II decision integrates factors such as the strength of efficacy evidence, the acceptability of the safety profile relative to the disease burden, and the identification of an optimal dose for Phase III. Additionally, sponsors must evaluate commercial viability and the drug’s position within the competitive clinical environment.
The statistics of Phase II are sobering. Approximately 70% of drugs fail to advance from Phase II to Phase III (Biotechnology Innovation Organization, QLS Advisors, and Informa Pharma Intelligence 2021). For some therapeutic areas—central nervous system disorders, for example—the failure rate is even higher. These failures represent not just scientific disappointments but also substantial financial losses.
Yet these failures serve a clear purpose. They prevent even costlier failures in Phase III. A rigorous Phase II program that kills an ineffective drug is doing exactly what it should do—providing a mechanism for learning that the drug does not work before hundreds of millions of additional dollars are spent.
8.5 The Critical Role of Biomarkers
Throughout Phase II, biomarkers provide mechanistic insight that complements clinical endpoints. A biomarker might measure target engagement (is the drug hitting its intended target?), pathway modulation (is the biological pathway being affected as expected?), or early disease response (are the first signs of therapeutic effect appearing?).
Biomarkers are particularly valuable when clinical endpoints take time to manifest. In Alzheimer’s disease, for example, cognitive decline occurs slowly. Biomarkers of amyloid plaque or tau pathology may provide earlier signals that a drug is having biological effect, even if clinical improvement is not yet evident.
The challenge is that biomarker changes do not always translate to clinical benefit. A drug might successfully engage its target without producing meaningful therapeutic improvement. The ultimate test remains clinical outcomes—but biomarkers help illuminate the path.