Critical Evaluation of Clinical Trials


6

Critical Evaluation of Clinical Trials



Elliott M. Antman


Despite many decades of advances in diagnosis and management, cardiovascular disease (CVD) remains the leading cause of death in the United States and other high-income countries, as well as many developing countries.1 Managing the burden of CVD consumes 16% of overall national health care expenditures in the United States; interventions to treat CVD are therefore a major focus of contemporary clinical research. Therapeutic recommendations are no longer based on nonquantitative pathophysiologic reasoning but instead are evidence-based. Rigorously performed trials are required before regulatory approval and clinical acceptance of new treatments (drugs, devices, and biologics) and biomarkers.2 Thus the design, conduct, analysis, interpretation, and presentation of clinical trials constitute a central feature of the professional life of the contemporary cardiovascular specialist.3,4 Case-control studies and analyses from registries are integral to epidemiologic and outcomes research but are not strictly clinical trials and are not discussed in this chapter.5,6



Constructing the Research Question


Before embarking on a clinical trial, investigators should review the FINER criteria for a good research question (Table 6-1) and the phases of evaluation of new therapies (Table 6-2) and should familiarize themselves with the processes of designing and implementing a research project, good clinical practice, and drawing conclusions from the findings (imageFig. e6-1).3,4,610 A clinical trial may be designed to test for superiority of the investigational treatment over the control therapy but also may be designed to show therapeutic similarity between the investigational and the control treatments (noninferiority design) (Fig. 6-1; Table 6-3).






TABLE 6-3


Trial Designs to Replace Standard of Care




















































PARAMETER SUPERIORITY NONINFERIORITY
Objective 1 Objective 2
Goal Test beats control Test beats placebo Test as good as standard
HO
HA
Ptest = Pcontrol
Ptest < Pcontrol
Assessment of test made against putative placebo Ptest ≥ Pstandard + M
Ptest < Pstandard + M
Source of data Trial Historical data Trial
Type I error Set by regulatory authorities, typically 0.05 Set by regulatory authorities, typically 0.05 Set by regulatory authorities, typically 0.05
Type II error (power) Set by investigator N/A Set by investigator
Major threats to validity Assay sensitivity; bias Assay constancy Assay sensitivity; bias
Inferential reasoning from trial Results in study cohort yield estimate of Ptest − Pcontrol in population of patients with same clinical characteristics and disease state Combining results from the trial (Ptest − Pstandard) and historical data (Pstandard − Pplacebo) yields estimate of (Ptest − Pplacebo) in population of patients with same clinical characteristics and disease state Results in study cohort yield estimate of Ptest − Pstandard in population of patients with same clinical characteristics and disease state
Generalizability to universe of all patients with the disease state Related to enrollment criteria; the more restrictive they are, the less generalizable are the results to the entire universe of patients with the disease state Enrollment criteria of prior trials and medical practice concurrent with those trials determines generalizability of estimate of Pstandard − Pplacebo to contemporary practice Related to enrollment criteria; the more restrictive they are, the less generalizable are the results to the entire universe of patients with the disease state


image





In a noninferiority trial, investigators specify a noninferiority criterion (M) and consider the investigational treatment to be therapeutically similar to control (standard) therapy if, with a high degree of confidence, the true difference in treatment effects is less than M (see Fig. 6-1).11,12 Specification of the noninferiority margin M involves considerable discussion between the investigators (advocating for clinical perception of minimally important difference) and regulatory authorities (advocating for assurance that the investigational treatment maintains a reasonable fraction of the efficacy of the standard treatment based on previous trials).11,12 The investigational therapy may satisfy the definition of noninferiority but may or may not also show superiority over the control therapy.13 Thus superiority can be considered a special case of noninferiority, in which the entire confidence interval for the difference in treatments falls in favor of the investigational treatment (see Fig. 6-1). Investigators can stipulate that a trial is being designed to test both noninferiority and superiority (see Table 6-3). For a trial that is configured as a noninferiority trial, it is acceptable to test for superiority conditional on having demonstrated noninferiority.14 Because of the subjective nature of the choice of M, the reverse is not true—trials configured for superiority cannot later test for noninferiority unless the margin M was prespecified.


Regardless of the design of the trial, it is essential that investigators provide a statement of the hypothesis being examined, using a format that permits biostatistical assessment of the results (see Fig. e6-1). Typically, a null hypothesis (H0) is specified (e.g., no difference exists between the treatments being studied) and the trial is designed to provide evidence leading to rejection of H0 in favor of an alternative hypothesis (HA) (a difference exists between treatments). To determine whether H0 may be rejected, investigators specify type I (α) and type II (β) errors, referred to as the false-positive and false-negative rates, respectively. By convention, α is set at 5%, indicating a willingness to accept a 5% probability that a significant difference will occur by chance when there is no true difference in efficacy. Regulatory authorities may on occasion demand a more stringent level of α—for example, when a single large trial is being proposed rather than two smaller trials—to gain approval of a new treatment. The value of β represents the probability that a specific difference in treatment efficacy might be missed, so that the investigators incorrectly fail to reject H0 when there is a true difference in efficacy. The power of the trial is given by the quantity (1 − β) and is selected by the investigators—typically, between 80% and 90%.7 Using the quantities α, β, and the estimated event rates in the control group, the sample size of the trial can be calculated with formulas for comparison of dichotomous outcomes or for a comparison of the rate of development of events over a follow-up period (time to failure). Table 6-3 summarizes the major features and concepts for superiority and noninferiority trials designed to change the standard of care for patients with a cardiovascular condition.



Clinical Trial Design


Controlled Trials


The randomized controlled trial (RCT) is considered the gold standard for the evaluation of new treatments (Fig. 6-2). The allocation of subjects to control and test treatments is not determined but is based on an impartial scheme (usually a computer algorithm). Randomization reduces the likelihood of patient selection bias in allocation of treatment, enhances the likelihood that any baseline differences between groups are random so that comparable groups of subjects can be compared, and validates the use of common statistical tests. Randomization may be fixed over the course of the trial or may be adaptive, based on the distribution of treatment assignments in the trial to a given point, baseline characteristics, or observed outcomes (see Fig. 6-2A).15 Fixed randomization schemes are more common and are specified further according to the allocation ratio (equal or unequal assignment to study groups), stratification levels, and block size (i.e., constraining the randomization of patients to ensure a balanced number of assignments to the study groups, especially if stratification [e.g., based on enrollment characteristics] is used in the trial). During the course of a trial, investigators may find it necessary to modify one or more treatments in response to evolving data (internal or external to the trial) or a recommendation from the trial’s data safety monitoring board (DSMB)—that is, to implement an adaptive design (see Fig. 6-2B).15 Adaptive designs are most readily implemented during phase II of therapeutic development. Regulatory authorities are concerned about protection of the trial integrity and the studywise alpha level when adaptive designs are used in registration pathway trials.15 The most desirable situation is for the control group to be studied concurrently and to comprise subjects distinct from those of the treatment group. Other trial formats that have been used in cardiovascular investigations include nonrandomized concurrent and historical controls (Fig. 6-3A, B), crossover designs (see Fig. 6-3C), withdrawal trials (see Fig. 6-3D), and group or cluster allocations (groups of subjects or investigative sites are assigned as a block to test or control). Depending on the clinical circumstances, the control agent may be a placebo or a drug or other intervention used in active treatment (standard of care).





Other Forms of Controlled Studies


Trials in which the investigator selects the subjects to be allocated to the control and treatment groups are nonrandomized, concurrent control studies (see Fig. 6-3A). In this type of trial design, clinicians do not leave the allocation of treatment in each patient to chance, and patients are not required to accept the concept of randomization. It is, however, difficult for investigators to match subjects in the test and control groups for all relevant baseline characteristics, introducing the possibility of selection bias, which could influence the conclusions of the trial. Clinical trials that use historical controls compare a test intervention with data obtained earlier in a nonconcurrent, nonrandomized control group (see Fig. 6-3B). Potential sources for historical controls include previously published trials in cardiovascular medicine and electronic data bases of clinic populations or registries. The use of historical controls allows investigators to offer the treatment(s) being investigated to all subjects enrolled in the trial. The major drawbacks are the potential for bias in the selection of the control population and failure of the historical controls to reflect accurately the contemporary picture of the disease under study.


The crossover design is a special case of the RCT in that each subject serves as his or her own control (see Fig. 6-3C). The appeal of this design is that the same subject is used for both test and control groups, thereby diminishing the influence of interindividual variability and allowing a smaller sample size. However, important limitations to a crossover design are the assumptions that the effects of the treatment assigned during the first period have no residual effect on the treatment assigned during the second period, and that the patient’s condition does not change during the periods being compared.


In a fixed sample size design, the trialists specify the necessary sample size before patient recruitment, whereas in an open or closed sequential design, subjects are enrolled only if the evolving test-control difference from previous subjects remains within prespecified boundaries.15,16 Trials with a fixed design can be configured to continue until the requisite number of endpoints is reached (event driven), thus ensuring that enough endpoints will occur to provide intended power to evaluate the null and alternative hypotheses. When both the patient and the investigator are aware of the treatment assignment, the trial is said to be unblinded. Single-blind trials mask the treatment from the patient but permit it to be known by the investigator, double-blind trials mask the treatment assignment from both the patient and the investigator, and triple-blind trials also mask the actual treatment assignment from the DSMB and provide data only in the form of group A and group B categories.



Jun 4, 2016 | Posted by in CARDIOLOGY | Comments Off on Critical Evaluation of Clinical Trials

Full access? Get Clinical Tree

Get Clinical Tree app for offline access