A clinical trial is a research study in consenting human beings to answer specific health questions. One type of trial is a treatment trial, which tests the effectiveness of an experimental treatment. An example is a planned experiment designed to assess the efficacy of a treatment in humans by comparing the outcomes in a group of patients who receive the test treatment with the outcomes in a comparable group of patients who receive a placebo control treatment, where patients in both groups are enrolled, treated, and followed over the same time period.
A clinical trial is conducted according to a plan called a protocol. The protocol provides detailed description of the study. For a fixed-sample trial, the study protocol contains detailed information such as the null hypothesis, the one-sided or two-sided test, and the Type I and II error probability levels. It also includes the test statistic and its associated critical values in the hypothesis testing.
Generally, the efficacy of a new treatment is demonstrated by testing a hypothesis in a clinical trial, where is the parameter of interest. For example, to test whether a population mean is greater than a specified value , can be used with an alternative .
A one-sided test is a test of the hypothesis with either an upper (greater) or a lower (lesser) alternative, and a two-sided test is a test of the hypothesis with a two-sided alternative. The drug industry often prefers to use a one-sided test to demonstrate clinical superiority based on the argument that a study should not be run if the test drug would be worse (Chow, Shao, and Wang, 2003, p. 28). But in practice, two-sided tests are commonly performed in drug development (Senn, 1997, p. 161). For a fixed Type I error probability , the sample sizes required by one-sided and two-sided tests are different. See Senn (1997, pp. 161–167) for a detailed description of issues involving one-sided and two-sided tests.
For independent and identically distributed observations of a random variable, the likelihood function for is
|
where is the population parameter and is the probability or probability density of . Using the likelihood function, two statistics can be derived that are useful for inference: the maximum likelihood estimator and the score statistic.
The maximum likelihood estimate (MLE) of is the value that maximizes the likelihood function for . Under mild regularity conditions, is an asymptotically unbiased estimate of with variance , where is the Fisher information
|
and is the expected Fisher information (Diggle et al., 2002, p. 340)
|
The score function for is defined as
|
and usually, the MLE can be derived by solving the likelihood equation . Asymptotically, the MLE is normally distributed (Lindgren, 1976, p. 272):
|
If the Fisher information does not depend on , then is known. Otherwise, either the expected information evaluated at the MLE () or the observed information can be used for the Fisher information (Cox and Hinkley 1974, p. 302; Efron and Hinkley 1978, p. 458), where the observed Fisher information
|
If the Fisher information does depend on , the observed Fisher information is recommended for the variance of the maximum likelihood estimator (Efron and Hinkley, 1978, p. 457).
Thus, asymptotically, for large n,
|
where I is the information, either the expected Fisher information or the observed Fisher information .
So to test versus , you can use the standardized Z test statistic
|
and the two-sided p-value is given by
|
where is the cumulative standard normal distribution function and is the observed Z statistic.
If the BOUNDARYSCALE=SCORE is specified in the SEQDESIGN procedure, the boundary values for the test statistic are displayed in the score statistic scale. With the standardized Z statistic, the score statistic and
|
The score statistic is based on the score function for ,
|
Under the null hypothesis , the score statistic is the first derivative of the log likelihood evaluated at the null reference 0:
|
Under regularity conditions, is asymptotically normally distributed with mean zero and variance , the expected Fisher information evaluated at the null hypothesis (Kalbfleisch and Prentice, 1980, p. 45), where is the Fisher information
|
That is, for large n,
|
Asymptotically, the variance of the score statistic , , can also be replaced by the expected Fisher information evaluated at the MLE (), the observed Fisher information evaluated at the null hypothesis (, or the observed Fisher information evaluated at the MLE () (Kalbfleisch and Prentice, 1980, p. 46), where
|
|
Thus, asymptotically, for large n,
|
where I is the information, either an expected Fisher information ( or ) or a observed Fisher information ( or ).
So to test versus , you can use the standardized Z test statistic
|
If the BOUNDARYSCALE=MLE is specified in the SEQDESIGN procedure, the boundary values for the test statistic are displayed in the MLE scale. With the standardized Z statistic, the MLE statistic and
|
The following one-sample test for mean is used to demonstrate fixed-sample clinical trials in the section One-Sided Fixed-Sample Tests in Clinical Trials and the section Two-Sided Fixed-Sample Tests in Clinical Trials.
Suppose are n observations of a response variable Y
from a normal distribution
|
where is the unknown mean and is the known variance.
Then the log likelihood function for is
|
where c is a constant. The first derivative is
|
where is the sample mean.
Setting the first derivative to zero, the MLE of is , the sample mean. The variance for can be derived from the Fisher information
|
Since the Fisher information does not depend on in this case, is used as the variance for . Thus the sample mean has a normal distribution with mean and variance :
|
Under the null hypothesis , the score statistic
|
has a mean zero and variance
|
With the MLE , the corresponding standardized statistic is computed as , which has a normal distribution with variance 1:
|
Also, the corresponding score statistic is computed as and
|
which is identical to computed under the null hypothesis .
Note that if the variable Y
does not have a normal distribution, then it is assumed that the sample size n is large such that the sample mean has an approximately normal distribution.