[Skip to Navigation]
Sign In
Image description not available.
Figure 1.—Time to publication from start of enrollment (A), time to publication from completion of follow-up (B), and time to completion of follow-up from start of enrollment (C) for studies with positive and negative results. Panel D shows the time to completion of follow-up from start of enrollment for trials with significant results for either an experimental or a control arm and for trials with nonsignificant results. Only the 66 completed studies are considered.
Image description not available.
Figure 2.—Time from completion of follow-up to first submission among 66 completed studies (A) and time from first submission to publication among 45 submitted studies (B) for studies with positive and negative results.
Table 1.—Characteristics of Trials*
Image description not available.
Table 2.—Predictors of Time to Study Completion and Publication*
Image description not available.
Table 3.—Examples of Positive and Negative Trials on the Same Regimen Starting at Approximately the Same Time, but Published With a Time Lag*
Image description not available.
1.
Easterbrook P, Berlin JA, Gopalan R, Matthews DR. Publication bias in clinical research.  Lancet.1991;337:867-872.Google Scholar
2.
Dickersin K, Min YI. NIH clinical trials and publication bias.  Online J Curr Clin Trials.April 28, 1993;doc 50.Google Scholar
3.
Dickersin K. The existence of publication bias and risk factors for its occurrence.  JAMA.1990;263:1385-1389.Google Scholar
4.
Dickersin K, Min Y-I. Publication bias: the problem that won't go away.  Ann N Y Acad Sci.1993;703:135-148.Google Scholar
5.
Dickersin K, Min Y-I, Meinert CL. Factors influencing publication of research results: follow-up of applications submitted to two institutional review boards.  JAMA.1992;267:374-378.Google Scholar
6.
Hetherington J, Dickersin K, Chalmers I, Meinert C. Retrospective and prospective identification of unpublished controlled trials: lessons from a survey of obstetricians and pediatricians.  Pediatrics.1989;84:374-380.Google Scholar
7.
Scherer RW, Dickersin K, Langenberg P. Full publication of results initially presented as abstracts: a meta-analysis.  JAMA.1994;272:158-162.Google Scholar
8.
Simes RJ. Publication bias: the case for an international registry of clinical trials.  J Clin Oncol.1986;4:1529-1541.Google Scholar
9.
Ioannidis JPA, Cappelleri JC, Schmid CH, Lau J. Impact of epidemic and individual heterogeneity on the population distribution of disease progression rates: an example from populations of trials of human immunodeficiency virus infection.  Am J Epidemiol.1996;144:1074-1085.Google Scholar
10.
Koren G. Bias against negative studies in newspaper reports of medical research.  JAMA.1991;266:1824-1826.Google Scholar
11.
Koren G, Graham K, Shear H, Einarson T. Bias against the null hypothesis: the reproductive hazards of cocaine.  Lancet.1989;2:1440-1442.Google Scholar
12.
Cooper DA, Gatell JM, Kroon SA.  et al.  Zidovudine in persons with asymptomatic HIV infection and CD4+ cell counts greater than 400 per cubic millimeter.  N Engl J Med.1993;329:297-303.Google Scholar
13.
Volberding PA, Lagakos SW, Koch MA.  et al.  Zidovudine in asymptomatic human immunodeficiency virus infection: a controlled trial in persons with fewer than 500 CD4-positive cells per cubic millimeter.  N Engl J Med.1990;322:941-949.Google Scholar
14.
Concorde Coordinating Committee.  Concorde: MRC/ANRS randomised double-blind controlled trial of immediate and deferred zidovudine in symptom-free HIV infection.  Lancet.1994;343:871-881.Google Scholar
15.
Youle MS, Gazzard BG, Johnson MA.  et al.  Effects of high-dose oral acyclovir on herpesvirus disease and survival in patients with advanced HIV disease: a double-blind, placebo controlled study.  AIDS.1994;8:641-649.Google Scholar
16.
Hersh EM, Brewton G, Abrams D.  et al.  Ditiocarb sodium (diethyldithiocarbamate) therapy in patients with symptomatic HIV infection and AIDS: a randomised, double-blind, placebo-controlled multicenter study.  JAMA.1991;265:1538-1544.Google Scholar
17.
The HIV87 Study Group.  Multicenter, randomized, placebo-controlled study of ditiocarb (Imuthiol) in human immunodeficiency virus-infected asymptomatic and minimally symptomatic patients.  AIDS Res Hum Retroviruses.1993;9:83-89.Google Scholar
18.
Spector SA, McKinley GF, Lalezari JP.  et al.  Oral ganciclovir for the prevention of cytomegalovirus disease in persons with AIDS.  N Engl J Med.1996;334:1491-1497.Google Scholar
19.
Ioannidis JPA, Cappelleri JC, Sacks HS, Lau J. The relationship between study design, results and reporting of randomized clinical trials of HIV infection.  Control Clin Trials.1997;18:431-444.Google Scholar
20.
Ioannidis JPA, Cappelleri JC, Lau J.  et al.  Early or deferred zidovudine therapy in HIV-infected patients without an AIDS-defining illness: a meta-analysis.  Ann Intern Med.1995;122:856-866.Google Scholar
21.
DeMets DL, Fleming TR, Whitley RJ.  et al.  The Data and Safety Monitoring Board and acquired immune deficiency syndrome trials.  Control Clin Trials.1995;16:408-421.Google Scholar
22.
O'Brien PC, Fleming TR. A multiple testing procedure for clinical trials.  Biometrics.1979;35:549-556.Google Scholar
23.
Blumenthal D, Campbell EG, Anderson MS, Causino N, Louis KS. Withholding research results in academic life sciences: evidence from a national survey of faculty.  JAMA.1997;277:1224-1228.Google Scholar
24.
Rennie D. Thyroid storm.  JAMA.1997;277:1238-1243. [published correction appears in JAMA1997;277;1762].Google Scholar
25.
Cappelleri JC, Ioannidis JPA, Schmid CH.  et al.  Large trials vs meta-analysis of smaller trials: how do their results compare?  JAMA.1996;276:1332-1338.Google Scholar
26.
Ioannidis JP, Lau J. The impact of high risk patients on the results of clinical trials.  J Clin Epidemiol.1997;50:1089-1098.Google Scholar
27.
Lau J, Antman EM, Jimenez-Silva J, Kupelnick B, Mosteller F, Chalmers TC. Cumulative meta-analysis of therapeutic trials for myocardial infarction.  N Engl J Med.1992;327:248-254.Google Scholar
28.
McIntosh M. The population risk as an explanatory variable in research synthesis of clinical trials.  Stat Med.1996;15:1713-1728.Google Scholar
29.
Ioannidis JP, Lau J. On meta-analyses of meta-analyses.  Lancet.1996;348:756.Google Scholar
Original Contribution
January 28, 1998

Effect of the Statistical Significance of Results on the Time to Completion and Publication of Randomized Efficacy Trials

Author Affiliations

From the HIV Research Branch, Division of AIDS, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Bethesda, Md.

JAMA. 1998;279(4):281-286. doi:10.1001/jama.279.4.281
Abstract

Context.— Medical evidence may be biased over time if completion and publication of randomized efficacy trials are delayed when results are not statistically significant.

Objective.— To evaluate whether the time to completion and the time to publication of randomized phase 2 and phase 3 trials are affected by the statistical significance of results and to describe the natural history of such trials.

Design.— Prospective cohort of randomized efficacy trials conducted by 2 trialist groups from 1986 to 1996.

Setting.— Multicenter trial groups in human immunodeficiency virus infection sponsored by the National Institutes of Health.

Patients.— A total of 109 efficacy trials (total enrollment, 43708 patients).

Main Outcome Measures.— Time from start of enrollment to completion of follow-up and time from completion of follow-up to peer-reviewed publication assessed with survival analysis.

Results.— The median time from start of enrollment to publication was 5.5 years and was substantially longer for negative trials than for results favoring an experimental arm (6.5 vs 4.3 years, respectively; P<.001; hazard ratio for time to publication for positive vs negative trials, 3.7; 95% confidence interval [CI], 1.8-7.7). This difference was mostly attributable to differences in the time from completion to publication (median, 3.0 vs 1.7 years for negative vs positive trials; P<.001). On average, trials with significant results favoring any arm completed follow-up slightly earlier than trials with nonsignificant results (median, 2.3 vs 2.5 years; P=.045), but long-protracted trials often had low event rates and failed to reach statistical significance, while trials that were terminated early had significant results. Positive trials were submitted for publication significantly more rapidly after completion than were negative trials (median, 1.0 vs 1.6 years; P=.001) and were published more rapidly after submission (median, 0.8 vs 1.1 years; P=.04).

Conclusion.— Among randomized efficacy trials, there is a time lag in the publication of negative findings that occurs mostly after the completion of the trial follow-up.

SEVERAL INVESTIGATORS have raised concerns that clinical studies with negative results may never be published and their failure to appear in the literature may distort the picture we obtain from clinical experiments about the optimal practice of medicine.1-5 However, ascertaining the extent of this bias is difficult. Retrieving information about lost studies is a challenge.3-6 Prior research has been based on retrospective interviews about the fate of research protocols located through questionnaires,6 meeting abstracts,7 or archives of institutional boards or funding organizations.1,2,5 Prospective evaluation of the phenomenon with detailed trial registries8 gathering information on all the implementation milestones of randomized trials (onset, completion of enrollment, completion of follow-up, submission, and publication) has not been accomplished. Also most prior investigations have applied the term publication bias indistinguishably to phase 1, 2, and 3 trials and to both randomized and nonrandomized studies. However, the loss of information from a small, unrevealing pilot or observational study is not comparable to the disappearance of the results of randomized efficacy trials, which form the mainstream of evidence for medical practice. Such loss of information could affect systematic reviews, the decisions of funding agencies, and the outcomes of patients.

From retrospective investigations, it remains unclear whether publication bias affects specifically phase 2 and 3 trials to a substantial extent.1,4 Moreover, since many efficacy trials perform interim analyses during their conduct and may be prematurely interrupted if significant results are seen, it is unknown whether the results of a study would affect not only the time to publication after completion, but also the time a trial takes to be completed. Until now it has not been possible to address this last question with retrospectively constructed databases. Knowledge of the exact natural history of efficacy trials is needed to provide an accurate perspective on how medical evidence is obtained and whether there is a time lag between studies with "positive" and "negative" results.

The best insight into the fate of clinical trials can be gained from trial databases that collect prospectively information about the conduct of all studies in a given domain. Such an approach can offer a perspective on the natural history of clinical trials from inception through completion to publication. In this article detailed information from a large database of clinical trial protocols was used with data on the implementation milestones and publication dates of registered trials. This allowed assessment of the natural history of randomized efficacy trials in the domain of human immunodeficiency virus (HIV) infection and its complications, a discipline of rapidly expanding therapeutics with intense clinical research activity.

Methods

All efficacy clinical trials conducted from 1986 until 1996 by the AIDS [acquired immunodeficiency syndrome] Clinical Trials Group (ACTG) and by the Terry Beirn Community Programs for Clinical Research on AIDS (CPCRA) were considered in the analysis. These trial groups sponsored by the Division of AIDS of the National Institute of Allergy and Infectious Diseases (NIAID) represent globally the largest networks for the conduct of clinical trials on HIV infection and its complications. ACTG uses the resources of 30 university sites across the United States as well as other collaborative clinical units. The CPCRA is a community-based program encompassing more than 160 clinical practices across the United States. All protocols and detailed dated information on their implementation and presentation are archived by the Division of AIDS. Supplemental information about recently analyzed trials and clarifications on unclear or missing data were obtained from investigators and medical officers and staff responsible for the protocols.

The analysis was limited to randomized trials addressing the efficacy of the compared interventions (phase 2 and phase 3 trials). Observational, nonrandomized, pharmacokinetic, and safety phase 1 and phase 1/2 studies were excluded as well as substudies of the main protocols. Qualification for inclusion and recording of study design factors (including target sample size, end points, designation of phase, blinding, and data management) was based on examination of the complete protocols archived in the Division of AIDS, which offered a distinct advantage to avoid bias in recording as compared with interviews, questionnaires, and data extraction from abstracts that have been used in previous investigations of publication bias.1-7 Trials were selected regardless of whether they compared a regimen with placebo, different regimens, or different doses of the same medication. All protocols that enrolled any patients have been registered. The considered implementation milestones included the dates of starting enrollment, completion of follow-up, submission for peer-reviewed publication, and publication. For studies that continued follow-up beyond their primary analysis and publication, follow-up was censored at the time of the primary analysis. For 9 early studies, exact dates for completion of follow-up were not available but could be approximated with high accuracy from data on closure to enrollment and trial follow-up. All data were censored for analysis on October 10, 1996.

In this article, a trial is called "positive" if a statistically significant finding (denoted by P<.05) had been found in the analysis of the data for a main efficacy end point defined in the protocol in favor of an experimental therapy arm. Trials with nonstatistically significant findings or favoring the control arm are called "negative." Whenever there was no distinct control (traditional therapy) arm, a study was called positive if it showed a statistically significant finding in favor of any arm. When multiple major efficacy end points were available, the trial was considered positive if any major efficacy end point reached statistical significance. In 2 trials, significance favored different arms for survival and another end point; these trials were classified according to the direction of the survival results. The availability of the complete archived protocols and trial reports minimized the chances for subjective interpretation of end points and trial results. For further quality assurance, a random selection of 12 protocols and trial reports was evaluated by another colleague who was blinded to the study milestones. There was agreement in determination of pertinent end points and classification of significance between the author and the second independent observer in all cases.

Time-to-event analyses were performed with the Kaplan-Meier method, and comparisons used the log-rank test. The significance levels of the findings and other trial characteristics were used as covariates for the risk of publication in Cox proportional hazards regressions. Trial characteristics included the actual sample size, the ratio of accrual compared with originally anticipated (target) enrollment (typically based on power calculations), the trialist group, the age of the population (adult or pediatric), the trial domain (antiretroviral therapy vs complications of HIV), the presence or not of double blinding, and the place where data were managed (pharmaceutical industry or other). Both univariate and multivariate models and interactions between variables were assessed, but only univariate regressions are reported since the results of the multivariate regressions were similar and no significant interactions were identified. Statistical analyses were run on SPSS software, version 6.0 (SPSS, Inc, Chicago, Ill). All reported P values are 2 tailed.

Results
Characteristics of Registered Trials

A total of 109 randomized efficacy trials with total enrollment of 43708 patients qualified for the analysis. Of these, 8 were closed, having failed to accrue more than 20 patients, and are excluded from all subsequent analyses. Typically the conduct of these 8 trials became unfeasible or futile, and no publishable evidence materialized. Of the remaining 101 trials, 25 were still open to accrual, 10 were closed to accrual and continuing follow-up, and 66 trials with 30715 patients had been completed; 36 of the 66 completed trials had been published (18 at the New England Journal of Medicine, 6 in the Annals of Internal Medicine, 5 in the Journal of Infectious Diseases, and 1 each in 7 other journals) at the time data were censored for analysis. Of the 30 completed but unpublished trials, 9 had been submitted at least once for publication, and 8 more had been completed less than 1 year ago. Characteristics of the trials are shown in Table 1.

Survival Analysis for Time to Completion and Time to Publication

In a Kaplan-Meier analysis, the median time to publication among the 101 analyzed trials was estimated to be 5.5 years from the time a trial started enrollment (interquartile range, 3.9-7.0 years). On average, the time it took to conduct a study was of similar magnitude as the time it took to publish the results after its completion. The median time from starting until completing follow-up was 2.6 years (interquartile range, 2.0-3.8 years). Among the 66 completed trials, the median time from completion of follow-up to publication was 2.4 years (interquartile range, 1.6-3.8 years).

Overall, positive trials were published significantly earlier than trials with negative findings (Figure 1, A; log-rank P<.001). The median time from starting enrollment to publication was 4.3 years for positive trials vs 6.5 years for negative trials (mean, 4.2 vs 6.4 years). As shown in Figure 1, B, this time lag was largely attributable to differences between positive and negative trials in the time from completion of follow-up to publication (median, 1.7 vs 3.0 years; mean, 1.8 vs 3.6 years; log-rank P <.001).

Conversely, as shown in Figure 1, C, positive and negative trials differed little in the time they took to complete their follow-up (log-rank P=.17). However, when all trials with statistically significant results were considered regardless of whether the experimental or control arm was favored, these trials completed follow-up slightly faster than trials with nonsignificant results (Figure 1, D; log-rank P=.045). The absolute difference was not large (median, 2.3 vs 2.5 years; mean, 2.3 vs 2.8 years). To avoid the possible bias that trials still continuing follow-up at the censoring date may be more likely to be negative, a separate analysis included only the 50 trials that had started before June 30, 1992, and had all been completed by the time of analysis. Trials with significant results were completed slightly earlier (median, 2.3 vs 2.7 years; mean, 2.4 vs 3.0 years; log-rank P =.02).

The time-to-completion difference was probably important mostly for some early interrupted trials and some trials that had protracted enrollment and follow-up. Ten trials completed follow-up within less than 18 months from starting: 4 ran their prespecified course within this period of time; 1 was interrupted prematurely because of the surfacing of the significant results of a similar trial; and 5 trials were stopped early because of significant differences in either survival or clinical outcomes (n=4) or both efficacy and toxicity (n=1). On the other end, 12 trials had taken more than 4 years to complete follow-up: 2 were still trying to accrue patients, and another 9 were protracted because of lower than anticipated event rates in interim analyses despite full accrual; only 1 trial had an anticipated event rate.

Of the 8 trials not published at 6 years after starting enrollment, 5 had a prolonged enrollment and follow-up period (3.8-6.5 years) because of lower than anticipated event rates. Seven of the 8 trials had negative results. In contrast, the 5 trials published within less than 3 years from starting enrollment were all interrupted prematurely on the basis of early differences in efficacy between the arms in interim analyses (2 favoring the defined experimental arm, 2 favoring 1 of the arms, 1 favoring the defined control arm) and were all subsequently published in prestigious journals (New England Journal of Medicine, n=4; Annals of Internal Medicine, n=1).

Other Determinants of Time to Completion and Publication

Table 2 shows that significance of results was the only major determinant of the time from starting enrollment to publication for a clinical trial. The rate of publication was 3.7 (95% confidence interval [CI], 1.8-7.7) times higher for positive trials compared with negative ones, and the difference was explained again mostly by differences in the rapidity of publication after trials had been completed. The magnitude of the effect was unchanged in multivariate analyses adjusting for other trial characteristics (odds ratio, 4.9 [95% CI, 2.1-11.0]). Interestingly, large trials with more than 1000 patients took the same time to be published (if not less) than smaller trials. Large trials took probably a longer time to complete than smaller trials (P=.12), but their time to publication after completion was significantly shorter (P=.02). Using study accrual as a continuous variable, the rate at which a trial was completed decreased 1.8-fold (95% CI, 1.0-3.3) per 1000 patients, but the rate at which a trial was subsequently published after completion increased 2.5-fold (95% CI, 1.3-4.5) per each 1000 patients. Another interesting feature is that trials with data management performed by the pharmaceutical industry were of shorter duration (P<.001), but this did not expedite their overall time to appearance in the peer-reviewed literature (P=.33). There was no significant correlation between the presence of statistical significance, study accrual, and whether the data were managed by the industry or not (all correlation coefficients <0.2 in absolute value), but it should be remembered that trials were still monitored by NIH.

Underpowered trials that accrued less than half of their required prespecified target sample size (which was typically based on a priori power calculations) were published as fast as trials that had reached closer to their target sample size, although a moderate difference may have been missed in this analysis (odds ratio, 1.2 [95% CI, 0.4-3.5]). Among the 43 completed trials that had enrolled at least 90% of their prespecified target sample size, positive trials were still published substantially more rapidly than negative ones (log-rank P <.001).

Components of the Time From Completion to Publication

Among the 66 completed trials, 45 had been submitted for publication by the time data were censored. The median time to first submission was 1.4 years after completion (interquartile range, 0.7-2.3 years), and the median time to publication was 0.8 year after submission (interquartile range, 0.6-1.4 years). Positive trials were submitted significantly more rapidly compared with trials with negative results (mean, 1.0 vs 2.4 years; median, 1.0 vs 1.6 years; log-rank P =.001; Figure 2, A). A similar time lag was observed when trials with significant results favoring any arm were compared with trials with nonsignificant results (mean, 1.3 vs 2.4 years; median, 1.0 vs 1.6 years; log-rank P =.008). Data were managed by the industry in 2 of the 3 negative studies that had not yet been submitted despite a lapse of more than 3.5 years after their completion. After submission, positive trials were also published more rapidly than negative trials, but the time lag was relatively smaller (for both mean and median, 0.8 vs 1.1 years; log-rank P=.04; Figure 2, B).

Of the 45 submitted trials, 17 were rejected by at least 1 journal. At least 4 negative trials with over 300 patients each were rejected 2 or 3 times, while no positive trial was rejected multiple times. There was a trend that positive results might increase the odds of acceptance on the first submission (odds ratio, 1.6 [95% CI, 0.5-5.6]), but this was far from being formally significant (P=.54 by the Fisher exact test).

Among the 36 published trials, there was a moderately strong correlation between the time from completion to submission and the time from submission to publication (r=0.50; P =.002). These 2 time intervals were not significantly different across the published trials (mean difference, 0.2 year; P=.09 by t test).

Comment

This analysis shows that, even within multicenter trial groups of high efficiency, randomized efficacy trials are published more rapidly when results reach traditional levels of statistical significance. Negative studies suffer a substantial time lag. With some exceptions, most of this lag is generated after a trial has been completed. On the average, the time it takes to publish the results of efficacy trials, positive or negative, is of the same magnitude as the time it takes to actually conduct them.

This analysis extends the observations of earlier retrospective investigations of publication bias that claimed that investigators were not interested in the publication of small, negative trials.1,3,4,6 Earlier investigations have not agreed on the extent of publication bias for randomized trials,1,2 and the distinction between efficacy trials and early phase 1 and 1/2 trials could not be made clearly with retrospective approaches. For efficacy trials, publication lag may be a more exact term than publication bias, and this is most accurately described with a survival analysis approach. The effect of this publication lag can still be important, however. Trial results may be outdated when published belatedly. Rapid advances in technology and changes in the patient populations9 and clinical practice decrease the importance of the conveyed information. This is particularly true in rapidly evolving fields such as HIV therapeutics, where even significant findings become rapidly outdated. Publication lag drastically reduces the value of the provided evidence and introduces bias over time since the publication of positive results tends to antedate the surfacing of negative findings.

Besides proper peer-reviewed publication, information is disseminated by meetings, preliminary non–peer-reviewed presentations, and pharmaceutical advertisement. The extent of bias in non–peer-reviewed mechanisms of disseminating information is unknown, but unavoidably advertisement is fed preferentially by positive results,10 and prior investigators have suggested that meetings favor positive results.11 Beyond doubt, medical advances should draw prompt, strong attention. However, there is a concern that, in the rapidly changing face of medicine, positive trials may one-sidedly dominate our information system, as their negative counterparts take longer to appear, typically with less or no advertising glare around them. There are several examples where, on the same topic, positive trials caused early excitement, followed later by disappointing or less promising counterparts. Typical examples in HIV disease (Table 3) include the use of early zidovudine monotherapy in asymptomatic patients,12-14 acyclovir,15 ditiocarb (Imuthiol),16,17 and oral ganciclovir prophylaxis,18 where positive and negative trials started at about the same time, but negative studies appeared later14,17 or are still unpublished. Thus, a wave of enthusiasm is sometimes followed by a wave of disappointment or skepticism. Early systematic reviews of the accumulating evidence may give misleading results. An investigation in HIV-related trials, in particular, has shown that meta-analyses including only the published evidence would have found sizable treatment benefits for several controversial or even abandoned treatments,19 while in the case of zidovudine, a meta-analysis of the early published, short-term trials would suggest markedly more favorable results compared with the pooled treatment effect of the trials with longer follow-up that appeared later.20

Many phase 3 trials use sequential preplanned interim analyses to ensure that early differences between arms are not missed,21,22 and thus, positive trials may be interrupted prematurely while negative trials may be further protracted. This analysis shows that, with some exceptions, this approach does not have a major direct impact on the timing of most clinical trials. Exceptions do exist, mostly for trials at the 2 extremes: all the efficacy trials published within less than 3 years from opening to enrollment had been prematurely interrupted on the basis of interim analyses, while prolonged enrollment and follow-up, often because of slowly accrued events, was common in trials languishing for more than 6 years.

Publication delays for negative trials occur largely after the trial completion. Delays can be substantial even in multicenter trial groups with strict publication policies and high efficiency. As part of the ACTG and CPCRA standards of practice, strong incentives to publication are offered, including credit in the competitive renewal of funding for the site of the principal investigator. Similarly, disincentives to negligent investigators include potential dismissal from authorship and exclusion from chairmanship of future protocols. It is possible that publication lag may be more pronounced for trials conducted by groups not dependent on federal funding and for trials where the sponsor has a financial interest in the outcome.23,24 Future analyses should try to address prospectively whether the natural history of efficacy trials is different under these less ideal circumstances.

The observed average delay for negative findings after submission to peer review seemed small in this analysis, but it should be acknowledged that it could be an underestimate. Among published trials, the time to submit correlated fairly strongly with the time to be published. It is unknown whether negative trials whose submission had been deferred by the censoring date in this analysis might have similarly been deferred publication after an eventual submission. Also, multiple rejections seemed to cluster among negative trials.

Three more points warrant discussion. First, this analysis shows that, under the current circumstances, evidence from large trials becomes available in the peer-reviewed literature as fast as evidence from smaller trials. Large trials take longer to complete, but then they surface much more rapidly. The relative merits of small vs large trials have often been debated.25 With limited resources and rapid changes in medicine, large trials with prolonged follow-up are a challenge to design and conduct in many medical disciplines, including HIV disease. However, if small trials are to offer an advantage over larger trials, their results should be submitted to peer review and published promptly. Otherwise, a few large trials may be preferable to small trials presented in non–peer-reviewed sources and potentially affected by publication lag and bias.

Second, publication lag and bias may increase under circumstances where positive results accumulate rapidly, leaving even less interest for studies with inconclusive or less spectacular findings. Currently, with a geometric increase in the available therapeutic options in many medical disciplines, including HIV infection, it is possible that the light of publicity may fall even earlier and more heavily on the most positive results, probably even for trials considering the same or comparable regimens. The extent of publication lag should be addressed with similar methods in other domains besides HIV infection to examine its extent in less rapidly evolving specialties.

Third, publication lag may affect evidence-based medicine and systematic reviews and may lead to spuriously larger treatment effects in early meta-analyses of the available evidence. Meta-analyses should be aware of ongoing studies that may influence their conclusions. Late-appearing trials are more likely to show nonsignificant results. Lack of significance could be attributable either to smaller treatment effect estimates or to lower event rates resulting in wider CIs. In this registry, delay in completion and publication because of low event rates and slow accrual of events was a common occurrence. Differences between early- and late-appearing trials could reflect heterogeneity in the trial protocols and the disease risk of the studied patient populations.26 Cumulative meta-analysis may be used to evaluate systematically the extent of the impact of late negative trials,27 and reasons for discrepancies between early meta-analyses and late-appearing trials should be studied systematically.28

Enthusiasm about the results of clinical research should not be based on its P value.29 Clinical trials and meta-analyses provide a continuum of evidence and nondefinitive trials may provide as important information as trials with high levels of statistical significance when seen in the context of other pieces of evidence. It is still very questionable what really optimizes clinical care, but the results of properly conducted trials are definitely a cornerstone in the process. Equipoise in the presentation of all results is important for an objective evaluation of the accumulated evidence.

References
1.
Easterbrook P, Berlin JA, Gopalan R, Matthews DR. Publication bias in clinical research.  Lancet.1991;337:867-872.Google Scholar
2.
Dickersin K, Min YI. NIH clinical trials and publication bias.  Online J Curr Clin Trials.April 28, 1993;doc 50.Google Scholar
3.
Dickersin K. The existence of publication bias and risk factors for its occurrence.  JAMA.1990;263:1385-1389.Google Scholar
4.
Dickersin K, Min Y-I. Publication bias: the problem that won't go away.  Ann N Y Acad Sci.1993;703:135-148.Google Scholar
5.
Dickersin K, Min Y-I, Meinert CL. Factors influencing publication of research results: follow-up of applications submitted to two institutional review boards.  JAMA.1992;267:374-378.Google Scholar
6.
Hetherington J, Dickersin K, Chalmers I, Meinert C. Retrospective and prospective identification of unpublished controlled trials: lessons from a survey of obstetricians and pediatricians.  Pediatrics.1989;84:374-380.Google Scholar
7.
Scherer RW, Dickersin K, Langenberg P. Full publication of results initially presented as abstracts: a meta-analysis.  JAMA.1994;272:158-162.Google Scholar
8.
Simes RJ. Publication bias: the case for an international registry of clinical trials.  J Clin Oncol.1986;4:1529-1541.Google Scholar
9.
Ioannidis JPA, Cappelleri JC, Schmid CH, Lau J. Impact of epidemic and individual heterogeneity on the population distribution of disease progression rates: an example from populations of trials of human immunodeficiency virus infection.  Am J Epidemiol.1996;144:1074-1085.Google Scholar
10.
Koren G. Bias against negative studies in newspaper reports of medical research.  JAMA.1991;266:1824-1826.Google Scholar
11.
Koren G, Graham K, Shear H, Einarson T. Bias against the null hypothesis: the reproductive hazards of cocaine.  Lancet.1989;2:1440-1442.Google Scholar
12.
Cooper DA, Gatell JM, Kroon SA.  et al.  Zidovudine in persons with asymptomatic HIV infection and CD4+ cell counts greater than 400 per cubic millimeter.  N Engl J Med.1993;329:297-303.Google Scholar
13.
Volberding PA, Lagakos SW, Koch MA.  et al.  Zidovudine in asymptomatic human immunodeficiency virus infection: a controlled trial in persons with fewer than 500 CD4-positive cells per cubic millimeter.  N Engl J Med.1990;322:941-949.Google Scholar
14.
Concorde Coordinating Committee.  Concorde: MRC/ANRS randomised double-blind controlled trial of immediate and deferred zidovudine in symptom-free HIV infection.  Lancet.1994;343:871-881.Google Scholar
15.
Youle MS, Gazzard BG, Johnson MA.  et al.  Effects of high-dose oral acyclovir on herpesvirus disease and survival in patients with advanced HIV disease: a double-blind, placebo controlled study.  AIDS.1994;8:641-649.Google Scholar
16.
Hersh EM, Brewton G, Abrams D.  et al.  Ditiocarb sodium (diethyldithiocarbamate) therapy in patients with symptomatic HIV infection and AIDS: a randomised, double-blind, placebo-controlled multicenter study.  JAMA.1991;265:1538-1544.Google Scholar
17.
The HIV87 Study Group.  Multicenter, randomized, placebo-controlled study of ditiocarb (Imuthiol) in human immunodeficiency virus-infected asymptomatic and minimally symptomatic patients.  AIDS Res Hum Retroviruses.1993;9:83-89.Google Scholar
18.
Spector SA, McKinley GF, Lalezari JP.  et al.  Oral ganciclovir for the prevention of cytomegalovirus disease in persons with AIDS.  N Engl J Med.1996;334:1491-1497.Google Scholar
19.
Ioannidis JPA, Cappelleri JC, Sacks HS, Lau J. The relationship between study design, results and reporting of randomized clinical trials of HIV infection.  Control Clin Trials.1997;18:431-444.Google Scholar
20.
Ioannidis JPA, Cappelleri JC, Lau J.  et al.  Early or deferred zidovudine therapy in HIV-infected patients without an AIDS-defining illness: a meta-analysis.  Ann Intern Med.1995;122:856-866.Google Scholar
21.
DeMets DL, Fleming TR, Whitley RJ.  et al.  The Data and Safety Monitoring Board and acquired immune deficiency syndrome trials.  Control Clin Trials.1995;16:408-421.Google Scholar
22.
O'Brien PC, Fleming TR. A multiple testing procedure for clinical trials.  Biometrics.1979;35:549-556.Google Scholar
23.
Blumenthal D, Campbell EG, Anderson MS, Causino N, Louis KS. Withholding research results in academic life sciences: evidence from a national survey of faculty.  JAMA.1997;277:1224-1228.Google Scholar
24.
Rennie D. Thyroid storm.  JAMA.1997;277:1238-1243. [published correction appears in JAMA1997;277;1762].Google Scholar
25.
Cappelleri JC, Ioannidis JPA, Schmid CH.  et al.  Large trials vs meta-analysis of smaller trials: how do their results compare?  JAMA.1996;276:1332-1338.Google Scholar
26.
Ioannidis JP, Lau J. The impact of high risk patients on the results of clinical trials.  J Clin Epidemiol.1997;50:1089-1098.Google Scholar
27.
Lau J, Antman EM, Jimenez-Silva J, Kupelnick B, Mosteller F, Chalmers TC. Cumulative meta-analysis of therapeutic trials for myocardial infarction.  N Engl J Med.1992;327:248-254.Google Scholar
28.
McIntosh M. The population risk as an explanatory variable in research synthesis of clinical trials.  Stat Med.1996;15:1713-1728.Google Scholar
29.
Ioannidis JP, Lau J. On meta-analyses of meta-analyses.  Lancet.1996;348:756.Google Scholar
×