A scientific debate about the effectiveness of second-generation antipsychotics, compared to conventional antipsychotics, has been going on for several years. Although all questions have not as yet been answered, second-generation antipsychotics are now defined as the gold standard in most aspects of treatment, at least in highly industrialized countries. As a result, so-called head-to-head comparisons, i.e., randomized, controlled clinical trials with two or more active second-generation antipsychotic comparators, have become increasingly important as new drugs enter the market.
Somewhat confusing is the fact that different trials comparing the same two drugs have had contradictory conclusions (1, 2). This effect may not be totally unrelated to the funding sources of the trials. Conflicts of interest arising from a pharmaceutical company’s sponsorship of clinical trials of a drug it manufactures are obvious (3), and the association of funding and conclusions is found in numerous medical specialties (4). In this article, we present a summary of head-to-head comparison studies in psychiatry in which we focus on various aspects of potential bias that may arise from such conflicts of interest. To our knowledge, this work is the first examination of potential bias related to study sponsorship of head-to-head comparison studies of antipsychotic medications. We also examined the association of the conclusions of head-to-head comparison studies with the source of funding. Consequently this study is not a review or a meta-analysis in which the efficacy or tolerability of different second-generation antipsychotics is examined but an exploratory approach to clarifying partly contradictory study results in the field of schizophrenia treatment.
We searched MEDLINE (1966–September 2003) for randomized, controlled trials comparing the second-generation antipsychotics aripiprazole, amisulpride, clozapine, olanzapine, quetiapine, risperidone, sertindole, and ziprasidone. The search terms were paired combinations of the second-generation antipsychotics and the term "rand*" (for "random," "randomized," etc.). We excluded reviews, meta-analyses, reports focused solely on laboratory or electrophysiological data, trials with combined drug treatment, and reports on patient populations with diagnoses other than schizophrenia or schizoaffective disorder. Reports on drug efficacy were considered to be the primary publication of a trial, unless the abstract stated otherwise. Secondary publications were excluded in order to avoid multiple inclusions of the source trial in the analysis. We also screened proceedings of selected conferences for the period from 1999 to February 2004. The conference reports we reviewed were limited to materials from events attended by members of our work group.
Blinded Rating of Abstracts
On the basis of the hypothesis that funding by a pharmaceutical company may influence the outcome of a trial, we checked the reports for information on sponsorship by a "profit-making organization." The abstract of each study was modified to mask the names and doses of the drugs used in the trial, and two physicians (a psychiatrist [K.M.] and an internist [E.J.]), both of whom were blinded to the funding source for the trial and were not involved in the design of the evaluation, read the complete abstract and rated which drug was favored in the overall conclusion. The ratings were made on a 6-point scale proposed by Gilbert et al. (5) and previously used in studies evaluating the association of funding and conclusions in drug trials (4, 6). The scoring method is described in the footnote to F1. For blinding, the second-generation antipsychotic names in the abstracts were replaced by "DRUG A" and "DRUG B" ("DRUG A" was not always the sponsor’s drug and vice versa), and the total dose/dose range was replaced by "x." A separate sensitivity analysis that included only peer-reviewed publications was carried out. Two-sided binomial sign tests were used to test the hypothesis of potential influence of the sponsor on the study outcome, and Cohen’s kappa was used for measuring interrater reliability. Statistical significance was defined at an alpha level of <0.05.
Identifying Potential Sources of Bias
The trial reports were read independently by two authors who were not blinded to the sponsor of the trial (S.H., S.L.) to identify potential sources of bias that could have influenced the results in favor of the sponsor’s drug. We focused on several factors that have been discussed as potential sources of bias, including features of study design, dose ranges, titration schedules, statistics, reporting of results, and wording of findings (4, 7, 8). If the conclusions of the two reviewers differed, consensus was achieved by discussion. The second author (J.D.) checked and approved the findings. As a reference for dose ranges, we used the following range recommendations included in the American Psychiatric Association Practice Guideline for the Treatment of Patients With Schizophrenia, second edition (9): 10–30 mg/day of aripiprazole, 150–600 mg/day of clozapine, 10–30 mg/day of olanzapine, 300–800 mg/day of quetiapine, 2–8 mg/day of risperidone, 120–200 mg/day of ziprasidone, and 5–20 mg/day of haloperidol. For amisulpride, we used the following dose ranges suggested in the drug company’s product information: 400–800 mg/day for acutely ill patients and 50–300 mg/day for patients with predominantly negative symptoms.
From 146 publications found in the MEDLINE search, we excluded 61 reviews, 22 reports of additional data from previously published trials or preliminary results, 17 reports of laboratory or electrophysiological data, five reports of add-on therapy with other drugs, four reports on alternative diagnoses, 11 reports of studies that did not include a direct head-to-head comparison, and one report on combined antipsychotic treatment, which left 25 publications for analysis. The complete trial report for one of the 25 publications could not be obtained, and that study was excluded. Thirteen conference presentations of head-to-head drug comparisons were identified, and during the analysis, another four publications and one report in press were identified, for a total of 42 trial reports. Of the 42 reports, 32 were fully or partly funded by a pharmaceutical company that manufactured one of the drugs used in the trial (1, 2, 10–39). One of the 42 studies was conducted with supplemental funding from a pharmaceutical company, although the acquisition and reporting of the data were implemented with no input from the company (40); this study was not included in the blinded rating of abstracts, but it was included in the analysis of sources of bias. Nine of the 42 studies were not funded by a pharmaceutical company (41–49). Two reports of sponsored studies did not include an abstract (10, 36). Thus, 30 trials were included in the blinded rating of study abstracts.
Sponsorship and Outcome as Reported in Study Abstracts
According to the ratings by the two physicians, the overall outcome reported in the study abstracts was in favor of the sponsor’s drug in 90.0% of the abstracts (N=27 of 30) (p<0.001, binomial sign test) (F1). For each abstract, the scores of the two raters were the same or differed by only 1 point, and the two raters did not differ in whether the outcome was judged to be in favor of the sponsor’s drug (a score of 4, 5, or 6) or the comparator (a score of 1, 2, or 3). According to the criteria of Landis and Koch (50), the interrater agreement was "moderate" (kappa=0.44, p≤0.001) for the numeric rating and "almost perfect" (kappa=1.0, p<0.001) for the outcome category. F1 shows the distribution of the scores for both raters. In the sensitivity analysis that included only the abstracts that underwent peer review (N=21), the result was virtually identical, with 90.5% (19 of 21) rated as having an outcome in favor of the sponsor’s drug (p<0.001, binomial sign test). The interrater agreement was "substantial" (kappa=0.61, p<0.001) for the numeric rating and "almost perfect" (kappa=1.0, p<0.001) for the outcome category. t1 summarizes the ratings for studies comparing pairs of drugs by whether one or the other manufacturer sponsored the study. Only three of these 21 reports did not favor the sponsor’s drug. These pairwise comparisons revealed contrasting outcomes, depending on the sponsor of the study, although the outcomes were derived from trials involving the same two drugs.
Possible Effects of Sponsorship on Trial Outcome and Reporting
Two authors who were not blinded to the sponsor of the trial reviewed the study reports and identified potential sources of bias in the following areas: dose and dose escalation, entry criteria and study population, statistics and methods, and reporting and wording of results. The characteristics of the individual trials and the potential sources of bias are summarized in a separate table (Data Supplement 1) available from the first author and available with the online version of this article at http://ajp.psychiatryonline.org. We identified potential sources of bias as debatable or clear. For example, in several instances, we identified debatable sources of bias in dose ranges for risperidone, for which the appropriate range may still be arguable. We identified clear sources of bias in instances involving obviously inappropriate choices of dose, design, reporting, etc. We emphasize that although at least some of the biases we identified seemed very obvious, our analysis remains speculative, and there is no proof that the factors we identified really influenced the results. The biases we identified are described in the following sections.
Doses and Dose Escalation
Dose ranges and dose escalation are crucial factors that potentially influence trial outcome. In numerous trials, dose ranges are scheduled according to the manufacturer’s package insert, which is problematic with antipsychotic drugs. For example, in trials with risperidone, doses up to 10 mg/day or even 12 mg/day are frequently possible in flexible titration schedules, although this dosing level may diminish both the efficacy and tolerability of the drug. After the introduction of risperidone to the market, several studies in the mid-1990s yielded evidence of an optimal dose range of 4–8 mg/day, with an increasing risk of extrapyramidal side effects at higher doses without any gain in efficacy (51, 52). At the time of the earliest studies included in this summary (1), these data were presumably not yet accessible, but in more recent trials, the dose ranges should have been adapted to maintain a fair level of comparison. Trials that did not include the 4-mg/day dose, recently referred to as the advisable dose (53), and trials that allowed doses of up to 12 mg/day (10, 12, 34, 37, 40) are problematic. Choosing 4 mg/day as the lower limit of the dose range is also problematic, as downward dose adjustment in case of side effects is not possible. Although a dose range of 2–6 mg/day was used in trials sponsored by the manufacturer (2, 18), and even lower doses were used in elderly patients with schizophrenia in trials sponsored by the manufacturer (19, 21), competitors consistently used higher doses.
Dose ranges are also problematic in comparisons involving other drugs. Dose ranges of clozapine, especially in trials that included patients with treatment-resistant schizophrenia, often appear to be too strictly limited (53), resulting in relatively low mean daily doses (<400 mg/day) (13, 14, 39). These levels are in contrast to data revealing that doses up to 600 mg/day (54) or even 900 mg/day (55, 56) of clozapine proved highly efficacious in treatment-resistant schizophrenia. In comparisons involving olanzapine, the upper limit of the dose range is often set at 15 mg/day (16, 20, 38), thus excluding the most effective 20-mg/day dose. Use of this limited dose range possibly reduces olanzapine’s efficacy and may result in a misleading conclusion of the competitor’s therapeutic superiority or equality. The optimum dose range of amisulpride in patients with predominantly negative symptoms ranges from 50 to 300 mg/day (57), but in a study comparing amisulpride with another antipsychotic, it should have been ensured that the patients did not have significant positive symptoms at study entry because higher amisulpride doses (400–800 mg/day) are necessary for treatment of positive symptoms (30).
Finding the optimum dose escalation schedules for both compounds in a study is difficult and may be another source of bias (2, 12, 16, 18–20, 24, 28, 34, 40, 58). In some cases, the bias may derive from the fact that titration is mandatory for some drugs (risperidone, clozapine, sertindole), while the comparator (for example, olanzapine) does not require a stepwise dose escalation. Slow titration can prolong the time to the full onset of the therapeutic effect of a drug, and the optimal dose of the comparator may be reached earlier. This difference plays a major role in studies evaluating efficacy over a brief period of time. On the other hand, side effects might be more likely to appear with fast dose escalation. The attempt to escape the escalation problem by using a fixed-dose regimen raises other problems. Studies with fixed-dose regimens lack naturalistic plausibility because the unrealistic limits imposed do not reflect the therapeutic flexibility required in the treatment of schizophrenia (16, 23, 32, 33, 44, 45).
Entry Criteria and Study Population
Because the second-generation antipsychotics became available on the market one by one over the last decade, a trial’s entry criteria with respect to previous drug treatment have to be chosen carefully. Risperidone had been in use for more than 5 years when newer drugs such as amisulpride (32, 37), quetiapine (24, 29), olanzapine (17), sertindole (11), and ziprasidone (10) became comparators in trials. Exclusion of patients who previously were nonresponders to risperidone or any other comparator (16) is seldom explicitly stated in reports of head-to-head trials, although this feature could have a critical effect on observations of the efficacy of or response to antipsychotic treatment.
For trials involving schizophrenic patients with predominantly negative symptoms, questions about the accurate definition of the study population may be raised. Even if appropriate scales for measuring negative symptoms, such as the negative syndrome subscale of the Positive and Negative Syndrome Scale (PANSS) or the Scale for the Assessment of Negative Symptoms, are applied, there is still the need for information on positive symptoms, as they might also be present at study entry. An entry criterion of a difference of 6 points between the PANSS negative and positive subscale scores may ensure that subjects have a predominance of negative symptoms, but it leaves room for speculation about the effect of positive symptoms if baseline information about positive symptoms is not presented (30). Correspondingly, in trials involving patients with treatment-resistant illness, transparent criteria for inclusion and exclusion of participants are also required (54), although no universally accepted definition of treatment-resistant schizophrenia exists (59). Studies in which antipsychotic treatment nonresponse and intolerance are allowed as alternative entry criteria (14) may have results that are difficult to interpret. If results derived from such studies are presented in terms of efficacy in treatment-resistant patients, even if the study is not explicitly focused on this population, misunderstandings are foreseeable (13).
In recent years, studies with a noninferiority design have become a reasonable alternative to placebo-controlled trials for comparison of the efficacy of antipsychotic agents (60). In a study designed to prove a drug’s superiority over an active comparator, large sample sizes are usually required. However, equivalence can be shown in a one-sided noninferiority design with less effort, depending on the predicted threshold for equivalence, although it is important to note that in a noninferiority design with a narrow range of equivalence, the sample size required may exceed that necessary for a superiority design. Consequently, a basic requirement is to define a priori the extent of the difference between the treatments that is considered acceptable for declaring noninferiority (61). It seems very arguable to assume an equivalent antipsychotic efficacy of a drug at a threshold of just over 60% of the treatment effect achieved by the active comparator as measured by the reduction in the PANSS total score (10) or the PANSS negative subscale score (30). Other equivalence thresholds yield findings of more clinical relevance, but the thresholds differ between comparable studies (28, 32, 37, 39).
For multiple comparisons, such as those that occur with the use of test batteries in cognition studies, an adjustment for multiple testing may be necessary, but no generally accepted approach toward this statistical problem exists. One work group may confuse the reader by applying an adjustment for multiple testing in one study (18, 20) and not in a comparable trial (19). In some studies, the application of an adjustment was not explicitly mentioned or adequately discussed, despite the presence of multiple comparisons (1, 16, 31, 62).
Another source of potential bias is a study design in which an acute-phase trial of up to 8 weeks is followed by a continuation phase of up to 12 months that is focused on long-term maintenance of the treatment effect. After the acute phase, patients who are nonresponders are discontinued from the study and only those who meet the response criteria are included in the maintenance phase (63). This design may be acceptable for relapse studies but leads to problems in response trials. Selecting only responders for continuation in a trial that is focused on response (as measured, for example, with the mean reduction of the PANSS score from baseline to endpoint) as well as further improvement alters the study population radically, necessitating careful interpretation of the results in the follow-up (10).
Reporting and Wording of Results
A complete disclosure of all results of the head-to-head comparison would appear to be mandatory but is not always provided. Results favoring the drug manufactured by the sponsor are often presented in detail, and unfavorable results often are mentioned in a brief sentence at the very end of the report’s results section or not mentioned at all (1, 12). Accordingly, the report’s authors may choose to present only data from observed cases or only data from a last-observation-carried-forward analysis, depending on the resulting outcomes. If the last-observation-carried-forward design showed no significant difference between drugs, the results from the observed cases may be displayed in detail and presented as a significant outcome of the study (11). The relevant population for evaluation of the primary outcome should be stated a priori in the protocol and made transparent to the reader.
Furthermore, reporting of adverse events seems to be selective (34, 36, 38, 62), and the corresponding level of significance for comparisons of rates of adverse events may not be consistently stated (21, 29). Information on side effects that are very likely to occur, such as sedation and weight gain with olanzapine (15, 64) or elevation of prolactin levels with amisulpride (28), may be lacking. In addition, in reports of extrapyramidal symptoms, detailed information on the mean daily dose of anticholinergic medication and the number of patients who received at least one dose of anticholinergic medication should be provided. If this information is omitted, the reported frequency of occurrence of extrapyramidal symptoms gives only a vague impression of the likelihood of these side effects (23, 28).
Poster Reports and Multiple Publishing
Phrasing of abstracts is difficult, because much information has to be made transparent to the reader in only a few lines. Although the abstracts of many head-to-head studies adhere to widely accepted structural standards (65), the results stated are often highly selective. For example, in the abstract of one study (29), a significant difference in rates of extrapyramidal symptoms that favored the sponsor’s drug is reported in detail, but the side effects unfavorable to the drug were mentioned without corresponding levels of significance.
Preliminary results of trials are often presented as poster reports at conferences. Presentation of multiple poster reports on the same trial with different first authors can lead to the impression that independent studies have been conducted (10, 66). If data from a previously published trial are later used as the basis for reports focusing on subpopulations or secondary objectives, the abstracts of the later studies should contain a cross-reference to disclose the source of the data at a glance (62–64, 67, 68). Stand-alone publication of data deriving from another trial without a reference to the earlier trial gives the impression that separate trials have been conducted (18, 19).
The first part of our analysis revealed a clear link between sponsorship and study outcome as reported in the abstract, as 90.0% of the abstracts were rated as showing an overall superiority of the sponsor’s drug. This finding is in accordance with numerous previous reports of a similar effect in other medical fields (3, 4, 6, 69). Even more striking were our findings for pair-wise comparison of different trials that examined the effects of the same two drugs (t1). We found that different comparisons of the same two antipsychotic drugs led to contradictory overall conclusions, depending on the sponsor of the study. On the basis of these contrasting findings in head-to-head trials, it appears that whichever company sponsors the trial produces the better antipsychotic drug. This peculiar result led us to take a closer look at various design and reporting features. Indeed, a number of potential reasons for the association between drug-company-sponsored trials and favorable results were identified.
Limitations to Our Approach
A first limitation is that we did not retrieve all trials that were presented at conferences. Because no databases for such presentations exist, we were limited to the posters from conferences attended by members of our work group. The conference presentations we included are therefore not necessarily representative of all conference publications. We did not, however, want to exclude this material completely, because conferences are an important way for companies to distribute information. We made no selection among the available reports. The main limitation of our exploratory analysis is that it must remain speculative by nature. Although in some cases—for example, the trial in which the optimal risperidone dose of 4 mg/day was explicitly excluded (10)—it is quite obvious that the factor we identified may have biased the results, there is no proof that it really did. Only a "remake" of the study factoring out the source of bias could test the impact. Furthermore, other readers may have different opinions, especially about the more subtle potential sources of bias. Finally, we emphasize that most of the identified factors were indeed rather subtle and did not reflect an attempt by the drug trial sponsors to intentionally misinterpret their findings or to willfully mislead readers.
Benefit From Industry-Sponsored Trials
In many respects the industry-sponsored studies included in our review met high methodological standards (26, 27) and often surpassed non-industry-sponsored trials in the quality of research methods (6, 70). Industry-independent studies are not necessarily free of bias and are often too underpowered to find statistically significant differences or to allow any generalization (46, 47, 71). In our review, the sample size per group of the nine studies not funded by a for-profit organization ranged from nine to 113 patients. Other factors that contribute to the excellent methodological standards of industry-sponsored trials are valid central randomization, the high quality of data acquisition and management, regular auditing processes, and the pharmaceutical company’s researchers’ detailed knowledge about the drug (6, 70). There is also no doubt that the development of the second-generation antipsychotics was a major step forward. For the first time antipsychotic drugs with clearly defined dose ranges were made available, while the optimum dose, even of the standard conventional antipsychotic haloperidol, is still in doubt. Industry-organized trials also markedly improved our knowledge about general clinical questions such as medication switching strategies (72), the treatment of patients with refractory disorders (34), and the overall effectiveness of new and conventional antipsychotics for treatment of negative symptoms (73). However, if all studies by drug companies report positive outcomes, the findings may lose credibility.
Suggestions for Potential Improvement
Given the unique opportunities of industry for organizing methodologically sound, large-scale trials, the association between outcome and sponsor found in the rating of abstracts in our study is unsatisfactory. We believe, however, that in the case of many of the problematic points raised in the Results section, relatively simple measures could improve the situation to an appreciable extent.
Sponsorship and outcome as reported in the abstract
Our results show that reading only the abstract of a study is insufficient for a complete understanding of the study findings. However, lack of time makes it difficult even for scientific experts to read all trial reports in detail. Therefore, peer reviewers of studies being considered for publication should pay close attention to the conclusions stated in study abstracts. Overall, we found that the structure of the abstracts in the current review adhered to widely accepted standards (65), but the selection of the results and the phrasing used to convey the results needed to be carefully scrutinized. To avoid bias in this crucial section of trial reporting, we suggest that peer reviewers verify whether the abstract really summarizes the overall results of the trial in a balanced way. Detailed guidelines in this area for peer reviewers would be useful.
In head-to-head trials, dose ranges and escalation schemes have a major effect on the outcome. To avoid potential bias, study initiators could ask the competitor to provide a suggested dose range and titration schedule for its compound, as the manufacturer of a drug knows its properties best. Alternatively, external experts could function as independent advisers, but they should then be named in the report as a source of information on the dosing regimen. In addition, responsible agencies such as the U.S. Food and Drug Administration (FDA) or the European Medicines Agency (EMEA) might be given the chance to look at the protocol before the study is begun in order to allow the correction of obvious flaws.
Entry criteria and study population
Regarding study population and inclusion criteria, study initiators should follow broadly accepted standards in the characterization of the eligible patients. Diagnostic validity is hardly ever mentioned in sponsored trials, and theoretically heterogeneous outcomes may be partly due to the heterogeneity of the study population. The use of structured clinical interviews may help identify the proper study population. For example, a characterization of patients with predominantly negative symptoms has been proposed (73). Defining a valid study population is essential in studies of patients with treatment-resistant illness that focus on the efficacy of antipsychotics, and other aspects of previous treatment discontinuation, such as medication intolerance, should not be used as alternative inclusion criteria. Otherwise it is unclear which aspect is related to the superiority of a compound (14).
A comprehensive assessment of the statistical methods applied in the studies we reviewed is beyond the scope of this article. We therefore comment only on two points that came up several times during our review. In the last 5 years, noninferiority designs have become more common, leading to a major problem with the threshold of equivalence (74). It is hardly acceptable to consider the lower margin of the 95% confidence interval at a level of only 60% of the efficacy of the competitor to be a sign of noninferiority. As the trend toward this type of statistical design is likely to endure, an expert consensus on methods for setting the thresholds is needed. Other confusing aspects include the use of various test methods and lack of the correction for multiple statistical tests in trials in which effects on cognitive function are examined. Recently, a guideline for standard test batteries for measuring cognition became available (75), and it could soon be followed by a consensus on the statistical methods that should be used in this field of research. In general, study initiators should define outcome parameters a priori and choose the appropriate correction method for multiple testing. If the correction method is applied to a subset of tests only, this fact should be explained.
Reporting and wording of results
Wording and phrasing of study results are surely the most debatable sources of bias. The CONSORT (consolidated standards of reporting trials) statement, developed in the mid-1990s, proposed a checklist to ensure completeness of reporting and assessment of the validity of trial results (76). In addition, the International Committee of Medical Journal Editors set up a list of uniform requirements for manuscripts, including trial registration and complete reporting of all acquired data (77). The recommendations leave a considerable margin for wording and interpretation of the findings. Therefore, it is again the responsibility of peer reviewers for scientific journals to demand balanced reporting of the results.
Readers of the trial reports should pay close attention to the choice of the primary outcome variables and to the presentation of the results in order to obtain a realistic impression of whether a new and unknown aspect of drug treatment, following the "uncertainty principle" (6), was observed or whether the study was designed to yield predictable results in favor of the sponsor’s drug. The uncertainty principle states that a patient should be enrolled in a randomized, controlled trial only if there is substantial uncertainty about which of the treatments would benefit the patient most. For example, the appropriateness of a trial focused on weight gain is debatable if a sponsor’s drug that is already known for its minor impact on weight is compared to a treatment previously shown to be highly likely to cause weight gain.
The observation that only studies with significant findings tend to be published led Melander et al. (78) to coin the phrase "evidence b(i)ased medicine." It is noteworthy that a guideline for "good publication practice" has been proposed to help avoid further publication bias (79). Each protocol registered with the European Clinical Trial Database is issued a unique number, making trials traceable and missing reports conspicuous. Unfortunately, access to this information is limited to the study initiator and EMEA staff. The international Current Controlled Trials meta-register (www.controlled-trials.com) combines national as well as disease-specific registers, and each trial included in the register is assigned a specific number. The U.S. Freedom of Information Act mandates publicly accessible "electronic reading rooms" for materials available through the Freedom of Information Act, such as, for example, information about studies registered with the FDA. However, in our experience, the registers are not easy to browse.
Poster reports and multiple publishing
Publication of findings on different aspects of the same trial in several reports has been criticized as the "salami strategy" of scientific reporting. This criticism may not always be justified, because it is simply not feasible to report in one publication all the data from a large trial with several aspects of interest or a huge sample size. Readers’ understanding of the different aspects covered by the study can be enhanced if the masses of data are split into several reports. However, authors should always clearly state the source reference of the data that are presented (78). Otherwise, the reader might get the impression that several trials were undertaken, although in fact there was only one. A similar problem occurs if different researchers from the same trial are listed as the first author of various conference presentations or publications by the work group. Because many scientists have only limited time and choose the abstract as the primary information source, the underlying core study should always be mentioned in the abstract. Moreover, data presented exclusively in conference poster sessions or symposia, which normally do not undergo peer review, must be considered problematic (70).
Is It All a Matter of Sponsoring?
The need for more industry-independent studies has been recognized, and some have already been conducted and published (80). Although reports from industry-independent trials may not include biased reporting and wording, specific design features such as dose ranges and study populations can still remain problematic. For example, the design of a recent industry-independent study of Alzheimer’s disease patients (81) has been criticized (82, 83). The treatment of schizophrenia has many different aspects, and numerous studies will be needed to advance treatment. It is unlikely that public funding will cover them all. We therefore believe that the chance for further improvement of current industry-supported trials should not be passed up.
Received Nov. 22, 2004; revisions received April 28 and May 31, 2005; accepted Sept. 19, 2005. From Klinik und Poliklinik für Psychiatrie und Psychotherapie der Technischen Universität München am Klinikum rechts der Isar; the Department of Psychiatry, University of Illinois at Chicago; the Department of Psychiatry, Ludwig-Maximilian Universität, München, Germany; and the Department of Internal Medicine, Technische Universität München, Germany. Address correspondence and reprint requests to Dr. Heres, Klinik und Poliklinik für Psychiatrie und Psychotherapie der Technischen Universität München am Klinikum rechts der Isar, Moehlstrasse 26, 81675 München, Germany; email@example.com (e-mail).No source of funding or any grant was used to finance this study.Additional information on this study accompanies the online version of the article.
Scores Assigned by Two Physicians in Blind Ratings of 30 Abstracts Reporting the Outcome of Head-to-Head Comparison Studies of Second-Generation Antipsychoticsa
aThe abstracts were modified for blind ratings by replacing the drug names with "DRUG A" and "DRUG B" and replacing the total dose/dose range by "x." Scores from 4 to 6 favored the sponsor’s drug, and scores from 1 to 3 favored the comparator. Scores were assigned according to the following scale: 1=DRUG B is highly preferred and is the best alternative; should be considered the standard intervention in all patients, or the like; 2=DRUG B preferred to DRUG A, but DRUG A might be promising under certain circumstances or the like; 3=DRUG A and DRUG B about equal, but DRUG A is disappointing, as DRUG B had some minor advantages; 4=DRUG A and DRUG B about equal, but DRUG A is successful because of minor advantages; 5=DRUG A preferred to DRUG B, but further trials still indicated; may be more costly or similar disclaimer; 6=DRUG A highly preferred and should be considered the standard intervention for all patients, or the like. In this example, "DRUG A" designates the study sponsor’s drug and "DRUG B" designates the comparator, although in the actual abstracts modified for blind ratings, "DRUG A" was not always used to designate the sponsor’s drug and vice versa.