Скачать книгу

interventions appear to have a larger effect size than they do in reality, and will sometimes make the effect seem smaller than it really is [79]. Inflated effect sizes are more likely to be significant and reduced ones less so. The result is that small trials have much more heterogeneous effect sizes than large ones [88].

      To ensure that the trial has adequate power, researchers should carry out a formal sample size calculation (specifying the likely size of the treatment effect as well as the required power and an estimate of variance). The frequency of reporting sample size calculations is often low. For example, only 41% of low back pain trials [89] and 35% neurosurgical trials [90] reported this calculation. Even when the sample size calculation is reported, researchers often overestimate the possible benefit of the treatment and end up with sample sizes that are too small to detect a clinically realistic effect [89, 91].

      Industry‐Funded Trials

      In general drug company studies are at not at a higher risk of bias in methods than other trials [93], so the explanation for the higher frequency of positive results must lie elsewhere. Other possibilities have been suggested. These include the more frequent use of surrogate outcomes, and publication bias, in which trials with negative findings are not published [93, 102, 103]. An analysis of head‐to‐head trials (which directly compare drug versus drug, rather than drug versus placebo) showed that 96.5% of trials favoured the drug manufactured by the company funding the study [104]. This may provide evidence of industry manipulation. An analysis of internal documents from the industry found that suppression of negative studies and spinning of negative findings were recognised techniques [105]. As one author suggested, allowing drug companies to generate evidence ‘is akin to letting a politician count their own votes’ [103].

      This chapter has explored the many sources of bias that commonly afflict randomised controlled trials. The randomised controlled trial is held to be the gold standard for evidence on treatment effectiveness, but that gold is more than a little tarnished. One commentator concluded that randomised controlled trials are: ‘often flawed, mostly useless, clearly indispensable’ [106].

      The two key questions for clinical trials are: how frequently do these flaws occur; and how big an effect do they exert on estimated effect sizes? The frequency of flaws varies across individual review studies and by the type of deficiency, so there is not a specific estimate of how often bias occurs. Instead we can put the frequencies on a scale from very rare to very common. As most of the estimates from review articles are in the fairly common or very common area, flaws are a serious problem.

      Some of the weaknesses in clinical trials may simply be due to lack of knowledge or experience. This could explain deficiencies in the handling loss to follow‐up or the lack of blinding of outcome assessment. Tampering with the randomisation sequence, and the replacement of primary outcomes, suggests a less innocent explanation. The deficiencies described in this chapter, which may result from inadvertent mistakes or deliberate actions, pose a serious threat to the integrity of medical evidence. The next chapter describes weaknesses in study design and conduct that lead to wasted and unhelpful trials.

      1 1. Riechelmann, R.P., Peron, J., Seruga, B. et al. (2018). Meta‐research on oncology trials: a toolkit for researchers with limited resources. Oncologist 23: 1467–1473.

      2 2. Page, M.J., Higgins, J.P., Clayton, G. et al. (2016). Empirical evidence of study design biases in randomized trials: systematic review of meta‐epidemiological studies. PLoS One https://doi.org/10.1371/journal.pone.0159267.

      3 3. Adie, S., Harris, I.A., Naylor, J.M. et al. (2017). The quality of surgical versus non‐surgical randomized controlled trials. Contemp. Clin. Trials Commun. 5: 63–66.

      4 4. Savovic, J., Jones, H., Altman, D. et al. (2012). Influence of reported study design characteristics on intervention effect estimates from randomised controlled trials: combined analysis of meta‐epidemiological studies. Health Technol. Assess. 16: 1–82.

      5 5. Dechartres, A., Trinquart, L., Atal, I. et al. (2017). Evolution of poor reporting and inadequate methods over time in 20 920 randomised controlled trials included in Cochrane reviews: research on research study. BMJ https://doi.org/10.1136/bmj.j2490.

      6 6. Wuytack, F., Regan, M., Biesty, L. et al. (2019). Risk of bias assessment of sequence generation: a study of 100 systematic reviews of trials. Syst. Rev. https://doi.org/10.1186/s13643‐018‐0924‐1.

      7 7. Savovic, J., Turner, R.M., Mawdsley, D. et al. (2018). Association between risk‐of‐bias assessments and results of randomized trials in Cochrane reviews: the ROBES meta‐epidemiologic study. Am. J. Epidemiol. 187: 1113–1122.

      8 8. Zhai, X., Cui, J., Wang, Y. et al. (2017). Quality of reporting randomized controlled trials in five leading neurology journals in 2008 and 2013 using the modified “risk of bias” tool. World Neurosurg. 99: 687–694.

      9 9. Rikos, D., Dardiotis, E., Tsivgoulis, G. et al. (2016). Reporting quality of randomized‐controlled trials in multiple sclerosis from 2000 to 2015, based on CONSORT statement. Mult. Scler. Relat. Disord. 9: 135–139.

      10 10. Saltaji, H., Armijo‐Olivo, S., Cummings, G.G. et al. (2018). Impact of selection bias on treatment effect size estimates in randomized trials of Oral health interventions: a meta‐epidemiological study. J. Dent. Res. 97: 5–13.

      11 11. Schulz, K.F. and Grimes, D.A. (2002). Unequal group sizes in randomised trials: guarding against guessing. Lancet 359: 966–970.

      12 12. Clark, L., Fairhurst, C., Hewitt, C.E. et al. (2014). A methodological review of recent meta‐analyses has found significant heterogeneity in age between randomized groups. J. Clin. Epidemiol. 67: 1016–1024.

      13 13. Clark, L., Fairhurst, C., Cook, E. et al. (2015). Important outcome predictors showed greater baseline heterogeneity than age in two systematic reviews. J. Clin. Epidemiol. 68: 175–181.

      14 14. Schulz, KF. (1995). Subverting randomization in controlled trials. JAMA 274: 1456–1458.

      15 15. Paludan‐Muller, A., Laursen, D.R.T., and Hrobjartsson, A. (2016). Mechanisms and direction of allocation bias in randomised clinical trials. BMC Med. Res. Methodol. https://doi.org/10.1186/s12874‐016‐0235‐y.

      16 16.

Скачать книгу