Crossref Citations
This article has been cited by the following publications. This list is generated based on data provided by Crossref.
Meinlschmidt, Gunther
Lee, Jong-Hwan
Stalujanis, Esther
Belardi, Angelo
Oh, Minkyung
Jung, Eun Kyung
Kim, Hyun-Chul
Alfano, Janine
Yoo, Seung-Schik
and
Tegethoff, Marion
2016.
Smartphone-Based Psychotherapeutic Micro-Interventions to Improve Mood in a Real-World Setting.
Frontiers in Psychology,
Vol. 7,
Issue. ,
Markozannes, Georgios
Aretouli, Eleni
Rintou, Evangelia
Dragioti, Elena
Damigos, Dimitrios
Ntzani, Evangelia
Evangelou, Evangelos
and
Tsilidis, Konstantinos K.
2017.
An umbrella review of the literature on the effectiveness of psychological interventions for pain reduction.
BMC Psychology,
Vol. 5,
Issue. 1,
Gold, Stefan M
Enck, Paul
Hasselmann, Helge
Friede, Tim
Hegerl, Ulrich
Mohr, David C
and
Otte, Christian
2017.
Control conditions for randomised trials of behavioural interventions in psychiatry: a decision framework.
The Lancet Psychiatry,
Vol. 4,
Issue. 9,
p.
725.
Wagner, Christoph J.
Dintsios, Charalabos Markos
Metzger, Florian G.
L'Hoest, Helmut
Marschall, Ursula
Stollenwerk, Bjoern
and
Stock, Stephanie
2018.
Longterm persistence and nonrecurrence of depression treatment in Germany: a four‐year retrospective follow‐up using linked claims data.
International Journal of Methods in Psychiatric Research,
Vol. 27,
Issue. 2,
Cristea, Ioana A.
and
Naudet, Florian
2019.
Increase value and reduce waste in research on psychological therapies.
Behaviour Research and Therapy,
Vol. 123,
Issue. ,
p.
103479.
Yoder, Whitney Rose
Karyotaki, Eirini
Cristea, Ioana-Alina
van Duin, Daniëlle
and
Cuijpers, Pim
2019.
Researcher allegiance in research on psychosocial interventions: meta-research study protocol and pilot study.
BMJ Open,
Vol. 9,
Issue. 2,
p.
e024622.
Cristea, Ioana A.
2019.
The waiting list is an inadequate benchmark for estimating the effectiveness of psychotherapy for depression.
Epidemiology and Psychiatric Sciences,
Vol. 28,
Issue. 03,
p.
278.
Rabeyron, Thomas
2020.
Why Most Research Findings About Psi Are False: The Replicability Crisis, the Psi Paradox and the Myth of Sisyphus.
Frontiers in Psychology,
Vol. 11,
Issue. ,
Rabeyron, Thomas
2021.
L’évaluation et l’efficacité des psychothérapies psychanalytiques et de la psychanalyse.
L'Évolution Psychiatrique,
Vol. 86,
Issue. 3,
p.
455.
Price, Larry
Briley, Josh
Haltiwanger, Steve
and
Hitching, Rita
2021.
A meta-analysis of cranial electrotherapy stimulation in the treatment of depression.
Journal of Psychiatric Research,
Vol. 135,
Issue. ,
p.
119.
Pile, Victoria
Smith, Patrick
Leamy, Mary
Oliver, Abigail
Bennett, Eleanor
Blackwell, Simon E.
Meiser-Stedman, Richard
Stringer, Dominic
Dunn, Barnaby D.
Holmes, Emily A.
and
Lau, Jennifer Y.F.
2021.
A feasibility randomised controlled trial of a brief early intervention for adolescent depression that targets emotional mental images and memory specificity (IMAGINE).
Behaviour Research and Therapy,
Vol. 143,
Issue. ,
p.
103876.
Haun, Markus W.
Tönnies, Justus
Krisam, Regina
Kronsteiner, Dorothea
Wensing, Michel
Szecsenyi, Joachim
Vomhof, Markus
Icks, Andrea
Wild, Beate
Hartmann, Mechthild
and
Friederich, Hans-Christoph
2021.
Mental health specialist video consultations versus treatment as usual in patients with depression or anxiety disorders in primary care: study protocol for an individually randomised superiority trial (the PROVIDE-C trial).
Trials,
Vol. 22,
Issue. 1,
Blackwell, Simon E
and
Woud, Marcella L
2022.
Making the leap: From experimental psychopathology to clinical trials.
Journal of Experimental Psychopathology,
Vol. 13,
Issue. 1,
Ware, John W
Venere, Kenny
Miller, Stephanie A
Freeman, Heather
Scalzitti, David A
and
Hoogeboom, Thomas
2023.
A Systematic Appraisal of Conflicts of Interest and Researcher Allegiance in Clinical Studies of Dry Needling for Musculoskeletal Pain Disorders.
Physical Therapy,
Vol. 103,
Issue. 6,
Lawless, William
and
Moskowitz, Ira S.
2024.
Shannon Holes, Black Holes, and Knowledge: The Essential Tension for Autonomous Human–Machine Teams Facing Uncertainty.
Knowledge,
Vol. 4,
Issue. 3,
p.
331.
Cuipers and Cristea (Reference Cuipers and Cristeathis issue) present a smart Decalogue on how to prove that psychotherapy is effective even when it is not. Given such an impressive list of ‘techniques’ for biasing results and empirical estimates for their prevalence, most current evidence on psychotherapies is probably heavily distorted. No one wishes this situation perpetuated. Therefore, using this Decalogue in inverse should decrease bias. However, here I argue that not all of these bias ‘techniques’ should be eliminated. Some ‘techniques’ are always detrimental, while some others may have a niche in the development of psychotherapies. Psychotherapies may need to be tested under biased conditions, but bias should be of the right type.
Improper biases
Let me start with the five ‘techniques’ that are always bad. I will call them ‘improper biases’. There is nothing to say in defence of these research practices. Using the ‘weak spots’ of randomised trials, not concealing treatment allocation to assessors of outcome, analysing only the participants who completed the intervention and ignoring dropouts, using multiple outcome instruments and selectively reporting only the significant ones and not publishing results unless positive represent clear cheating. Any treatment that does not work may seem to work when these recipes are followed. Empirical evidence suggests that this is happening frequently, as summarised by Cuipers and Cristea. Given the prevalence of these ‘improper biases’ alone (Ioannidis et al. Reference Ioannidis, Munafò, Fusar-Poli, Nosek and David2014), most psychotherapies probably have no or little benefit, even when published literature suggests they are very effective. For example, published psychotherapies for depression have an average d = 0.69 (Cuipers et al. Reference Cuipers, Vanstraten, Warmerdam and Smits2008), but the true average effect may be d = 0.2–0.3 or less. For many other conditions the situation may be even worse.
These ‘improper biases’ vary in the extent to which they are avoidable. Preemptive action is preferable, but some biases are difficult to eliminate entirely. In particular, concealing allocation to assessors of outcomes is sometimes difficult. Even minimal contact with the assessor may reveal what psychotherapy the patient has been on. Moreover, treatment dropouts and losses to follow-up are frequent even in short-term studies and, indeed, they often reflect lack of effectiveness or poor tolerability. However, this is difficult to judge. Imputation methods are better than ignoring missing observations, but still leave substantial uncertainty. What this all means is that at least the other improper biases (those possible to deal thoroughly with) should be eliminated. There is absolutely no reason nowadays for a trial not to be performed with robust randomisation, allocation concealment and pre-specified outcomes and not to get published as pre-specified. I leave some room only about the need to occasionally modify the analysis plan if something ensues during the study conduct that could not be anticipated upfront. Then this still needs to be transparently acknowledged, the modified analysis plan justified and results interpreted with caution.
Potentially proper biases
The other five ‘techniques’ listed by Cuipers and Cristea are not necessarily always bad. In fact, under some circumstances, they are very appropriate. I will call them ‘potentially proper biases’. Let me explain.
When researchers test their own pet intervention themselves they may do everything possible to increase expectations in the participants. Allegiance bias has been documented to be powerful, although its ability to change results varies across interventions (Munder et al. Reference Munder, Brütsch, Leonhart, Gerger and Barth2013). We should set apart the components of allegiance bias that operate through the five ‘improper biases, as discussed above. If we safeguard that this cheating does not occur, it is not unreasonable to have the researcher who developed a method to do the initial testing. If indeed most psychotherapies do not work or have only weak effects, it makes sense that the best experts should perform the first studies on how to implement the therapy. Conversely, administration of therapies by inexperienced students is the best way to decrease the observed effect (Cuipers et al. Reference Cuipers, Vanstraten, Warmerdam and Smits2008). If the best experts can get no clinically meaningful effect, they can safely quit further testing, since the intervention will do worse in the hands of others. If a psychotherapy cannot attain d = 0.3 under optimised circumstances, it is unlikely to attain d = 0.1 when used more widely. The incentives system has to be such that a researcher is rewarded for showing that his idea for a new psychotherapy did not work. Given that most new psychotherapies do not really work, perhaps journals should always accept for publication well-done studies with ‘negative’ results, and may request further, additional, replication evidence (Open Science Collaboration, 2015) by independent investigators before they publish significant, ‘positive’ results (Ioannidis, Reference Ioannidis2006).
This is different to the testing of drugs for diseases such as cancer or cardiovascular disease, where I have argued that the entire clinical testing agenda should be kept independent from the manufacturer (Naci & Ioannidis, Reference Naci and Ioannidis2015). For a 5 mg pill, we do not need to have the CEO of the company by the bedside to administer it. Conversely, a new psychotherapy may need its developer to implement it initially. This gives it the best possible shot. Moreover, most drugs go through a long and expensive screening preclinical process of testing (in vitro, in vivo, in animals). Even though biases at these stages are prominent (Ioannidis & Begley, Reference Begley and Ioannidis2015), the whole process, even if inefficient, ends up eliminating many candidate drugs that have weak or unfavourable preclinical evidence (Goodman and Gerson, Reference Goodman and Gerson2013). Conversely, psychotherapies can hardly be assessed in test-tubes, cell cultures or animals. They emerge from theoretical speculation and currently hit patients with little pre-screening. Thus the odds of success are probably weaker for psychotherapies than for drugs. (Of course, even for drugs, pre-clinical testing may have varying screening ability depending on how well the disease is reproduced in the test-tube, cell or animal. A preclinical model of depression is more elusive than one of clot formation).
Raising the expectations of the participants by boosting the placebo effect is also not a bad idea. Anyhow, most psychotherapies that do work, probably work to a good extent through the placebo effect. Some interventions may be more amenable by boosting the placebo effect. If this can be done efficiently, reproducibly and without cost, that is great! A treatment that exploits and magnifies the placebo effect from d = 0.2 to d = 0.5 is not worse than the one that achieves the same benefit through a different, more sophisticated mechanism. Conversely, a treatment is undesirable if it affects some favourable mechanism but concomitantly diminishes the placebo effect by a greater amount.
Small sample size is also not a vice on its own. The problem is that small studies are more susceptible to the five ‘improper biases’ than larger ones. A trial of n = 20 participants is easier to hide quietly in a file drawer than a trial of n = 2000 participants. However, if improper biases are eliminated (as they should), performing a few well-done studies of modest size is a reasonable choice (Inthout et al. Reference Inthout, Ioannidis and Borm2012). If most psychotherapies do not work, it makes sense to test them initially with optimised studies of modest size, barely sufficient to exclude an effect of clinical interest. Otherwise, we waste resources to run large trials on low-yield experimental therapies. Large studies at an early stage make sense only if there is a reasonable chance to see a clinically meaningful effect and that clinically meaningful effect is small (thus needing a large study for sufficient power). This is probably more common for new drugs emerging from extensive preclinical screening than for new psychotherapies, as discussed above.
Similarly, using a waiting list control group and not comparing the tested intervention against an existing effective intervention are not necessarily vices on their own. Both of these approaches create a more discernible effect size. This effect size is not spurious (as in the case of the five ‘improper biases’). It simply reflects the comparative effect against nothing, rather than the incremental effect against something else that is already effective. If the few psychotherapies that are effective have small (but still clinically meaningful) effects, testing them against absolutely nothing can help discern their benefit without having to resort to huge sample sizes. If this initial screening is successful, then the relative benefits and overall merits of the new treatment should still be compared against other existing ones with large trials. However most psychotherapies that do not work even against nothing will be quickly screened out with small trials, failing even this favourably biased test. Again, incentives should reward publishing such ‘negative’ results and save the field from wasting effort chasing spurious claims.
Moving from pre-screening trials to large pragmatic trials
To summarise, some biases are always bad and should be eliminated, whenever possible. Other biases may be desirable. Carefully properly biased studies may show which psychotherapies might work. The few best survivors can then move to large-scale testing in pragmatic randomised trials of many thousands participants and many practitioners. The even fewer survivors at this second stage can be considered for wide, everyday use. From this viewpoint, most current trials of psychotherapies represent merely early stage, pre-screening research that has been done poorly to-date and can be done much better.
Financial Support
The work of John Ioannidis is supported by an unrestricted gift from Sue and Bob O'Donnell.