Nature | Editorial

STAP retracted

Two retractions highlight long-standing issues of trust and sloppiness that must be addressed.

Article tools

This week, Nature publishes retractions of two high-profile papers that claimed a major advance in the field of stem cells (see page 112). Between them, the two papers seemed to demonstrate that a physical perturbation could do what had previously been achieved only by genetic manipulation: transform adult cells into pluripotent stem cells able to differentiate into almost any other cell type. The acronym STAP (stimulus-triggered acquisition of pluri­potency) became instantly famous.

Soon after the papers were published on 30 January, cracks appeared in the edifice. As various media outlets including Nature’s independent news team reported, errors were found in the figures, parts of the methods descriptions were found to be plagiarized and early attempts to replicate the work failed.

The problems that initially emerged did not fundamentally undermine the papers’ conclusions. Moreover, replication of such work is not necessarily straightforward or quick, and the ability to use some techniques can be very sensitive to aspects of the experimental protocol.

Nevertheless, the RIKEN research centre in Japan, one of the institutions in which most of the work was carried out, promptly organized an inquiry and found inadequacies in data management, record-keeping and oversight (see go.nature.com/2vrjxs). One author was found guilty of misconduct — a charge that RIKEN reaffirmed following an appeal (see go.nature.com/tnxuhy).

We at Nature have considered what lessons we can derive from these flaws. When figures often involve many panels, panels duplicated between figures may, in practice, be impossible for journals to police routinely without disproportionate editorial effort. By contrast, image manipulation is easier to detect. Our policies have always discouraged inappropriate manipulation. However, our approach to policing it was never to do more than to check a small proportion of accepted papers. We are now reviewing our practices to increase such checking greatly, and we will announce our policies when the review is completed.

But only since the RIKEN investigation has it become clear that data that were an essential part of the authors’ claims had been misrepresented. Figures that were described as representing different cells and different embryos were in fact describing the same cells and the same embryos.

All co-authors of both papers have finally concluded that they cannot stand behind the papers, and have decided to retract them.

The papers themselves have now been clearly watermarked to highlight their retracted status, but will remain hosted on Nature’s website, as is consistent with our retraction policy. (In our opinion, to take down retracted papers from journal websites amounts to an attempt to rewrite history, and makes life needlessly difficult for those wishing to learn from such episodes.)

“We at Nature have considered what lessons we can derive from these flaws.”

We at Nature have examined the reports about the two papers from our referees and our own editorial records. Before publishing, we had checked that the results had been independently replicated in the laboratories of the co-authors, and we regret that we did not capture the authors’ assurances in the author-contributions statements.

We have concluded that we and the referees could not have detected the problems that fatally undermined the papers. The referees’ rigorous reports quite rightly took on trust what was presented in the papers.

For more than two years, Nature has been publishing articles that highlight the problems that can arise in the analysis and publication of the results of biomedical research. We have also launched several initiatives to improve our own rigour. For a collection of relevant content, see go.nature.com/huhbyr.

Underlying these issues, often, is sloppiness, whether in the handling of data, in their analysis, or in the inadequate keeping of laboratory notes. As a result, the conclusions of such papers can seem misleadingly robust. Another contributory factor lies in selection bias behind the data presented, whether implicit because the experiment was not randomized or blinded, or explicit in the deliberate selection of data that, usually with honest good intentions, are judged to be representative. (This is not to say that randomizing and blinding is always required, but more is needed than currently occurs.)

A manifestation of these problems has been a growth in the number of corrections reported in journals in recent years. It is hoped that the extension of our methods sections, the addition of a checklist intended to improve the standards of reporting, and our use of statistical advisers will reduce these problems in Nature.

In short: although editors and referees could not have detected the fatal faults in this work, the episode has further highlighted flaws in Nature’s procedures and in the procedures of institutions that publish with us. We — research funders, research practitioners, institutions and journals — need to put quality assurance and laboratory professionalism ever higher on our agendas, to ensure that the money entrusted by governments is not squandered, and that citizens’ trust in science is not betrayed.

Journal name:
Nature
Volume:
511,
Pages:
5–6
Date published:
()
DOI:
doi:10.1038/511005b

For the best commenting experience, please login or register as a user and agree to our Community Guidelines. You will be re-directed back to this page where you will see comments updating in real-time and have the ability to recommend comments to other users.

Comments for this thread are now closed.

Comments

9 comments Subscribe to comments

  1. Avatar for Sanal Madhusudana Girija
    Sanal Madhusudana Girija
    More than retraction of a paper.... retracted someones life too- Sasai. Share responsibility! A reasonable Editorial process would have perhaps prevented a death!
  2. Avatar for Taranjit Singh Rai
    Taranjit Singh Rai
    We scientists do experiments, Its time top journals such as Nature and Cell should also do an experiment. Experiment on anonymous peer review. Anonymous authors, no self citing, everything blinded. Just like computing science and economics. The progress in computing and economics fields have no parallels for last decade or so. Reviewers are anonymous so why can't the authors be anonymous. I can see pitfalls but it's an experiment begging to be done. I believe this would be one step of many to bring fairer system.
  3. Avatar for Sanal Madhusudana Girija
    Sanal Madhusudana Girija
    This is a business-I mean publishing business where little editors strive to thrive in their little kingdom below the senior management and the chief Editor. These princesses (quite often) have little interest in science when they struggle for a little raise! Who has interest in double blind peer-review in big journals or big grants? Neither big scientists or big journals!
  4. Avatar for Donald Newgreen
    Donald Newgreen
    We read that "our (Nature's) approach to policing it was never to do more than to check a small proportion of accepted papers". While checking here appears to refer to image manipulation, it does seem like more Editorial diligence should have been exercised. Everyone, and I mean everyone, with whom I discussed the papers as published had their TGTBT Reflex activated--too good to be true! And in a few minutes they had questions formulated. A knee jerk reflex of itself is no grounds for rejection, but given that Nature Editors were well aware of the potential importance of the papers, it is grounds for including such manuscripts in that small proportion subject to more intensive scrutiny. Regarding reviewers being influenced by the standing of some authors; this is an all too human failing. The phenomenon of "guest authorship" is a reflection of the perceived effectiveness of this as an aid to acceptance. Reviewers should be routinely reminded of this failing, and advised to judge strictly on the matter at hand. This would help but not totally prevent this: in the present case reviewers clearly stated that they were influenced by author reputation. Anonymising manuscripts for peer review might also help (although educated guesses would of course still be made), and on-line availability of the review process (named or anonymous) would also be useful in maintaining the quality of reviews.
  5. Avatar for Harold Dresden
    Harold Dresden
    Clearly Nature is trying to wash its hand in innocence while trying to milk as much out of this story as possible. The only reason this editorial has been written is because ONE scientist got caught. Right now I could name five Nature papers from the last two years which are generally considered to be bullshit in the community. Yet, nobody will say anything because this is something you just don't do. Instead nobody will cite those papers, the authors still use them to secure more funding or positions and everybody is happy. I believe you that you are sorry. But not because you published fraudulent data but you finally got caught for once. You know exactly that in a month nobody will talk about this anymore and from then on it is business as usual.
  6. Avatar for Mick G
    Mick G
    There are still many investigators that believe that professional statisticians are not necessary. I believe that's part of the problem. On the contrary, I personally think that there should be at least one MS or Ph.D. professional statistician clearly mentioned in each submission (similarly as for the corresponding author). They would be responsible for the integrity of the data management procedures and presentation of the results of the analyses. Indeed, I think that the current state of affairs may not be due only to "misconduct", but also to poor judgement or competence. Disclaimer: I say this as a professional statistician myself, who has been able to appreciate how easy it might be sometime to get "fooled" by data.
  7. Avatar for Fillip Port
    Fillip Port
    It is commendable that Nature takes this case as an occasion to critically question the quality control procedures it has in place for the research it publishes. However, I find the suggested measures fairly limited. The STAP case also nicely illustrates other problems with the current peer review system (which are by no means specific to Nature). An important part of the answer to the question why reviewers and coauthors did not spot the problems with the two papers appears to be that their judgement was influenced by the reputation of the scientists involved and their affiliations. For example, one of the reviewers of an earlier version of the papers is quoted as saying that he did not carefully analyse the figures for signs of manipulation, "because Hitoshi (Niwa), Teruhiko (Wakayama), and Yoshiki (Sasai) were on the paper; there are some people you 100% trust" (http://www.sciencemagazinedigital.org/sciencemagazine/13_june_2014/m3/Page.action?lm=1402605805000&pg=20). Similarly, one of the coauthors suggested that part of the reason why he did not more rigorously check the reported results was that the first author had "an impressive educational background, had trained at Harvard, and was known as a superior researcher" (http://news.sciencemag.org/asiapacific/2014/06/evidence-mounts-against-new-stem-cell-method). Together these quotes illustrate a problem in peer review that is likely to be much more general. Evaluation of a manuscript is likely to be influenced by the subconscious opinion the reviewer holds about the authors. Therefore, making the reviewer unaware or at least uncertain of the identity of the authors would surely reduce overall bias in the peer review system. It is often pointed out that this is not possible as expert reviewers know what labs produce which work in their field. Although that will be true for a good number of studies, it is certainly not true for all papers and all reviewers and these would benefit from double-blind peer review. The STAP case also highlights the usefulness of publishing the peer review files along with the accepted manuscript, as is standard in journals such as EMBO Journal or eLife. How can we have a discussion of what went wrong without knowing what issues were raised during review? The knowledge that the reviews will be published is also likely to present an additional incentive for reviewers to provide quality reviews and these files are an important resource for junior scientists to learn how to critically evaluate papers. Nevertheless, all measures discussed here and in the editorial above are just gradual improvements of the current system. To really address the reported rise in papers that contain non-representative data and manipulated images it is paramount to ask what are the underlying causes. Insufficient training for junior scientists is likely part of the answer. However, it would be naive to suggest that the relentless pressure to publish in the most prestigious journals is not an important contributing factor. Only if the scientific community can establish a culture in which authors are convinced that what matters for their career is if their paper turns out to be true, rather than where it has been published, is when we will see a drop in reports that contain consciously included inaccuracies.
  8. Avatar for T Roll
    T Roll
    While it is right that the papers have been retracted, two statements in this editorial stand out to me: 1. "When figures often involve many panels, panels duplicated between figures may, in practice, be impossible for journals to police routinely without disproportionate editorial effort." Does disproportionate effort represent actually having someone read the article to notice duplicated figures? It seems to me that as a hugely profitable publisher, Nature could afford to have someone on staff who makes sure basic things like duplicated figures are noticed prior to publication. 2. "We have concluded that we and the referees could not have detected the problems that fatally undermined the papers." Surely the many issues with this paper, highlighted almost immediately by scientists around the world by social media are indication that were these articles peer reviewed properly, this whole debacle could have been avoided. To me this editorial smacks of Nature trying to wash its hands of the situation and shift any blame away from its own editorial and peer review process. The cynical person would suggest that the journal may have been trying to court controversy with these papers, knowing how much publicity the ensuing debate, and eventual retraction would create.
  9. Avatar for Yiding Zhao
    Yiding Zhao
    Even with retraction of STAP papers, no one will ask NPG to refund cost of ordering articles containing fake data or waste of time and resource on replication. On the other hand if more staff is hired to look out errors like "KC1" instead of "KCl", the profit margin will go down and company owner will not happy! Remember staffs of NPG report to his/her boss that demand a return from their stock, not to some imaginary god of science.

Top Story

Retina

Next-generation stem cells cleared for human trial

Researchers hope to treat macular degeneration of the retina with induced pluripotent stem cells, a method that has generated enormous expectations.

Science jobs from naturejobs