Showing posts with label research philosophy. Show all posts
Showing posts with label research philosophy. Show all posts

Saturday, March 31, 2012

What Should Be Done about Reproducibility

A recent Commentary and linked editorial in Nature regarding reproducible science (or rather the lack thereof in science) has been troubling me for a few days now. The article brings to light a huge problem in the current academic science enterprise.

What am I talking about?

ResearchBlogging.org

In the comment, two former Amgen researchers describe some of the efforts of that company to reproduce "landmark studies" in cancer biology. Amgen had a team of about a hundred researchers called the reproducibility team and their job was to test new basic science findings prior to investing in following up these targets. Shockingly, according to the authors, only 6/53 of these landmark studies were actually reproduced. When things were not reproduced they contacted the authors to attempt to work through the potential problems. This is an incredibly dismal 11% reproducibility rate!

Could it really be that bad?

The first problem is what exactly is meant by reproducibility. In the commentary the authors acknowledge that they did attempt to use additional models in the validation process and that technical issues may have under-lied some of these differences. They also point out that their sample set is biased with respect to the findings. These were often novel and cutting edge type findings and typically more surprising than the general research finding. Also, their definition of reproducibility is unclear. If researcher says drug X has a 10 fold effect on something and the Amgen guys say it has a 3X effect on the process is that a reproducible finding. My initial reaction was that the 89% were thing where the papers said something like thing X does thing Y and there was no evidence supporting that. We don't know, and in a bit of an ironic twist, since no data is provided (either which papers were good and which were bad, or within those, which findings were good and bad) this commentary could be considered both unscientific and non-reproducible itself (also we are awfully close to April Fools Day).


So there is some bad papers out there, who cares?

Reproducibility is at the heart of everything we do as scientists. No one cares if you did something once and for reasons you cant really explain, were never able to do it again. If something is not replicable and reproducble for all intents and purposes it should be ignored. We need measures of these to be able to evaluate research claims, and we need context specificity to understand the breadth of claims. Ill toss out few reasons why this problem really matters both to those of us who do science, and to everyone else.

This is a massive waste of time and money

From the commentary:
Some non-reproducible preclinical papers had spawned an entire field, with hundreds of secondary publications that expanded on elements of the original observation, but did not actually seek to confirm or falsify its fundamental basis.
Wow, really? Whole fields have been built on these? In a way I don't feel bad for these fields at all. If you are going to work in a field, and are never going to bother even indirectly testing the axioms on which your field is built then you are really not so good at the science. If you are going to rely on everyone else being correct and never test it then your entire research enterprise might as well be made from tissue paper. More importantly, if you are on top of these things you are going to waste time and money figuring out not to follow this up. Hopefully this is the more common case. This really goes back to the difficulty in publishing negative data to let people know which conditions work and which don't.

The reward system for science is not in sync with the goals of the enterprise

Why are people publishing things that they know only happen one out of six times? Why are they over-extending their hypotheses and why are they reluctant to back away from their previous findings? All of these things are because we are judged for jobs and for tenure and for grants on our ability to do these things. The person who spends 3 years proving that a knockout mouse model does not actually extend lifespan walks away with nothing, the one who shows it (even if done incorrectly) gets a high impact paper and a job. Even if it didn't take an unreasonable amount time and effort to publish non-reproducible data, the risk of insulting another researcher or not contributing anything new might be enough to prevent this. Until the rewards of publishing negative or contravening data are on par with the effort, people just won't do it.

This reflects really poorly on science and scientists

Science is always and probably has always been under some type of "attack". Science as an entity and scientists as their representatives need to not shirk this off or ignore it. We have to deal with this problem head-on, whether it be at the review level or at the post-publication level. People who are distrustful of science are rightful to point at this and say, why are we giving tens of billions of dollars to the NIH when they are 89% wrong. Why not just give that money to Amgen, who seem to be the ones actually searching for the truth (not that they will share that data with anyone else).

Can anything be done?

The short answer is its really going to be difficult and its going to rely on a lot of moving parts. Reviewers should (and in my experience do) ask for explicit reprodicibility statements in the papers. This can go farther, if someone says this blot is representative of 5 experiments then there is no reason the other 4 couldnt be put in the supplement. If they looked at 100 cells and show just one, then why cant the rest be quantified in some way. Post-publication, there should be open (ie not just in lab meetings) discussion of papers and the problems and where they match or mismatch with the rest of the literature. Things like blogs and the Faculty of 1000 are great, but how often have you seen a negative F1000 review? Finally, eventually there ought to be some type of network of research findings. If I am reading a paper, and I would like to know what other results agree or disagree with this, it would be fantastic to get there in a reasonable way. This is probably the most complicated, as it requires not only publication of disagreeing findings, but also some network to link them together.



Begley, C., & Ellis, L. (2012). Drug development: Raise standards for preclinical cancer research Nature, 483 (7391), 531-533 DOI: 10.1038/483531a

Creative Commons License
What Should Be Done about Reproducibility by Dave Bridges is licensed under a Creative Commons Attribution 3.0 Unported License.

Sunday, February 26, 2012

Future Bridges Lab Rules version 0.1

Lab Rules


Version 0.1.3 on July 14, 2012 by Dave Bridges

Remember when you were growing up and you would say, well when I’m older I will (or won’t) do that.
I have been thinking of that, for my future when I run my own group.
It is fairly easy (and as a bit of a blowhard I do this all the time) to say I would do this, or I would do that.
I think posting this publicly will encourage me to stick to these rules.
Below are some roles and a bit of rationale and caveats.
This is the first version of this post but future versions will include links to the previous versions.
The version numbering is described in the Lab Policies README.
Check out the GitHub Repository for more granular changes.

Supervision of Trainees

Trainee-Advisor Contract
Both myself and trainees (either mine, or co-supervised trainees) will read, discuss and sign a contract describing our roles and responsibilities both as a trainee/employee and a mentor. This will include data dissemination/publishing rules, expectations of productivity, note keeping and time commitment, rules for dealing with other members both in my group and in collaborations, rules for sharing of reagents and data, rules for adjucating disagreements and grounds and procedures for termination. These rules will be in conformity with any institutional rules. Exceptions can be discussed and the agreement can be modified throughout the term of the relationship. I will post a generic version of this agreement in a publicly viewable location.
Online Presence
All trainees will appear on the laboratory website and write a blurb about their research interests and goals. Trainees will be strongly encouraged to blog, tweet and otherwise engage in social networking tools regarding their research and the work of others, but this is not required. Links to their publicly available social network profiles will be posted on the laboratory website.
Open Access Policy
Trainees will be made aware of the open publishing, dissemination, software and data/reagent sharing policies of the laboratory at the outset and will have to agree to these standards.

Reagents, Software and Tools

Software Usage
Wherever possible, free open source software will be used for data acquisition, analysis and dissemination. Exceptions will be made if necessary, but trainees will be encouraged to use/incorporate/development free tools.
Software Development
If software, scripts or the like are generated they will be released under a permissible open source license such as CC-BY and the license will be attached explicitly to the source code. Scripts and programs will be uploaded to a public revision control database such as GitHub or similar (my GitHub profile is here).
Publishing of Protocols and Scripts
When not present in the published article, detailed step by step protocols, data analysis scripts and other things which cannot fit into either methods and materials sections or supplementary materials will be posted online and linked to the publication’s online presence (post or as a comment on the paper’s website).
Protocol Sharing
Protocols will be made available online in a wiki format in a publicly available location, whether they have been published on or not. Editing will be restricted to laboratory members and collaborators.
Reagent and Tool Sharing
Reagents generated by my group will be shared upon request without condition (aside from potential restrictions placed by other collaborators, funding agencies and the institution). These reagents will be shipped with an explicit statement of free use/sharing/modification. Once a reagent sharing license is generated/identified it will be linked to in this document. This policy includes unpublished reagents and will never require attribution as a condition. If a reagent is obtained from another group and modified, we will offer the modified reagent back to the originator immediately.

Publishing and Data Dissemination

Open Access Journals
I believe that all work should be available to the public to read, evaluate and discuss. I am strongly against the mentality that data/knowledge should be restricted to experts and the like. I will therefore send all papers in which I am corresponding author and have supervised the majority of the work to journals (or their equivalent) which are publicly available. The two major caveats will be for work in which I am a minor (less than 50% effort) collaborator and the primary group leader wants to submit the work elsewhere. This will not exempt any potential major impact publications, no matter how awesome they may be. Delayed open access does not count in this respect.
Open Peer Review
Journals will be selected which publish non-anonymous reviewer comments alongside the articles whenever possible. If this is not done, and if permissible by the publisher and/or reviewers I will re-post the reviewer comments online without any modifications.
Public Forum for Article Discussion
Although I will encourage discussion of articles to occur at the point of publication (for example via the posting of comments directly at the website of the publisher), I will also provide a publicly available summary of every published finding from which I am an author (corresponding or not) and allow commenting at that point too. This discussion post will also link to or contain the reviewer and editor comments where possible. This summary might be a blog post, a facebook post or a google plus post or anything else that might come up in the future. If I am not the primary author or corresponding author I will encourage the first or corresponding author to write the post and link/quote that directly.
Presentations
All presentations of published data will be posted on an online repository such as Slideshare or something similar. My slideshare profile is here. If unpublished or preliminary data is presented privately and then is later published, then those slides will be presented upon publication. Similar to papers, an online blog post or the like will also accompany that upload. If audio or video of the presentation is available, that will be uploaded as well.
Data Sets
All datasets, once published will be made available in manipulable (preferably non-proprietary) formats for further analysis. Based on the scheme set out by the Linked Data Research Center Laboratory, all data will be provided at level 2 or above.