The Case for Preregistration in Cell Biology: A Practical Guide

January 23, 2026 at 5:00 pm | | literature, science community, scientific integrity

UPDATE: I recently saw this great blog post from Arjun Raj.

The basics of the scientific method—hypothesize, test, revise—rely on prediction. By forming a hypothesis only after seeing the data, we fail to severely test our ideas. I encourage cell biologists to preregister their study designs and statistical analyses. I also provide practical tutorials for making this a reality.

Preregistration will Bolster Cell Biology

Preregistration has helped add rigor to various fields, including psychology and clinical trials (Nosek et al., 2018). But in cell biology, molecular biology, and other basic research, preregistration has not become popular, at least partially due to a general belief that it would be overly burdensome and impractical in exploratory experiments. Here, I attempt to dispel those myths and provide practical steps for cell biologists to preregister their experiments and data analysis. Instead of hindering research activity, preregistration can aid experimentalists in designing better experiments, reducing waste, and adding rigor to their science.

While preregistration may sound daunting, it need not be. Most of us already write down a basic experimental plan in our lab notebook before we step into the lab, so uploading that to a repository (public or private) is a small effort. But that simple gesture could encourage researchers to flesh out their plans and more carefully consider potential pitfalls, like not having a sufficient sample size. There is nothing more annoying than completing weeks in the lab only to realize that you forgot an essential control or really needed an additional replicate to get a statistically significant result. Better planning can help avoid wasted effort, and preregistration helps. Getting feedback on your plan with labmates and colleagues is easier with a formal plan, and preregistration ensures that the PI and all the researchers on the project are on the same page.

We all know that arranging the data analysis procedure before seeing the results is good scientific hygiene. Preregistration forces us to make it a habit to plan the analysis, thereby reducing the temptation to p- and N-hack our way to “significance.” Furthermore, listing a priori plans and any deviations in the final paper adds transparency and trust to scientific publishing.

Ancillary Benefits

Taking the time to plan an adequate study design before stepping into the lab is of course beneficial in itself. But preregistering an analysis plan could actually increase statistical power and even reduce the number of required samples. For instance, if the hypothesis predicts a directional effect of a treatment, researchers can plan for a one-sided t-test, which requires fewer samples to achieve the same power (Lakens, 2022). While it wouldn’t be fair to opt for a one-sided test after collecting the data and learning the direction of the effect, documenting the bold prediction of a directional effect beforehand is in some ways superior to the standard two-sided test. (Convincing reviewers of this may be a challenge, but they should look favorably on the transparency and rigor of preregistration.)

Preregistration can further reduce the burden on cell biologists by allowing them to “peek” at their data partway through completion of an experiment, providing the opportunity to stop data collection with fewer replicates if the effect is sufficiently large. “N-hacking” is when researchers arbitrarily add samples until they reach a “significant” result, but with proper planning using sequential analysis—correcting the alpha level for multiple peeks—one can possibly stop early while maintaining the desired false positive rate (Lakens, 2022; Schott et al., 2019). A major benefit of preregistering a sequential analysis is that the expected number of biological replicates is actually smaller than the maximum sample size you would plan, because peeking allows the possibility of stopping the experiment early with an already significant p-value (Lakens, 2022). This is a common approach in clinical trials, and cell biologists could benefit from borrowing this statistically rigorous way to make our lives easier.

Encouraging Preregistration

Doing robust science is a reward in itself, and the vast majority of researchers already strive to do science right. If preregistration became a norm—just like displaying scale bars on microscopy images, providing raw data, uploading structural data to PDB, sharing code on GitHub, reporting sample sizes, including loading controls for Westerns, and showing uncropped gels in the supporting information—then scientists would eventually preregister as a habit.  

But some external encouragement couldn’t hurt. Journals and reviewers could ask for preregistration documentation, and reward authors who can produce such plans. For example, journals could offer some reassurance that null findings will be welcome, and when reviewer 2 asks for additional experiments or more replicates, editors should be emboldened by a preregistration to politely disregard such demands (when they are indeed unreasonable). And journals should prominently display a link or badge on papers that provide preregistration documentation.

A Practical Guide to Preregistering

The OSF repository provides resources for recording preregistration reports (Open Science Framework, 2023), including what to do when plans change, and others have expanded on the benefits and practical application of preregistration in small studies (Johnson and Cook, 2019). Here, we provide a crude template (Box 1) for a simple preregistration. See Box 2 for example paragraphs describing the preregistration in the final manuscript. As your plans develop while you traverse a project, you can augment the preregistered plan with additional experiments. Because most cell biology papers include several smaller experiments, the preregistered plan can include descriptions of multiple experiments, or you may opt to submit a separate short plan for each experiment.

Common Objections

I don’t want my competitors knowing my experiments before I even start.

Preregistrations need not be public to be effective. Depositing a preregistration privately (or even simply writing it down, timestamping it, and storing it on a local computer) still keeps researchers accountable and the research design more robust against post-hoc statistical exploration. The plan can eventually be made public to peer reviewers and readers of the ultimate publication. That said, making a plan public—or at least sending it to some trusted colleagues—offers an opportunity to get helpful feedback on an experiment before you even enter the lab.

My experiments are too exploratory to plan ahead.

It probably doesn’t make much sense to preregister pilot studies and research that is still in the beginning, exploratory stages. So maybe 90% of the work you do in the lab—culturing and observing cells, developing assays, purifying proteins, making and characterizing reagents, performing chemical analyses, and running pilot experiments—doesn’t need a preregistration. But before you sit down to start an actual experiment, it is worth planning: write down, discuss, and preregister how you will tackle the project. It is challenging to recognize precisely when research shifts from exploratory to a confirmatory experiment, and I don’t have a perfect prescription for identifying which work deserves preregistration, but most scientists know in their gut when their efforts are destined for a figure in a paper. When in doubt, just fill out the preregistration template (Box 1); you can always discard it if it turns out it wasn’t needed or update it as necessary as the project progresses. That shouldn’t take too much time, and it might help focus your attention on deficits in your plan early on.

This is too much work.

Preregistering adds little administrative burden, since the hypothesis, research plan, and often even the statistical analyses have already been laid out in research grant applications, lab notebooks, or thesis committee presentations. But taking the simple steps of formally organizing and uploading these plans to a repository ensures you have a clear strategy and helps avoid common pitfalls of poor experimental planning. Each experiment can be preregistered on its own, so there’s no need to decide what the ultimate paper will be during the early planning stages.

This will lock me into a plan that might send me down the wrong path.

Preregistration is a plan, not a prison (Open Science Framework, 2023). If plans change before collecting the data, it’s as simple as updating the preregistered documents. If you perform unplanned tests or data analysis during or after the experiment, authors can transparently report deviations from the plan and justifications in the final paper. And they can do additional experiments not listed in their original preregistration and either update the preregistration or simply note in the final manuscript that that particular experiment had not been listed in the original preregistered plan. This adds transparency and empowers the reader with useful information. Ultimately, if you completely ditch your preregistered plan and don’t even mention it in your paper, while this is dishonest, it’s no worse off than the status quo. So there’s really no harm in experimenting with preregistration. 

My Personal Experience with Preregistration

This is where I am seeking input from anyone who has ever preregistered their experiments before. For my next experiment, I will preregister the plan and the analysis approach and include that when we submit to a journal. But that will be months away. Has anyone already done this and want to expound upon the process and the outcome?

Conclusion

The value of preregistration has been demonstrated for years in clinical research and other fields, and it is time to apply this proven technique to cell biology. Of course, preregistration is not magic: it will not singlehandedly solve the replication crisis, eliminate fraud, or make all research findings true. Improving science will take a concerted effort by journals, authors, and funders to better incentivize transparent protocols and more rigorous testing of findings (Lord et al., 2023; Forstmeier et al., 2017; Kohrs et al., 2023). But this small yet powerful step would increase rigor in scientific research and help strengthen the literature.

References

  • Forstmeier, W., E.-J. Wagenmakers, and T.H. Parker. 2017. Detecting and avoiding likely false-positive findings – a practical guide. Biol. Rev. Camb. Philos. Soc. 92:1941–1968. doi:10.1111/brv.12315.
  • Johnson, A.H., and B.G. Cook. 2019. Preregistration in Single-Case Design Research. Except. Child. 86:95–112. doi:10.1177/0014402919868529.
  • Kohrs, F.E., S. Auer, A. Bannach-Brown, S. Fiedler, T.L. Haven, V. Heise, C. Holman, F. Azevedo, R. Bernard, A. Bleier, N. Bössel, B.P. Cahill, L.J. Castro, A. Ehrenhofer, K. Eichel, M. Frank, C. Frick, M. Friese, A. Gärtner, K. Gierend, and T.L. Weissgerber. 2023. Eleven strategies for making reproducible research and open science training the norm at research institutions. eLife. 12. doi:10.7554/eLife.89736.
  • Lakens, D. 2022. Improving Your Statistical Inferences.
  • Lord, S.J., A. Charles-Orszag, K. Skruber, R.D. Mullins, and A. Rehfeld. 2023. Peer Replication. Zenodo. doi:10.5281/zenodo.10067391.
  • Nosek, B.A., C.R. Ebersole, A.C. DeHaven, and D.T. Mellor. 2018. The preregistration revolution. Proc Natl Acad Sci USA. 115:2600–2606. doi:10.1073/pnas.1708274114.
  • Open Science Framework. 2023. Preregistration – OSF Support. Schott, E., M. Rhemtulla, and K. Byers-Heinlein. 2019. Should I test more babies? Solutions for transparent data peeking. Infant Behav. Dev. 54:166–176. doi:10.1016/j.infbeh.2018.09.010.

____________________________________________________

____________________________________________________

Box 2: Example paragraphs for eventual manuscript

We preregistered our research plans and statistical analyses (link to plans).

Based on a small pilot experiment, we predicted that abc drug treatment would substantially reduce cell speeds relative to control, so we proposed a one-tailed t-test with α=0.05. After the proposed n=6 biological replicates (3 control and 3 treated), the p-value was 0.026. However, we noticed after the data was collected that there was substantial variation in the speed of the control cells on different days. Because control and treated samples were split from the same flask of cells for each day’s experiment, we subsequently concluded that a paired t-test was the more appropriate approach and therefore departed from our preregistered analysis plan. The one-sided paired t-test resulted in p=0.011. Results are shown in Fig 3.

Also, after observing the full results, we noted that the nuclei seemed smaller in the drug-treated samples. Although not in our original preregistered plan, we decided to also compare the size of the nuclei in control vs. treated. We blinded the filenames and measured the area of the nuclei of all the cells in the dataset. The two-tailed paired t-test resulted in a p=0.034. Results are shown in Fig 3.

For the neuronal aging experiment, we proposed using a paired design, so we planned to use male mice from the same litter. We proposed to sacrifice mice at different times, fix and stain brain tissue, and image the average intensity of xyz marker in 100 neurons from each mouse. Sample collection was performed blinded with respect to the identity/age of the sample. We proposed a two-tailed t-test to calculate the p-value, because we did not know whether the marker would increase or decrease with age. To potentially reduce the number of mice needed, we planned a sequential analysis where we would calculate the p-value at n=4 and 8 mouse pairs; to account for multiple assessments of the data, we proposed the Pocock correction of α=0.029 for each look at the data, for a cumulative α of 0.05. After n=4 mice, we calculated p=0.022, so we halted the experiment according to our plan. Results are shown in Fig 5.

PromoPlot

January 23, 2026 at 4:23 pm | | crazy figure contest, literature, nerd

This is hilarious: https://arxiv.org/abs/2503.24254

end grant writing; award prizes for papers instead

January 20, 2026 at 4:52 pm | | literature, science and the public, science community, scientific integrity

My proposal for funding academic research science: instead of funding grants, just award financial prizes for good papers.

Implementation

Award half the prize upon publication (or preprint); award the second half upon an independent replication of the results, 25% to the original authors and 25% to the replicators. (If a good-faith replication attempt fails, the replicators would still receive their portion, but the original authors would not receive their replication bonus. The measure for the quality of the replication would be whether the work could be published in a reputable journal.)

Prize winning papers would be chosen by rotating committees of professors following a rubric announced prior to the prize committee meeting (replacing grant study sections). Different committees could be assembled for different research topics, with appropriately tailored rubrics. Any paper published in the prior 12 months would be eligible during each prize committee meeting (there could be multiple committees convened each year, giving any given paper multiple chances at winning). Some committees would be assembled to identify “sleeper” papers: works that are up to 5 years old that had been overlooked by previous committees but have proven to be especially valuable.

The prize amount could vary by field or even by paper. A typical R01 grant is $150k+ annually for 5 years. So prizes could vary from $200k up to $1M+ for a major tour-de-force paper with substantial impact. The goal would be to distribute the same pot of money to approximately the same number of labs in roughly the same proportion, but with a much smaller bureaucratic burden and with better incentives for reproducibility.

There would probably need to be some cap on the total prize money an individual PI could receive at any one time. If they are over the cap, then either their papers would simply not be eligible for a prize that round or their prize money would flow to the collaborators on that paper or department.

Prize money would still go to the PI and lab via the university, not as a blank check to an individual professor. Indirects/overhead would still be applied. The goal would be to have approximately the same amount of money flowing to the university and the labs, but with a less onerous process.

Advantages

The prize approach accomplishes the following:

  • Rewards actual outputs rather than promises.
  • Creates strong incentives for reproducible work.
  • Incentivizes impactful, innovative results that are convincing and accepted by the scientific community.
  • Frees up professors’ time to focus on paper writing and mentoring instead of grant writing.
  • Reduces the workload for university grant offices, which currently spend considerable time reviewing and approving proposals before submission, thereby reducing costs and freeing administrative staff at universities and funding agencies to focus on other support functions.
  • Removes back-and-forth about budgets, justifications, biosketches, data-sharing policies, etc.
  • Does away with byzantine formatting requirements.
  • Obviates1 the need for progress reports and renewals throughout the grant period.

Professors spend a huge amount of time writing grants, and university grant offices dedicate substantial resources helping researchers navigate complex formatting and compliance requirements, including page lengths, fonts, hyperlinks, sections, etc. Academic researchers currently focus on publishing their work, so grants should reinforce that drive, not distract PIs with hundreds of hours of grant writing. Prize committees would replace grant review panels (roughly equivalent work), while reducing the administrative burden on both university grant offices and granting agencies.

The current grant funding structure purports to be forward-looking, providing funding for good ideas to be implemented. But actual grants require not only a good track record for the PI, but also a large amount of preliminary data. So, in effect, grants fund a lot of work that’s already been completed, and the awarded money usually gets spent on lab work outside the scope of the original proposal. So, while appearing to be prospective, grants are often retrospective in practice. Granting prizes for good papers simply makes official how funding already flows in reality.

How to fund research before winning prizes

The central challenge would be providing runway funding, especially for early-career researchers who have no eligible papers yet. With fewer resources needed for proposal preparation support, some overhead funds could be redirected to departmental slush funds. Departments could then provide startup packages, bridge funding between prizes, or support for labs that demonstrate progress toward publishable work.

This would shift evaluation of proposed research from granting agencies to local departments, where a lab’s potential is more knowable. However, departments are often rife with unhealthy politics and lack sufficient guardrails against favoritism and bias. Whether departmental evaluation would be less burdensome than current grants – and whether freed-up overhead would provide sufficient funding – remains an open question.

Alternatively, granting agencies could dedicate some funds for prizes awarded specifically to a professor’s first paper, or maintain a smaller pot of traditional proposal-style funding for early-career researchers. But as it stands, professors who don’t publish within their first few years are unlikely to receive traditional grants or be awarded tenure anyway, so the prize approach doesn’t necessarily create more pressure than the current system.

Downsides

Not every good paper would win a prize. Yes, of course there would still be disappointing results: inevitably, some deserving papers would fail to attract money because the subject matter is out of fashion or reviewer bias or a host of other unfair reasons. But with a limited amount of money and an effectively unlimited number of professors, it’s impossible to fund everyone, so someone will always be neglected. Winning a prize would be no more unpredictable than receiving a good score on a grant. My proposal is no less fair than the whims of the current granting process.

Incentivizes short-term work. Scientists would be less motivated to tackle a longer-term project, even if it is promising. The department or other funding agency would need to step in to float the project for years until it comes to fruition. And if whoever funds that float requires proposals to have the same level of detail and complexity as current grant proposals, we’d be back to square one. But proposals for future work need not be as onerous as current NIH regulations, especially if funded by the department or by private organizations like HHMI. Another source of funding for longer-term projects may actually come from paper prizes: professors who have won multiple prizes in the recent past would have ample money to fund collaborations and longer projects.

Disincentivizes high-risk research. Researchers might be wary of taking on high-risk/high-reward projects, because, if the experiments fail, they couldn’t expect a prize. That said, if such a project succeeds, then the prize money would likely be larger. Such is the nature of high-risk/high-reward. Therefore, the prize money for impactful work would need to be sufficiently large to encourage at least a portion of researchers to take on projects that are higher risk. Furthermore, if researchers are able to publish details of a failed “moonshot” that advanced the field despite not succeeding, such a paper would be eligible for a prize. The flexibility of prize funding might actually enable more ambitious research, since successful researchers would have discretion over how to deploy their winnings without the constraints of grant-specified aims.

Disincentivizes collaborations. Collaborators would likely negotiate some cut of a potential prize ahead of time, or there would be a standard formula (e.g. 75% to the last author and the remainder of the prize to the other PI collaborators on the paper). Maybe PIs would be less willing to add collaborators to a project, for fear that they may need to share some of the prize money with other PIs. Maybe there could be a bonus to help support collaborators.

Expensive projects. This would not work for huge projects like clinical trials or massive particle physics consortiums, which would need ample funding before the experiments or construction actually begin. But those could continue to be funded in the current fashion.

What will professors do with their freed up time? Part of the job of a professor is to write grants. If they suddenly don’t need to write grants, will they actually spend more time teaching and mentoring, or will they just work less? Not sure. That’s a valid point.

Moving forward

The next step would be to run a pilot. Initially, 20% of NIH and NSF funding could be converted to this model in select fields that would be amenable, then evaluated after 5 years. At that time, the percentage of funding provided via the prize mechanism might be increased or decreased based on the outcome of the pilot. It’s unlikely that granting agencies would ever completely eliminate the traditional grant proposal, but the health of academic science could be strengthened if even a minority of the funding rewarded impactful, reproducible papers.

___________________

  1. Thesaurus, don’t fail me now. ↩︎

eLife’s new publishing policy

October 21, 2022 at 10:25 am | | literature, science community, scientific integrity

Let me preface this post with the admission that I’m likely wrong about my concerns about eLife’s new model. I was actually opposed to preprints when I first heard about them in 2006!

The Journal eLife and Mike Eisen announced it’s new model for publishing papers:

  • Authors post preprint.
  • Authors submit preprint to eLife.
  • eLife editorial board decides whether to reject the manuscript or send out for review.
  • After reviews, the paper will be published no matter what the reviews say. The reviews and an eLife Assessment will be published alongside the paper. At this point, the paper has a DOI and is citable.
  • Authors then have a choice of whether to revise their manuscript or just publish as-is.
  • When the authors decide the paper is finalized, that will become the “Version of Record” and the paper will be indexed on Pubmed.

Very interesting and bold move. The goal is to make eLife and its peer review not a gatekeeper of truth, but instead a system of evaluating and summarizing papers. Mike Eisen hopes that readers will see “eLife” and no longer think “that’s probably good science” and instead think “oh, I should read the reviews to see if that’s a good paper.”

Potential problems

But here are my primary concerns

  1. This puts even more power in the hands of the editors to effectively accept/reject papers. And this process is currently opaque, bias-laden, and authors have no recourse when editors make bad decisions.
  2. The idea that the eLife label will no longer have prestige is naive. The journal has built a strong reputation as a great alternative to the glam journal (Science, Nature, Cell) and that’s not going away. For example, people brag when their papers are reviewed in F1000, and I think the same will apply to eLife Assessments: readers will automatically assume that a paper published in eLife is high-impact, regardless of what the Assessment says.
  3. The value that eLife is adding to the process is diminishing, and the price tag is steep ($2000).
  4. The primary problem I have with peer review is that it is simultaneously overly burdensome and not sufficiently rigorous. This model doesn’t substantially reduce the burden on authors to jump through hoops held by the reviewers (or risk a bad eLife Assessment). It also is less rigorous by lowering the bar to “publication.”

Solutions

Concern #1: I think it’s a step in the wrong direction to grant editors even more power. Over the years, editors haven’t exactly proven themselves to be ideal gatekeepers. How can we ensure that the editors will act fairly and don’t get attracted by shiny objects? That said, this policy might actually put more of a spotlight on the desk-rejection step and yield change. eLife could address this concern in various ways:

  • The selection process could be a lottery (granted, this isn’t ideal because finding editors and reviewers for a crappy preprint will be hard).
  • Editors could be required to apply a checklist or algorithmic selection process.
  • The editorial process could be made transparent by publishing the desk rejection/acceptace along with the reasons.

Concern #2 might resolve itself with time. Dunno. Hard to predict how sentiment will change. But I do worry that eLife is trying to change the entire system, while failing to modify any of the perverse incentives that drive the problems in the first place. But maybe it’s better to try something than to do nothing.

Concern #3 is real, but I’m sure that Mike Eisen would love it if a bunch of other journals adopted this model as well and introduced competition. And honestly, collating and publishing all the reviews and writing a summary assessment of the paper is more than what most journals do now.

Journals should be better gatekeepers

But #4 is pretty serious. The peer review process has always had to balance being sufficiently rigorous to avoid publishing junk science with the need to disseminate new information on a reasonable timescale. Now that preprinting is widely accepted and distributing results immediately is super easy, I am less concerned with latter. I believe that the new role of journals should be as more exacting gatekeepers. But it feels like eLife’s policy was crafted exclusively by editors and authors to give themselves more control, reduce the burden for authors, and shirk the responsibility of producing and vetting good science.

There are simply too many low-quality papers. The general public, naive to the vagaries of scientific publishing, often take “peer-reviewed” papers as being true, which is partially why we have a booming supplement industry. Most published research findings are false. Most papers cannot be replicated. Far too many papers rely on pseudoreplication to get low p-values or fail to show multiple biological replicates. And when was the last time you read a paper where the authors blinded their data acquisition or analysis?

For these reasons, I think that the role of a journal in the age of preprints is to better weed out low-quality science. At minimum, editors and peer reviewers should ensure that authors followed the 3Rs (randomize, reduce bias, repeat) before publishing. And there should be a rigorous checklist to ensure that the basics of the scientific process were followed.

Personally, I think the greatest “value-add” that journals could offer would be to arrange a convincing replication of the findings before publishing (peer replication), then just do away with the annoying peer review dog-and-pony show altogether.

Conclusion

We’ll have to wait and see how this new model plays out, and how eLife corrects stumbling blocks along the way. I have hope that, with good editorial team and good practices/rules around the selection process, eLife might be able to pull this off. Not sure if it’s a model that will scale to other, less trustworthy journals.

But just because this isn’t my personal favorite solution to the problem of scientific publishing, that doesn’t mean that eLife’s efforts won’t help make a better world. I changed my mind about the value of preprints, and I’ll be happy to change my mind about eLife’s new publishing model if it turns out to be a net good!

Checklist for Experimental Design

October 25, 2021 at 9:42 am | | everyday science, scientific integrity

One of the worst feelings as a scientist is to realize that you performed n-1 of the controls you needed to really answer the question you’re trying to solve.

My #1 top recommendation when starting a new set of experiments is to write down a step-by-step plan. Include what controls you plan to do and how many replicates. Exact details like concentrations are less important than the overarching experimental plan. Ask for feedback from your PI, but also from a collaborator or labmate.

Here are some things I consider when starting an experiment. Feel free to leave comments about things I’ve missed.

Randomization and Independence

  • Consider what “independent” means in the context of your experiment.
  • Calculate p-value using independent samples (biological replicates), not the total number of cells, unless you use advanced hierarchical analyses (see: previous blog post and Lord et al.: https://doi.org/10.1083/jcb.202001064).
  • Pairing or “blocking” samples (e.g. splitting one aliquot into treatment and control) helps reduce confounding parameters, such as passage number, location in incubator, cell confluence, etc.

Power and Statistics

  • Statistical power is to the ability to distinguish small but real differences between conditions/populations.
  • Increasing the number of independent experimental rounds (AKA “biological replicates”) typically has a much larger influence on power than the cells or number of measurements per sample (see: Blainey et al.: https://doi.org/10.1038/nmeth.3091).
  • If the power of an assay is known, you can calculate the number of samples required to be confident you will be able to observe an effect.
  • Consider preregistering
    • Planning the experimental and analysis design before starting the experiment substantially reduces the chances false positives.
    • Although formal preregistration is not typically required for cell biology studies, simply writing the plan down for yourself in your notebook is far better than winging it as you go.
    • Plan the number of biological replicates before running statistical analysis. If you instead check if a signal is “significant” between rounds of experiment and stop when p < 0.05, you’re all but guaranteed to find a false result.
    • Similarly, don’t try several tests until you stumble upon one that gives a “significant” p-value.

Blocking

  • “Blocking” means to subdivide samples into similar units and running those sets together. For example, splitting a single flask of cells into two, one treatment and one control.
  • Blocking can help reveal effects even when experimental error or sample-to-sample variability is large.
  • Blocked analyses include paired t-test or normalizing the treatment signal to the control within each block.

Controls

  • Sledgehammer controls
    • These controls are virtually guaranteed to give you zero or full signal, and are a nice simple test of the system.
    • Examples include wild type cells, treating with DMSO, imaging autofluorescence of cells not expressing GFP, etc.
  • Subtle controls
    • These controls are more subtle than the strong controls, and might reveal some unexpected failures.
    • Examples include: using secondary antibody only, checking for bleed-through and crosstalk between fluorescence channels, and using scrambled siRNA.
  • Positive & negative controls
    • The “assay window” is a measure of the range between the maximum expected signal (positive control) and the baseline (negative control).
    • A quantitative measure of the assay window could be a standardized effect size, like Cohen’s d, calculated with multiple positive and negative controls.
    • In practice, few cell biologist perform multiple control runs before an experiment. So a qualitative estimate of the assay window should be considered using the expected signal and expected variability sample to sample. In other words, consider carefully if an experiment can possibly work
  • Concurrent vs historical controls
    • Running positive and/or negative control in the same day’s experimental run as the samples that receive real treatment helps eliminate additional variability.
  • Internal controls
    • “Internal” controls are cells within the same sample that randomly receive treatment or control. For example, during a transient transfection, only a portion of the cells may actually end up expressing, while those that aren’t can act as a negative control.
    • Because cells with the same sample experience the same perturbations (such as position in incubator, passage number, media age) except for the treatment of interest, internal controls can remove many spurious variables and make analysis more straightforward.

Bias

  • Blinding acquisition
    • Often as simple as having a labmate put tape over labels on your samples and label them with a dummy index. Confirm that your coworker actually writes down the key, so later you can decode the dummy index back to the true sample information.
    • In cases where true blinding is impractical, the selection of cells to image/collect should be randomized (e.g. set random coordinates for the microscope stage) or otherwise designed to avoid bias (e.g. selecting cells using transmitted light or DAPI).
  • Blinding analysis
    • Ideally, image analysis would be done entirely by algorithms and computers, but often the most practical and effective approach is old-fashioned human eye.
    • Ensuring your manual analysis isn’t biased is usually as simple as scrambling filenames. For microscopy data, Steve Royle‘s macro, which works well: https://github.com/quantixed/imagej-macros#blind-analysis
    • I would highly recommend copying all the data to a new folder before you perform any filename changes. Then test the program forward and backwards to confirm everything works as expected. Maybe perform analysis in batches, so in case something goes awry, you don’t lose all that work.

Resources

Great primer on experimental design and analysis, especially for the cell biologist or microscopist: Stephen Royle, “The Digital Cell: Cell Biology as a Data Science” https://cshlpress.com/default.tpl?action=full&–eqskudatarq=1282

Advanced, detailed (but easily digestible) book on experimental design and statistics: Stanley Lazic, “Experimental Design for Laboratory Biologists” https://stanlazic.github.io/EDLB.html

I like this very useful and easy-to-follow stats book: Whitlock & Schluter, “The Analysis of Biological Data” https://whitlockschluter.zoology.ubc.ca

Alex Reinhart, “Statistics Done Wrong”

Tomek & Eisner, “Basic Statistics for Life Scientists”

SuperPlots: Communicating reproducibility and variability in cell biology. (HTMLPDF)
Lord, S. J.; Velle, K. B.; Mullins, R. D.; Fritz-Laylin, L. K. J. Cell Biol.2020219(6), e202001064.

Replace Peer Review with “Peer Replication”

October 13, 2021 at 1:35 pm | | literature, science and the public, science community, scientific integrity

UPDATE: EMBO Reports article here and a write-up in Nature here.

As I’ve posted before and many others have noted, there is a serious problem with lack of adequate replication in many fields of science. The current peer review process is a dreadful combination of being both very fallible and also a huge hurdle to communicating important science.

Instead of waiting for a few experts in the field to read and apply their stamp of approval to a manuscript, the real test of a paper should be the ability to reproduce its findings in the real world. (As Andy York has pointed out, the best test of a new method is not a peer reviewing your paper, but a peer actually using your technique.) But almost no published papers are subsequently replicated by independent labs, because there is very little incentive for anyone to spend time and resources testing an already published finding. That is precisely the opposite of how science should ideally operate.

Let’s Replace Traditional Peer Review with “Peer Replication”

Instead of sending out a manuscript to anonymous referees to read and review, preprints should be sent to other labs to actually replicate the findings. Once the key findings are replicated, the manuscript would be accepted and published.

(Of course, as many of us do with preprints, authors can solicit comments from colleagues and revise a manuscript based on that feedback. The difference is that editors would neither seek such feedback nor require revisions.)

Along with the original data, the results of the attempted replication would be presented, for example as a table that includes which reagents/techniques were identical. The more parameters that are different between the original experiment and the replication, the more robust the ultimate finding if the referees get similar results.

A purely hypothetical example of the findings after referees attempt to replicate. Of course in reality, my results would always have green checkmarks.

Incentives

What incentive would any professor have to volunteer their time (or their trainees’ time) to try to reproduce someone else’s experiment? Simple: credit. Traditional peer review requires a lot of time and effort to do well, but with zero reward except a warm fuzzy feeling (if that). For papers published after peer replication, the names of researchers who undertook the replication work will be included in the published paper (on a separate line). Unlike peer review, the referees will actually receive compensation for their work in the form of citations and another paper to include on their CV.

Why would authors be willing to have their precious findings put through the wringer of real-world replication? First and foremost, because most scientists value finding truth, and would love to show that their findings hold up even after rigorous testing. Secondly, the process should actually be more rewarding than traditional peer review, which puts a huge burden on the authors to perform additional experiments and defend their work against armchair reviewers. Peer replication turns the process on its head: the referees would do the work of defending the manuscript’s findings.

Feasible Experiments

There are serious impediments to actually reproducing a lot of findings that use seriously advanced scientific techniques or require long times or a lot of resources (e.g. mouse work). It will be the job of editors—in collaboration with the authors and referees—to determine the set of experiments that will be undertaken, balancing rigor and feasibility. Of course, this might leave some of the most complex experiments unreplicated, but then it would be up to the readers to decide for themselves how to judge the paper as a whole.

What if all the experiments in the paper are too complicated to replicate? Then you can submit to JOOT.

Ancillary Benefits

Peer replication transforms the adversarial process of peer review into a cooperation among colleagues to seek the truth. Another set of eyes and brains on an experiment could introduce additional controls or alternative experimental approaches that would bolster the original finding.

This approach also encourages sharing experimental procedures among labs in a manner that can foster future collaborations, inspire novel approaches, and train students and postdocs in a wider range of techniques. Too often, valuable hands-on knowledge is sequestered in individual labs; peer replication would offer an avenue to disseminate those skills.

Peer replication would reduce fraud. Often, the other authors on an ultimately retracted paper only later discover that their coworker fabricated data. It would be nearly impossible for a researcher to pass off fabricated data or manipulated images as real if other researchers actually attempt to reproduce the experimental results. 

Potential Problems

One serious problem with peer replication is the additional time it may take between submission and ultimate publication. On the other hand, it often takes many months to go through the traditional peer review process, and replicating experiments may not actually add any time in many cases. Still this could be mitigated by authors submitting segments of stories as they go. Instead of waiting until the entire manuscript is polished, authors or editors could start arranging replications while the manuscript is still in preparation. Ideally, there would even be a  journal-blind mechanism (like ReviewCommons) to arrange reproducing these piecewise findings.

Another problem is what to do when the replications fail. There would still need to be a judgement call as to whether the failed replication is essential to the manuscript and/or if the attempt at replication was adequately undertaken. Going a second round at attempting a replication may be warranted, but editors would have to be wary of just repeating until something works and then stopping. Pre-registering the replication plan could help with that. Also, including details of the failed replications in the published paper would be a must.

Finally, there would still be the problem of authors “shopping” their manuscript. If the replications fail and the manuscript is rejected, the authors could simply submit to another journal. I think the rejected papers would need to be archived in some fashion to maintain transparency and accountability. This would also allow some mechanism for the peer replicators to get credit for their efforts.

Summary of Roles:

  • Editor:
    • Screen submissions and reject manuscripts with obviously flawed science, experiments not worth replicating, essential controls missing, or seriously boring results.
    • Find appropriate referees.
    • With authors and referees, collaboratively decide which experiments the referees should attempt to replicate and how.
    • Ultimately conclude, in consultation with referees, whether the findings in the papers are sufficiently reproducible to warrant full publication.
  • Authors:
    • Write the manuscript, seek feedback (e.g. via bioRxiv), and make revisions before submitting to the journal.
    • Assist referees with experimental design, reagents, and even access to personnel or specialized equipment if necessary.
  • Referees:
    • Faithfully attempt to reproduce the experimental results core to the manuscript.
    • Optional: Perform any necessary additional experiments or controls to close any substantial flaws in the work.
    • Collate results.
  • Readers:
    • Read the published paper and decide for themselves if the evidence supports the claims, with the confidence that the key experiments have been independently replicated by another lab.
    • Cite reproducible science.

How to Get Started

While it would be great if a journal like eLife simply piloted a peer replication pathway, I don’t think we can wait for Big Publication to initiate the shift away from traditional peer review. Maybe the quickest route would be for an organization like Review Commons to organize a trial of this new approach. They could identify some good candidates from bioRxiv and, with the authors, recruit referees to undertake the replications. Then the entire package could be shopped to journals.

I suspect that once scientists see peer replication in print, it will be hard to take seriously papers vetted only by peer review. Better science will outcompete unreproduced findings.

(Thanks Arthur Charles-Orszag for the fruitful discussions!)

Nobel predictions over the years

October 6, 2021 at 1:16 pm | | nobel

2008 Tsien (predicted 2008)

2010 Suzuki & Heck (predicted among others in 2010)

2012 Kobilka (among my 6 predictions in 2012)

2013 Higgs (among my 2 predictions in 2013)

2014 Moerner (among my 3 predictions in 2013); Betzig & Hell (predicted among others in 2010)

2018 Allison (predicted in 2018)

2019 Goodenough (predicted 2019)

2020 Doudna & Charpentier (predicted 2020)

2021 Nobel Prize Predictions

September 16, 2021 at 8:07 am | | nobel

Medicine: Adenovirus vector vaccines

Medicine (alternate): A second Nobel Prize for Ivermectin ;)

Chemistry: Katalin Karikó, Drew Weissman, Ugur Sahin, Özlem Türeci, & Robert Malone, for developing mRNA & liposome vaccine technology.

Peace: Mike Pence, for following his oath of office and standing up to pressure and physical threats to carry out his Constitutional duties on January 6.

2020 Nobel prize predictions

September 20, 2020 at 4:11 pm | | nobel

CRISPR: Jennifer Doudna, Emmanuelle Charpentier, Francisco Mojica, Virginijus Siksnys EDIT: Announcement.

(This year, I’m only predicting the chemistry prize.)

avoiding bias by blinding

July 3, 2020 at 11:15 am | | literature, scientific integrity

Key components of the scientific process are: controls, avoiding bias, and replication. Most scientists are great at controls, but without the other two, we’re simply not doing science.

The lack independent samples (and thus improper inflation of n) and the failure to blind experiments are too common. The implications of these mistakes, especially when combined in one study, mean that many published cell biology results are likely artifact. Generating large datasets, even with a slight bias, can quickly yield “significant” results out of noise. For example, see the “NHST is unsuitable for large datasets” section of:

Szucs D, Ioannidis JPA. When Null Hypothesis Significance Testing Is Unsuitable for Research: A ReassessmentFront Hum Neurosci. 2017;11:390. https://pubmed.ncbi.nlm.nih.gov/28824397/

Now combine these common false positives with the inclination to publish flashy results, and we’ve made a recipe for unreliable scientific literature.

I do not condemn authors for these problems. Most of us have made one or all of these mistakes. I have. And I probably will again in the future. Science is hard. There is no shame in making honest mistakes. But we can all strive to be better (see my last section).

Failing to perform the data collection blinded

Blinding is just basic scientific rigor, and skipping this should be considered almost as bad as skipping controls.

Blinding samples during data collection and analysis is ideal. For data collection, it is usually as simple as having a labmate put tape over labels on your samples and label them with a dummy index. Insist that your coworker writes down the key, so later you can decode the dummy index back to the true sample information.

In cases where true blinding is impractical, the selection of cells to image/collect should be randomized (e.g. set random coordinates for the microscope stage) or otherwise designed to avoid bias (e.g. selecting cells using transmitted light if fluorescence the readout).

Failing to perform the data analysis blinded

Blinding during data analysis is generally very practical, even when the original data was not collected in a bias-free fashion. Ideally, image analysis would be done entirely by algorithms and computers, but often the most practical and effective approach is old-fashioned human eye. Ensuring your manual analysis isn’t biased is usually as simple as scrambling the image filenames.

I stumbled upon these ImageJ macros for randomizing and derandomizing image filenames, written by Martin Höhne: http://imagej.1557.x6.nabble.com/Macro-for-Blind-Analyses-td3687632.html

More recently, Christophe Leterrier directed me to Steve Royle‘s macro, which works very well: https://github.com/quantixed/imagej-macros#blind-analysis

There are probably some excellent solution using Python. Regardless of the approach you take, I would highly recommend copying all the data to a new folder before you perform any filename changes. Then test the program forward and backwards to confirm everything works as expected. Maybe perform analysis in batches, so in case something goes awry, you don’t lose all that work.

My “blind and dumb” experiment

There are many stories about unintended bias leading to false conclusions. Here’s mine: I was testing to see whether a drug treatment inhibited cells from crawling through a porous barrier by counting the number of cells that made it through the barrier to an adjacent well.

My partner in crime had labeled the samples with dummy indices, so I didn’t know which wells were treated and which were control. But I immediately could tell that there were more cells in one set of wells, so I presumed those were the control set. Fortunately, I had taken the extra precaution of randomizing the stage positions, so I didn’t let my bias alter the data collection. We then blinded the analysis by relabeling the microscopy images. I manually counted all the cells in each image.

We then unblinded the samples. At first, we were disappointed that the wells I had assumed were control turned out to be treated. Then we looked at the results. SURPRISE! My snap judgement at the beginning of the experiment had been precisely backwards: the wells I thought looked like they had sparser cells actually had significantly more on average. So it turned out that the drug treatment had indeed worked. Thankfully, I didn’t rely on my snap judgement nor allow that bias to influence the results.

Treating each cell as an n in the statistical analysis

This error plagues the majority of cell biology papers. Go scan a recent issue of your favorite journal and count the number of papers that have minuscule P values; invariably, the authors aggregated all the cell measurements from multiple experiments and calculated the t-test or ANOVA based on those dozens or hundreds of measurements. This is a fatal error.

It is patently absurd to consider neighboring cells in the same dish all treated simultaneously with the same drug as independent tests of a hypothesis.

If your neighbor told you that he ate a banana peal and it reversed his balding, you might be a little skeptical. If he further explained that he measured 1000 of his hair follicles before and after eating a banana peel and measured a P < 0.05 difference in growth rate, would you be convinced? Maybe it was just a fluke or noise that his hairs started growing faster. You would want him to repeat the experiment a few times (maybe even with different people) before you started believing.

Similarly, there are many reasons two dishes of cells might be different. To start believing that a treatment is truly effective, we all understand that we should repeat the experiment a few times and get similar results. Counting each cell measurement as the sample size n all but guarantees a small—but meaningless—P value.

Observe how dramatically different scenarios (on the right) yield the same plot and P value when you assume each cell is a separate n (on the left):

Elegant solutions include “hierarchical” or “nested” or “mixed effect” statistics. A simple approach is to separately pool the cell-level data from each experiment, then compare experiment-level means (n the case above, the n for each condition would be 3, not 300). For more details, please read my previous blog post or our paper:

Lord SJ, Velle KB, Mullins RD, Fritz-Laylin LK. SuperPlots: Communicating reproducibility and variability in cell biology. J Cell Biol. 2020;219(6):e202001064.
https://pubmed.ncbi.nlm.nih.gov/32346721

How do we fix this?

Professors need to teach their trainees the very basics about how to design experiments (see Stan Lazic’s book: Experimental Design for Laboratory Biologists) and perform analysis (see Mike Whitlock and Dolph Schluter’s book: The Analysis of Biological Data). PIs need to provide researchers with the tools to blind their experiments or otherwise remove bias. They need to ask for multiple biological replicates and correctly calculated P values. This does not require advanced understanding of statistics, just the basic understanding of the importance of repeating an experiment multiple times to ensure an observation is real.

Editors and referees need to demand correct data analysis. While asking researchers to redo an experiment isn’t really acceptable, requiring a reanalysis of the data after blinding or recalculating P values based on biological replicates seems fair. Editors should not even send manuscripts to referees if the above errors are not corrected or at least addressed in some fashion. Editors can offer the simple solutions listed above.

UPDATE: Also read my proposal to replace peer review with peer “replication.”

unbelievably small P values?

November 18, 2019 at 9:56 am | | literature, scientific integrity

Check out our newest preprint at arXiv:

If your P value looks too good to be true, it probably is: Communicating reproducibility and variability in cell biology

Lord, S. J.; Velle, K. B.; Mullins, R. D.; Fritz-Laylin, L. K. arXiv 2019, 1911.03509. https://arxiv.org/abs/1911.03509

UPDATE: Now published in JCB: https://doi.org/10.1083/jcb.202001064

I’ve noticed a promising trend away from bar graphs in the cell biology literature. That’s great, because reporting simply the average and SD or SEM or an entire dataset conceals a lot of information. So it’s nice to see column scatter, beeswarm, violin, and other plots that show the distribution of the data.

But a concerning outcome of this trend is that, when authors decide to plot every measurement or every cell as a separate datapoint, it seems to trick people into thinking that each cell is an independent sample. Clearly, two cells in the same flask treated with a drug are not independent tests of whether the drug works: there are many reasons the cells in that particular flask might be different from those in other flasks. To really test a hypothesis that the drug influences the cells, one must repeat the drug treatment multiple times and check if the observed effect happens repeatably.

I scanned the latest issues of popular cell biology journals and found that over half the papers counted each cell as a separate N and calculated P values and SEM using that inflated count.

Notice that bar graphs—and even beeswarm plots—fail to capture the sample-to-sample variability in the data. This can have huge consequences: in C, the data is really random, but counting each cell as its own independent sample results in minuscule error bars and a laughably small P value.

But that’s not to say the the variability cell-to-cell is unimportant! The fact that some cells in a flask react dramatically to a treatment and others carry on just fine might have very important implications in an actual body.

So we proposed “SuperPlots,” which superimpose sample-to-sample summary data on top of the cell-level distribution. This is a simple way to convey both variability of the underlying data and the repeatability of the experiment. It doesn’t really require any complicated plotting or programming skills. On the simplest level, you can simply paste two (or more!) plots in Illustrator and overlay them. Play around with colors and transparency to make it visually appealing, and you’re done! (We also give a tutorial on how we made the plots above in Graphpad Prism.)

Let me know what you think!

UPDATE: We simplified the figure:

Figure 1. Drastically different experimental outcomes can result in the same plots and statistics unless experiment-to-experiment variability is considered. (A) Problematic plots treat N as the number of cells, resulting in tiny error bars and P values. These plots also conceal any systematic run- to-run error, mixing it with cell-to-cell variability. To illustrate this, we simulated three different scenarios that all have identical underlying cell-level values but are clustered differently by experiment: (B) shows highly repeatable, unclustered data, (C) shows day-to-day variability, but a consistent trend in each experiment, and (D) is dominated by one random run. Note that the plots in (A) that treat each cell as its own N fail to distinguish the three scenarios, claiming a significant difference after drug treatment, even when the experiments are not actually repeatable. To correct that, “SuperPlots” superimpose summary statistics from biological replicates consisting of independent experiments on top of data from all cells, and P values were calculated using an N of three, not 300. In this case, the cell-level values were separately pooled for each biological replicate and the mean calculated for each pool; those three means were then used to calculate the average (horizontal bar), standard error of the mean (error bars), and P value. While the dot plots in the “OK” column ensure that the P values are calculated correctly, they still fail to convey the experiment-to-experiment differences. In the SuperPlots, each biological replicate is color-coded: the averages from one experimental run are yellow dots, another independent experiment is represented by gray triangles, and a third experiment is shown as blue squares. This helps convey whether the trend is observed within each experimental run, as well as for the dataset as a whole. The beeswarm SuperPlots in the rightmost column represent each cell with a dot that is color coded according to the biological replicate it came from. The P values represent an unpaired two-tailed t-test (A) and a paired two-tailed t-test for (B-D). For tutorials on making SuperPlots in Prism, R, Python, and Excel, see the supporting information.

2019 Nobel prize prediction

October 2, 2019 at 2:19 pm | | nobel

OK, it’s Nobel season again, and time for my annual blog post. (Quite literally this year, unfortunately.)

Chemistry: Lithium-ion batteries (John Goodenough) EDIT: Yay! Finally!

Medicine: DNA fingerprinting and blotting (Edwin Southern, Alec Jefferys, George Stark, Harry Towbin)

Physics: Two-photon microscopy (Watt Webb, Winfried Denk, Jim Strickler)

Last year, I correctly predicted Jim Allison. And I think I have to stop predicting Vale/Sheetz/Spudich, or my posts will look like copy-pastes. With the lawsuits in apparent stand-still, maybe this is the year for CRISPR, but I think the committee will wait a few more years until treatments come out of clinical trials.

Many sites are predicting metal-organic frameworks winning, namely Omar Yaghi. Maybe. Yaghi won the Wolf prize in 2018. And a Nobel for this work would put a spotlight on carbon capture and catalysis that might help fight global warming. But since no wide-scale efforts have actually been made to capture carbon or produce alternative fuels, I doubt MOFs will win.

The other climate-change related chemistry prize could conceivably be battery technology, especially John Goodenough, for the development of the lithium-ion battery. Given that he is 97, this would be the year to award it. Given that Alfred Nobel intended his prize to go to those who convey the “greatest benefit on mankind,” I think batteries would be fitting. (Note that I also thought it would be fitting back in 2016. I never learn.)

Another perennial prediction is DNA blotting and fingerprinting. In 2014, I predicted Southern, Jefferys, and Burnette. I’m repeating the prediction again this year. Southern and Jefferys won the Lasker award way back in 2005, and their techniques are widely used in the lab and in forensics. I’m tweaking my prediction to include Towbin, who more accurately invented Western blotting (although Neal Burnette was a genius at naming). I think these techniques have proved themselves invaluable to so many medical researchers, it would be a shame to not recognize their originators. (I acknowledge that the Nobel committee will not award this to 4 people. But I don’t know who to leave out.)

Another technique ubiquitous in biomedical imaging is two-photon microscopy. While super-resolution imaging won a few years ago, I think it would be OK to recognize the central importance of microscopy in many fields of science. By offering the capability to image deep into tissues and even live organisms, two-photon microscopy has given researchers amazing views of what would otherwise be unseeable. It is a powerful and popular technique that clearly should be recognized.

Well, that’s it for 2019. I’ve made many predictions in the past, which you can browse here.

2018 nobel prize predictions

September 20, 2018 at 2:01 pm | | nobel

It’s approaching Nobel season again, and here are my predictions:

Chemistry: Cytoskeletal motor proteins (Ron Vale, Mike Sheetz, Jim Spudich)

Medicine: T-cell and cancer immunotherapy (Jim Allison, Stephen Rosenberg, Philippa Marrack)

Physics: Dark matter (Sandra Faber, Margaret Geller, Jerry Ostriker, Helen Quinn)

I know I’ve made this prediction before, but I think it’s high time that the discovery of kinesin and early observations of single myosin activity is recognized by the Nobel committee. UPDATE: Darn. Wrong again. Phage display and protein engineering by directed evolution. Cool!

In 2016, I predicted T-cell receptor, but in the meantime cancer immunotherapy has continued to grow, so I’m tweaking the predicted winners a little. I’m not naive enough to think that we’re on the cusp of curing cancer, but it’s the first time that I thought it might be possible to—someday—conquer the disease. UPDATE: I got 1/2 of the prize correct.

Unfortunately, Vera Rubin was never awarded a Nobel Prize, but the committee could honor her memory by awarding some other deserving astrophysicists with the prize this year. UPDATE: Nope. It went to laser tweezers and ultrafast laser pulses. Interesting: these are reminiscent of some late-90s awards to Zewail and Chu.

OTHER PREDICTIONS:

My past predictions: I’ve made (partially) correct predictions in 2008, 2010, 2012, 2013, and 2017. Other predictions ended up coming to fruition in subsequent years, such as gravitational waves, super-resolution and single-molecule microscopy,

Citation Laureates

C&E News

ChemistryWorld

ChemistryViews voting

PhysicsWorld

InsideScience

update on Nikon objective immersion oils

August 30, 2018 at 8:41 am | | everyday science, hardware, review

A few years ago, I compared different immersion oils. I concluded that Nikon A was the best for routine fluorescence because: (A) it had low autofluorescence, (B) it didn’t smell, (C) it was low viscosity, and (D) the small plastic dropper bottles allowed for easy and clean application.

Unfortunately, my two favorites, Nikon A and NF, were both discontinued. The oil Nikon replaced these with is called F. But I don’t love this oil for a few reasons. First, it’s fairly stinky. Not offensive, but I still don’t want my microscopes smelling if I can help it. Second, I’ve heard complaints from others that Nikon F can have microbubbles (or maybe crystals?) in the oil, making image quality worse. Finally, dried F oil hardens over time, and can form a lacquer unless it is cleaned off surfaces very well. That said, F does have very low fluorescence, so that’s a good thing.

I explored some alternatives. Cargille LDF has the same optical properties as Nikon F (index of refraction = 1.518 and Abbe Ve = 41). But LDF smells terrible. I refuse to have my microscope room smell like that! Cargille HF doesn’t smell and has similar optical properties, but HF is autofluorescent at 488 and 405 nm excitation, so it adds significant background and isn’t usable for sensitive imaging.

At the recommendation of Kari in the UCSF microscopy core (and Caroline Mrejen at Olympus), I tried Olympus Type F, which also has an index of refraction of 1.518 and an Abbe number of 40.8, which is compatible with Nikon. The Olympus oil had very low autofluorescence, on par with Nikon A, NF, and F. (I also tested low-fluorescence oils Leica F and Zeiss 518F, but their dispersion numbers are higher (Ve = 45-46), which can cause chromatic aberration and may interfere with Perfect Focus.)

I used to love the low viscosity of Nikon A (150 cSt), because it allowed faster settling after the stage moved and was less likely to cause Perfect Focus cycling due to mechanical coupling to thin or light samples, plus it was easier to apply and clean. Nikon NF was higher viscosity (800 cSt). Olympus F is higher than Nikon A (450 cSt), but acceptable.

Finally, Olympus F comes is an easy to use applicator bottle: instead of a glass rod that can drip down the side of the vial if you’re not careful, the Olympus F is in a plastic bottle with a dropper. It’s not quite as nice as the 8 cc dropper bottles that Nikon A used to come in, and I don’t love the capping mechanism on the Olympus F, but I’ll survive.

I plan to finish up our last bottle of Nikon A, then switch over to Olympus F. We also have a couple bottles of Nikon NF remaining, which I will save for 37C work (the higher viscosity is useful at higher temperatures).

UPDATE:

Some people claim that type A was simply renamed type N. I don’t think that’s true. First of all, I couldn’t get Perfect Focus on our Ti2 to work with Nikon type N oil. Second, the autofluorescence of Nikon type N (right) was way higher than Olympus type F (left) or the old Nikon type A, at least at 405 and 488 nm:

So I’ll stick with Olympus type F. :)

UPDATE 2:

Here are some example images. These are excited with 640 (red), 561 (red), 488 (green), and 405 nm (blue) and the display ranges are the same for each sample. (The dots are single fluorophores on the glass.) You can see that Cargille HF is slightly more autofluorescent (especially at 405 nm) than either the old Nikon A or Olympus type F. This matches what Cargille states for HF: “Slightly more fluorescent than Type LDF.”

Belkin Conserve Switch is great for scopes

February 23, 2018 at 10:47 am | | hardware, review

I thought I’d pass along one of my favorite tips: I have several of these Belkin Conserve Switch power strips in lab. I use them to turn on a scope and all its peripherals with a flip of one single switch!

You can set one switch to power multiple strips. It’s so much better than flipping on each piece of equipment (and inevitably forgetting one)! You just need to check with the equipment manufacturer if it’s safe to power on/off the item by basically unplugging it.

Next Page >

Powered by WordPress, Theme Based on "Pool" by Borja Fernandez
Entries and comments feeds. Valid XHTML and CSS.
^Top^