<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.8.4">Jekyll</generator><link href="https://bitsandbrains.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://bitsandbrains.io/" rel="alternate" type="text/html" /><updated>2020-10-14T03:25:03+00:00</updated><id>https://bitsandbrains.io/feed.xml</id><title type="html">Bits and Brains</title><subtitle>NeuroData: enabling data-driven neuroscience</subtitle><entry><title type="html">Writing Effective Pull Requests</title><link href="https://bitsandbrains.io/2020/10/05/PR-checklist.html" rel="alternate" type="text/html" title="Writing Effective Pull Requests" /><published>2020-10-05T00:00:00+00:00</published><updated>2020-10-05T00:00:00+00:00</updated><id>https://bitsandbrains.io/2020/10/05/PR-checklist</id><content type="html" xml:base="https://bitsandbrains.io/2020/10/05/PR-checklist.html">&lt;p&gt;So, you have changes to code that the world would be better off having. Great! This checklist
describes how to make sure that your pull request (PR) sails through the review process.&lt;/p&gt;

&lt;h2 id=&quot;why-care&quot;&gt;Why care?&lt;/h2&gt;

&lt;p&gt;Developing and maintaining open source software is a lot of work. Ralf Gommers has a &lt;a href=&quot;https://rgommers.github.io/2019/06/the-cost-of-an-open-source-contribution/&quot;&gt;blog
post on this topic&lt;/a&gt;.
Part of the “cost” of each open source contribution is that someone has to review each change,
which is often time consuming.&lt;/p&gt;

&lt;p&gt;Following the checklist below for your PR (prior to submitting it or requesting a review)
will make life easier for you as the author and for anyone reviewing your code by reducing the amount of modifications that are necessary to merge your PR. This will also likely help 
your PR get merged faster so you can move on to other things.&lt;/p&gt;

&lt;h2 id=&quot;the-checklist&quot;&gt;The checklist&lt;/h2&gt;
&lt;p&gt;Note: the guidelines below were written explicitly for our package &lt;a href=&quot;https://github.com/microsoft/graspologic&quot;&gt;graspologic&lt;/a&gt;. 
Every project will have its own policies and practices that may be slightly different from 
what you see here - for instance, the location where documentation files are located may vary, 
or the project may use some service other than &lt;code class=&quot;highlighter-rouge&quot;&gt;Netlify&lt;/code&gt; for hosting documentation, or they may
have slightly different style conventions, etc. Most repos will have a contribution guideline
that should get you up to speed on these things.&lt;/p&gt;

&lt;p&gt;Before requesting a review on your pull request, make sure that:&lt;/p&gt;

&lt;h3 id=&quot;pr-itself-on-github&quot;&gt;PR itself (on GitHub)&lt;/h3&gt;
&lt;ul&gt;
  &lt;li&gt;PR has a descriptive and succinct title.&lt;/li&gt;
  &lt;li&gt;PR has a &lt;em&gt;brief&lt;/em&gt; description of what was changed.&lt;/li&gt;
  &lt;li&gt;PR is addressing a change that has already been brought up in an issue. This will ensure that the change you propose is desired by the maintainers, and may provide a chance to discuss implementation details prior to PRing which will save everyone time.&lt;/li&gt;
  &lt;li&gt;PR cites any related issues in the PR description and uses &lt;a href=&quot;https://docs.github.com/en/free-pro-team@latest/github/managing-your-work-on-github/linking-a-pull-request-to-an-issue#linking-a-pull-request-to-an-issue-using-a-keyword&quot;&gt;closing keywords&lt;/a&gt; for any issues that should be closed as a result of the PR.&lt;/li&gt;
  &lt;li&gt;PR does not have any extraneous file changes associated with it. One common example of an extraneous change is one that was already made in the target branch, but is showing up as part of this PR under &lt;code class=&quot;highlighter-rouge&quot;&gt;Files changed&lt;/code&gt;.&lt;/li&gt;
  &lt;li&gt;PR does not have any unresolved merge conflicts.&lt;/li&gt;
  &lt;li&gt;If implementing a major new feature or algorithm, a notebook demonstrating the proof of effectiveness is linked in the PR description. Depending on the feature, this kind of validation may take too long to run as a test or in a tutorial notebook.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;style&quot;&gt;Style&lt;/h3&gt;
&lt;ul&gt;
  &lt;li&gt;Code follows Python variable naming conventions, one description is &lt;a href=&quot;https://visualgit.readthedocs.io/en/latest/pages/naming_convention.html&quot;&gt;here&lt;/a&gt;. Most variables should be &lt;code class=&quot;highlighter-rouge&quot;&gt;snake_case&lt;/code&gt; and classes shoud be defined as &lt;code class=&quot;highlighter-rouge&quot;&gt;UpperCamelCase&lt;/code&gt;, for example.&lt;/li&gt;
  &lt;li&gt;Code uses descriptive variable names throughout.&lt;/li&gt;
  &lt;li&gt;Code has no commented out “junk code.”&lt;/li&gt;
  &lt;li&gt;Line lengths are short (recommended &amp;lt;88 chars). This includes docstrings. Note:  &lt;a href=&quot;https://black.readthedocs.io/en/stable/&quot;&gt;black&lt;/a&gt; tries to shorten line lengths but sometimes it is unable to do so automatically, especially for docstrings.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;documentation&quot;&gt;Documentation&lt;/h3&gt;
&lt;ul&gt;
  &lt;li&gt;Any new public classes/functions have docstrings.&lt;/li&gt;
  &lt;li&gt;Any new public classes/functions have been added to the appropriate &lt;code class=&quot;highlighter-rouge&quot;&gt;.rst&lt;/code&gt; file &lt;a href=&quot;https://github.com/microsoft/graspologic/tree/dev/docs/reference&quot;&gt;here&lt;/a&gt; to be rendered by &lt;code class=&quot;highlighter-rouge&quot;&gt;Sphinx&lt;/code&gt;.&lt;/li&gt;
  &lt;li&gt;Any major new features (like a new algorithm) are accompanied by a succinct tutorial notebook. Or, if it is agreed upon with the maintainers that this is out of scope for the current PR, a new issue has been created specifying that we need a tutorial notebook for this functionality.&lt;/li&gt;
  &lt;li&gt;Any new tutorial notebooks have been added to the appropriate folder &lt;a href=&quot;https://github.com/microsoft/graspologic/tree/dev/docs/tutorials&quot;&gt;here&lt;/a&gt; and included in the &lt;code class=&quot;highlighter-rouge&quot;&gt;.rst&lt;/code&gt; file &lt;a href=&quot;https://github.com/microsoft/graspologic/blob/dev/docs/tutorial.rst&quot;&gt;here&lt;/a&gt; to be rendered by &lt;code class=&quot;highlighter-rouge&quot;&gt;Sphinx&lt;/code&gt;.&lt;/li&gt;
  &lt;li&gt;Modifications to the documentation render appropriately in the &lt;code class=&quot;highlighter-rouge&quot;&gt;Netlify&lt;/code&gt; build.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3 id=&quot;testing&quot;&gt;Testing&lt;/h3&gt;
&lt;ul&gt;
  &lt;li&gt;New public classes/functions are tested to ensure they achieve the desired output.&lt;/li&gt;
  &lt;li&gt;New public classes/functions are tested to ensure proper errors are thrown when invalid inputs are passed.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2 id=&quot;acknowledgements&quot;&gt;Acknowledgements&lt;/h2&gt;
&lt;p&gt;Thanks to Adam Li and Ariel Rokem for feedback on an earlier version of this checklist 
(&lt;a href=&quot;https://twitter.com/bpedigod/status/1313827628965597190&quot;&gt;see Twitter&lt;/a&gt;).&lt;/p&gt;</content><author><name>Ben Pedigo</name></author><category term="checklist" /><summary type="html">So, you have changes to code that the world would be better off having. Great! This checklist describes how to make sure that your pull request (PR) sails through the review process.</summary></entry><entry><title type="html">10 Simple Rules to Developing a New Method</title><link href="https://bitsandbrains.io/2019/11/21/new-method.html" rel="alternate" type="text/html" title="10 Simple Rules to Developing a New Method" /><published>2019-11-21T00:00:00+00:00</published><updated>2019-11-21T00:00:00+00:00</updated><id>https://bitsandbrains.io/2019/11/21/new-method</id><content type="html" xml:base="https://bitsandbrains.io/2019/11/21/new-method.html">&lt;p&gt;So you want to develop a new method.  I recommend you do the following:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Find something about the world that you’d like to be better&lt;/strong&gt;
My recommendation is to read &lt;a href=&quot;https://bitsandbrains.io/2018/08/31/sig-and-feas.html&quot;&gt;how to choose a project&lt;/a&gt;, and find the project that maximizes feasibility and significance for you, given you intrinsic motivation.  This is the hardest part, but once you have done this, you’ll hopefully have both purpose and direction.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Write down the mathematical problem you are trying to solve that will improve the world for the above described project&lt;/strong&gt;
Using as formal notation as possible, write down the goal of your method.  Is it a classification problem? Are there any particular constraints?  Which of &lt;a href=&quot;https://bitsandbrains.io/2018/09/24/modeling-desiderata.html&quot;&gt;these things&lt;/a&gt; do you particularly care about, and how much, and what constraints do you have.  I particularly like the example provided in my &lt;a href=&quot;https://dl.acm.org/citation.cfm?id=2498823&quot;&gt;signal subgraph paper&lt;/a&gt;.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Identify Reference Methods&lt;/strong&gt; Identify the current state-of-the-art (SoA) methods for solving this problem.  Make a list of references (up to five articles) describing SoA, and highlight those for which you can identify reference implementations.  For python users, compare to at least  one relevant algorithm in &lt;a href=&quot;https://scikit-learn.org/stable/&quot;&gt;sklearn&lt;/a&gt;, and probably a few others (sklearn only includes algorithms that are at least three years old, so it never has the most up-to-date algorithms). For R users, search through the &lt;a href=&quot;https://cran.r-project.org/web/views/&quot;&gt;CRAN Task Views&lt;/a&gt;.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Identify Benchmark Settings&lt;/strong&gt;
The previous SoA must have run their code on some datasets, including both simulated and real.  Identify those settings.  We will be comparing our algorithm’s performance on those settings, so as to not be cherry-picking settings.  Also consider a few (like one or two) very simple real datasets, such as &lt;a href=&quot;https://scikit-learn.org/stable/datasets/index.html#toy-datasets&quot;&gt;sklearn’s toy datasets&lt;/a&gt;.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Identify Metrics&lt;/strong&gt;
The best metrics will depend on your goals.  For example, in classification, held-out classification error might be the best metric, but other metrics can also be useful, for example, see the metrics we used in our &lt;a href=&quot;https://arxiv.org/abs/1506.03410&quot;&gt;sporf paper&lt;/a&gt;, which include a normalized effect size variant of Cohen’s kappa.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Understand Geometry&lt;/strong&gt;
Describe in detail (as formally as you can) the simplest parametric model that you will be using to evaluate this algorithm. This model should have the property that you can change a one-dimensional parameter to make vary the performance of your algorithm from better thant the current SoA to worse than the current SoA. For example, e.g., k-means will work better than GMM when two clusters are all spherically symmetric Gaussians and each cluster is equally likely, but will work worse than GMM when one cluster is much more likely than the other. Therefore, the probability of a cluster is the one-dimensional parameter that illustrates the phase transition in for the algorithms.  If you can’t figure out a single simple illustration for this, come up with two different simple settings: one for which you expect your new algorithm to outperform the SoA, and one for which you expect your new algorithm to under-perform the SoA. In this case, vary some property of both settings (for example, number of dimensions, or variance), to demonstrate the robustness of the result across this dimension. To the extent possible, use previously proposed settings, such as the ones in sklearn’s &lt;a href=&quot;https://scikit-learn.org/stable/modules/clustering.html&quot;&gt;clustering&lt;/a&gt; or &lt;a href=&quot;https://scikit-learn.org/stable/modules/classes.html#module-sklearn.datasets&quot;&gt;sample generators&lt;/a&gt;; we have also proposed several in our papers, such as &lt;a href=&quot;https://elifesciences.org/articles/41690&quot;&gt;mgc&lt;/a&gt;, &lt;a href=&quot;https://arxiv.org/abs/1506.03410&quot;&gt;sporf&lt;/a&gt;, and &lt;a href=&quot;https://arxiv.org/abs/1907.02844&quot;&gt;geodesic forests&lt;/a&gt;.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Simulate&lt;/strong&gt;
Write code to simulate those scenarios, and look at the results.  Confirm that your simulations are in fact simulating what the equations describing the simulations meant to describe.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Run Previous SoA&lt;/strong&gt;
    &lt;ol&gt;
      &lt;li&gt;Make a Jupyter notebook running previous SoA on the simulations and data from their paper.  Make sure your results at least approximately match their reported results (like, within 1%); ideally your results are identical to the published results, but that is not typical.  This is especially important if you are re-implementing the previous SoA, rather than simply using their results.&lt;/li&gt;
      &lt;li&gt;Make a Jupyter notebook on any additional simulation settings.&lt;/li&gt;
      &lt;li&gt;PR the part of this Jupyter notebook that includes an sklearn algorithm.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Pseudocode, Code, and Test&lt;/strong&gt;
    &lt;ol&gt;
      &lt;li&gt;Write pseudocode (as formally as you can) for your algorithm.  I particularly like the text explanation pseudocode that we wrote for our &lt;a href=&quot;https://elifesciences.org/articles/41690&quot;&gt;mgc paper&lt;/a&gt;.&lt;/li&gt;
      &lt;li&gt;Write the code for your algorithm.  If you are in Python, to the extent possible, follow the &lt;a href=&quot;https://scikit-learn.org/stable/developers/contributing.html#apis-of-scikit-learn-objects&quot;&gt;sklearn API&lt;/a&gt;.&lt;/li&gt;
      &lt;li&gt;Test your code on very simple experimental settings.  Settings which should result in an obvious result.  I recommend the following three tests:
        &lt;ol&gt;
          &lt;li&gt;A setting in which it is so easy that for sure any reasonable algorithm will get the answer perfect.&lt;/li&gt;
          &lt;li&gt;A setting which is impossible for your algorithm to get right.&lt;/li&gt;
          &lt;li&gt;A reasonable setting, in which as sample size increases, accuracy should improve.
If your implementation works as expected in all of these settings, proceed.  Otherwise, iterate.&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Evaluate Algorithm on Simulations&lt;/strong&gt;
Make a Jupyter notebook evaluating all the  algorithms on the simulations. For each simulation setting:&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Generate 10x random samples of the data&lt;/li&gt;
      &lt;li&gt;Run all relevant algorithms.&lt;/li&gt;
      &lt;li&gt;Compare performance (using the above established metrics) on each run of the simulation.  That is, do not compare the average, but rather, make 10 comparisons.  Comparing each run is more powerful and informative.&lt;/li&gt;
      &lt;li&gt;Plot all 10x comparisons for each algorithm compared to yours.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Evaluate Algorithm on Real Data&lt;/strong&gt;
Make a Jupyter notebook evaluating all the  algorithms on the simulations. For each real data setting:&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Run 5-fold or 10-fold cross-validation. When appropriate, use stratified cross-validation.  Do cross-validation such that the data are divided into k (approximately) equal sizes subsets of data.  Train on the k-1 subsets and test on the remaining subset, so that each sample is used for testing exactly once.&lt;/li&gt;
      &lt;li&gt;Run all relevant algorithms.&lt;/li&gt;
      &lt;li&gt;Compare performance (using the above established metrics) on each fold.  That is, do not compare the average, but rather, make k comparisons.  Comparing each run is more powerful and informative.&lt;/li&gt;
      &lt;li&gt;Plot all 10x comparisons for each algorithm compared to yours.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;hr /&gt;

&lt;p&gt;If you follow the above plan, you will carefully demonstrated the value of your new algorithm in an unambiguous fashion.&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="10" /><category term="bits" /><summary type="html">So you want to develop a new method. I recommend you do the following:</summary></entry><entry><title type="html">An Email to Andrew Gelman About Evaluating Journals</title><link href="https://bitsandbrains.io/2019/08/03/journal-evaluation.html" rel="alternate" type="text/html" title="An Email to Andrew Gelman About Evaluating Journals" /><published>2019-08-03T00:00:00+00:00</published><updated>2019-08-03T00:00:00+00:00</updated><id>https://bitsandbrains.io/2019/08/03/journal-evaluation</id><content type="html" xml:base="https://bitsandbrains.io/2019/08/03/journal-evaluation.html">&lt;p&gt;I wrote the below transcribed email to Andrew Gelman today.  He often publishes my emails on his blog with a 6 month time delay, so I figured I’d post it here just in case.  If he bloggs about it, I’ll link to it here.&lt;/p&gt;

&lt;hr /&gt;

&lt;p&gt;i noticed you disparage a number of journals quite frequently on your blog.
i wonder what metric you are using implicitly to make such evaluations?
is it the number of articles that they publish that end up being bogus?
or the fraction of articles that they publish that end up being bogus?
or the fraction of articles that get through their review process that end up being bogus?
or the number of articles that they publish that end up being bogus AND enough people read them and care about them to identify the problems in those articles.&lt;/p&gt;

&lt;p&gt;my guess (without actually having any data), is that Nature, Science, and PNAS are the best journals when scored on the metric of fraction of bogus articles that pass through their review process. in other words, i bet all the other journals publish a larger fraction of the false claims that are sent to them than Nature, Science, or PNAS.&lt;/p&gt;

&lt;p&gt;the only data i know on it is described [here]
(https://www.nature.com/articles/s41562-018-0399-z)
according to the article, 62% of social-science articles in Science and Nature published from 2010-2015 replicated
an &lt;a href=&quot;https://science.sciencemag.org/content/351/6280/1433.long&quot;&gt;earlier paper from the same group&lt;/a&gt; found that 61% of papers from specialty journals published between 2011 and 2014 replicated.&lt;/p&gt;

&lt;p&gt;i’d suspect that the fraction of articles on social sciences that pass the review criteria for Science and Nature is much smaller than that of the specialty journals, 
implying that the fraction of articles that get through peer review in Science and Nature that replicate is much higher than the specialty journals.&lt;/p&gt;

&lt;p&gt;curious to here your thoughts….&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><summary type="html">I wrote the below transcribed email to Andrew Gelman today. He often publishes my emails on his blog with a 6 month time delay, so I figured I’d post it here just in case. If he bloggs about it, I’ll link to it here.</summary></entry><entry><title type="html">How to Join Us</title><link href="https://bitsandbrains.io/2019/07/28/join-us.html" rel="alternate" type="text/html" title="How to Join Us" /><published>2019-07-28T00:00:00+00:00</published><updated>2019-07-28T00:00:00+00:00</updated><id>https://bitsandbrains.io/2019/07/28/join-us</id><content type="html" xml:base="https://bitsandbrains.io/2019/07/28/join-us.html">&lt;p&gt;NeuroData is literally always hiring exceptional individuals.  If you think you’d like to work with us in any capacity, please &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt; now!  Some thoughts that might be useful prior to emailing us to consider further:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;We are interested in &lt;em&gt;all&lt;/em&gt; kinds of people with &lt;em&gt;all&lt;/em&gt; kinds of backgrounds, we do not discriminate for any reason, and we are all students at different stages.  Although this should go without saying, this includes women and minorities.&lt;/li&gt;
  &lt;li&gt;We believe that to maximally flourish requires being in the appropriate environment specifically for &lt;em&gt;you&lt;/em&gt;.  Each individual benefits from unique circumstances.  We wrote this &lt;a href=&quot;https://bitsandbrains.io/2018/08/31/sig-and-feas.html&quot;&gt;blog post&lt;/a&gt; which describes out thoughts on choosing projects, but it is also relevant for choosing teams. If you are reading this far, it seems likely that the kind of work we do is intrinsically motivating for you.  Even if not, and you think we could help find a team better aligned with your dreams, please &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt;.&lt;/li&gt;
  &lt;li&gt;To create the most peaceful and productive environment, our team has constructed these &lt;a href=&quot;https://neurodata.io/about/agreements/&quot;&gt;agreements&lt;/a&gt;, which we all adhere to. This includes  treating everyone with respect on the team, and at any team events.  See our &lt;a href=&quot;https://neurodata.io/about/agreements/#respect&quot;&gt;code of conduct&lt;/a&gt; for details.  If it seems like that kind of environment would suit you well, please &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt;.&lt;/li&gt;
  &lt;li&gt;If you are interested in pursuing graduate studies with us, there are many different programs that are available.  The &lt;a href=&quot;https://www.bme.jhu.edu/graduate/overview/&quot;&gt;Johns Hopkins University Biomedical Engineering graduate programs&lt;/a&gt; are the most common approach, which includes the &lt;a href=&quot;https://www.bme.jhu.edu/graduate/mse/prospective-students/&quot;&gt;MSE&lt;/a&gt; and &lt;a href=&quot;https://www.bme.jhu.edu/graduate/phd/overview/&quot;&gt;PhD&lt;/a&gt; programs, as well as a few others (&lt;a href=&quot;https://www.bme.jhu.edu/graduate/tsinghua-jhu-dual-degree-ms-program/overview/&quot;&gt;Tsinghua JHU-BME dual degree&lt;/a&gt; and &lt;a href=&quot;https://www.bme.jhu.edu/graduate/md-phd/&quot;&gt;MD/PhD&lt;/a&gt;).  Note that BME@JHU is the #1 rated BME graduate school in the world according to &lt;a href=&quot;https://www.usnews.com/best-graduate-schools/top-engineering-schools/biomedical-rankings&quot;&gt;US News &amp;amp; World Reports&lt;/a&gt;. I also sometimes advise students in &lt;a href=&quot;https://www.cs.jhu.edu/graduate-studies/&quot;&gt;Computer Science&lt;/a&gt;,  &lt;a href=&quot;https://engineering.jhu.edu/ams/graduate-studies/overview/&quot;&gt;Applied Mathematics &amp;amp; Statistics&lt;/a&gt;, &lt;a href=&quot;https://www.jhsph.edu/departments/biostatistics/academics-and-student-life/degree-programs/&quot;&gt;Biostatistics&lt;/a&gt;, and &lt;a href=&quot;http://neuroscience.jhu.edu/graduate&quot;&gt;Neuroscience&lt;/a&gt;. If you are interested in graduate studies with us in a different program (e.g., &lt;a href=&quot;https://www.janelia.org/you-janelia/students-postdocs/joint-graduate-program&quot;&gt;JHU/Janelia&lt;/a&gt;, or &lt;a href=&quot;https://engineering.jhu.edu/ece/graduate-studies/&quot;&gt;ECE@JHU&lt;/a&gt;), please also feel free to &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt;.&lt;/li&gt;
  &lt;li&gt;If you are already enrolled at JHU and are interested in joining the lab, the simplest route is to enroll in the year-long project-based course we offer called &lt;a href=&quot;https://neurodatadesign.github.io/Syllabus/&quot;&gt;NeuroData Design&lt;/a&gt;.  A key motivation for offering this course was to ensure that we could work with as many amazing individuals as possible.  If for some reason, enrolling in the course does not seem feasible for you, please &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt; and we will try to figure out some other arrangement.&lt;/li&gt;
  &lt;li&gt;If you are a visiting or exchange student, please see &lt;a href=&quot;https://apply.jhu.edu/visiting-and-exchange-students/&quot;&gt;JHU’s official exchange program overview&lt;/a&gt;, and go through the official process; after you do this, &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt;  to say you applied.  It is complicated for legal and bureaucratic reasons to accept visiting/exchange students outside these official programs, although it is possible. So, if those programs do not work for any reason, please  &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;email us&lt;/a&gt;.&lt;/li&gt;
  &lt;li&gt;In general, we find part time work, and remote work, inefficient, and unlikely to be in your or our best interest (though there are exceptions). If that is the only option for you, we’d be happy to discuss your interests with you to find a more suitable position given your constraints.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;OK, given all the above, we recommend that you email us.  Below we provide an example email.  Note that we will respond within a week in almost all cases.&lt;/p&gt;

&lt;p&gt;Dear Prof. X,&lt;/p&gt;

&lt;p&gt;I have been studying [something quite relevant to your research] for the last [X duration], and I expect to graduate from [Y institution] in [Z expected month].
I recently became interested with your work, specifically, the article “[some article from our group]”.&lt;br /&gt;
I am particularly intrigued by this work because of [X], which is one of the things I am most excited about studying next.
I would therefore be thrilled to do [X] for a period of [Y duration] in your group.
For your convenience, I have attached my CV, and and an unofficial copy of my transcript.&lt;/p&gt;

&lt;p&gt;I look forward to hearing from you at your earliest convenience.&lt;/p&gt;

&lt;p&gt;Best,
[Some Awesome Person.]&lt;/p&gt;

&lt;p&gt;In conclusion, a brief checklist that may help you of what to include in your email to faculty:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;Excitement about working on something specific relevant to the individual you are contacting&lt;/li&gt;
  &lt;li&gt;Relevant background experience&lt;/li&gt;
  &lt;li&gt;Desired duration and title of position&lt;/li&gt;
  &lt;li&gt;Up to date cv&lt;/li&gt;
  &lt;li&gt;Remember, the shorter the better, this is just to get your foot in the door.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Good luck!&lt;/p&gt;

&lt;!-- I get many emails from potential trainees inquiring about a job. --&gt;
&lt;!-- This ranges from high school students to people that have already completed a few postdocs to retirees. --&gt;

&lt;!-- We love working with many different people, from all walks of life and at all stages of education (we are in many of them ourselves). This blog post explains our thought process in terms of how we choose who to be on our team.  We hope it is useful to you, and look forward to your feedback and working with you! --&gt;

&lt;!-- Firstly, if you  --&gt;
&lt;!-- Secondly, We have a few basic requirements for being on the team: --&gt;

&lt;!-- 1. You are more excited about working on the stuff we work on than anything else in the world. We believe maximally thrive/flourish  their [zone of genius](https://www.amazon.com/Big-Leap-Conquer-Hidden-Level/dp/0061735361), and we desire to support people to  maximally thrive. --&gt;
&lt;!-- 2. You are excited to specifically work with and hang out with us, and we feel the same way about you. We care deeply about the quality of our lives, which, for us, is typically largely determined by our relationships with the people around us.   --&gt;
&lt;!-- 3. You agree to all our [agreements](https://github.com/neurodata/about/blob/master/agreements.md). We collectively generated these agreements, and frequently update them, to reflect our core values. If they are not cool with you, it is unlikely that you will be a good cultural fit on our team.   --&gt;

&lt;!-- 
If you meet all the above requirements, than there is a pretty good chance we'll get to do something awesome together :)  However, there are some actual skills and background knowledge that are quite useful.  These are not quite required, because we often work with high school students, for example, who lack these skills, but they are suggested.
For college and post-college students, some background in either theoretical mathematics/statistics and/or numerical programming is essentially required to contribute.  

Something else that is required, but is basically impossible to assess without ever talking to us, is an ability to communicate technical content to **us** effectively.  This is important because we work on teams, and the efficiency of our team is partially determined by our ability to communicate effectively with one another.  We acknoweldge that this is highly subjective, a given individual might be able to communicate effectively with some people, and not others.

One last thing: the longer you are interested in working on our team, the better.  For example, PhD students stick around for five years, we can train you and do a whole bunch of amazing things together in that time. Postdocs can stick around for 2 years, which is nice if they already are up to speed on crucial data science skills.  Master's students are more complicated, as they have so little time and so much to learn.  Though we often take summer students, they are the most difficult to do right by.  We want them to learn and have an awesome experience, but we are a small team, and summer is not long enough to contribute, just long enough to get some basic training in.  So, summer positions are particularly selective. In any case, we hope to work with you soon, and if only for a short time now, we hope a much longer time in the future.


OK, now we are in good shape, and we're actually looking forward to an email from you!  I try to respond to every single request I receive, but some I judge as spam (or my spam filter does), and are therefore ignored.  If you have not received a response from me within a week, please email again (probably change the text because the first one I thought was spam)!  Below is a template of an email that gets me excited to respond,  mostly probably because it caters to my ego.
Nonetheless, it is effective, and potentially helpful, so I thought I'd share. I've written it as if this person is finishing grad school, so make appropriate changes if you are finishing something else. --&gt;</content><author><name>Joshua Vogelstein</name></author><category term="education" /><summary type="html">NeuroData is literally always hiring exceptional individuals. If you think you’d like to work with us in any capacity, please email us now! Some thoughts that might be useful prior to emailing us to consider further:</summary></entry><entry><title type="html">How to Succeed in a Biomedical Data Science Lab (specifically ours)</title><link href="https://bitsandbrains.io/2019/07/27/succeed.html" rel="alternate" type="text/html" title="How to Succeed in a Biomedical Data Science Lab (specifically ours)" /><published>2019-07-27T00:00:00+00:00</published><updated>2019-07-27T00:00:00+00:00</updated><id>https://bitsandbrains.io/2019/07/27/succeed</id><content type="html" xml:base="https://bitsandbrains.io/2019/07/27/succeed.html">&lt;p&gt;The degree of success one can achieve in any environment, specifically a lab, depends upon appropriately matching individual talents with an environment.  There are many web resources that describe how to succeed in graduate school, or in a job [in a subsequent update to this post, I’ll include links to my favorite ones].  Here, we try to list the qualities of individuals that we believe tend to be positively related to “success” in our lab.  We believe these qualities are also likely to be positively related to success in other positions, but are relatively data poor for the other positions.  On the other hand, we’ve interacted with between 50-100 individuals so far, and have therefore amassed a fair bit of anecdotal evidence.&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;Mission Aligned&lt;/strong&gt; Our mission is to &lt;code class=&quot;highlighter-rouge&quot;&gt;Understand and improve animal and machine intelligences worldwide&lt;/code&gt;.  If you are motivated to solve these problems, over and above any other problems, you are in the right place.  Read our &lt;a href=&quot;https://neurodata.io/about/&quot;&gt;mission&lt;/a&gt; for more details.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Interest in Data Science&lt;/strong&gt;   Our work is at the intersection of statistical machine learning, brain science, and mental health.  If you are maximally intrinsically motivated to study this intersection to achieve our mission, read on.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Personality Traits&lt;/strong&gt; Each institution/lab is a unique environment, with unique quirks, ours included.  The traits that seem most important to us include:
    &lt;ol&gt;
      &lt;li&gt;enthusiasm for solving these problems;&lt;/li&gt;
      &lt;li&gt;humility to realize that we are all often wrong, and the goal is to learn and understand, rather than be “correct”;&lt;/li&gt;
      &lt;li&gt;gratitude, both for the opportunity for learning and growth that comes from our errors, and because we get to work in the best environment with the best in the world;&lt;/li&gt;
      &lt;li&gt;patience, because these things are hard and take time;&lt;/li&gt;
      &lt;li&gt;simplicity, because the fewer parts the easier to understand and the less that can go wrong; and&lt;/li&gt;
      &lt;li&gt;trust in each other, as each member of the team has unique talents.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Data Science Chops&lt;/strong&gt; To succeed, you’ll essentially need to learn or already know graduate level probability, statistics, matrix analysis,  network science, and some numerical programming and brain science. If you plan on developing methods, strong numerical programming skills are required.  If you plan on applying methods, strong brain science understanding is required.  If you plan on proving theorems, more theory is required&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Technical Communication&lt;/strong&gt;  Something that is particularly difficult to quantify, but as important as the above properties,  is an ability to communicate technical content to &lt;strong&gt;us&lt;/strong&gt; effectively.  This is important because we work on teams, and the efficiency of our team is partially determined by our ability to communicate effectively with one another.  We acknowledge that this is highly subjective, a given individual might be able to communicate effectively with some people, and not others. This does &lt;em&gt;not&lt;/em&gt; require fluency in English per se, but does require being able to speak, make slides, and write reports about technical content in a coherent fashion.  This is something we all learn by doing, it is not expected that anybody joins the team knowing these things.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Agreements&lt;/strong&gt; Our team made a set of &lt;a href=&quot;https://neurodata.io/about/agreements/&quot;&gt;agreements&lt;/a&gt;, that we continue to update as appropriate. Agreeing to these agreements is a &lt;em&gt;prerequisite&lt;/em&gt; to joining the team, and therefore, is required for succeeding on the team.  If you have recommendations for how to further improve our agreements, please &lt;a href=&quot;mailto:join@neurodata.io&quot;&gt;let us know&lt;/a&gt;. Note that these agreements include both personal and professional activities.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Learn Context&lt;/strong&gt; To make a meaningful contribution to the literature, it is helpful to know what is already known in the literature, and what are the biggest feasible open challenges.  There are many kinds of activities that help you gain context including:
    &lt;ol&gt;
      &lt;li&gt;Take relevant classes.  We specifically recommend classes taught by Jovo, Carey Priebe, Randal Burns, Rene Vidal, and Raman Arora.&lt;/li&gt;
      &lt;li&gt;Attend conferences.  One to two conferences a year, for example, one large general conference and one smaller specialized conference, can be invaluable for learning.  At those conferences, literally attend all possible conference activities, including all talks, posters, social activities, and workshops.  For multi-track conferences, as long as one track has an activity, go to it.  Expect to be learning essentially 12 hours a day. Take breaks as appropriate, probably nearly hourly, to decompress.  It helps to attend talks/posters/etc. with other people who are more senior, so that you can discuss the contents afterwards and digest the most important points.  Go a day early, or stay a day late, to be able to enjoy the city you are in; there is no need to do those activities &lt;em&gt;during&lt;/em&gt; the conference, there is plenty of time after.  If at all possible, do the fun activities with other conference attendees, ideally not from your lab.&lt;/li&gt;
      &lt;li&gt;Read papers.  There is no limit to the number of papers one can read.  If you find a really interesting article, read all the references of that paper.  A good rule is read an average of one paper a day.  The depth with which you read it can vary as appropriately, from only 30 minutes to get the gist, to 8 hours to understand the details.  If you think you want to read a paper for 8 hours, first read it for 30 minutes, and then read the key references each for 30 minutes.  Then do the 8 hour read.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Plan on Being Here a While&lt;/strong&gt; Many people approach us and are interested for a summer internship, or even a year long fellowship.  It is quite difficult to make a positive contribution in less than 1 year, for the simple reason that there is a lot of specific background knowledge and context that the lab has that is nearly impossible to get anywhere else (this is likely true for most labs).  PhD’s take about 5 years, that is long enough to be able to make a contribution.  Realistically, after about two years, expect to make a useful contribution, that is, your research output more than exceeds the output the team would have had had you not been there (which means you’ve overcome the research debt accrued by virtue of the energy spent training you).  The efficiency with which you can make contributions continues to increase with training.  Ideally, the best trainees would never leave, meaning, they would either take a postdoc position with us, and/or get a faculty position at JHU, or a neighboring university if that makes more sense for various reasons.&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Meet the Experts&lt;/strong&gt; There are many brilliant people working on related questions all over the world.  To the extent possible, meet them.  I know of three ways:
    &lt;ol&gt;
      &lt;li&gt;Sign up for relevant seminars and attend them.  At JHU, this includes the following seminar: &lt;a href=&quot;http://www.cis.jhu.edu/seminars/&quot;&gt;CIS&lt;/a&gt;, &lt;a href=&quot;http://www.math.jhu.edu/~data/&quot;&gt;data&lt;/a&gt;, &lt;a href=&quot;https://www.cs.jhu.edu/seminars/2019/fall&quot;&gt;CS&lt;/a&gt;, &lt;a href=&quot;https://www.bme.jhu.edu/news-events/bme-events/&quot;&gt;BME&lt;/a&gt;, &lt;a href=&quot;https://www.cs.jhu.edu/~mdinitz/theory/seminar/&quot;&gt;CS theory&lt;/a&gt;, &lt;a href=&quot;https://engineering.jhu.edu/ams/news-and-events/calendar/action~agenda/page_offset~2/request_format~json/&quot;&gt;AMS&lt;/a&gt;.  Please &lt;a href=&quot;mailto:support@neurodata.io&quot;&gt;email us&lt;/a&gt; if we forgot any.  At the seminars, if you don’t understand something, raise your hand and ask a question.  The speaker’s job is to make you understand, either during the talk, or if they deem it appropriate, after the talk.  Let them choose when to answer, rather than choosing for them.  Also, when possible, sign up to speak with the speakers.  For many speakers there are lunches with students, go to as many of them as you can.  If there is no lunch, either sign-up to speak with them yourself, or with a small team of other students, or ask your PI to join their meeting.&lt;/li&gt;
      &lt;li&gt;At conferences.  Prior to the conference, say about 1 week before, find out which faculty that interest you will be at the conference (if they are speaking, they will be there). Email them all, offer to meet them at &lt;em&gt;literally any time and place&lt;/em&gt; that is convenient for them.  Offer to meet them at the airport, or go with them to the airport, if necessary.  Faculty tend to love this kind of thing, as it caters to our egos.  Also, it makes the rides more enjoyable. After their talk, go up to them and ask them questions, or simply tell them how much you learned and how much you appreciate their work. Don’t forget to introduce yourself, and in 30 seconds, tell them whose lab you are in, and what your main project is focused on.&lt;/li&gt;
      &lt;li&gt;Invite faculty of interest to seminars.  In many seminars there are slots for student invites.  In others, faculty decide, but should always be willing/interested in inviting faculty that are interesting to the students.  So, bring the faculty you want to you, and then be their host for the day, take them around to their meetings, join them for lunch and dinner, etc.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;I imagine there are other qualities that I’ve forgotten.  If so, please let us know in the comments below, and we can update this list.&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="education" /><summary type="html">The degree of success one can achieve in any environment, specifically a lab, depends upon appropriately matching individual talents with an environment. There are many web resources that describe how to succeed in graduate school, or in a job [in a subsequent update to this post, I’ll include links to my favorite ones]. Here, we try to list the qualities of individuals that we believe tend to be positively related to “success” in our lab. We believe these qualities are also likely to be positively related to success in other positions, but are relatively data poor for the other positions. On the other hand, we’ve interacted with between 50-100 individuals so far, and have therefore amassed a fair bit of anecdotal evidence.</summary></entry><entry><title type="html">10 Simple Rules to Write a Paper from Start to Finish</title><link href="https://bitsandbrains.io/2019/02/10/how-to-write-a-paper.html" rel="alternate" type="text/html" title="10 Simple Rules to Write a Paper from Start to Finish" /><published>2019-02-10T18:27:57+00:00</published><updated>2019-02-10T18:27:57+00:00</updated><id>https://bitsandbrains.io/2019/02/10/how-to-write-a-paper</id><content type="html" xml:base="https://bitsandbrains.io/2019/02/10/how-to-write-a-paper.html">&lt;ol&gt;
  &lt;li&gt;&lt;strong&gt;Learn how to write&lt;/strong&gt; by reading &lt;a href=&quot;https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1005619&quot;&gt;how to structure a paper&lt;/a&gt; and &lt;a href=&quot;https://www.amazon.com/Writing-Science-Papers-Proposals-Funded/dp/0199760241&quot;&gt;Writing Science&lt;/a&gt; (sorry that I am recommending purchasing a book, though it should be available for free from any academic library, and I recommend all PIs own a copy for themselves and their students, so it does not cost students to read it at a minimum).&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Main result&lt;/strong&gt; write draft paper title and  draft killer fig that visually makes the main point of your story&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Abstract&lt;/strong&gt; write title and abstract based on OCAR story structure (described in &lt;a href=&quot;https://www.amazon.com/Writing-Science-Papers-Proposals-Funded/dp/0199760241&quot;&gt;Writing Science&lt;/a&gt;)&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Figures&lt;/strong&gt; draft the remaining figures&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Outline&lt;/strong&gt; generate a 1 sentence per paragraph outline&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Venue&lt;/strong&gt; choose the publication venue&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Draft&lt;/strong&gt; flesh out details of paper
&lt;!-- 7. **References** add them and check that they are right --&gt;
&lt;!-- 6. **wordsmithing** check  [paragraph guidelines](https://bitsandbrains.io/2018/10/14/paragraphs.html), [words guidelines](https://bitsandbrains.io/2018/10/14/words.html), and remove redundant works --&gt;
&lt;!-- 7. **feedback** send to co-authors for review &amp; detailed critical feedback --&gt;&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Revision&lt;/strong&gt; revise&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Submission&lt;/strong&gt; post to arxiv, submit to journal, tweet to world&lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Rebuttal&lt;/strong&gt; respond to reviewers&lt;/li&gt;
&lt;/ol&gt;

&lt;hr /&gt;

&lt;p&gt;Upon believing that you have completed work sufficient to write a peer reviewed manuscript, follow the below steps in order. If you are simply writing an abstract (for a conference, for example), just do the section entitled “Outline and Abstract”&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Learn how to write&lt;/strong&gt;
Before you write any paper, I recommend reading  &lt;a href=&quot;https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1005619&quot;&gt;How to structure a paper&lt;/a&gt;. I also recommend reading &lt;a href=&quot;https://www.amazon.com/Writing-Science-Papers-Proposals-Funded/dp/0199760241&quot;&gt;Writing Science&lt;/a&gt;, if care about writing clear compelling stories.  You only need to do this once ever, not once per paper. Once you’ve read these things, follow these steps:&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Main Result&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Write a one sentence summary of your work (will become your &lt;strong&gt;title&lt;/strong&gt;;  ~5 min). This sentences describes the main take home message / main result.  Avoide jargon to the extent possible, and grab the attention of your readership with strong substantiated claims. It is often required to have  less than 88 characters.&lt;/li&gt;
      &lt;li&gt;Draw (or make) the “killer fig” that makes the point as clearly and concisely as possible.  Guidance for making paper quality figures is &lt;a href=&quot;https://bitsandbrains.io/2018/09/08/figures.html&quot;&gt;here&lt;/a&gt;.  (~5 min)&lt;/li&gt;
      &lt;li&gt;Get feedback from your PI and other principle authors.  If the author list is unclear, send this summary and figure to everyone you think might believe they deserve to be authors on the manuscript, and invite them to be co-authors.  Tell them that if they agree (by responding to you and the corresponding author), they can expect to get emails from you with updates, each of which will request feedback.  And if they provide feedback, it will be carefully considered.  Explain to them that the goal of feedback at this particular stage is that you’d like to know what they gathered was the main point of your paper, based on reading the title and looking at the killer fig (without sending them the caption).  If they don’t get it exactly right, iterate until they do (don’t tell them the answer, just update the title/figure).&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Choose a writing medium.  I recommend google docs unless there are a large number of equations, in which case I recommend overleaf. In either case, you want people to be able to comment easily directly on the draft, in real time, as to avoid the possibility of collisions.  Prior to sending it to anybody, make a local copy, so they can’t screw it up (and update your local copy prior to each round of feedback).&lt;/li&gt;
      &lt;li&gt;Describe the other results, typically 3-5 additional figures or theorems.  The goal of each of these is to support the main claim, for example, by further refining, adding controls, etc.  Ideally, they are sequenced together in a logical chain, like a proof, each building on the next, to tell the story. (~1 hr)&lt;/li&gt;
      &lt;li&gt;Write a one paragraph summary (will become your &lt;strong&gt;abstract&lt;/strong&gt;;  ~30 min).  This will be about 250 - 300 words, more than 500 words is a page, not an abstract.  To include:
        &lt;ol&gt;
          &lt;li&gt;Big opportunity sentence: what is the grandest opportunity that this work is addressing? In other words, what is wrong in the world, impacting people (not just scientists), that you are working on righting?  Potentialy answers include disease, world hunger, data deluge, etc.&lt;/li&gt;
          &lt;li&gt;Specific opportunity: what opportuntity specifically will this manuscript address? This is filtering down from the above, we are not going to try to solve all of world hunger, maybe just identify the primary causal factors contributing to hunger in India, for example.&lt;/li&gt;
          &lt;li&gt;Challenge sentence: what is hard about addressing this opportunity? Think about the limitations of approaches that others have taken, they are great in some ways, but incomplete because something was a challenge that they did not overcome. Make sure to comment how great the previous approaches are, the people who developed them are our readers, reviewers, etc., and we all have egos.&lt;/li&gt;
          &lt;li&gt;Gap sentence: what is currently missing? What is the key insight/innovation that enables us to overcome the challenge that others did not yet overcome.&lt;/li&gt;
          &lt;li&gt;Action sentence: what did you do to address the gap, overcome the challenge, and therefore meet the opportunity? it should provide the &lt;em&gt;key&lt;/em&gt; intuition/insight, the magic that makes this work, where others failed. This is the place where one can write something like “In this paper, we….”, although I would instead simply write “We….” (in this paper should be obvious from context).&lt;/li&gt;
          &lt;li&gt;Resolution sentence: what changes for the reader now that you have met this challenge?&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Note that the above structure is called “OCAR” in &lt;a href=&quot;https://www.amazon.com/Writing-Science-Papers-Proposals-Funded/dp/0199760241&quot;&gt;Writing Science&lt;/a&gt;, for &lt;em&gt;O&lt;/em&gt;pportunity, &lt;em&gt;C&lt;/em&gt;hallenge, &lt;em&gt;A&lt;/em&gt;ction, and &lt;em&gt;R&lt;/em&gt;esolution (tehcnically they use “Opening” instead of “Opportunity”, but I like “Opportunity” more).  Steps 1 and 2 form the opportunity, steps 3 and 4 form the challenge, 5 is action and 6 is resolution.&lt;/p&gt;

&lt;p&gt;When considering which results to mention in the abstract, consider the following: the killer fig gets two sentences.  Any other figure could get up to one sentence, and result that is not a figure does not get mentioned.&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;Generate a “take home message” for each of the figures you plan to create.  These become the first sentence of the figure caption.&lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;Ask for feedback from senior and other co-first authors. Explain that the role of feedback at this stage is for them to tell you whether they think there are any glaring  flaws with the basic setup, eg, do they have a different idea about what is the biggest challenge.  There is no need to knit-pick about grammar at this stage.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Figures&lt;/strong&gt;
    &lt;ol&gt;
      &lt;li&gt;Make a first draft of all the figures and tables with detailed captions (~ 1 week).  Captions should each be about a paragraph long.  At this point, the figures need not be “camera ready”, but should have all the main points made.&lt;/li&gt;
      &lt;li&gt;Get feedback on figures from co-authors and close colleagues (~1 week).  Show them the figures, do not show them the captions, and ask them to tell you what the main point of each figure is.  If they don’t get it right, don’t worry about it, take notes on what they thought, and ask them to go to the next figure.  Then, spend another week updating the figures, and repeat.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Outline&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Write a “long-form” outline of the &lt;strong&gt;intro&lt;/strong&gt; (~1 hr).   This is essentially an expansion of the abstrast, and is therefore structured as follows:
        &lt;ol&gt;
          &lt;li&gt;a bulleted list of  ~3-5 main factors that create an opportunity for your work, filtering from most general to most specific, and not including anything ancillary (~20 min)&lt;/li&gt;
          &lt;li&gt;bulleted list of the ~3-5 main challenges that must be overcome (~20 min)&lt;/li&gt;
          &lt;li&gt;1 sentence summary of the &lt;strong&gt;gap&lt;/strong&gt;, that is, the key ingredient that is missing (~5 min)&lt;/li&gt;
          &lt;li&gt;2-3 sentence summary of what you did (~5 min)&lt;/li&gt;
          &lt;li&gt;2-3 sentence summary on how your work changes the world (~5 min)&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
      &lt;li&gt;
        &lt;p&gt;Outline the methods.  (~20 min)&lt;/p&gt;
      &lt;/li&gt;
      &lt;li&gt;Outline the results results (~20 min). Allow for at least one paragraph per figure, table, and/or theorem.  These paragraphs follow the following form (no need for them to be proper “paragraphs” at this point):
        &lt;ol&gt;
          &lt;li&gt;a sentence describing the setting,&lt;/li&gt;
          &lt;li&gt;a few sentences providing additional details, and&lt;/li&gt;
          &lt;li&gt;a concluding sentence providing the take home summary.&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
      &lt;li&gt;Outline the  discussion (~1 hr), to include (not a summary)
        &lt;ol&gt;
          &lt;li&gt;bulleted list of previous related work (~20 min)&lt;/li&gt;
          &lt;li&gt;bulleted list of potential extensions (~20 min)&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
      &lt;li&gt;Get feedback from your co-authors.  At this stage, the question is whether the logic is sound, meaing, based on the challenge you proposed, the sequence of results provides compelling evidence that you’ve satisfactorily addressed it. If they disagree, iterate until there is agreement.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Venue&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Read the top 3-5 most cited/downloaded articles from each potential venue to submit to, to see what the readership of that community likes.  This will help you both choose the appropriate venue, and write effectively for that venue.&lt;/li&gt;
      &lt;li&gt;Choose the actual journal/conference you will be submitting to.  Until this point, it is irrelevant, since you’ve only been focusing on the logical structure of your story.  However, before you start writing, you want to make sure you are writing for a particular target audience, and write for them.  This is because different venues have different expectations, which you’d like to meet.&lt;/li&gt;
      &lt;li&gt;Note the structure and approximate length of the venue of choice for the top articles published there.  Do &lt;em&gt;not&lt;/em&gt; worry about length at this time, just note down how many pages and figures they have, if they have an explicit methods section, if they have a preferred outline structure.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Draft&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Expand the outline into a full draft. Do &lt;em&gt;not&lt;/em&gt; concern yourself at this point with the details of what the publisher wants, we will deal with that later.  The point of this is to simply have a draft &lt;em&gt;at all&lt;/em&gt;.&lt;/li&gt;
      &lt;li&gt;Check that paper follows all relevant checklists, including:
        &lt;ol&gt;
          &lt;li&gt;&lt;a href=&quot;https://bitsandbrains.io/2018/09/08/figures.html&quot;&gt;figure guidelines&lt;/a&gt;&lt;/li&gt;
          &lt;li&gt;&lt;a href=&quot;https://bitsandbrains.io/2018/10/14/paragraphs.html&quot;&gt;paragraph guidelines&lt;/a&gt;.&lt;/li&gt;
          &lt;li&gt;&lt;a href=&quot;https://bitsandbrains.io/2018/10/14/words.html&quot;&gt;words guidelines&lt;/a&gt;.&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
      &lt;li&gt;Read the paper carefully outloud, and remove any words that are unnecessary.&lt;/li&gt;
      &lt;li&gt;Make sure each display item (e.g., figure, table, or theorem) is enumerated and explicitly refered to in the main text.  If you are using LaTeX, use &lt;code class=&quot;highlighter-rouge&quot;&gt;\ref{fig:&amp;lt;informative_name&amp;gt;}&lt;/code&gt; to refer to each display item. Recall that when referring to sections or figures, etc., the name of the section/figure is a proper noun, and is therefore capitalized.&lt;/li&gt;
      &lt;li&gt;Make sure you have sufficiently cited the literature to place your work in context.  For conferences, it is typical to have about 1 page of citations (10-20).  For journal articles, 30-50 is more typical.  Recall, the authors of these papers are likely to be the reviewers and readers for this paper. So, it is important that you highlight all the important work, and say how great it is. In particular, you want your readers to feel good about themselves while they are reading your work, which you can facilitate by citing their work, and explaining why it is so great and important.  This is, of course, actually true, since you are building on this work, and your work would likely not even be possible without the work you are citing. If you are using Google docs, I recommend using the &lt;a href=&quot;https://paperpile.com/blog/free-google-docs-add-on/&quot;&gt;Paperpile “Add-On”&lt;/a&gt;.&lt;/li&gt;
      &lt;li&gt;Get feedback again, ideally from a professional editor.  At this stage, feedback on logic/etc. is no longer appropriate, so be clear when asking for feedback that you are asking for feedback whether the individual sentences/paragraphs are clear, and random grammar/spelling mistakes.  The opportunity to provide feedback on logic has passed.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Revision&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Update abstract and introduction to finalize draft on text (~1 day).&lt;/li&gt;
      &lt;li&gt;Revise manuscript addressing each and every  concern you were made aware of by any of your readers at this point (~1 week).  This does not necessarily mean making new figures, rather, it might mean clarifying various points of confusion.&lt;/li&gt;
      &lt;li&gt;Do another round of feedback, give them another week.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;&lt;strong&gt;Submission&lt;/strong&gt;&lt;/p&gt;

    &lt;ol&gt;
      &lt;li&gt;Finalize manuscript (~1 wk).  I recommend that you fully &lt;em&gt;ignore&lt;/em&gt; all guidelines provided by the journal, and you submit that which you believe will be easiest for the editors/reviewers to read and understand.  I have almost always done it this way, and the submission is almost always reviewed anyway.  If it gets good reviews, we can modify it so it fits their rules.  An exception is conferences, where they actually care.&lt;/li&gt;
      &lt;li&gt;If your code is not yet open source, make it so, following the &lt;a href=&quot;https://bitsandbrains.io/2018/10/21/numerical-packages.html&quot;&gt;FIRM&lt;/a&gt; principles&lt;/li&gt;
      &lt;li&gt;If your data are not yet anonymized and open access, make it so, following the &lt;a href=&quot;https://www.go-fair.org/fair-principles/&quot;&gt;FAIR&lt;/a&gt; principles.&lt;/li&gt;
      &lt;li&gt;Draft a cover letter (in google docs if you are writing in google docs, or in the same overleaf repo as the manuscript if you are using overleaf).  The cover letter has the following form:
        &lt;ol&gt;
          &lt;li&gt;It is on institutional letterhead&lt;/li&gt;
          &lt;li&gt;Dear [name of editor that will be reading the letter],&lt;/li&gt;
          &lt;li&gt;Paragraph 1
            &lt;ol&gt;
              &lt;li&gt;We are delighted to submit to you our manuscript submission titled, “[title],” for publication as a [type of article, assuming there are multiple types in the journal].&lt;/li&gt;
              &lt;li&gt;Establish the &lt;em&gt;gap&lt;/em&gt; that this paper is filling is clear non-technical terms&lt;/li&gt;
              &lt;li&gt;State how one could address this gap.&lt;/li&gt;
            &lt;/ol&gt;
          &lt;/li&gt;
          &lt;li&gt;Pargraph 2:
            &lt;ol&gt;
              &lt;li&gt;Summary of main contributions of the manuscript, in non-technical terms. One sentence per contribution.&lt;/li&gt;
              &lt;li&gt;Conclude by stating what we expect the implications will be for their readership, and more broadly&lt;/li&gt;
            &lt;/ol&gt;
          &lt;/li&gt;
          &lt;li&gt;Signed by the corresponding author, “on behalf of my co-authors”&lt;/li&gt;
          &lt;li&gt;Add a ps - that all the code is open source and all the data is open access, in accordance with the above mentioned principles.&lt;/li&gt;
          &lt;li&gt;The language can be more flowery/confident about the expected importance of the contribution than you would write in the article itself.&lt;/li&gt;
        &lt;/ol&gt;
      &lt;/li&gt;
      &lt;li&gt;Submit to journal.&lt;/li&gt;
      &lt;li&gt;Post to pre-print server.&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;&lt;strong&gt;Rebuttal&lt;/strong&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Great, you’ve now submitted the work, waiting some number of weeks/months, and received detailed feedback from the editor and reviewers.  How to respond? Note, no matter what they say with regard to accept, minor revision, major revision, or reject, we essentially respond in the same way.&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;Respond immediately, unless you have a grant deadline, revising this paper is now your #1 top priority bar none.  All upcoming conference deadlines, and anything else you are working on now take a back seat.  The sooner you get the responses, the sooner they read them, and the less they forget what they read/wrote, and the less likely they are to re-read your whole paper, and give us a whole new set of complaints.&lt;/li&gt;
  &lt;li&gt;Make a google doc, copy the entirety of the comments you received into it. Change the font color for all of their comments to red.&lt;/li&gt;
  &lt;li&gt;After each “complaint”, make a line break, and respond directly to it. (Basically) Never disagree with them.   The game now is simply to address the comments (ego) of the reviewers.  Any comment they made is useful, at a minimum, to inform us as to places where readers might be confused about why what we did is so awesome.  Sometimes this will mean making your paper slightly worse to appease them.  I find it is typically worthwhile. In the response,
    &lt;ol&gt;
      &lt;li&gt;Tell them how great they are for identifying this problem with the manuscript, and express gratitude for them finding it, really try to feel grateful, your response will be more productive;&lt;/li&gt;
      &lt;li&gt;Tell them how you have altered the manuscript to address the issue. Yes, this means modifying the manuscript to address each and every one of their complaints, no matter how big/small they are.  This includes certainly citing everything they recommend you cite, and also praising that work in the actual text.&lt;/li&gt;
      &lt;li&gt;For any big changes, such as a modification/addition of a figure, or a new paragraph, directly append the new/revised content into this document.  The goal is to make sure the reviewer does not look back at the original manuscript, lest they might find additional limitations that they want addressed. Quotes from the paper should be in blue, to highlight that they are quotes.  Note: send the google doc proposing changes (but without implementing any yet) to your PI to get the green light on how to response, once you have the green light, proceed apace).&lt;/li&gt;
    &lt;/ol&gt;
  &lt;/li&gt;
  &lt;li&gt;Send google doc to your co-authors with track changes on.  Often a good idea to also make a back-up copy of the rebuttal prior to sharing with them just in case.  Specifically ask for feedback of the following form: “if you were the referee, and you saw that I responded thusly, would you be satisfied or not?”&lt;/li&gt;
&lt;/ol&gt;

&lt;hr /&gt;

&lt;p&gt;If you follow the above plan, you will have a manuscript ready to submit two to three months after you start writing.  Note that it includes many stages of feedback, and at each stage, a specific kind of feedback is explicitly requested.  This procedure helps streamline the amount of work you do between iterations, and streamlines the entire process. Good luck!&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="writing" /><category term="10" /><summary type="html">Learn how to write by reading how to structure a paper and Writing Science (sorry that I am recommending purchasing a book, though it should be available for free from any academic library, and I recommend all PIs own a copy for themselves and their students, so it does not cost students to read it at a minimum). Main result write draft paper title and draft killer fig that visually makes the main point of your story Abstract write title and abstract based on OCAR story structure (described in Writing Science) Figures draft the remaining figures Outline generate a 1 sentence per paragraph outline Venue choose the publication venue Draft flesh out details of paper Revision revise Submission post to arxiv, submit to journal, tweet to world Rebuttal respond to reviewers</summary></entry><entry><title type="html">What We Do</title><link href="https://bitsandbrains.io/2019/02/10/what-we-do.html" rel="alternate" type="text/html" title="What We Do" /><published>2019-02-10T18:27:57+00:00</published><updated>2019-02-10T18:27:57+00:00</updated><id>https://bitsandbrains.io/2019/02/10/what-we-do</id><content type="html" xml:base="https://bitsandbrains.io/2019/02/10/what-we-do.html">&lt;p&gt;This blog post is our attempt to characterize what we, at &lt;a href=&quot;https://neurodata.io/&quot;&gt;NeuroData&lt;/a&gt;, do. All of our actions are motivated by our mission, &lt;code class=&quot;highlighter-rouge&quot;&gt;to understand and improve intelligences&lt;/code&gt;; the motivation behind our approach is described in our &lt;a href=&quot;https://neurodata.io/about/&quot;&gt;about page&lt;/a&gt;.  This post summarizes the bulk or our actual work, which can be divided into three complementary threads:&lt;/p&gt;
&lt;ol&gt;
  &lt;li&gt;big data systems&lt;/li&gt;
  &lt;li&gt;statistics / machine learning / artificial intelligence&lt;/li&gt;
  &lt;li&gt;applications&lt;/li&gt;
&lt;/ol&gt;

&lt;h3 id=&quot;big-data-systems&quot;&gt;Big Data Systems&lt;/h3&gt;

&lt;p&gt;A big data system is a computational ecosystem, including hardware and software, design to support analysis of “big data”, operationally defined as data too big to fit into a single machine’s working memory.  The operations of any big data system include: storing, interfacing, pipelining, and visualizing. Existing solutions have been inadequate to support efficient hypothesis generation and scientific discovery.  We therefore build, extend, and deploy systems that generalize the functionality of existing systems to support scientific inquiry.  We first developed the Open Connectome Project stack in 2011 to host the first ever big dataset in neuroscience, a 10 terabyte electron microscopy dataset from Davi Bock and Clay Read (&lt;a href=&quot;https://doi.org/10.1145/2484838.2484870&quot;&gt;ocp&lt;/a&gt;).  As the data size and complexity increased, the development needs extended beyond our capabilities, so we began collaborating extensively with other teams, including scientists and engineers at &lt;a href=&quot;https://ai.google/research/people/VirenJain&quot;&gt;Google&lt;/a&gt;, &lt;a href=&quot;https://alleninstitute.org/what-we-do/brain-science/&quot;&gt;Allen Institute for Brain Science&lt;/a&gt;, and &lt;a href=&quot;https://www.janelia.org/&quot;&gt;Janelia Research Campus&lt;/a&gt;.  This collaboration led to our existing open source, community developed, computational ecosystem (&lt;a href=&quot;https://www.nature.com/articles/s41592-018-0181-1&quot;&gt;ndcloud&lt;/a&gt;). The core components of this work include &lt;a href=&quot;https://www.biorxiv.org/content/10.1101/217745v1&quot;&gt;bossDB&lt;/a&gt; for big data storage, &lt;a href=&quot;https://github.com/neurodata/ndwebtools&quot;&gt;NDWebtools&lt;/a&gt; for interfacing with the data, &lt;a href=&quot;https://github.com/google/neuroglancer&quot;&gt;neuroglancer&lt;/a&gt; for visualizing, and various pipelines for different modalities, including &lt;a href=&quot;https://neurodata.io/ndmg/&quot;&gt;ndmg&lt;/a&gt; for functional, structural, and diffusion magnetic resonance imaging and &lt;a href=&quot;https://neurodata.io/reg/&quot;&gt;reg&lt;/a&gt; for registration of 2D and 3D volumes.  We continue to extend our collaboration network and ecosystem, to support an ever growing need for big data systems in brain sciences across scales and modalities, ranging from electron microscopy, to whole clear brains, to human and non-human magnetic resonance imaging.&lt;/p&gt;

&lt;h3 id=&quot;statistics--machine-learning--artificial-intelligence&quot;&gt;Statistics / Machine Learning / Artificial Intelligence&lt;/h3&gt;

&lt;p&gt;Statistics, machine learning, and artificial intelligence are terms that describe complementary approaches to solving overlapping sets of problems.  Central to all of them is the existence of some data samples, and a question; these approaches then build tools  designed to learn an answer to the question from the data.  Existing tools, however, have severe limitations that we must overcome in order to obtain answers to the questions of interest.   First, raw data in neuroscience tends to be very high-dimensional (e.g., images can be petabytes), and exhibit many nonlinear relationships (e.g., the input/output function of a neuron).  We have therefore developed a number computational statistics tools for such settings. This includes state of the art methods for dimensionality reduction (&lt;a href=&quot;https://arxiv.org/abs/1709.01233&quot;&gt;LOL&lt;/a&gt;), classification (&lt;a href=&quot;http://arxiv.org/abs/1506.03410&quot;&gt;rerf&lt;/a&gt;), time-series modeling [mr. sid]](https://doi.org/10.1016/J.PATREC.2016.12.012),  clustering (&lt;a href=&quot;https://arxiv.org/abs/1710.09859&quot;&gt;eclust&lt;/a&gt;), and hypothesis testing (&lt;a href=&quot;https://elifesciences.org/articles/41690&quot;&gt;mgc&lt;/a&gt;).  Second, the eventual representation of data  most interesting to us is networks in the brain, or connectomes. We have therefore spent much of the last 10 years building foundational statistical estimators, theories, and algorithms for modeling populations of networks with graph, vertex, and edge attributes.  A  survey summarizing much of our work was recently published in JMLR (&lt;a href=&quot;http://jmlr.org/papers/v18/17-448.html&quot;&gt;rdpg&lt;/a&gt;).  We subsequently developed a python package that implements all of our theoretical developments (&lt;a href=&quot;https://neurodata.io/graspy/&quot;&gt;graspy&lt;/a&gt;), and wrote a review article on our approach to modeling connectomes, called &lt;a href=&quot;https://doi.org/10.1016/j.conb.2019.04.005&quot;&gt;&lt;code class=&quot;highlighter-rouge&quot;&gt;connectal coding&lt;/code&gt;&lt;/a&gt;.&lt;/p&gt;

&lt;h3 id=&quot;applications&quot;&gt;Applications&lt;/h3&gt;

&lt;p&gt;The potential applications of our work are widespread.  We illustrate a couple applications spanning our most relevant work, including model (non-human) systems and  human variation.  First, we characterized which sets of neurons were causally involved in which behaviors, in &lt;a href=&quot;https://doi.org/10.1126/science.1250298&quot;&gt;larval Drosophila&lt;/a&gt;. This led to an exploratory analysis of the mushroom body of the larval Drosophila, which we describe in detail in a &lt;a href=&quot;http://arxiv.org/abs/1705.03297&quot;&gt;technical report&lt;/a&gt;, with further analysis available in our &lt;a href=&quot;http://jmlr.org/papers/v18/17-448.html&quot;&gt;survey JMLR paper&lt;/a&gt;, and our &lt;a href=&quot;https://doi.org/10.1016/j.conb.2019.04.005&quot;&gt;connectal coding paper&lt;/a&gt;.  Second, to characterize human variation, we developed a cloud pipeline for estimation and analysis of human connectomes (&lt;a href=&quot;https://doi.org/10.1093/gigascience/gix013&quot;&gt;sic&lt;/a&gt;), and used it to quantify variability present within and across individuals and studies (&lt;a href=&quot;https://doi.org/10.1101/188706&quot;&gt;ndmg&lt;/a&gt;).&lt;/p&gt;

&lt;h3 id=&quot;open-science&quot;&gt;Open Science&lt;/h3&gt;

&lt;p&gt;A key principle underlying our work is the democratization of science: we desire that anybody, regardless of resources, is able to both access and contribute to the cutting edge of scientific discovery.  To that end, all of the work we do is open science, including both open source code and open access data.  Over the last 8 years, our website, https://neurodata.io, has grown from about 100 unique visitors each week to &amp;gt;1,000 unique visitors each month.  In the last calendar year,  &amp;gt;13,000 unique visitors browsed the site.  These visitors span every inhabited continent, including over 2,000 different cities (see image below).  Over the 8 years, approximately 80,000 unique individuals have visited our site, over double the number of people at a typical Society for Neuroscience conference, suggesting that many non-neuroscientists have visited our site. We hope that those people visiting the site feel the same sense of awe and inspiration as we do, associated with unraveling the secrets of mental function in these beautiful images of brains.&lt;/p&gt;

&lt;p&gt;&lt;img src=&quot;/assets/post_images/neurodata-visitors.png&quot; alt=&quot;significance&quot; /&gt;&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><summary type="html">This blog post is our attempt to characterize what we, at NeuroData, do. All of our actions are motivated by our mission, to understand and improve intelligences; the motivation behind our approach is described in our about page. This post summarizes the bulk or our actual work, which can be divided into three complementary threads: big data systems statistics / machine learning / artificial intelligence applications</summary></entry><entry><title type="html">Some thoughts on building interdisciplinary biological/data science teams</title><link href="https://bitsandbrains.io/2019/02/10/cross-disciplinary-teams.html" rel="alternate" type="text/html" title="Some thoughts on building interdisciplinary biological/data science teams" /><published>2019-02-10T18:27:57+00:00</published><updated>2019-02-10T18:27:57+00:00</updated><id>https://bitsandbrains.io/2019/02/10/cross-disciplinary-teams</id><content type="html" xml:base="https://bitsandbrains.io/2019/02/10/cross-disciplinary-teams.html">&lt;p&gt;I recently was asked a few questions on this topic, and thought I’d share my answers.  I did, tweeted it, and then updated this post based on public and private responses.  For the twitter discussion, including contributors, please see &lt;a href=&quot;https://twitter.com/neuro_data/status/1146830916863631361&quot;&gt;here&lt;/a&gt;.  I’m grateful to everyone who has contributed to this dialog. If anybody has contrary opinions, I’d love to know about them.&lt;/p&gt;

&lt;p&gt;What do effective biologist/data scientist teams look like?&lt;/p&gt;

&lt;p&gt;While I’m sure there are many different flavors of succesful teams, &lt;em&gt;we&lt;/em&gt; have been most succesful when team  have the following properties:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;A relatively concrete biological question, with a careful experimental design.  Good data scientists can find patterns in anything, but not all patterns are scientifically interesting.&lt;/li&gt;
  &lt;li&gt;The data scientist is integrated into the scientific process as early as possible.  For example, there are often choices the experimentalist could make that would result in much easier or more difficult challenges for the data scientist. Only by virtue of the data scientist being engaged before the data are collected is it even possible for them to provide input into these decisions.&lt;/li&gt;
  &lt;li&gt;Both the data scientist and the neurobiologist have deep respect for the other’s time and expertise. In particular, both realize that the challenges the other faces are difficult to overcome, and may take time.  At the same time, both realize that they are working to answer a particular question, rather than solve some “generic” problem.  For example, the data scientist must develop an algorithm that detects cell bodies with sufficiently high sensitivity and specificity to answer the particular neurobiological question of interest, rather than “the best cell body detection method ever”.  This may require either or both individuals to go outside their comfort zones at time for the success of the project.&lt;/li&gt;
  &lt;li&gt;If novelty and complexity are not required for the methods, even if such properties would be sexy, neither party insists on pursuing such approaches, which can be quite time consuming and typically not a dramatic improvement on existing technology.&lt;/li&gt;
  &lt;li&gt;Often making authorship agreements up front can avoid future clashes.  One workable approach is a single manuscript approach, with  first-author and corresponding-author rights shared by both labs.  Another reasonable approach is a two-manuscipt approach, with one manuscript describing the data science technique, and another describing the biology.  A potential downside to this approach is that often the data science technique does not have a natural “peer reviewed” home, because it is a “trivial” extension of previous work. That is not to say that it did not require a Herculean amount of work to get it working sufficiently well, but rather, that many stats/AI/ML peer-reviewed venues do not highly value this kind of work.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;One caveat to the above, is that it assumed a legitimate collaboration.  Sometimes, the biologists simply needs a consultant, to check their work, suggest a complementary strategy, etc.  Those can often be quite fruitful for the biologist, although the data scientist rarely gets any “academic credit” for such efforts.  On the other hand, for serious collaborations, getting the data scientist up to speed on the biology can often take quite some time, for example one to two years even.  Doing so is a considerable energetic investment for all parties involved.  In these cases, it can be advantageious for everyone if there are smaller “wins” in the interm period that indicate to all parties that continued investment is likely worthwhile.&lt;/p&gt;

&lt;p&gt;How deeply are they integrated?&lt;/p&gt;

&lt;p&gt;In our experience, weekly (potentially remote) meetings between both neurobiologist and data scientist are incredibly helpful to discuss current results and get further direction for next steps.  The data scientist also visits the neurobiologist at least once in the beginning of the project to witness a day of experiment, to understand what the data are.  Quarterly or semi-annual in person visits are also quite beneficial, in our experience.&lt;/p&gt;

&lt;p&gt;What is the management structure?&lt;/p&gt;

&lt;p&gt;Usually one trainee acquiring data and one trainee developing/applying methods, each supervised by their respective PI.  Once the method is fairly well established and routinely able to run by the data science trainee, it is often helpful to engage a software engineer on the data science side, to transition the code from “gradware” to at least “labware”, meaning that other people in the data science lab could also run the analysis, thereby freeing up the trainee to write a manuscript, further develop the method, or transition to other projects.  Another complementary option is to engage a younger trainee to transition the project to them.&lt;/p&gt;

&lt;p&gt;Do you have a sense for if the data scientists in the teams will continue in tackling tough biological problems, or is this perceived more as a training opportunity for junior researchers/scientist who may be looking at industry careers?&lt;/p&gt;

&lt;p&gt;With the artificial intelligence industry sky rocketing, a large fraction of trainees are interested in transitioning to industry.  Industry can be start-ups, big tech companies, the financial sector, or biotech. In each of those industries, there are now opportunities to work in data science as applied to biological problems, particularly in healthcare.  Those trainees that are motivated specifically to tackle the biology question are more likely to continue doing so when they graduate.  And in fact, those that are inspired to actually solve the problem, rather than transition to a different industry, I find to be much more effective at solving the biological question.&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="collaboration" /><summary type="html">I recently was asked a few questions on this topic, and thought I’d share my answers. I did, tweeted it, and then updated this post based on public and private responses. For the twitter discussion, including contributors, please see here. I’m grateful to everyone who has contributed to this dialog. If anybody has contrary opinions, I’d love to know about them.</summary></entry><entry><title type="html">Bilateral Homology</title><link href="https://bitsandbrains.io/2019/02/10/bilateral-homology.html" rel="alternate" type="text/html" title="Bilateral Homology" /><published>2019-02-10T18:27:57+00:00</published><updated>2019-02-10T18:27:57+00:00</updated><id>https://bitsandbrains.io/2019/02/10/bilateral-homology</id><content type="html" xml:base="https://bitsandbrains.io/2019/02/10/bilateral-homology.html">&lt;p&gt;Homology, means having the same or similar relation, relative position, or structure.
In biology, &lt;a href=&quot;https://en.wikipedia.org/wiki/Homology_(biology)&quot;&gt;homology&lt;/a&gt; usually refers to comparison across taxa, like human vs non-human primate. 
Two biological structures are homologous in different taxa if they are derived from a common ancestor, such as feet and flippers.&lt;/p&gt;

&lt;p&gt;Within an individual, people speak of &lt;a href=&quot;https://en.wikipedia.org/wiki/Symmetry_in_biology#Bilateral_symmetry&quot;&gt;bilateral symmetry&lt;/a&gt;, 
which means there is an approximate reflection symmetry about some body axis. 
For example, the human cortex is bilaterally symmetric. This is despite the fact that the left side has regions of the brain that are not present in the right side.&lt;br /&gt;
In other words, if one were to quantify the degree of differences between the two sides, bilateral symmetry only requires that the difference between the two sides is less than some threshold.  Where one puts the threshold is a subjective matter.&lt;/p&gt;

&lt;p&gt;Bilateral homology is a relatively new concept.  For now, we say that two biological structures are bilaterally homologous if, under “normal” environmental conditions, the degree of bilateral symmetric is &lt;script type=&quot;math/tex&quot;&gt;c&lt;/script&gt;, 
and under certain perturbations of the conditions, the degree of bilateral symmetry is &lt;script type=&quot;math/tex&quot;&gt;c'&lt;/script&gt;, and &lt;script type=&quot;math/tex&quot;&gt;c' &gt; c&lt;/script&gt;. Which means that, by definition, any “normal” connectome cannot exhibit bilateral homology.&lt;/p&gt;

&lt;p&gt;To be a bit more concrete, for a particular individual, we observe n points from one side, &lt;script type=&quot;math/tex&quot;&gt;x_1,\ldots, x_n&lt;/script&gt;, 
and m points from the other side, &lt;script type=&quot;math/tex&quot;&gt;y_1,\ldots, y_m&lt;/script&gt;. 
Assume that each observation is sampled independently from the others, and that those from the first side are sampled from some true but unknown distribution &lt;script type=&quot;math/tex&quot;&gt;F&lt;/script&gt;, and those from the other side are sampled from some  true but unknown distribution &lt;script type=&quot;math/tex&quot;&gt;G&lt;/script&gt;.  We can then posit a formal hypothesis test:&lt;/p&gt;

&lt;script type=&quot;math/tex; mode=display&quot;&gt;H0: F = G, \qquad  HA: F \neq G&lt;/script&gt;

&lt;p&gt;Implementing this test requires choosing a test statistic&lt;/p&gt;

&lt;script type=&quot;math/tex; mode=display&quot;&gt;t=t( x_1,...,x_n, y_1,..., y_m) \in \mathcal{R},&lt;/script&gt;

&lt;p&gt;and we reject bilateral homology if and only if the observed &lt;script type=&quot;math/tex&quot;&gt;t&lt;/script&gt; is more extreme than &lt;script type=&quot;math/tex&quot;&gt;(1 - \alpha)\%&lt;/script&gt; of the time under the null.&lt;/p&gt;

&lt;p&gt;This test, like all other tests, requires a distribution of the test statistic &lt;script type=&quot;math/tex&quot;&gt;t&lt;/script&gt; under the null. One way to proceed would be to acquire S “typical” individuals, and compute &lt;script type=&quot;math/tex&quot;&gt;t&lt;/script&gt; for each of them. Then, given 1 individual that experienced an experimental perturbation, we could accurately ascertain how extreme this individuals test statistic is relative to the “typical” ones.  This approach, of course, is sensitive to both the choose of test statistic and model: &lt;script type=&quot;math/tex&quot;&gt;F, G \in \mathcal{F}&lt;/script&gt;.  Nonetheless, reasonable choices for both are available in many situations.&lt;/p&gt;

&lt;p&gt;On the other hand, in the absence of S typical individuals, how might one proceed?  One could sample S times from an estimated null.  This would depend &lt;em&gt;strongly&lt;/em&gt; on &lt;script type=&quot;math/tex&quot;&gt;\mathcal{F}&lt;/script&gt;, but is still possible.&lt;/p&gt;

&lt;p&gt;Now, assume we observe only a single typical individual. The question we are interested in asking is: are the left and the right “more different than one would expect by chance?”.  Since we’ve only observed 1 individual, how might we go about estimating the null? One option would be to fit a series of increasingly complex models to the one side, &lt;script type=&quot;math/tex&quot;&gt;\mathcal{F}_1, \mathcal{F}_2, \ldots&lt;/script&gt;, 
and ascertain their model fit the other side. Then, do the same training models on the right, and ascertaining them on the left. Assuming for simplicity that both experiments revealed that the same model was the best fit across sides, we can operate under the assumption that this is a reasonable model.  Now, we can proceed as proposed above: using this selected model, estimate the null distribution by sampling, which provides a p-value.&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="brains" /><category term="bits" /><summary type="html">Homology, means having the same or similar relation, relative position, or structure. In biology, homology usually refers to comparison across taxa, like human vs non-human primate. Two biological structures are homologous in different taxa if they are derived from a common ancestor, such as feet and flippers.</summary></entry><entry><title type="html">Tips for Getting into a Top Graduate Program</title><link href="https://bitsandbrains.io/2018/10/21/getting-into-grad-school.html" rel="alternate" type="text/html" title="Tips for Getting into a Top Graduate Program" /><published>2018-10-21T18:27:57+00:00</published><updated>2018-10-21T18:27:57+00:00</updated><id>https://bitsandbrains.io/2018/10/21/getting-into-grad-school</id><content type="html" xml:base="https://bitsandbrains.io/2018/10/21/getting-into-grad-school.html">&lt;p&gt;As faculty in the &lt;a href=&quot;https://www.bme.jhu.edu/&quot;&gt;Department of Biomedical Engineering&lt;/a&gt; at Johns Hopkins University, the best BME department in the world, both in terms of &lt;a href=&quot;https://www.usnews.com/best-colleges/rankings/engineering-doctorate-biological-biomedical&quot;&gt;undergraduate&lt;/a&gt; and &lt;a href=&quot;https://www.usnews.com/best-graduate-schools/top-engineering-schools/biomedical-rankings&quot;&gt;graduate&lt;/a&gt; schools, I have learned what I and other faculty are looking for in applicants. Before getting into what the most important factors are, I believe it is important to understand our goals, which motivate those factors.  From the perspective of the graduate admissions committee, our goal is to estimate whether we believe that BME@JHU is the best place for &lt;em&gt;you&lt;/em&gt; to thrive to achieve you ultimate dreams.  In other words, we try to ascertain whether the environment that we create at JHU will be maximally supportive of both your strengths and weaknesses.  As it turns out, this does not mean necessarily that we accept the best students in some abstract sense (as defined by some arbitrary metric), but rather, we try to accept the students for which we believe that &lt;em&gt;we&lt;/em&gt; will be the best mentors for you. Of course, this is a complicated objective function, and one for which we will most likely sometimes make errors.  Nonetheless, it is our goal.  To make such estimations, we look for the following:&lt;/p&gt;

&lt;ol&gt;
  &lt;li&gt;
    &lt;p&gt;Research Experience: First and foremost, we are a research university.  So, the best way for us to determine whether our research environment will support you to flourish is to understand your previous research experience, and in which settings you flourished more than others.  Although successful research is difficult to quantify, research artifacts provide some data with which we can evaluate your achievements.  Such artifacts include poster presentations, conference proceedings, pre-prints, journal publications, numerical packages, and even patents sometimes.  If you are the first (or co-first) author on any of these, we typically assume that much of the work is yours, and thus first author research artifacts are most informative.  Middle author works are also informative, especially if you clarify your role in the research in your personal statement. Note, however, strong research experience is &lt;em&gt;not&lt;/em&gt; a pre-requisite for admission. Rather, it is an information-rich piece of data for us.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;Grades: JHU is not just a research university, we are also a teaching university, and we take our teach responsibilities quite seriously.  Moreover, many of our graduate level BME courses are also serious and time consuming.  It is important to us that you perform well in them, because they provide the necessary background upon which our research programs are based.  It is &lt;em&gt;not&lt;/em&gt; important to us that you got straight A’s, very few applicants have.  Rather, we care that you perform well in the courses that will be the most relevant for your research during your PhD, typically quantitative and biology classes for us.  It is also not crucial that you performed well in every semester. The grades in the most recent semesters, in the most relevant courses, are most important.  Aim for getting A’s in them, but GPA alone neither gets students admission nor rejection. We understand that life happens, and certain things are more important than coursework (family, health, well-being, etc.). Finally, we appreciate that not everybody gets the same opportunities, in life, in high school, etc., and therefore not everybody is equally well prepared for our coursework. That is ok, we are trying to estimate whether you will be successful in our program.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;Recommendations: While these do not come directly from you, they are quite important to us.  BME@JHU is like a big extended family. We work closely with one another, sit near each other (typically), we have been doing so for a long time, and plan to continue doing so for many years to come.  Therefore, our community is quite important to us, and our success comes largely from surrounding ourselves not just with the smartest people in the world, but more importantly, really good people.  So, the recommendation letters are a way for us to get information about how pleasant it is to work with you.  I particularly look for recommendations from other faculty with successful research programs, as they are the most informative with regards to what it takes to have a successful PhD. Recommendations from industry can be somewhat informative, but less so. In other words, the number of PhD students somebody has mentored matters in our assessment.  In terms of content, we are looking for recommendations that write that you are pleasant to work with, and amongst the best of his/her previous students along &lt;em&gt;some&lt;/em&gt; dimensions, such as productivity, passion, drive, creativity, organization, etc. In other words, you excel in the kinds of personality traits that we think contribute to successful PhDs.  Much like research experience and grades, a good or bad recommendation cannot determine your acceptance.&lt;/p&gt;
  &lt;/li&gt;
  &lt;li&gt;
    &lt;p&gt;Personal Statement: Your personal statement is your opportunity to express yourself. The most important aspect of a personal statement for me is &lt;strong&gt;passion&lt;/strong&gt;.  Success in our field, I believe, is strongly correlated with passion.  Even if that is not the case, it is more fun for me to work with people that are passionate about solving some problems. So, express yourself freely and passionately. And be specific. Find a few faculty members in the department that you are applying to, and write about what, in particular, you find most exciting about their work.  In this way, we’ll be able to align your passions with ours in the review process.  If you’ve reached out to any of the faculty, or anybody else associated with the department prior to application, mention it, and how it has informed your decision to apply.  I recommend that you do reach out to faculty in advance, if possible. And don’t forget to spell/grammar check it.&lt;/p&gt;
  &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;There are a few things that people invest a bunch of energy in, that do not matter hardly at all. First is the GRE.  Evidence is building that it is classist, racist, and sexist (see for example, &lt;a href=&quot;http://www.takepart.com/article/2015/11/07/gre-bias&quot;&gt;here&lt;/a&gt;, though see counter-points &lt;a href=&quot;https://www.insidehighered.com/admissions/article/2018/09/17/decision-penns-philosophy-department-renews-debate-about-gre&quot;&gt;here&lt;/a&gt;).  Several schools have stopped using them, but not all (there are lists online).   As it currently stands at BME@JHU, only if somebody does quite poorly on the quantitative aspect of the GRE (say, below 70%), does his/her GRE score even typically come up for discussion. In certain cases, the GRE can be waived, so we encourage you to email and ask. Second, is fellowships. In general, if we have not heard of them, do not understand the criteria for winning them, who applies, or what is achieved, it is hard to evaluate their value.  I’ve literally never ever heard them come up in discussing any applicant, and I’ve now been privvy to discussion literally hundreds, maybe over 1,000 applicants across multiple different departments.&lt;/p&gt;

&lt;p&gt;My lab, as well as many other successful labs, are &lt;em&gt;always&lt;/em&gt; accepting exceptional graduate students.  The success of our labs’ depends on the success of excellent students, so we are always searching for and hoping to find people whose passions align with ours, and whose abilities either align with or complement our own.&lt;/p&gt;

&lt;p&gt;Finally, we strongly encourage applications from diverse individuals. We try our best to evaluate each individual in the context of his/her/their background.  We believe that a more diverse and inclusive academic environment leads to better science, and a better society.&lt;/p&gt;

&lt;p&gt;I hope this is helpful. If anybody disagrees with my assessment, or has other recommendations, or further questions, I’d love to hear from you in the comments.&lt;/p&gt;</content><author><name>Joshua Vogelstein</name></author><category term="education" /><summary type="html">As faculty in the Department of Biomedical Engineering at Johns Hopkins University, the best BME department in the world, both in terms of undergraduate and graduate schools, I have learned what I and other faculty are looking for in applicants. Before getting into what the most important factors are, I believe it is important to understand our goals, which motivate those factors. From the perspective of the graduate admissions committee, our goal is to estimate whether we believe that BME@JHU is the best place for you to thrive to achieve you ultimate dreams. In other words, we try to ascertain whether the environment that we create at JHU will be maximally supportive of both your strengths and weaknesses. As it turns out, this does not mean necessarily that we accept the best students in some abstract sense (as defined by some arbitrary metric), but rather, we try to accept the students for which we believe that we will be the best mentors for you. Of course, this is a complicated objective function, and one for which we will most likely sometimes make errors. Nonetheless, it is our goal. To make such estimations, we look for the following:</summary></entry></feed>