Code Monkey home page Code Monkey logo

jump-scope-analysis's Introduction

JUMP-Scope

Data analysis for the JUMP-Scope project. The preprint can be viewed here: https://www.biorxiv.org/content/10.1101/2023.02.15.528711v1.

Images and profiles examined here are part of the Cell Painting Gallery, dataset name cpg0002-jump-scope.

To download the profiles specifically, you can run the following command:

aws s3 sync s3://cellpainting-gallery/cpg0002-jump-scope/source_4/workspace/profiles/ ./profiles/ --no-sign-request

jump-scope-analysis's People

Contributors

callum-jpg avatar nasimj avatar

Watchers

Anne Carpenter avatar David Logan avatar Arnaud Ogier avatar Daniel Kuhn avatar Beth Cimini avatar  avatar  avatar Niranj Chandrasekaran avatar  avatar

jump-scope-analysis's Issues

Preliminary analysis

Golas:

  • identify optimal settings for each instrument
  • quantify the magnitude of impact of these variations (are they relatively small, or substantial?)
  • assess the ability to match signatures across samples imaged by different instruments

Conclusions:


Scope Vendors:

  • Molecular Devices
  • Nikon
  • Perkin Elmer
  • Yokogawa US
  • Yokogawa Japan

Additional information on scopes and imaging details can be found here.


Experimental details:

Plate map and compounds: JUMP-MOA plate map and compounds were used (90 compounds from 47 distinct MOA classes, with 4 replicates per compound)

Test effects of FOV number and/or cell count on profile strength

We have a qualitative difference between conditions that have higher FOV counts (20X, typically) those with lower FOV counts (10X, typically, though not exclusively). This matches previous work

A couple of complicating factors here though -

  1. 4 fields of 10X vs 9 fields of 20X may have equal numbers of cells, even with lower numbers of fields. We've always speculated that more fields is better data because more cells, but maybe it isn't - maybe it's more chances to get a really good segmentation, or some other thing we haven't thought of. Our previous experiments were all tested at the same magnification, with cell count and FOV count essentially perfectly correlated, so we have no data addressing this directly.
  2. It may truly be that other magnifications really are worse than our standard 20X, but right now we can't tell because we know that FOV count MAY be a contributing factor.

To address these, I think we need to do 2 things -

  • Subsample all the existing batches to all the applicable FOV counts we can - we have batches with 1, 2, 3, 4, and 9 sites, so we should subsample 9 site batches to 4, 3, 2, and 1 sites, subsample for site batches to 3, 2, 1 sites, etc. Because we are currently pretty sure that we can use the same recipe run for all the batches being subsampled to 4 sites, another one for all the batches subsampled to 3, etc, this is actually only 4 config files to generate and run.
  • Once we've done the above, for each vendor look at a 2x2 matrix - percent matching / percent replicating vs FOV count/ cell count. I think this will help us be more definitive about both questions above.

Look at distributions of features in each batch

The question was raised as to whether the feature composition of each batch is the same; I think it's going to be hard to do this on a per feature level because of the random dropout of highly correlated features, but I do think there are a few metrics we can quite easily generate for each batch based just on the columns present in each CSV:

  • How many total features did this batch use?
  • What percent of features are Cells vs Nuclei vs Cytoplasm? (These should add to 100)
  • What percent of features are Texture vs Neighbors vs AreaShape etc? (These should add to 100)
  • What percent of features are RNA vs DNA vs ER vs Mito vs AGP vs BF? (These should not typically add to 100 but may coincidentally, since AreaShape features have no channels and Colocalization have 2)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.