Team:Berkeley/Project/Automation

From 2012.igem.org

(Difference between revisions)
Line 219: Line 219:
<img src="https://static.igem.org/mediawiki/2012/6/62/Circled.jpg" width="400" hspace = "10">
<img src="https://static.igem.org/mediawiki/2012/6/62/Circled.jpg" width="400" hspace = "10">
</div>
</div>
-
<div class = "col12">
+
<div class = "col3-2">
<img src="https://static.igem.org/mediawiki/2012/4/47/FormFactor.jpg" width="400" hspace = "10">
<img src="https://static.igem.org/mediawiki/2012/4/47/FormFactor.jpg" width="400" hspace = "10">
</div>
</div>

Revision as of 02:00, 27 October 2012

header
iGEM Berkeley iGEMBerkeley iGEMBerkeley

Mercury

We can accommodate libraries of over a million members through our MiCodes design scheme. With a data set this large, we needed to develop methods to make MiCodes a high-throughput screening technique. Because this library utilizes visual phenotypes, two aspects of microscopy needed to be automatable: image acquisition and image processing.


If a MiCode library size is very large, it is important to find a time-efficient method of taking many pictures of yeast cells. This summer we dealt with libraries on the order of 10^3 members, and our images were taken with a regular fluorescence microscope. Larger libraries would most likely make use of automated stages to speed up image acquisition.


The segmentation of cells from an image allows for cell-by-cell analysis in downstream steps. In order to perform such analyses, we wrote cell segmentation software using MATLAB to recognize an individual cell from its background. We first used edge detection with the Sobel operator and several filtering options to approximate possible cell outlines in the image. Then we performed a series of dilation and erosion steps to clear background pixels. We also added an additional filtering routine to refine the overall cell segmentation algorithm. This refining algorithm uses several geometric criteria, which we collected through testing of approximately 400 sample images. Our cell identification (cellID) file can be downloaded here.

On the left we see the original image that requires cell segmentation. On the right we have the processed image where each cell has
been identified as a separate object.


The cellID program is currently optimized for identifying yeast cells and works best for images in which cell clumping is at a minimum. We avoided cell clumping by experimentally altering the cell density of the samples we imaged. Overall, we have seen a successful identification rate of 96 percent for images we have tested. Images of each individual cell were generated, as shown below, and saved at the end of this process for use with downstream analysis. These saved images have several uses. For example, the cell images can be converted into a binary mask to sort out the organelles of a cell for several fluorescent channels. The number of images generated can also be used to count the number of cells identified from a certain population.


An essential portion of automation is determining morphological features. In the context of organelle detection, we developed pipelines to collect geometric measurements characteristic of each organelle (actin, plasma membrane, vacuolar membrane, and nucleus), and defined specific measurements and cutoff values as our feature set for each organelle.

To achieve this, we imaged a small sample size of about 100 cells with varying phenotypes, and determined feature sets associated with each of the four organelles. We used CellProfiler 2.0, an open-source software designed for quantitative analysis of cellular phenotypes. We used this software to build pipelines that demonstrate the feasibility of automated identification of cellular phenotypes. The nucleus and actin MiCodes identification pipelines can be viewed here, in .cp format (for CellProfiler 2.0).

Pipeline procedure:



1) Acquire single-celled images of each channel.

2) Convert cells to grayscale, and apply a thresholding algorithm, removing background pixels and retaining foreground.
3) Determine distinct blobs of high intensity pixels. These established as undefined "objects".
4) Take geometric and intensity measurements for each distinct "object".

In step (2), the thresholding algorithm we used was a Robust Background Adaptive operation, which is a threshold technique available in CellProfiler. We found that this method most effective in high background images, which is often the case when analyzing subcellular phenotypes over cytosolic background.




Establishing geometric features:

In order to train the computer to recognize blobs as organelles, we needed to develop a comprehensive set of criteria that would indicate the identity of a blob recognizable by the computer. To do this, we took geometric measurements of positive control images (only one type of organelle present per cell). We used these measurements to build feature sets associated with each organelle. The measurements we gathered are displayed below:



To better refine the organelle detection, we also intend to analyze the following additional properties to improve the classification:

Geometric Property Description
Solidity Area/Convex Area, or how "solid" a blob is. 
Eccentricity Ratio of the distance between foci of an ellipse and its major axis length. 


4. Use the measurements taken to classify each "object" as a particular organelle.



Results: Our software proved capable of successfully identifying MiCoded nuclei 95% of the time when processing our sample image size (100 cells). At this stage, our pipelines are primarily designed to differentiate between fluorescent membranes and punctate structures, while taking measurements. In the future, with a large, already-constructed dataset (a complete MiCode library), we would likely employ machine learning techniques to improve the automated identification.