1 / 45

Central dogma of biology DNA  RNA  pre-mRNA  mRNA Protein

Central dogma. Central dogma of biology DNA  RNA  pre-mRNA  mRNA Protein. DNA:. CGAACAAACCTCGAACCTGCT. Translation. Basic molecular biology. Transcription. mRNA:. GCU UGU UUA CGA. Polypeptide:. Ala Cys Leu Arg. Transcription. End modification. Splicing.

tannar
Télécharger la présentation

Central dogma of biology DNA  RNA  pre-mRNA  mRNA Protein

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Central dogma Central dogma of biology DNA  RNA  pre-mRNA  mRNA Protein

  2. DNA: CGAACAAACCTCGAACCTGCT Translation Basic molecular biology Transcription mRNA: GCU UGU UUA CGA Polypeptide: Ala Cys Leu Arg

  3. Transcription End modification Splicing Transport Translation Less basic molecular biology

  4. RNA Biological Sample Test Sample Test Sample Reference PE Cy3 Cy5 ARRAY ARRAY Oligonucleotide Synthesis cDNA Clone (LIBRARY) PCR Product Microarray technology Ramaswamy and Golub, JCO

  5. Microarray technology Oligonucleotide cDNA Lockhart and Winzler 2000

  6. Microarray experiment Yeast experiment

  7. Analytic challenge When the science is not well understood, resort to statistics: Infer cancer genetics by analyzing microarray data from tumors Ultimate goal: discover the genetic pathways of cancers Immediate goal: models that discriminate tumor types or treatment outcomes and determine genes used in model Basic difficulty: few examples 20-100, high-dimensionality 7,000-16,000 genes measured for each sample, ill-posed problem Curse of dimensionality: Far too few examples for so many dimensions to predict accurately

  8. Cancer Diagnosis Acute Myeloblastic Leukemia v Acute Lymphoblastic Leukemia

  9. Cancer Classification 38 examples of Myeloid and Lymphoblastic leukemias Affymetrix human 6800, (7128 genes including control genes) 34 examples to test classifier Results: 33/34 correct d perpendicular distance from hyperplane d Test data

  10. Two gene example: two genes measuring Sonic Hedgehog and TrkC Coregulation and kernels Coregulation: the expression of two genes must be correlated for a protein to be made, so we need to look at pairwise correlations as well as individual expression Size of feature space: if there are 7,000 genes, feature space is about 24 million features, so the fact that feature space is never computed is important

  11. Gene coregulation Nonlinear SVM helps when the most informative genes are removed, Informative as ranked using Signal to Noise (Golub et al). • Genes removed errors • 1st order 2nd order 3rd order polynomials • 0 1 1 1 • 10 2 1 1 • 20 3 2 1 • 30 3 3 2 • 40 3 3 2 • 50 3 2 2 • 100 3 3 2 • 200 3 3 3 • 1500 7 7 8

  12. Cancer g2 Reject Normal g1 Rejecting samples Golub et al classified 29 test points correctly, rejected 5 of which 2 were errors using 50 genes Need to introduce concept of rejects to SVM

  13. Rejecting samples

  14. Estimating a CDF

  15. The regularized solution

  16. 95% confidence or p = .05 d = .107 .95 Rejections for SVMs P(c=1 | d) 1/d

  17. Results with rejections Results: 31 correct, 3 rejected of which 1 is an error d Test data

  18. Gene selection SVMs as stated use all genes/features Molecular biologists/oncologists seem to be convinced that only a small subset of genes are responsible for particular biological properties, so they want the genes most important in discriminating Practical reasons, a clinical device with thousands of genes is not financially practical Possible performance improvement Wrapper method for gene/feature selection

  19. d d Test data Test data Results with gene selection AML vs ALL: 40 genes 34/34 correct, 0 rejects. 5 genes 31/31 correct, 3 rejects of which 1 is an error. B vs T cells for AML: 10 genes 33/33 correct, 0 rejects.

  20. Molecular classification of cancer • Hierarchy of difficulty: • Histological differences: normal vs. malignant, skin vs. brain • Morphologies: different leukemia types, ALL vs. AML • Lineage B-Cell vs. T-Cell, folicular vs. large B-cell lymphoma • Outcome: treatment outcome, elapse, or drug sensitivity.

  21. Morphology classification

  22. Outcome classification

  23. p-val = 0.00039 p-val = 0.0015 Outcome classification Error rates ignore temporal information such as when a patient dies. Survival analysis takes temporal information into account. The Kaplan-Meier survival plots and statistics for the above predictions show significance. Lymphoma Medulloblastoma

  24. Multi tumor classification Note that most of these tumors came from secondary sources and were not at the tissue of origin.

  25. Clustering is not accurate CNS, Lymphoma, Leukemia tumors separate Adenocarcinomas do not separate

  26. Multi tumor classification Combination approaches: All pairs One versus all (OVA)

  27. Supervised methodology

  28. Dataset Sample Type Validation Method Sample Number Total Accuracy Confidence High Low Fraction Accuracy Fraction Accuracy Train Well Differentiated Cross-val. 144 78% 80% 90% 20% 28% Train/cross - val. Test 1 Train/cross - val. Test 1 Test 1 Well Differentiated Train/Test 54 78% 78% 83% 22% 58% Train/ Test 1 Train/ Test 1 cross - val. cross - val. Accuracy Fraction of Calls Train/cross Train/cross - - val. val. Test 1 Test 1 5 5 1 1 1 1 Accuracy Fraction of Calls 0.9 0.9 4 4 1 1 0.8 0.8 0.8 0.8 3 3 0.7 0.7 0.8 0.8 0.6 0.6 0.6 0.6 2 2 0.6 0.5 0.5 0.6 0.4 0.4 0.4 0.4 0.4 0.4 1 1 0.3 0.3 Low High Low High Confidence Confidence 0.2 0.2 0.2 0.2 0.2 0.2 0 0 0.1 0.1 0 0 0 0 0 0 -1 -1 -1 0 1 2 3 4 -1 0 1 2 3 4 First First Top 2 Top 2 Top 3 Top 3 -1 0 1 2 3 4 -1 0 1 2 3 4 Correct Errors Correct Errors Correct Errors Correct Errors Prediction Calls Prediction Calls Low Low High High Low Low High High Confidence Confidence Confidence Confidence Well differentiated tumors

  29. Feature selection hurts performance

  30. Dataset Sample Type Validation Method Sample Number Total Accuracy Confidence High Low Fraction Accuracy Fraction Accuracy Test Poorly Differentiated Train/test 20 30% 50% 50% 50% 10% Accuracy Fraction of Calls 5 1 1 4 0.9 0.8 0.8 3 0.7 0.6 0.6 2 0.5 Low High Confidence 0.4 0.4 1 0.3 0.2 0.2 0 0.1 0 0 -1 Correct Errors -1 0 1 2 3 4 First Top 2 Top 3 Low High Prediction Calls Confidence Poorly differentiated tumors

  31. Morphing

  32. Morphing

  33. Talking faces

  34. Talking faces

  35. Talking faces

  36. Two feature selection algorithms Recursive feature elimination (RFE): based upon perturbation analysis, eliminate genes that perturb the margin the least Optimize leave-one out (LOO): based upon optimization of leave-one out error of a SVM, leave-one out error is unbiased

  37. Recursive feature elimination

  38. Optimizing the LOO Use leave-one-out (LOO) bounds for SVMs as a criterion to select features by searching over all possible subsets of n features for the ones that minimizes the bound. When such a search is impossible because of combinatorial explosion, scale each feature by a real value variable and compute this scaling via gradient descent on the leave-one-out bound. One can then keep the features corresponding to the largest scaling variables. The rescaling can be done in the input space or in a “Principal Components” space.

  39. R2/M2 =1 R2/M2 >1 R M = R M x2 x2 x1 Pictorial demonstration Rescale features to minimize the LOO bound R2/M2

  40. Three LOO bounds Radius margin bound: simple to compute, continuous very loose but often tracks LOO well Jaakkola Haussler bound: somewhat tighter, simple to compute, discontinuous so need to smooth, valid only for SVMs with no b term Span bound: tight complicated to compute, discontinuous so need to smooth

  41. We add a scaling parameter s to the SVM, which scales genes, genes corresponding to small sj are removed. The SVM function has the form: Classification function with scaling

  42. SVM and other functionals

  43. Algorithm

  44. Computing gradients

  45. Toy data Linear problem with 6 relevant dimensions of 202 Nonlinear problem with 2 relevant dimensions of 52 error rate error rate number of samples number of samples

More Related