1 / 70

Dirk Husmeier Adriano V. Werhli

Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks. Dirk Husmeier Adriano V. Werhli. +. +. +. +. +. +. +. …. Learning Bayesian networks. from data and prior knowledge. Bayesian networks.

Télécharger la présentation

Dirk Husmeier Adriano V. Werhli

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Bayesian integration of biological prior knowledge into the reconstruction of gene regulatory networks Dirk Husmeier Adriano V. Werhli

  2. +

  3. + +

  4. + + + + …

  5. Learning Bayesian networks from data and prior knowledge

  6. Bayesian networks • Marriage between graph theory and probability theory. • Directed acyclic graph (DAG) representing conditional independence relations. • It is possible to score a network in light of the data. • We can infer how well a particular network explains the observed data. NODES A B C EDGES D E F

  7. Bayesian networks versus causal networks Node A unknown A A True causal graph B C B C

  8. Bayesian networks versus causal networks A A A B C B C B C • Equivalence classes: networks with the same scores. • Equivalent networks cannot be distinguished in light of the data. • We can only learn the undirected graph. • Unless… we use interventions or prior knowledge. A B C

  9. Learning Bayesian networks from data P(M|D) = P(D|M) P(M) / Z

  10. Use TF binding motifs in promoter sequences

  11. Biological prior knowledge matrix Indicates some knowledge about the relationship between genes i and j Biological Prior Knowledge

  12. Biological prior knowledge matrix Indicates some knowledge about the relationship between genes i and j Biological Prior Knowledge Define the energy of a Graph G

  13. Energy of a network Prior distribution over networks

  14. Sample networks and hyperparameters • from the posterior distribution • Capture intrinsic inference uncertainty • Learn the trade-off parameters automatically P(M|D) = P(D|M) P(M) / Z

  15. Energy of a network Prior distribution over networks

  16. Energy of a network Rewriting the energy

  17. Approximation of the partition function

  18. Multiple sources of prior knowledge

  19. Energy of a network Rewriting the energy

  20. Approximation of the partition function

  21. MCMC sampling scheme

  22. Sample networks and hyperparameters from the posterior distribution Proposal probabilities Metropolis-Hastings scheme

  23. MCMC with one prior Sample graph and the parameter b. • Separate in two samples to improve the acceptance: • Sample graph with b fixed. • Sample b with graph fixed.

  24. MCMC with one prior Sample graph and the parameter b. • Separate in two samples to improve the acceptance: • Sample graph with b fixed. • Sample b with graph fixed. BGe BDe

  25. MCMC with one prior Sample graph and the parameter b. • Separate in two samples to improve the acceptance: • Sample graph with b fixed. • Sample b with graph fixed. BGe BDe

  26. MCMC with one prior Sample graph and the parameter b. • Separate in two samples to improve the acceptance: • Sample graph with b fixed. • Sample b with graph fixed. BGe BDe

  27. MCMC with one prior Sample graph and the parameter b. • Separate in two samples to improve the acceptance: • Sample graph with b fixed. • Sample b with graph fixed. BGe BDe

  28. Approximation of the partition function

  29. MCMC with two priors Sample graph and the parameters b1and b2 • Separate in three samples to improve the acceptance: • Sample graph with b1 and b2 fixed. • Sample b1 with graph and b2 fixed. • Sample b2 with graph and b1 fixed.

  30. Bayesian networkswith biological prior knowledge • Biological prior knowledge: Information about the interactions between the nodes. • We use two distinct sources of biological prior knowledge. • Each source of biological prior knowledge is associated with its own trade-off parameter:b1 and b2. • The trade off parameter indicates how much biological prior information is used. • The trade-off parameters are inferred. They are not set by the user!

  31. Bayesian networkswith two sources of prior Source 2 Source 1 Data BNs + MCMC b1 b2 Recovered Networks and trade off parameters

  32. Bayesian networkswith two sources of prior Source 2 Source 1 Data BNs + MCMC b1 b2 Recovered Networks and trade off parameters

  33. Bayesian networkswith two sources of prior Source 2 Source 1 Data BNs + MCMC b1 b2 Recovered Networks and trade off parameters

  34. Evaluation • Can the method automatically evaluate how useful the different sources of prior knowledge are? • Do we get an improvement in the regulatory network reconstruction? • Is this improvement optimal?

  35. Application to theRaf regulatory network

  36. Raf regulatory network From Sachs et al Science 2005

  37. Evaluation: Raf signalling pathway • Cellular signalling network of 11 phosphorylated proteins and phospholipids in human immune systems cell • Deregulation  carcinogenesis • Extensively studied in the literature  gold standard network

  38. DataPrior knowledge

  39. Flow cytometry data and KEGG • Intracellular multicolour flow cytometry. • Measured protein concentrations. • 11 proteins: 1200 concentration profiles. • We sample 5 separate subsets with 100 concentration profiles each.

  40. Microarray example Spellman et al (1998) Cell cycle 73 samples Tu et al (2005) Metabolic cycle 36 samples time time Genes Genes

  41. DataPrior knowledge

  42. Flow cytometry data and KEGG http://www.genome.jp/kegg/ KEGG PATHWAYS are a collection of manually drawn pathway maps representing our knowledge of molecular interactions and reaction networks.

  43. Prior knowledge from KEGG

  44. Prior distribution

  45. The data and the priors + KEGG + Random

More Related