1 / 57

Determinants of Distance Matrices of Trees

Determinants of Distance Matrices of Trees. Valerie Tillia. Outline. A look at graph theory Short biography of Ron Graham A look at determinants Exploring Condensation and the man who created it Overview of the proof by Yin and Yeh Conclusion Remaining questions. What is graph theory?.

erling
Télécharger la présentation

Determinants of Distance Matrices of Trees

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Determinants of Distance Matrices of Trees Valerie Tillia

  2. Outline • A look at graph theory • Short biography of Ron Graham • A look at determinants • Exploring Condensation and the man who created it • Overview of the proof by Yin and Yeh • Conclusion • Remaining questions

  3. What is graph theory? • Graph theory is the mathematical study of networks. • These networks are represented by nodes (vertices) and connections (edges). Who studies graph theory?? (and why?) The Petersen Graph

  4. Short biography of Ron Graham In the 1970’s, Ronald L. Graham, the man responsible for Graham’s Number (we’ll look at that later), published a paper using graph theory to study transmitting messages and calls efficiently at Bell System Tech. Ron Graham mathcircle.berkeley.edu

  5. Bell System Tech- graph theory in use! businessweek.com The U.S. National Archives' photo stream

  6. Ron Graham- Man of Many Hats! • Graham the Mathematician- began new areas of mathematical study such as worst-case analysis in scheduling theory, and Graham’s Number! Graham’s Number- largest number yet for application use. To write it out would take more ink than atoms in the universe (what we know of it)!

  7. More on Graham’s number… • The arrows are read “up”. • 3 up n means 3(3)(3)…(3) n times • 3 up up n is 3 up 3 up 3 up 3…. up 3 n times. • What does that mean?? This is like function composition. It is read, 3 up (3 up (3 up (3 up(…. up 3)…)) • Depending on how many arrows you have, it gets a lot worse.

  8. Ron Graham- Hat #2 • Graham the Juggler- he was at one point the president of the International Jugglers Association daviddarling.info/images/Graham_Ronald math.ucsd.edu

  9. Ron Graham- Hat #3 • The Circus Performer- part of the Bouncing Baers He has circus equipment installed in his home for practice. math.ucsd.edu

  10. Ron Graham- Hat #4 • The Scientist- Chief Scientist at California Institute for Telecommunication (more graph theory?!) and Information Technology www.reatlas.com physics.uci.edu

  11. Graham’s Formula • While working on the study transmitting messages and calls efficiently, he found and proved a formula.

  12. Graham’s Formula • While working on the study transmitting messages and calls efficiently, he found and proved a formula. • This formula is so simple, yet so intriguing it has inspired mathematicians from all sorts of backgrounds to create new and different proofs of it.

  13. Graham’s Formula • While working on the study transmitting messages and calls efficiently, he found and proved a formula. • This formula is so simple, yet so intriguing it has inspired mathematicians from all sorts of backgrounds to create new and different proofs of it. • This formula finds the determinant of the distance matrix of a tree.

  14. Graham’s Formula • While working on the study transmitting messages and calls efficiently, he found and proved a formula. • This formula is so simple, yet so intriguing it has inspired mathematicians from all sorts of backgrounds to create new and different proofs of it. • This formula finds the determinant of the distance matrix of a tree. • Not that kind of tree…

  15. Graham’s Formula • While working on the study transmitting messages and calls efficiently, he found and proved a formula. • This formula is so simple, yet so intriguing it has inspired mathematicians from all sorts of backgrounds to create new and different proofs of it. • This formula finds the determinant of the distance matrix of a tree. • Not that kind of tree… • Time for definitions!

  16. Definitions and concepts we need: • Trees • Distance and distance matrix • Determinants

  17. Trees • A tree is a graph that has no cycles: Cycle! Tree With bark? Not a tree

  18. Trees • A tree on n vertices always has n-1 edges. • A “terminating” vertex is also called a leaf. Don’t ask why there’s a leaf at the base of the trunk…

  19. Distance • The distance from one vertex to another is the length of the shortest path between them. You count the number of edges, or “steps” from one to the other. 2 1 3 The distance from x to y is 3.

  20. Distance Matrix • The distance matrix is a symmetric matrix that lists the distance between every possible pair of vertices in the graph. • Label the vertices in the tree and set up the matrix as in this example: As you can see, this matrix will always be symmetric about the diagonal.

  21. Determinant • In linear algebra, the determinant of a matrix turns out to be the product of the eigenvalues of the matrix. (But we won’t get into that) • To find the determinant of a 2x2 matrix, you take ad-bc.

  22. How to find harder determinants • To find the determinant of any matrix larger than this, a new idea is needed. • One way is called co-factor expansion. It requires taking determinants of smaller matrices sitting inside the big one. Here are a couple smaller 3x3’s sitting inside our 4x4

  23. How to do co-factor expansion • First, pick a row or column with the most zeros. This will make things easier for us. Might as well pick the top row here.

  24. How to do co-factor expansion • Next, you pick the first number in the row, and cross out the row and column it is sitting in. Take the number you circled, and multiply it by the determinant of the tiny left-over matrix. This is the first summand in the expression for the determinant.

  25. How to do co-factor expansion • Do this repeatedly for every entry in the row or column that you picked. See why zeros are nice? When you add them, use a +,-,+… pattern.

  26. Even larger matrices… • Although this isn’t so bad for 3x3’s or maybe even 4x4’s, trying to do this for a matrix of even 5x5 gets pretty hairy. Unless you have a lot of zeros, if you have an nxn matrix there are n!=n(n-1)(n-2)…(2) steps… n things in the first row which means n determinants. Then each of those is an (n-1)x (n-1) matrix. Each (n-1) x (n-1) has n-1 of (n-2) x (n-2) matrices…. yikes

  27. More yikes: • For a 16 by 16 matrix, that’s 2.09227899 × 1013 steps!! Outrageous. 20,922,789,900,000= twenty trillion, nine hundred twenty-two billion, seven hundred eighty-nine million, nine hundred thousand. • Is there another way besides technology? http://www.mathcats.com/explore/reallybignumbers.html

  28. Introducing: Charles Dodgson • Charles Dodgson, born in 1832 was also a man of many hats. He was an author, a reverend, a photographer, and a mathematician. • He came up with another way to find the determinant, and published it in the Proceedings of the Royal Society of London. • He also is known by an entirely different name…

  29. Lewis Carroll!! lewiscarroll.org Is he writing Alice, or doing math!?

  30. Lewis Carroll • Yes, Lewis Carroll is the man who wrote Alice in Wonderland AND came up with a new way to find • the determinant! • His way is based on an identity called the Desnanot-Jacobi identity. These people lived around the same time, but the identity became more commonly known • as Dodgson's Determinant Evaluation Rule.

  31. Condensation • Dodgson called his method condensation, because it “condenses” our large matrix into progressively smaller matrices until we are finally just left with the determinant. Lets see how this works. Condensed Soup.

  32. Condensation on a 4x4 matrix • Below is a nice 4x4 matrix to do condensation on. There are a lot of zeros, though. Although this is good in cofactor expansion, its bad in condensation. However, we can do row operation to make the “interior” zeros go away, and it won’t affect the determinant. Elementary row operations Zeros in the interior are not good

  33. Condensation on a 4x4 matrix • Our next step is to take the determinants of all the nestled 2x2’s. These are quick calculations you can usually do in your head.

  34. Condensation on a 4x4 matrix • Now you take all those determinants, and put them into a new, “condensed”, matrix:

  35. Condensation on a 4x4 matrix • Do the same thing for the second step, but this time, go back and find the corresponding “interior” entries from 2 steps back, and divide each entry in your newly condensed matrix by them. (hence the no-zero’s rule)

  36. Condensation on a 4x4 matrix • Continue this if you have a larger matrix (including the going back 2 steps ago to get some interior entries to divide by) until you arrive at a 2x2 matrix. • We’re there! The determinant of this matrix, divided by the interior of the matrix two steps back, is the determinant of the original matrix.

  37. Condensation vs. Cofactor Expansion • Condensation wasn’t exactly easy, and complications can occur if zeros spontaneously appear in the interiors of successive matrices. (In which case you basically have to start over) • However, for a 16x16 matrix, we’d be facing around 15 condensing steps, rather than 2.09227899 × 1013 . • I’m glad we have technology now!

  38. Graham's Formula • Now you are ready to appreciate Graham’s formula. The determinant of the distance matrix of a tree on n vertices is given by the formula, |D| Notice that the only parameter is n. The determinant does not care about the structure of the tree!!

  39. Regardless of the structure… • A little hard to believe, right? Let’s look at some examples, and see if we can convince ourselves. Tree 1 Tree 2 Here are two trees on 4 vertices, that clearly have different distance matrices. Tree 1 has 3’s and Tree 2 doesn’t. And they have the same determinant?

  40. Regardless of the structure… • Remember that matrix we did condensation on? That was actually the distance matrix for Tree 1, and we found its determinant was -12. Lets use co-factor expansion on Tree 2’s matrix.

  41. Regardless of the structure… • Remember that matrix we did condensation on? That was actually the distance matrix for Tree 1, and we found its determinant was -12. Lets use co-factor expansion on Tree 2’s matrix. Wow!

  42. Regardless of the structure… • Remember that matrix we did condensation on? That was actually the distance matrix for Tree 1, and we found its determinant was -12. Lets use co-factor expansion on Tree 2’s matrix. Wow! Wow!

  43. Overview of the proof by Yan and Yeh • Their proof uses a clever combination of the determinant evaluation rule, co-factor expansion, and a system of equations! • You now have all the knowledge you need to understand this proof, so I will give you a quick look at how it goes.

  44. Overview of the proof by Yan and Yeh • The proof is by induction, so a base case is shown that the formula does indeed work for a tree on 3 vertices. The distance matrix is

  45. Overview of the proof by Yan and Yeh • The proof is by induction, so a base case is shown that the formula does indeed work for a tree on 3 vertices. The distance matrix is

  46. Overview of the proof by Yan and Yeh • We now assume the formula holds for trees on less than n vertices. Next by using some slick elementary column operations on an arbitrary distance matrix, we are able to get it to have a lot of zeros in one column. Which has the same determinant before we changed it with row operations. Zeros are good!!!

  47. Overview of the proof by Yan and Yeh • Now we will use cofactor expansion along the first column (rather than row) of D. See all the zeros?

  48. Overview of the proof by Yan and Yeh • But this is an induction proof, so if we can get our expression for a determinant in terms of determinants of trees of smaller than n vertices, we can assume it works for them!

  49. Overview of the proof by Yan and Yeh • Now is where Dodgson’s Rule comes in. Condensation is based on the fact that for any matrix, • If we think of each of those determinants in that formula as the determinant of a distance matrix, for the ones on less than n vertices, we can apply the formula.

  50. Overview of the proof by Yan and Yeh • What we have done, is effectively gotten two different expressions for the determinant of our matrix D.

More Related