1 / 18

Random Variables and Probability Distributions Chapter 4

Random Variables and Probability Distributions Chapter 4. “Never draw to an inside straight.” from Maxims Learned at My Mother’s Knee. Goals for Chapter 4. Define Random Variable, Probability Distribution Understand and Calculate Expected Value

gita
Télécharger la présentation

Random Variables and Probability Distributions Chapter 4

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Random Variables and Probability DistributionsChapter 4 “Never draw to an inside straight.” from Maxims Learned at My Mother’s Knee MGMT 242 Spring, 1999

  2. Goals for Chapter 4 • Define Random Variable, Probability Distribution • Understand and Calculate Expected Value • Understand and Calculate Variance, Standard Deviation • Two Random Variables--Understand • Statistical Independence of Two Random Variables • Covariance & Correlation of Two Random Variables • Applications of the Above MGMT 242 Spring, 1999

  3. Random Variables • Refers to possible numerical values that will be the outcome of an experiment or trial and that will vary randomly; • Notation: uppercase letters for last part of alphabet--X, Y, Z • Derived from hypothetical population (infinite number), the members of which will have different values of the random variable • Example--let Y = height of females between 18 and 20 years of age; population is infinite number of females between 18 and 20 • Actual measurements are carried out on a sample, which is randomly chosen from population. MGMT 242 Spring, 1999

  4. Probability Distributions • A probability distribution gives the probability for a specific value of the random variable: • P(Y= y) gives the probability distribution when values for P are specified for specific values of y • example: tossed a coin twice(fair coin); Y is the number of heads that are tossed; possible events: TT, TH, HT, HH; each event is equally probable if coin is a fair coin, so probability of Y=0 (event: TT) is 1/4; probability of Y =2 is 1/4; probability of Y = 1 is 1/4 +1/4 = 1/2; • Thus, for example: • P(Y=0) = 1/4 • P(Y=1) = 1/2 • P(Y=2) = 1/4 MGMT 242 Spring, 1999

  5. Probability Distributions--Discrete Variables • Notation: P(Y=y) or, occasionally, PY(y) (with y being some specific value; • PY(y) is some value when Y=y and 0 otherwise • Example--(Ex. 4.1) 8 people in a business group, 5 men, 3 women. Two people sent out on a recruiting trip. If people randomly chosen, find PY(y) for Y being the number of women sent out on the recruiting trip. • Solution (see board work): PY(2) = 3/28; PY(1) = 15/28; PY(0) = 10 /28 MGMT 242 Spring, 1999

  6. Cumulative Probability Distribution--Discrete Variables • Notation: P(Y y) means the probability that Y is less than or equal to y; FY(y) is the same. • Complement rule: P(Y > y) = 1 - FY(y). • Example (previous scenario, 5 men, 3 women, interview trips). What is the probability that at least one woman will be sent on an interview trip? (Note: “at least” means 1 or greater than 1) • = P (Y> 0) = 1 - FY(0) = 1 - PY(0) = 1 - 10/28 = 18/28 • In this example, it’s just as easy to calculate P(Y 1) = P(Y=1) + P(Y=2) = 15 /28 + 3 / 28, but many times it’s not as easy. MGMT 242 Spring, 1999

  7. Expectation Values, Mean Values • The expectation value of a discrete random variable Y is defined by the relation E(Y) = iPY( yi) yi , that is to say, the sum of all possible values of Y, with each value weighted by the probability of the value. • Note that E(Y) is the mean value of Y; E(Y) is also denoted as < Y > . • Example (for previous case, 8 people, 5 men, 3 women,…) If Y is the number of women on an interview trip (Y= 0, 1, or 2), then E(Y) = (3/28) x 2 + (15/28) x 1 + (10/28)x0 =21/28 = 3/4 MGMT 242 Spring, 1999

  8. Variance of a Discrete Random Variable • The variance of a discrete random variable, V(Y), is the expectation of the square of the deviation from the mean (expected value); V(Y) is also denoted as Var(Y) • V(Y) = E[(Y- E(Y))2] = iPY( yi) [yi - E(Y)]2 ; • V(Y) = E(Y2) - (E(Y))2 • The second formula for V(Y) is derived as follows: • V(Y) = iPY( yi) [yi2 - 2 yi E(y) + (E(Y))2] • or V(Y) = E(Y2) - 2 E(y) E(y) + (E(Y))2 = E(Y2) - (E(Y))2 • Example (previous, 5 men, 3 women, etc..) • V(Y) = {(3/28)(2- 3/4)2 + (15/28) (1- 3/4)2 + (10/28) (0- 3/4)2, or V(Y) = (3/28) 22 + (15/28) 12 + 0 - (3/4)2 = 27/28 MGMT 242 Spring, 1999

  9. Expectation Values--Another ExampleExercise 4.25, Investment in 2 Apartment Houses MGMT 242 Spring, 1999

  10. Continuous Random Variables • Reasons for using a continuous variable rather than discrete • Many, many values (e.g. salaries)--too many to take as discrete; • Model for probability distribution makes it convenient or necessary to use a continuous variable-- • Uniform Distribution (any value between set limits equally likely) • Exponential Distribution (waiting times, delay times) • Normal (Gaussian) Distribution, the “Bell Shaped Curve” (many measurement values follow a normal distribution either directly or after an appropriate transformation of variables; also mean values of samples follow a normal distribution, generally.) MGMT 242 Spring, 1999

  11. Probability Density and Cumulative Density Functions for Continuous Variables • Probability density function, fX(x) defined: • P(x X  x+dx) = fX(x) dx, that is, the probability that the random variable X is between x and x+dx is given by fX(x) dx • Cumulative density function, FX(x), defined: • P(X  x ) = FX(x) • FX(x’) =  fX(x)dx, where the integral is taken from the lowest possible value of the random variable X to the value x’. MGMT 242 Spring, 1999

  12. Probability Density and Cumulative Density Functions for Continuous Variables, Example: Exercise 4.12 • Model for time, t, between successive job applications to a large computer is given by FT(t) = 1 - exp(-0.5t). • Note that FT(t) = 0 for t =0 and that FT(t) approaches 1 for t approaching infinity. • Also, fT(t) = the derivative of FT(t), or fT(t) = 0.5exp(-0.5t) MGMT 242 Spring, 1999

  13. Probability Density and Cumulative Density Functions for Continuous Variables--Example, Problem 4.12 MGMT 242 Spring, 1999

  14. Expectation Values for Continuous Variables • The expectation value for a continuous variable is taken by weighting the quantity by the probability density function, fY(y), and then integrating over the range of the random variable • E(Y) =  y fY(y) dy; • E(Y), the mean value of Y, is also denoted as Y • E(Y2) =  y fY(y) dy; • The variance is given by V(Y) = E(Y2) - (Y)2 MGMT 242 Spring, 1999

  15. Continuous Variables--Example • Ex. 4.18, text. An investment company is going to sell excess time on its computer; it has determined that a good model for the its own computer usage is given by the probability density function fY(y) = 0.0009375[40-0.1(y-100)2 ] for 80 < y < 120 fY(y) = 0, otherwise. The important things to note about this distribution function can be determined by inspection • there is a maximum in fY(y) at y = 100 • fY(y) is 0 at y=80 and y = 120 • fY(y) is symmetric about y=100 (therefore E(y) = 100 and FY(y)=1/2 at y = 100). • fY(y) is a curve that looks like a symmetric hump. MGMT 242 Spring, 1999

  16. Two Random Variables • The situation with two random variables, X and Y, is important because the analysis will often show if there is a relation between the two, for example, between height and weight; years of education and income; blood alcohol level and reaction time. • We will be concerned primarily with the quantities that show how strong (or weak) the relation is between X and Y: • The covariance of X and Y is defined by Cov(X,Y) = E[(X-X)(Y- Y)] = E(XY) -XY • The correlation of X and Y is defined by Cor(X,Y) = Cov(X,Y) / (V(X)V(Y) = Cov(X,Y)/(X Y) MGMT 242 Spring, 1999

  17. Properties of Covariance, Correlation • Cov(X,Y) is positive (>0) if X and Y both increase or both decrease together; Examples: • height and weight; years of education and salary; • Cov(X,Y) is zero if X and Y are statistically independent: [Cov(X,Y) = E(XY) -XY= E(X)E(Y)-XY = XY - XY= 0]Examples: adult hat size and IQ; • Cov(X,Y) is negative if Y increases while X decreases; Example: • annual income, number of bowling games per year. (no disrespect meant to bowlers). MGMT 242 Spring, 1999

  18. Statistical Independence of Two Random Variables • If two random variables, X and Y, are statistically independent, then • P(X=x, Y=y) = P(X=x) P(Y=y), that is to say, the joint probability density function can be written as the product of probability density functions for X and Y. • Cov(X, Y) = 0 This follows from the above relation: Cov(X,Y) = E[(X-µX) (Y-µY)] =[E(X-µX)][E(Y-µY)] = (µX-µX) (µY-µY) = 0 MGMT 242 Spring, 1999

More Related