Transcript Probability Essentials Chapter 3
Random Variables and Probability Distributions Chapter 4 “Never draw to an inside straight.” from
Maxims Learned at My Mother’s Knee
MGMT 242 Spring, 1999
Goals for Chapter 4
• • • • • Define Random Variable, Probability Distribution Understand and Calculate Expected Value Understand and Calculate Variance, Standard Deviation Two Random Variables--Understand – – Statistical Independence of Two Random Variables Covariance & Correlation of Two Random Variables Applications of the Above MGMT 242 Spring, 1999
Random Variables
• Refers to possible numerical values that will be the outcome of an experiment or trial and that will vary randomly; • Notation: uppercase letters for last part of alphabet--X, Y, Z • Derived from hypothetical population (infinite number), the members of which will have different values of the random variable • Example--let Y = height of females between 18 and 20 years of age; population is infinite number of females between 18 and 20 • Actual measurements are carried out on a sample, which is randomly chosen from population.
MGMT 242 Spring, 1999
Probability Distributions
• A probability distribution gives the probability for a specific value of the random variable: – P(Y= y) gives the probability distribution when values for P are specified for specific values of y – example: tossed a coin twice(fair coin); Y is the number of heads that are tossed; possible events: TT, TH, HT, HH; each event is equally probable if coin is a fair coin, so probability of Y=0 (event: TT) is 1/4; probability of Y =2 is 1/4; probability of Y = 1 is 1/4 +1/4 = 1/2; – Thus, for example: • P(Y=0) = 1/4 • P(Y=1) = 1/2 • P(Y=2) = 1/4 MGMT 242 Spring, 1999
Probability Distributions--Discrete Variables
• Notation: P(Y=y) or, occasionally, P Y (y) (with y being some specific value; • P Y (y) is some value when Y=y and 0 otherwise • Example--(Ex. 4.1) 8 people in a business group, 5 men, 3 women. Two people sent out on a recruiting trip. If people randomly chosen, find P Y (y) for Y being the number of women sent out on the recruiting trip.
• Solution (see board work): P Y (2) = 3/28; P Y (1) = 15/28; P Y (0) = 10 /28 MGMT 242 Spring, 1999
Cumulative Probability Distribution- Discrete Variables
• Notation: P(Y y) means the probability that Y is less than or equal to y; F Y (y) is the same.
• Complement rule: P(Y > y) = 1 - F Y (y).
• Example (previous scenario, 5 men, 3 women, interview trips). What is the probability that at least one woman will be sent on an interview trip? (Note: “at least” means 1 or greater than 1) – = P (Y> 0) = 1 - F Y (0) = 1 - P Y (0) = 1 - 10/28 = 18/28 – In this example, it’s just as easy to calculate P(Y 1) = P(Y=1) + P(Y=2) = 15 /28 + 3 / 28, but many times it’s not as easy. MGMT 242 Spring, 1999
Expectation Values, Mean Values
• The expectation value of a discrete random variable Y is defined by the relation E(Y) = i P Y ( y i ) y i , that is to say, the sum of all possible values of Y, with each value weighted by the probability of the value. • Note that E(Y) is the mean value of Y; E(Y) is also denoted as < Y > .
• Example (for previous case, 8 people, 5 men, 3 women,…) If Y is the number of women on an interview trip (Y= 0, 1, or 2), then E(Y) = (3/28) x 2 + (15/28) x 1 + (10/28)x0 =21/28 = 3/4 MGMT 242 Spring, 1999
Variance of a Discrete Random Variable
• The variance of a discrete random variable, V(Y), is the expectation of the square of the deviation from the mean (expected value); V(Y) is also denoted as Var(Y) • V(Y) = E[(Y- E(Y)) 2 ] = i P Y ( y i ) [y i - E(Y)] 2 ; • V(Y) = E(Y 2 ) (E(Y)) 2 • The second formula for V(Y) is derived as follows: – V(Y) = i P Y ( y i ) [y i 2 - 2 y i E(y) + (E(Y)) 2 ] – or V(Y) = E(Y 2 ) - 2 E(y) E(y) + (E(Y)) 2 = E(Y 2 ) - (E(Y)) 2 • Example (previous, 5 men, 3 women, etc..) – V(Y) = {(3/28)(2- 3/4) or V(Y) = (3/28) 2 2 2 + (15/28) (1- 3/4) 2 + (10/28) (0- 3/4) 2 , + (15/28) 1 2 + 0 - (3/4) 2 = 27/28 MGMT 242 Spring, 1999
Expectation Values--Another Example
Exercise 4.25, Investment in 2 Apartment Houses Return -50 0 50 100 150 200 250 exp.val.
variance st.dev.
0.02
0.03
0.2
0.5
0.2
0.03
0.02
0.15
0.1
0.1
0.1
0.3
0.2
0.05
100 2500 105 8475 50 92.05976
-1 0 10 50 30 6 5 450 300 500 0 500 300 450 sum exp1 sum var1 100 2500 MGMT 242 Spring, 1999
Continuous Random Variables
• Reasons for using a continuous variable rather than discrete – Many, many values (e.g. salaries)--too many to take as discrete; – Model for probability distribution makes it convenient or necessary to use a continuous variable- • Uniform Distribution (any value between set limits equally likely) • Exponential Distribution (waiting times, delay times) • Normal (Gaussian) Distribution, the “Bell Shaped Curve” (many measurement values follow a normal distribution either directly or after an appropriate transformation of variables; also mean values of samples follow a normal distribution, generally.) MGMT 242 Spring, 1999
Probability Density and Cumulative Density Functions for Continuous Variables
• Probability density function, f X (x) defined: – P(x X x+dx) = f X (x) dx, that is, the probability that the random variable X is between x and x+dx is given by f X (x) dx • Cumulative density function, F X (x), defined: – P(X x ) = F X (x) – F X (x’) = value x’. f X (x)dx, where the integral is taken from the lowest possible value of the random variable X to the MGMT 242 Spring, 1999
Probability Density and Cumulative Density Functions for Continuous Variables,
Example: Exercise 4.12
• Model for time, t, between successive job applications to a large computer is given by F T (t) = 1 - exp(-0.5t).
• Note that F T (t) = 0 for t =0 and that F T (t) approaches 1 for t approaching infinity.
• Also, f T (t) = the derivative of F T (t), or f T (t) = 0.5exp(-0.5t) MGMT 242 Spring, 1999
Probability Density and Cumulative Density Functions for Continuous Variables--
Example, Problem 4.12
FsubY 1.2
1 0.8
0.6
0.4
0.2
0 0 2 4 6 8 10 12 MGMT 242 Spring, 1999
Expectation Values for Continuous Variables
• The expectation value for a continuous variable is taken by weighting the quantity by the probability density function, f Y (y), and then integrating over the range of the random variable • E(Y) = y f Y (y) dy; – E(Y), the mean value of Y, is also denoted as
Y
• E(Y 2 ) = y f Y (y) dy; • The variance is given by V(Y) = E(Y 2 ) - ( Y ) 2 MGMT 242 Spring, 1999
Continuous Variables--Example
• Ex. 4.18, text. An investment company is going to sell excess time on its computer; it has determined that a good model for the its own computer usage is given by the probability density function f Y (y) = 0.0009375[40-0.1(y-100) 2 f Y (y) = 0, otherwise.
] for 80 < y < 120 The important things to note about this distribution function can be determined by inspection – there is a maximum in f Y (y) at y = 100 – f Y (y) is 0 at y=80 and y = 120 – f Y (y) is symmetric about y=100 (therefore E(y) = 100 and F Y (y)=1/2 at y = 100).
– f Y (y) is a curve that looks like a symmetric hump.
MGMT 242 Spring, 1999
Two Random Variables
• The situation with two random variables, X and Y, is important because the analysis will often show if there is a relation between the two, for example, between height and weight; years of education and income; blood alcohol level and reaction time.
• We will be concerned primarily with the quantities that show how strong (or weak) the relation is between X and Y: – The covariance of X and Y is defined by Cov(X,Y) = E[(X X )(Y Y )] = E(XY) X Y – The correlation of X and Y is defined by Cor(X,Y) = Cov(X,Y) / (V(X)V(Y) = Cov(X,Y)/( X Y ) MGMT 242 Spring, 1999
Properties of Covariance, Correlation
• Cov(X,Y) is positive (>0) if X and Y both increase or both decrease together; Examples: – height and weight; years of education and salary; • Cov(X,Y) is zero if X and Y are statistically independent: [ Cov(X,Y) = E(XY) X Y = E(X)E(Y) X Y = X Y X Y = 0] Examples: adult hat size and IQ; • Cov(X,Y) is negative if Y increases while X decreases; Example: – annual income, number of bowling games per year.
(no disrespect meant to bowlers).
MGMT 242 Spring, 1999
Statistical Independence of Two Random Variables
• If two random variables, X and Y, are statistically independent, then – P(X=x, Y=y) = P(X=x) P(Y=y), that is to say, the joint probability density function can be written as the product of probability density functions for X and Y.
– Cov(X, Y) = 0 This follows from the above relation: Cov(X,Y) = E[(X-µ X ) (Y-µ Y )] =[E(X-µ X )][E(Y-µ Y )] = (µ X -µ X ) (µ Y -µ Y ) = 0 MGMT 242 Spring, 1999