Document 7920980

Download Report

Transcript Document 7920980

730 Lecture 2
Today’s lecture:
Inequalities
Convergence
Taylor Series
5/27/2016
730 lecture 2
1
Inequalities
Chebychev :

P(| X   |  )  2

2
Jensen :
g[ E ( X )]  E[ g ( X )]
5/27/2016
730 lecture 2
2
Chebychev
Area Area
2/2
x  0,1,..., n
  +
5/27/2016
730 lecture 2
3
Chebychev: proof


( x   )2
 
f ( x)dx
2
2



2
 


 

(x  )

2
2
f ( x)dx +


 +
( x   )2

2
f ( x)dx

 f ( x)dx + +f ( x)dx
 P[ X     ] + P[ X   +  ]
 P[| X   |  ]
5/27/2016
730 lecture 2
4
Jensen
g(x)
g() + g’()(x-)
5/27/2016

730 lecture 2
5
Example : sample standard
deviation
• Graph of y=x:
5
4.5
4
3.5
3
2.5
2
1.5
1
0.5
0
0
5
10
15
20
25
2
5/27/2016
730 lecture 2
6
Std dev (cont)
• Slope at x=2 is 1/2
• Tangent at point (,2) lies above curve
• THUS
s2  2
s  +
2
Taking expectatio ns
E (s 2   2 )
E(s)   +

2
Sample sd BIASSED DOWN!
5/27/2016
730 lecture 2
7
Convergence
• Convergence in distribution
P( X n  x )  P ( X  x )
• Main use: approximation Eg
CLT
5/27/2016
730 lecture 2
8
Examples
• tn converges to normal
• Binomial converges to Poisson
• Gamma converges to normal
5/27/2016
730 lecture 2
9
Central limit Theorem
(CLT)
 n(X  )

P
 x    ( x)



5/27/2016
730 lecture 2
10
Example
Convergence of binomial to Normal:
Express Xn as
n
X n   Yi where Y1 ,..., Yn are independen t with
i 1
Yi  1 with probabilit y p,
 0 with probabilit y 1  p.
5/27/2016
730 lecture 2
11
Example (cont)
Then result is just CLT, as the Y’s have
mean p and variance p(1-p).
5/27/2016
730 lecture 2
12
Example
• Convergence of binomial to Poisson:
• Proved by showing convergence of the
prob. function since
[ x]
Fn ( x)   pn ( j )
j 0
5/27/2016
730 lecture 2
13
Example 5 (cont)
npn n
n n  1 n  y + 1 (npn ) y
y
.
...
.
.(1 
) .(1  pn)
n n
n
y!
n
5/27/2016
730 lecture 2
14
Convergence (2)
Convergence in probability: Xn
converges to X “in probability” if
P(| X n  X |  )  0
Eg proportion of heads in n
tosses 
5/27/2016
730 lecture 2
15
Example
Xn=proportion of heads in n tosses
E(Xn)=p=0.5, Var(Xn)=p(1-p)/n=1/4n
By Chebychev,
P(|Xn-0.5|>)  Var(Xn)/2 =1/(4n 2)  0
5/27/2016
730 lecture 2
16
More examples
• Sample mean converges in probability to
population mean as sample size gets
larger
(Weak law of large numbers)
• Sample variance converges to population
variance
• “Consistent” estimators
5/27/2016
730 lecture 2
17
Example: sample variance
n
1
2
s2 
(
X

X
)

i
n  1 i 1
n
1

{ ( X i   ) 2  n( X   ) 2}
n  1 i 1
n
1

{ Yi  n( X   ) 2}
n  1 i 1
n
n

Y
( X   )2
n 1
n 1
5/27/2016
730 lecture 2
18
Sample variance (cont)
Y  E (Yi )  E ( X i   )   ,
2
2
X    0, so ( X   )  0  0
2
2
Thus
s  1   1 0   .
2
5/27/2016
2
2
730 lecture 2
19
Example
X1,…Xn Uniform
Mn=max (X1,…Xn )
P(Mn  x) =xn (why?)
P(|Mn- 1|>)  P(Mn<1- ) = (1- )n  0
Thus Mn converges to 1 in probability
5/27/2016
730 lecture 2
20