Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
3 views37 pages

Lecture 3

good
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
3 views37 pages

Lecture 3

good
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 37

Statistical Modeling, Lecture 3

Melike Efe
July 16, 2025
Sabancı University
Table of contents

1. F-Distribution

2. Order Statistic

3. Point Estimation
Unbiased Estimator

1
F-Distribution
F-Distribution

What is the F-Distribution?

▶ The F-distribution is used to compare variances of two independent


samples.
▶ Defined as the ratio of two chi-square distributed random variables,
each divided by its degrees of freedom.

Definition: Let X and Y be independent chi-square distributed random


variables with ν1 and ν2 degrees of freedom. Then,

X /ν1
R=
Y /ν2

has an F-distribution with ν1 and ν2 degrees of freedom.

2
Theorem: The density of F-distribuiton with ν1 and ν2 degrees of
freedom is
ν1 + ν2   ν1 − 21 (ν1 +ν2 )
Γ( ) ν 2 ν1

ν1
f (r ) = ν1 2 1
r 2 −1
1+ r
Γ( 2 )Γ( ν22 ) ν2 ν2

for r > 0 and it is zero elsewhere.

3
Constructing an F-Distributed Variable

Understanding the F-Distribution Through Standard Normal


Variables:

▶ Consider two independent standard normal random variables:


d d
Z = N(0, 1), W = N(0, 1).

▶ Squaring each variable gives chi-square distributed variables with one


degree of freedom:
d d
Z 2 = χ21 , W 2 = χ21 .
▶ Taking their ratio results in an F-distributed variable:

Z2 d
R= = F1,1 .
W2

4
Inverse Property of the F-Distribution

Key Property: Relationship Between F and Its Reciprocal


▶ If R follows an F -distribution with degrees of freedom (ν1 , ν2 ):
d
R = Fν1 ,ν2
then its inverse follows an F-distribution with swapped degrees of
freedom:
1 d
= Fν2 ,ν1 .
R
▶ This property is useful when looking up F -tables and calculating
critical values.
Why does this hold?
▶ Recall that the F -distribution is based on the ratio of two chi-square
variables.
▶ Reversing the ratio effectively swaps the degrees of freedom, leading
to a new F -distribution.
5
6
Understanding Critical Points

▶ The critical points of the F-distribution are tabulated in statistical


tables.
▶ For an F-distributed random variable X with degrees of freedom
(ν1 ,ν2 ), the critical value fα,ν1 ,ν2 satisfies:

P(X > fα,ν1 ,ν2 ) = α

7
Critical points of the F distribution
Points fα,ν ,ν for α = 0.01.
1 2
ν2 ν1
1 2 3 4 5 6 7 8
1 4052.2 4999.5 5403.4 5624.6 5763.6 5859.0 5928.3 5981.1
2 98.50 99.00 99.17 99.25 99.30 99.33 99.36 99.37
3 34.12 30.82 29.46 28.71 28.24 27.91 27.67 27.49
4 21.20 18.00 16.69 15.98 15.52 15.21 14.98 14.80
5 16.26 13.27 12.06 11.39 10.97 10.67 10.46 10.29
6 13.75 10.92 9.78 9.15 8.75 8.47 8.26 8.10
7 12.25 9.55 8.45 7.85 7.46 7.19 6.99 6.84
8 11.26 8.65 7.59 7.01 6.63 6.37 6.18 6.03
9 10.56 8.02 6.99 6.42 6.06 5.80 5.61 5.47
10 10.04 7.56 6.55 5.99 5.64 5.39 5.20 5.06
11 9.65 7.21 6.22 5.67 5.32 5.07 4.89 4.74
12 9.33 6.93 5.95 5.41 5.06 4.82 4.64 4.50
13 9.07 6.70 5.74 5.21 4.86 4.62 4.44 4.30
14 8.86 6.51 5.56 5.04 4.69 4.46 4.28 4.14
15 8.68 6.36 5.42 4.89 4.56 4.32 4.14 4.00
16 8.53 6.23 5.29 4.77 4.44 4.20 4.03 3.89
17 8.40 6.11 5.19 4.67 4.34 4.10 3.93 3.79
18 8.29 6.01 5.09 4.58 4.25 4.01 3.84 3.71
19 8.19 5.93 5.01 4.50 4.17 3.94 3.77 3.63
20 8.10 5.85 4.94 4.43 4.10 3.87 3.70 3.56
30 7.56 5.39 4.51 4.02 3.70 3.47 3.30 3.17
40 7.31 5.18 4.31 3.83 3.51 3.29 3.12 2.99
50 7.17 5.06 4.20 3.72 3.41 3.19 3.02 2.89
60 7.08 4.98 4.13 3.65 3.34 3.12 2.95 2.82
70 7.01 4.92 4.07 3.60 3.29 3.07 2.91 2.78
80 6.96 4.88 4.04 3.56 3.26 3.04 2.87 2.74
90 6.93 4.85 4.01 3.54 3.23 3.01 2.84 2.72
100 6.90 4.82 3.98 3.51 3.21 2.99 2.82 2.69
∞ 6.63 4.61 3.78 3.32 3.02 2.80 2.64 2.51

8
Points fα,ν ,ν for α = 0.025.
1 2
ν2 ν1
1 2 3 4 5 6 7 8
1 647.79 799.50 864.16 899.58 921.85 937.11 948.22 956.65
2 38.51 39.00 39.17 39.25 39.30 39.33 39.36 39.37
3 17.44 16.04 15.44 15.10 14.88 14.73 14.62 14.54
4 12.22 10.65 9.98 9.60 9.36 9.20 9.07 8.98
5 10.01 8.43 7.76 7.39 7.15 6.98 6.85 6.76
6 8.81 7.26 6.60 6.23 5.99 5.82 5.70 5.60
7 8.07 6.54 5.89 5.52 5.29 5.12 4.99 4.90
8 7.57 6.06 5.42 5.05 4.82 4.65 4.53 4.43
9 7.21 5.71 5.08 4.72 4.48 4.32 4.20 4.10
10 6.94 5.46 4.83 4.47 4.24 4.07 3.95 3.85
11 6.72 5.26 4.63 4.28 4.04 3.88 3.76 3.66
12 6.55 5.10 4.47 4.12 3.89 3.73 3.61 3.51
13 6.41 4.97 4.35 4.00 3.77 3.60 3.48 3.39
14 6.30 4.86 4.24 3.89 3.66 3.50 3.38 3.29
15 6.20 4.77 4.15 3.80 3.58 3.41 3.29 3.20
16 6.12 4.69 4.08 3.73 3.50 3.34 3.22 3.12
17 6.04 4.62 4.01 3.66 3.44 3.28 3.16 3.06
18 5.98 4.56 3.95 3.61 3.38 3.22 3.10 3.01
19 5.92 4.51 3.90 3.56 3.33 3.17 3.05 2.96
20 5.87 4.46 3.86 3.51 3.29 3.13 3.01 2.91
30 5.57 4.18 3.59 3.25 3.03 2.87 2.75 2.65
40 5.42 4.05 3.46 3.13 2.90 2.74 2.62 2.53
50 5.34 3.97 3.39 3.05 2.83 2.67 2.55 2.46
60 5.29 3.93 3.34 3.01 2.79 2.63 2.51 2.41
70 5.25 3.89 3.31 2.97 2.75 2.59 2.47 2.38
80 5.22 3.86 3.28 2.95 2.73 2.57 2.45 2.35
90 5.20 3.84 3.26 2.93 2.71 2.55 2.43 2.34
100 5.18 3.83 3.25 2.92 2.70 2.54 2.42 2.32
∞ 5.02 3.69 3.12 2.79 2.57 2.41 2.29 2.19

9
Points fα,ν ,ν for α = 0.05.
1 2
ν2 ν1
10 12 15 20 24 30 40 100
1 241.88 243.90 245.95 248.01 249.05 250.09 251.14 253.04
2 19.40 19.41 19.43 19.45 19.45 19.46 19.47 19.49
3 8.79 8.74 8.70 8.66 8.64 8.62 8.59 8.55
4 5.96 5.91 5.86 5.80 5.77 5.75 5.72 5.66
5 4.74 4.68 4.62 4.56 4.53 4.50 4.46 4.41
6 4.06 4.00 3.94 3.87 3.84 3.81 3.77 3.71
7 3.64 3.57 3.51 3.44 3.41 3.38 3.34 3.27
8 3.35 3.28 3.22 3.15 3.12 3.08 3.04 2.97
9 3.14 3.07 3.01 2.94 2.90 2.86 2.83 2.76
10 2.98 2.91 2.85 2.77 2.74 2.70 2.66 2.59
11 2.85 2.79 2.72 2.65 2.61 2.57 2.53 2.46
12 2.75 2.69 2.62 2.54 2.51 2.47 2.43 2.35
13 2.67 2.60 2.53 2.46 2.42 2.38 2.34 2.26
14 2.60 2.53 2.46 2.39 2.35 2.31 2.27 2.19
15 2.54 2.48 2.40 2.33 2.29 2.25 2.20 2.12
16 2.49 2.42 2.35 2.28 2.24 2.19 2.15 2.07
17 2.45 2.38 2.31 2.23 2.19 2.15 2.10 2.02
18 2.41 2.34 2.27 2.19 2.15 2.11 2.06 1.98
19 2.38 2.31 2.23 2.16 2.11 2.07 2.03 1.94
20 2.35 2.28 2.20 2.12 2.08 2.04 1.99 1.91
30 2.16 2.09 2.01 1.93 1.89 1.84 1.79 1.70
40 2.08 2.00 1.92 1.84 1.79 1.74 1.69 1.59
50 2.03 1.95 1.87 1.78 1.74 1.69 1.63 1.52
60 1.99 1.92 1.84 1.75 1.70 1.65 1.59 1.48
70 1.97 1.89 1.81 1.72 1.67 1.62 1.57 1.45
80 1.95 1.88 1.79 1.70 1.65 1.60 1.54 1.43
90 1.94 1.86 1.78 1.69 1.64 1.59 1.53 1.41
100 1.93 1.85 1.77 1.68 1.63 1.57 1.52 1.39
∞ 1.83 1.75 1.67 1.57 1.52 1.46 1.39 1.24

10
Inverse Property of the F-Distribution

Show that the critical value of the F-distribution satisfies the identity:
1
f1−α,ν1 ,ν2 =
fα,ν1 ,ν2

▶ This property is particularly useful when working with F-tables and


hypothesis testing.

11
Distribution of the Ratio of Sample Variances

Theorem: If S12 and S22 are the variances of independent random


samples of sizes n1 and n2 from two normal populations with variances
σ12 and σ22 , respectively. Then

S12 /σ12
F =
S22 /σ22

is a random variable having a F-distribution with n1 − 1 and n2 − 1


degrees of freedom.

12
Example: Let S12 and S22 denote the variances of independent random
samples of size n1 = 8 and n2 = 12, taken from normal populations with
the same standard deviation. Estimate:
S12
P( < 4.89)
S22

13
Order Statistic
Order statistic

▶ Consider a random sample X1 , X2 , . . . , Xn from a continuous


population. We arrange the values in increasing order

Y1 , Y2 , . . . , Yn

and refer to them as order statistics. That is Y1 := min{X1 , X2 , . . . , Xn }


is the smallest and Yn := max{X1 , X2 , . . . , Xn } is the largest. The term
Yr is called r ’th order statistics.
▶ We denote the sample median X̃ . In a random sample of size
n = 2m + 1, it is Ym+1 . In a random sample of size n = 2m, it is defined
as 12 (Ym + Ym+1 ).

14
Example

Let f and F denote the population density and distribution function


respectively. In terms of f and F , derive the density of Y1 and Yn .

15
Theorem: The distribution of r ’th order statistic Yr is given by
n  
X n
FYr (y ) = [F (y )]k [1 − F (y )]n−k
k
k=r

and the density is


n!
fYr (y ) = f (y )[F (y )]r −1 [1 − F (y )]n−r
(n − 1)!(n − r )!

for −∞ < y < ∞.

16
Example: Let X1 , X2 , . . . , Xn be a random sample taken from a
continuous population with the density
(
1 −x/θ
e for x > 0
f (x) = θ
0 elsewhere

Find the density and expected value of Y1 and Yn .

17
Example: Let X1 , X2 , . . . , Xn be a random sample taken from a
continuous population with the density
(
1
for 0 < x < θ
f (x) = θ
0 elsewhere

Find the density and expected value of Y1 and Yn .

18
Example: Consider a random sample X1 , X2 , X3 , X4 from a uniform
distribution on the interval (0, 1). Calculate the expected value E(Y3 ) of
the third order statistics.

19
Point Estimation
Introduction

There are two major topics in statistics:

▶ Estimation: Point estimation (Ch. 10) and Interval estimation (Ch. 11).
▶ Testing hypotheses (Ch. 12).

First we discuss point estimation.

▶ What is point estimation?


▶ We have a random sample from a probability distribution.
▶ The distribution is known, except for one or more parameters, usually
denoted θ (or θ1 , . . . , θk ).
▶ Question: can we estimate the value of the parameter(s)?
▶ The “true” value remains usually unknown, even after taking a sample.
From the sample we calculate the estimate θ̂ of the parameter.

20
Statistics

Definition: Let X1 , X2 , . . . , Xn be a random sample. A statistic is a


function of our observations and also possibly of other known parameters.
That is, it is a random variable of the form h(X1 , X2 , . . . , Xn ) for a given
function h that is allowed to depend on known parameters as well.
Example of statistics:

▶ Sample mean: X
▶ Sample median: S 2
▶ Sample median
▶ Order statistics

21
Point Estimation

Definition: Using the value of a sample statistic to estimate the value of


a population parameter is called point estimation. We refer to the value
of the statistic as a point estimate. Correspondingly, we refer to the
statistics themselves as point estimators.
Remark: For some unknown parameters, there are natural candidates as
estimators. For example, when the population mean µ is unknown, the
sample mean X can be used as an estimator. When population variance
σ 2 must be estimated, the sample variance S 2 can be used as an
estimator. Sometimes there are more than one candidate; for example,
when µ is unknown, you can use the sample mean X , the sample median,
or X1 + Xn /2. In such cases, it is a good idea to compare the candidates
based on their properties.

22
Choosing the best estimator

What makes a good estimator? - We need properties that help us


determine which estimator performs best.
1. Unbiasedness
2. Minimum Variance (Efficiency)
3. Consistency
4. Sufficiency
5. Robustness
These properties may or may not be attained by a point estimator.
Since point estimators are random variables, they have their sampling
distributions. The above properties are directly related to their sampling
distributions. Sampling distributions are discussed in Ch 8.

23
Definition: A statistic Θ̂ is an unbiased estimator of the parameter θ of
a given distribution if and only if E[Θ̂] = θ for all possible values of θ.

24
Example: If X has the binomial distribution with the parameters n and
θ, show that the sample proportion, Xn , is an unbiased estimator of θ.

25
Example: Let X1 , . . . , Xn be a random sample from a continuous
population with the density function
(
e −(x−λ) for x > λ
f (x) =
0 elsewhere

for some unknown parameter λ > 0.


(a) Determine whether X is an unbiased estimator of λ.
(b) Determine whether X − 1 is an unbiased estimator of λ.

26
Asymptotically Unbiased Estimator

Defrinition: An estimator Θ̂ is biased if E[Θ̂] ̸= θ.


Definition: The bias of an estimator is given by:

bn (θ) = E[Θ̂] − θ

▶ If bn (θ) > 0, the estimator tends to overestimate.


▶ If bn (θ) < 0, the estimator tends to underestimate.
Definition: The estimator Θ̂ of θ is asymptotically unbiased if

lim bn (θ) = 0
n→∞

27
Example: Suppose X1 , X2 , . . . , Xn is a random sample from a uniform
distribuiton on (0, θ), where θ is an unknown parameter. Show that the
estimator Yn = max{X1 , . . . , Xn } is asymptotically unbiased.

28

You might also like