CHAPTER 13 Design and Analysis of Single-Factor Experiments: The Analysis of Variance

Download Report

Transcript CHAPTER 13 Design and Analysis of Single-Factor Experiments: The Analysis of Variance

CHAPTER 13 Design and Analysis of Single-Factor Experiments: The Analysis of Variance

Learning Objectives

• Design and conduct engineering experiments • Understand how the analysis of variance is used to analyze the data • Use multiple comparison procedures • Make decisions about sample size • Understand the difference between fixed and random factors • Estimate variance components • Understand the blocking principle • Design and conduct experiments involving the randomized complete block design

Engineering Experiments

• Experiments are a natural part of the engineering decision-making process • Designed to improve the performance of a subset of processes • Processes can be described in terms of controllable variables • Determine which subset has the greatest influence • Such analysis can lead to – Improved process yield – Reduced variability in the process and closer conformance to nominal or target requirements – Reduced cost of operation

Steps In Experimental Design

• Usually designed

sequentially

• Determine which variables are most important • Used to refine the information to determine the critical variables for improving the process • Determine which variables result in the best process performance

Single Factor Experiment

• Assume a parameter of interest • Consist of making up several specimens in two samples • Analyzed them using the statistical hypotheses methods • Can say an experiment with single factor – Has two levels of investigations – Levels are called treatments – Treatment has n observations or replicates

Designing Engineering Experiments

• More than two levels of the factor • This chapter shows – ANalysis Of VAriance (ANOVA) – Discuss r

andomization

of the experimental runs • Design and analyze experiments with several factors

Linear Statistical Model

• Following linear model

Y ij =

µ+ 2,…,n  i + 

ij

• i=1, 2,…,a, and j=1, • • •

Y ij

• µ is (ij)th observation called the overall mean  i called the

i

th treatment effect 

ij

is a random error component with mean zero and variance  2 • Each treatment defines a population • Mean µ i consisting of the overall mean µ • Plus an effect 

i

Pg. 471 Fig 13-1b

Completely Randomized Design

• Table shows the underlying model – Following observations are taken in random order – Treatments are used as uniform as possible • Called

completely randomized design

Fixed-effects and Random Models

• Chosen in two different ways – Experimenter chooses the

a

treatments • Called the fixed-effect model – Experimenter chooses the treatments from a larger population • Called random-effect model

Development of ANOVA

• Total of the observations and the average of the observations under the

i

th treatment • Grand total of all observations and the grand mean –

N=an

is the total number of observations – “dot” subscript notation implies summation

Hypothesis Testing

• • • Interested in testing the equality of the following

a

treatment means  1 ,  2 …..

 a Equivalent H 0 :  1 =  2 =…=  a =0 H 1 :  a #0 for at least one i If the null hypothesis is true, changing the levels of the factor has no effect on the mean response

Components of Total Variability

• Total variability in data is described by the total sum of squares • Partitions this total variability into two parts

i

 1

j n a

  1 (

y ij

y

..

) 2 

n i a

  1 (

y i

.

y

..

) 2 

i

 1

j n a

  1 (

y ij

y i

.

) 2 • Measure the differences between treatments • Measure the random error effect

Computational Formulas

• Efficient formulas • Total sum of squares

SS T

i

1

j n a

   1

y ij

2 

y

..

2

N

• Mean square for treatments

MS

Treatments =

SS

Treatments /(a-1) •

Error mean square

MS E

=

SS E

/[

a

(

n-

1)] • Treatment sum of squares

SS Treatments

a

 

y i

2 

y

..

2

i

1

n N

• Error sum of squares SS E =SS T - SS Treatment

ANOVA TABLE

Using Computer Software

• Packages have the capability to analyze data from designed experiments • Presents the output from the Minitab one-way analysis of variance routine

Example

• The tensile strength of a synthetic fiber is of interest to the manufacturer. It is suspected that strength is related to the percentage of cotton in the fiber. Five levels of cotton percentage are used, and five replicates are run in random order, resulting in the data below. Use α=0.05.

a) Does cotton percentage affect breaking strength?

Solution

• Use the general steps in hypothesis testing 1. Parameter of interest is the cotton percentage 2. H 0 :  1 =  2 =  3 =  4 =  5 =0 3. H 1 :  i #0 for at least one I 4.

α = 0.05

5. Test statistic F o = MS TR /MS E 6. Reject H o if f o > f α,(a-1)n(a-1) 7. Computations

Initial calculations

• Compute the last two columns Conc 15 20 25 30 35 1 7 12 14 19 7 2 7 17 18 25 10 3 15 12 18 22 11 4 11 18 19 19 15 5 9 18 19 23 11

y i

.

49 77 88 108 54

y

..

=376

y i

.

9.8

15.4

17.6

21.6

10.8

Solution - Cont.

7. Compute SS T , SS TR , SS E , MS TR , and MS E

SS T

i

1

j n a

   1

y ij

2 

y

..

2

N

=(7) 2 +(7) 2 + ….+(376) 2 /25= 636.96

SS Treatments

i a

  1

y i

2

n

y

..

2

N

=((49) 2 +(77) 2 +..+(54) 2 )/5 -376/5 =475.7

• SS E = 636.96-475.75 = 161.20

MS TR = SS TR /a-1 = 475.76/4 = 118.9

MS E = SS E /a(n-1)=161.20/5(5-1) = 8.0

Hence, the test statistic F o = MS TR /MS E = 118.96/8.06 = 14.75

8. Since f o =14.75> f 0.05,4,20 = 2.87, reject Ho

Solution

ANOVA results Source DF SS MS F

COTTON 4 475.76 118.94 14.76

P

0.000

Error 20 161.20 8.06

Total 24 636.96 • Reject H 0 and conclude that cotton percentage affects breaking strength

Multiple Comparisons Following the ANOVA

• When H 0 :  1 =  2 =…=  a =0 is rejected • Know that some of the treatment are different • Doesn’t identify which means are different • Called multiple comparisons methods • Called Fisher’s least significant difference (LSD) method

Fisher’s Least Significant Difference (LSD) Method

• Compares all pairs of means with the

H

0 : =

i

j

for all i

# j

• Test statistic

t o

y i

.

y j

.

2

MS E n

• Pair of means

i

and

j

would be different

y i

.

y j

.

LSD

• Least significant difference, LSD, is

LSD

t

 / 2 ,

a

(

n

 1 ) 2

MS E n

Example

• Use Fisher’s LSD method with α = 0.05 test to analyze the means of five different levels of cotton percentage content in the previous example • Recall H 0 was rejected and concluded that cotton percentage affects the breaking strength • Apply the Fisher’s LSD method to determine which treatment means are different

Solution

• Summarize – a =5 means, n=5, MS E t 0.025,20 =2.086

– Treatment means are = 8.06, and

y

1 .

y

2 .

y

3 .

y

4 .

y

5 .

9.8

15.4

17.6

21.6

10.8

Solution –Cont.

• Value of LSD

LSD

t

 / 2 ,

a

(

n

 1 ) • Comparisons 5 Vs. 1=I10.8

–9.8I=1 2

MS E n LSD

 2 .

086 2 ( 8 .

06 ) 5  3 .

7455 5 Vs. 2=I10.8-15.4I=4.6>3.74

5 Vs. 3=I10.8-17.6I=6.8>3.74

5 Vs. 4=I10.8-21.6I=10.8>3.74

4 Vs. 1=I21.6-9.8I=11.8>3.74

4 Vs. 2=I21.6-15.4I=6.2 > 3.74

4 Vs. 3=I21.6 –17.6I=4>3.74

3 Vs. 1=I17.6-9.8I=7.8>3.74

3 Vs. 2=I17.6 -15.4I=2.2

2 Vs. 1=I15.4-9.8I=5.6>3.74

• From this analysis, we see that there are significant differences between all pairs of means except 5 vs. 1 and 3 vs. 2

C.I. on Treatment Means

• Confidence interval on the mean of the ith treatment µ

i y i

.

t

 / 2 ,

a

(

n

 1 )

MS E n

 

i

y i

.

t

 / 2 ,

a

(

n

 1 )

MS E n

• Confidence interval on the difference in two treatment means

y i

.

y j

.

t

 / 2 ,

a

(

n

 1 ) 2

MS E n

 

i

 

j

y i

.

y j

.

t

 / 2 ,

a

(

n

 1 ) 2

MS E n

Determining Sample Size

• • Choice of the sample size to use is important

OC curves

selection provide guidance in making this • Power of the ANOVA test is 1 β=P( Reject H 0 | H 0 is false) =P(F 0 >

f α, a-1, a(n-1)

• Plot β against a parameter  | H 0 is false)  2 

n a

 

i a

1   2

i

2

Sample OC Curves

Example

• Suppose that four normal populations have common variance  2 =25 and means µ 1 =50, µ 2 =60, µ 3 =50, and µ 4 =60. How many observations should be taken on each population so that the probability of rejecting the hypothesis of equality of means is at least 0.90? Use α=0.05

Solution

• • Average mean   55  1  2 = -5,  2

a

n a i

  1   2

i

2 = 5,  3  2  = -5,  4

n

( 100 ) 4 ( 25 ) = 5 

n

,

a

 1  3 where

a

(

n

 1 )  4 (

n

 1 ) • Various choices: n 4 5 4 5  2  2 2.24

a(n-1) 12 16 Power=1  0.80

0.95

• Therefore, n = 5 is needed

The Random-effects Model

• A large number of possible levels • Experimenter randomly selects

a

of these levels from the population of factor levels – Called random-effect model • Valid for the entire population of factor levels

Linear Statistical Model

• • Following linear model 

= Y ij =

µ+  i + 

ij

1,2,….a, j=1,2,…n • • • •

Y ij

i

ij

i

is the (ij)th observation and 

ij

are independent random variables • Identical in structure to the fixed-effects case • Parameters have a different interpretation are with mean 0 and variance  2 are with mean zero and variance   2

Testing the Hypothesis

• Testing the hypothesis that the individual treatment effects are zero is meaningless • Appropriate to test a hypothesis about the variance of the treatment effect • H 0 :   2 =0 vs. H 1 :   2 >0   2 =0, all treatments are identical • There is variability between them • Total variability

SST=SS

Treatments +

SS E

• Expected values of the MS Eq 13 -21,22 E(MS Treatments )=  2 + n   2 and

E(MS E )=

 2 • Computational procedure and construction of the ANOVA table are identical to the fixed-effects case

Randomized Complete Block Design

• Desired to design an experiment so that the variability arising from a nuisance factor can be controlled • Recall about the paired • See the paired

t

nuisance factor effect extension of the paired

t t

-test • When all experimental runs cannot be made under homogeneous conditions -test as a method for reducing the noise in the experiment by blocking out a • Randomized block design can be viewed as an -test • Factor of interest has more than two levels • More than two treatments must be compared

Randomized Complete Block Design

• General procedure for a randomized complete block design consists of selecting

b

blocks • Data that result from running a randomized complete block design for investigating a single factor with

a

levels and

b

blocks

Linear Statistical Model

Following linear model • • • •

Y ij

j

is the (ij)th observation is the effect of the

j

th block • µ called the overall mean   i 

ij

called the

i

and variance th treatment effect is a random error component with mean zero  2

Hypothesis Testing

• • • Interested in testing the equality of the

a

treatment means  1 ,  2 …..

 a Equivalent H 0 :  1 =  2 =…=  a =0 H 1 :  a #0 for at least one i If the null hypothesis is true, changing the levels of the factor has no effect on the mean response

Displaying Data

Components of Total Variability

• Total variability in data • Partitions this total variability into three parts • Or symbolically, SST=SS treatments +SS blocks +SS errors

Computational Formulas

• Computing formulas for the sums of squares • Error sum of squares SS errors =SST-SS treatments -SS perform the analysis of variance blocks • Computer software package will be used to

Analysis of Variance

Next Agenda

• Ends our discussion with the analysis of variance when there are more then two levels of a single factor • In the next chapter, we will show how to design and analyze experiments with several factors with more than two levels