Hierarchical Linear Modeling

Download Report

Transcript Hierarchical Linear Modeling

1

Hierarchical Linear Modeling

David A. Hofmann Kenan-Flagler Business School University of North Carolina at Chapel Hill Academy of Management August, 2007

2

Overview of Session

• Overview – Why are multilevel methods critical for organizational research – What is HLM (conceptual level) – How does it differ from “traditional” methods of multilevel analysis – How do you do it (estimating models in HLM) – Critical decisions and other issues

3

Why Multilevel Methods

4

Why Multilevel Methods

• Hierarchical nature of organizational data – Individuals nested in work groups – Work groups in departments – Departments in organizations – Organizations in environments • Consequently, we have constructs that describe: – Individuals – Work groups – Departments – Organizations – Environments

5

Why Multilevel Methods

• Hierarchical nature of longitudinal data – Time series nested within individuals – Individuals – Individuals nested in groups • Consequently, we have constructs that describe: – Individuals over time – Individuals – Work groups

6

Why Multilevel Methods

• Meso Paradigm – Micro OB – Macro OB – Call for shifting focus: • Contextual variables into Micro theories • Behavioral variables into Macro theories • Longitudinal Paradigm – Intraindividual change – Interindividual differences in individual change

7

What is HLM

8

What is HLM

• Hierarchical linear modeling – The name of a software package – Used as a description for broader class of models • Random coefficient models • Models designed for hierarchically nested data structures • Typical application – Hierarchically nested data structures – Outcome at lowest level – Independent variables at the lowest + higher levels

9

What is HLM

• What might be some examples – Organizational climate predicting individual outcomes over and above individual factors – Organizational climate as a moderator of individual level processes – Individual characteristics (personality) predicting differences in change over time (e.g., learning) – Organizational structure moderating the relationship between individual characteristics – Industry characteristics moderating relationship between corporate strategy and performance

10

What is HLM

• Yes, but what

IS

it … – HLM models variance at two levels of analysis – At a conceptual level • Step 1 – Estimates separate regression equations within units – This summarizes relationships within units (intercept and slopes) • Step 2 – Uses these “summaries” of the within unit relationships as outcome variables regressing them on level-2 characteristics – Mathematically, not really a two step process, but this helps in understanding what is going on

11

What is HLM

Y ij

Level 1: Regression lines estimated separately for each unit X ij Level 2:

• •

Variance in Intercepts predicted by between unit variables Variance in slopes predicted by between unit variables

12

What is HLM

• For those who like equations … • Two-stage approach to multilevel modeling – Level 1: within unit relationships for each unit – Level 2: models variance in level-1 parameters (intercepts & slopes) with between unit variables Level 1: Level 2: Y ij ß 0j ß 1j = ß 0j =  00 + ß 1j +  01 X ij + r ij (Group j ) + U 0j =  10 +  11 (Group j ) + U 1j { j subscript indicates parameters that vary across groups}

13

What is HLM

• Think about a simple example – Individual variables • Helping behavior (DV) • Individual Mood (IV) – Group variable • Proximity of group members

14

What is HLM

• Hypotheses 1. Mood is positively related to helping 2. Proximity is positively related to helping after controlling for mood • On average, individuals who work in closer proximity are more likely to help; a group level main effect for proximity after controlling for mood 3. Proximity moderates mood-helping relationship • The relationship between mood and helping behavior is stronger in situations where group members are in closer proximity to one another

What is HLM

High Proximity Helping Low Proximity Mood

Overall, positive relationship between mood and helping (average regression line across all groups)

Overall, higher proximity groups have more helping that low proximity groups (average intercept green/solid vs. mean red/dotted line)

Average slope is steeper for high proximity vs. low proximity

15

What is HLM

• For those who like equations … • Here are the equations for this model Level 1: Helping ij Level 2: ß 0j =  ß 1j =  = ß 0j 00 +  10 +  + ß 1j 01 11 (Mood ij (Proximity (Proximity ) + r j j ij ) + U ) + U 1j 0j

ß 0j

16

17

How is HLM Different

18

How is HLM Different

• Ok, this all seems reasonable … • And clearly this seems different from “traditional” regression approaches – There are intercepts and slopes that vary across groups – There are level-1 and level-2 models • But, why is all this increased complexity necessary …

How is HLM Different

• “Traditional” regression analysis for our example – Compute average proximity score for the group – “Assign” this score down to everyone in the group

ID 1 2 3 4 5 6 7 8 Grp 1 1 1 1 2 2 2 2 Help 5 3 4 4 2 3 3 2 Mood 5 4 3 5 1 3 3 4 Prox 5 5 5 5 10 10 10 10

19

20

How is HLM Different

• Then you would run OLS regression • Regress helping onto mood and proximity • Equation: Help = b 0 + b 1 (Mood) + b 2 (Prox) + b 3 (Mood*Prox) + e ij • Independence of e ij component assumed – But, we know individuals are clustered into groups – Individuals within groups more similar than individuals in other groups – Violation of independence assumption

21

How is HLM Different

• OLS regression equation (main effect): Helping ij = b 0 + b 1 (Mood) + b 2 (Prox.) + e ij • The HLM equivalent model (ß 1j across groups): is fixed Level 1: Level 2: Helping ß ß 0j 1j ij =  =  00 10 = ß 0j +  + ß 1j (Mood) + r 01 (Prox.) + U 0j ij

22

How is HLM Different

• Form single equation from HLM models – Simple algebra – Take the level-1 formula and replace ß 0j equations for these variables and ß 1j with the level-2 Help = [  = =   00 00 00 + +   10 01 (Prox.) + U 0j (Mood) +  01 ] + [  10 ] (Mood) + r (Prox.) + U 0j + r ij ij + ß 1j (Mood) +  01 (Prox.) + [U 0j + r ij ] OLS = b 0 + b 1 (Mood) + b 2 (Prox.) + e ij Only difference – Instead of e ij you have [U 0j + r ij ] – No violation of independence, because different components are estimated instead of confounded

23

How is HLM Different

• HLM – Models variance at multiple levels – Analytically variables remain at their theoretical level – Controls for (accounts for) complex error term implicit in nested data structures – Weighted least square estimates at Level-2

24

Estimating Models in HLM

• Ok, I have a sense of what HLM is • I think I’m starting to understand that it is different that OLS regression … and more appropriate • So, how do I actually start modeling hypotheses in HLM

25

Estimating Models in HLM

26

Estimating Models in HLM

Some Preliminary definitions: – Random coefficients/effects • Coefficients/effects that are assumed to vary across units – Within unit intercepts; within unit slopes; Level 2 residual – Fixed effects • Effects that do

not

vary across units – Level 2 intercept, Level 2 slope Level 1: Level 2: Helping ij ß 0j =  00 = ß 0j +  ß 1j =  10 + ß 1j (Mood) + r 01 (Prox.) + U 0j ij

27

Estimating Models in HLM

• Estimates provided: – Level-2 parameters (intercepts, slopes)** – Variance of Level-2 residuals*** – Level 1 parameters (intercepts, slopes) – Variance of Level-1 residuals – Covariance of Level-2 residuals • Statistical tests: – t-test for parameter estimates (Level-2, fixed effects)** – Chi-Square for variance components (Level-2, random effects)***

28

Estimating Models in HLM

• Hypotheses for our simple example 1. Mood is positively related to helping 2.

Proximity is positively related to helping after controlling for mood • On average, individuals who work in closer proximity are more likely to help; a group level main effect for proximity after controlling for mood 3.

Proximity moderates mood-helping relationship • The relationship between mood and helping behavior is stronger in situations where group members are in closer proximity to one another

Estimating Models in HLM

• Necessary conditions – Systematic within and between group variance in helping behavior – Mean level-1 slopes significantly different from zero (Hypothesis 1) – Significant variance in level-1 intercepts (Hypothesis 2) – Variance in intercepts significantly related to Proximity (Hypothesis 2) – Significant variance in level-1 slopes (Hypothesis 3) – Variance in slopes significantly related to Proximity (Hypothesis 3) 29

Estimating Models in HLM

Helping High Proximity Low Proximity Mood

Overall, positive relationship between mood and helping (average regression line across all groups)

Overall, higher proximity groups have more helping that low proximity groups (average intercept green/solid vs. mean red/dotted line)

Average slope is steeper for high proximity vs. low proximity

30

Estimating Models in HLM

• Pop quiz – What do you get if you regress a variable onto a vector of 1’s and nothing else; equation: variable = b 1 (1s) + e – b-weight associated with 1’s equals mean of our variable – This is what regression programs do to model the intercept • How much variance can 1’s account for – Zero – All variance forced to residual

Variable 5 4 5 3 2 4 5 Mean = 4.0

1s 1 1 1 1 1 1 1

31

Estimating Models in HLM

• One-way ANOVA - no Level-1 or Level-2 predictors (null) Level 1: Helping ij Level 2: ß 0j =  00 = ß 0j + U + r 0j ij • where: ß 0j = mean helping for group j  00 = grand mean helping Var ( r ij ) =  2 = within group variance in helping Var ( U 0j ) =    between group variance in helping Var (Helping ij ) = Var ( U 0j + r ij ) =   +  2 ICC =   / (   +  2 ) 32

Estimating Models in HLM

• Random coefficient regression model – Add mood to Level-1 model ( no Level-2 predictors) Level 1: Helping ij Level 2: ß 0j =  = ß 0j + ß 00 ß 1j + U 0j =  10 1j (Mood + U 1j ij ) + r ij • where:  00 = mean (pooled) intercepts (t-test)  10 = mean (pooled) slopes (t-test; Hypothesis 1) Var ( r ij ) =  2 = Level-1 residual variance (R 2 , Hyp. 1) Var ( U 0j ) =    variance in intercepts (related Hyp. 2) Var (U 1j ) = variance in slopes (related Hyp. 3) R 2 = [σ 2 owa - σ 2 rrm] / σ 2 owa] R 2 = [(σ 2 owa +   owa) – (σ 2 rrm +   rrm)] / [σ 2 owa +   owa)] 33

34

Estimating Models in HLM

• Intercepts-as-outcomes - model Level-2 intercept (Hyp. 2) – Add Proximity to intercept model Level 1: Helping ij Level 2: ß 0j =  = ß 0j 00 ß 1j +  = 01  + ß 1j (Mood (Proximity 10 + U 1j ij j ) + r ) + U ij 0j • where:    00 01 = Level-2 intercept (t-test) = Level-2 slope (t-test; Hypothesis 2) 10 = mean (pooled) slopes (t-test; Hypothesis 1) Var ( r ij Var ( U 0j ) = Level-1 residual variance ) =   = residual inter. var (R 2 - Hyp. 2) Var (U 1j ) = variance in slopes (related Hyp. 3) R 2 = [   rrm   intercept ] / [   rrm ] R 2 = [(σ 2 rrm +   rrm) – (σ 2 inter +   inter)] / [σ 2 rrm +   rrm)

Estimating Models in HLM

• Slopes-as-outcomes - model Level-2 slope (Hyp. 3) – Add Proximity to slope model • where: Level 1: Helping ij Level 2: ß 0j =  ß 1j =  00 = ß 0j +  10 +  + ß 01 11 1j (Mood ij ) + r ij (Proximity (Proximity j j ) + U ) + U 0j 1j    00  01 10 11 = Level-2 intercept (t-test) = Level-2 slope (t-test; Hypothesis 2) = Level-2 intercept (t-test) = Level-2 slope (t-test; Hypothesis 3) Var ( r ij ) = Level-1 residual variance Var ( U 0j ) = residual intercepts variance Var (U 1j ) = residual slope var (R 2 - Hyp. 3) 35

36

Other Issues

37

Other Issues

• Assumptions • Statistical power • Centering level-1 predictors • Additional resources

38

Other Issues

• Statistical assumptions – Linear models – Level-1 predictors are independent of the level-1 residuals – Level-2 random elements are multivariate normal, each with mean zero, and variance 

qq

and covariance 

qq’

– Level-2 predictors are independent of the level-2 residuals – Level-1 and level-2 errors are independent.

– Each r ij is independent and normally distributed with a mean of zero and variance  2 for every level-1 unit i within each level-2 unit j (i.e., constant variance in level-1 residuals across units).

39

Other Issues

• Statistical Power – Kreft (1996) summarized several studies – .90 power to detect cross-level interactions 30 groups of 30 – Trade-off • Large number of groups, fewer individuals within • Small number of groups, more individuals per group • More recent paper (ORM) – 30/30 still somewhat holds – Factor in cost of sample level-1 and level-2 – Additional formulas for computing power

40

Other Issues

• Picture this scenario – Fall of 1990 – Just got in the mail the DOS version of HLM – Grad student computer lab at Penn State – Finally, get some multilevel data entered in the software and are ready to go – Then, we are confronted with …

41 Select your level-1 predictor(s): 1 1 for (job satisfaction) 2 for (pay satisfaction) How would you like to center your level-1 predictor Job Satisfaction?

1 for Raw Score 2 for Group Mean Centering 3 for Grand Mean Centering Please indicate your centering choice: ___

42

Other Issues

• HLM forces to you to make a choice in how to center your level-1 predictors • This is a critical decision – The wrong choice can result in you testing theoretical models that are inconsistent with your hypotheses – Incorrect centering choices can also result in spurious cross-level moderators • The results indicate a level-2 variable predicting a level-1 slope • But, this is not really what is going on

43

Centering Decisions

• Level-1 parameters are used as outcome variables at level-2 • Thus, one needs to understand the

meaning

parameters of these • Intercept term: expected value of Y when X is zero • Slope term: expected increase in Y for a unit increase in X • Raw metric form: X equals zero might not be meaningful

44

Centering Decisions

• 3 Options – Raw metric – Grand mean – Group mean • Kreft et al. (1995): raw metric and grand mean equivalent, group mean non-equivalent • Raw metric/Grand mean centering – intercept var = adjusted between group variance in Y • Group mean centering – intercept var =

between

group variance in Y [ Kreft, I.G.G., de Leeuw, J., & Aiken, L.S. (1995). The effect of different forms of centering in Hierarchical Linear Models.

Multivariate Behavioral Research, 30

, 1-21.]

45

Centering Decisions

• Bottom line – Grand mean centering and/or raw metric estimate incremental models • Controls for variance in level-1 variables prior to assessing level-2 variables – Group mean centering • Does NOT estimate incremental models – Does not control for level-1 variance before assessing level-1 variables – Separately estimates with group regression and between group regression

46

Centering Decisions

• An illustration from Hofmann & Gavin (1998): – 15 Groups / 10 Observations per group – Individual variables: A, B, C, D – Between Group Variable: G j • G =

f

(A • Thus, if between group variance in A & B (i.e., A is accounted for, G outcome j , B j ) j j & B j should not significantly predict the ) – Run the model: • Grand Mean • Group Mean • Group + mean at level-2

47

Centering Decisions

• Grand Mean Centering Null A ij A ij B ij A A ij ij B B ij ij

Variables included in: Level 1 Model Level 2 Model

C C ij ij D ij G j G j G j G j G j

Parameter Est. G j .445** .209** .064

.007

-.028

• What do you see happening here … what can we conclude?

Centering Decisions

• Group Mean Centering Null A ij - A j A ij - A j B ij - B j A A ij ij - A - A

Variables included in: Level 1 Model Level 2 Model

j j B B ij ij - B - B j j C C ij ij - C - C j j D ij - D j G j G j G j G j G j

Parameter Est. G j .445** .445** .445** .445** .445**

• What do you see happening here … what can we conclude?

48

49

Centering Decisions

• Group Mean Centering with A, B, C, D Means in Level-2 Model Null A ij - A j A A A ij ij ij - A - A - A

Variables included in: Level 1 Model Level 2 Model

j j j B B B ij ij ij - B - B - B j j j C C ij ij - C - C j j D ij - D j G j G j A j G j A j B j G j A j B j C j G j A j B j C j D j

Parameter Est. G j .445** .300* .132

.119

.099

• What do you see happening here … what can we conclude?

50

Centering Decisions

• Centering decisions are also important when investigating cross-level interactions • Consider the following model: Level 1: Y ij Level 2: ß 0j ß 1j = ß 0j =  =  00 + ß 10 1j + U (X 0j grand ) + r ij • The ß 1j does not provide an unbiased estimate of the pooled within group slope – It actually represents a mixture of both the within and between group slope – Thus, you might not get an accurate picture of cross-level interactions

51

Centering Decisions

• Bryk & Raudenbush make the distinction between cross-level interactions and between-group interactions – Cross-level: Group level predictor of level-1 slopes – Group-level: Two group level predictors interacting to predict the level-2 intercept • Only group-mean centering enables the investigation of both types of interaction • Illustration (Hofmann & Gavin, 1999, J. of Management) – Created two data sets • Cross-level interaction, no between-group interaction • Between-group interaction, no cross-level interaction

Model Level-1:Y ij =  0j +  1j (X ij - X..) + e ij Level-2:  0j =  00 +  01 G 1j +

u 0j

 1j =  10 +  11 G 1j +

u 1j

 00  01  10 

11

Level-1:Y Level-2:  0j =  00 +  01 X j +  02 G j +  03 ( X j G j ) +

u 0j

 1j =  10 ij = +   11 0j G + j  + 1j

u

(X

1j

ij - X j ) + r ij  00  01  02 

03

 10 

11

Cross-level Interaction Para t meter p -.27 ns 3.10 .91 ns -.71

3.51

-.73

7.24

ns

<.01

-.22 .19 1.59

-.19

-.52

4.52

-.03 .10 .44

-.22

-.49

8.43

ns ns

ns

ns

<.01

Between-group interaction Para t p meter 2.10 .63 ns 10.16 6.23 <.01 .01

.72

.01

3.80

ns

<.01

.10 .07 .80

3.77

-.20

.19

.07 .19 1.16

22.50

-.51

.95

ns ns ns

<.01

ns

ns

52

53

Centering Decision

• Incremental – group adds incremental prediction over and above individual variables – grand mean centering – group mean centering with means added in level-2 intercept model • Mediational – individual perceptions mediate relationship between contextual factors and individual outcomes – grand mean centering – group mean centering with means added in level-2 intercept model

54

Centering Decisions

• Moderational – group level variable moderates level-1 relationship – group mean centering provides clean estimate of within group slope – separates between group from cross-level interaction – Practical: If running grand mean centered, check final model group mean centered • Separate – group mean centering produces separate within and between group structural models

55

Do You Really Need HLM?

Alternatives for Estimating Hierarchical Models

56

SAS: Proc Mixed

• SAS Proc Mixed will estimate these models • Key components of Proc Mixed command language – Proc mixed • Class – Group identifier • Model – Regression equation including both individual, group, and interactions (if applicable) • Random – Specification of random effects (those allowed to vary across groups)

57

SAS: Proc Mixed

• Key components of Proc Mixed command language – Some options you might want to select • Class: noitprint ( suppresses interation history ) • Model: – solution ( prints solution for random effects ) – ddfm=bw ( specifies the “between/within” method for computing denominator degrees of freedom for tests of fixed effects ) • Random: – sub=

id

( how level-1 units are divided into level-2 units ) – type=un ( specifies unstructured variance-covariance matrix of intercepts and slopes; i.e., allows parameters to be determined by data )

SAS: Proc Mixed

proc means; run; data; set; {grand mean center mood} moodgrd = mood-5.8388700; data; set; proc mixed noitprint; class id; model helping = / solution; random intercept / sub=id; proc mixed noitprint; class id; model helping = moodgrd/ solution ddfm=bw ; random intercept moodgrd/ sub=id type=un; proc mixed noitprint; class id; model helping = moodgrd proxim / solution ddfm=bw ; random intercept moodgrd / sub=id type=un; proc mixed noitprint; class id; model helping = moodgrd proxim moodgrd*proxim / solution ddfm=bw ; random intercept moodgrd / sub=id type=un; run;

58

Model One-way ANOVA L1: Helping ij =  0j + r ij L2:  0j =  00 + U 0j HLM Proc Mixed Random coefficient regression L1: Helping ij =  0j +  1j (Mood ij ) + r ij L2:  0j =  00 + U 0j L2:  1j =  10 + U 1j HLM Proc Mixed Intercepts-as-outcomes L1: Helping ij =  0j +  1j (Mood ij ) + r ij L2:  0j =  00 +  01 (Proximity j ) + U 0j L2:  1j =  10 + U 1j HLM Proc Mixed Slopes-as-outcomes L1: Helping ij =  0j +  1j (Mood ij ) + r ij L2:  0j =  00 +  01 (Proximity j ) + U 0j HLM L2:  1j =  10 +  11 (Proximity j ) + U 1j Proc Mixed  00 31.39 31.39 31.42 31.43 24.92 24.91 25.14 25.14  01 -- -- -- -- 1.24 1.24 1.19 1.19  10 -- -- 3.01 3.01 3.01 3.01 2.06 2.06  11 -- -- -- -- -- -- .18 .18  31.76 31.76 99.82 99.81 5.61 5.61 5.61 5.61 2 5.61 5.61  00 45.63 45.64 41.68 41.68 42.95 42.94  11 -- -- .13 .13 .13 .13 .02 .02 59

SAS: Proc Mixed

• Key references – Singer, J. (1998). Using SAS PROC MIXED to fit multilevel models, hierarchical models, and individual growth models.

Journal of Educational and Behavioral Statistics, 23,

323 355.

60 • Available on her homepage – http://hugse1.harvard.edu/~faculty/singer/

Resources

• http://www.unc.edu/~dhofmann/hlm.html

– PowerPoint slides – Annotated output – Raw data + system files – Link to download student version of HLM – Follows chapter: • Hofmann, D.A., Griffin, M.A., & Gavin, M.B. (2000). The application of hierarchical linear modeling to management research. In K.J. Klein, & S.W.J. Kozlowski, (Eds.),

Multilevel theory, research, and methods in organizations: Foundations, extensions, and new directions

. Jossey-Bass, Inc. Publishers.

– Proximity = Group Cohesion • Also: http://www.ssicentral.com/hlm/hlmref.htm

61