Transcript X - UCLA

Otis Dudley Duncan Memorial Lecture

THE RESURRECTION OF DUNCANISM

Judea Pearl University of California Los Angeles (www.cs.ucla.edu/~judea/)

1

OUTLINE

1. Duncanism = Causally Assertive SEM 2. History: Oppression, Distortion, and Resurrection 3. The Old-New Logic of SEM 4. New Tools 4.1

Local testing 4.2

4.3

4.3

Non-parametric identification Logic of counterfactuals Non-linear mediation analysis 2

DUNCAN BOOK

3

DUNCANISM = ASSERTIVE SEM

SEM – A tool for deriving causal conclusion from data and assumptions: • • • • [

y = bx + u

] may be read as "a change in produces a change in

y

” ... or “

x

(ibid, p. viii) and

u x

or

u

are the causes of

y

” (Duncan, 1975, p. 1) “ [the disturbance]

u

stands for all other sources of variation in

y

” (ibid) “ doing the model consists largely in thinking about what kind of model one wants and can justify .” Assuming a model for sake of argument, we can express its properties in terms of correlations and (sometimes) find one or more conditions that must hold if the model is true“ (ibid, p. 20). 4

HISTORY: BIRTH, OPPRESSION, DISTORTION, RESURRECTION

• • • Birth and Development (1920 - 1980) Sewell Wright (1921), Haavelmo (1943), Simon (1950), Marschak (1950), Koopmans (1953), Wold and Strotz (1963), Goldberger (1973), Blalock (1964), and Duncan (1969, 1975) The regressional assault (1970-1990) Causality is meaningless, therefore, to be meaningful, SEM must be a regression technique, not “causal modeling.” Richard (1980) Cliff (1983), Dempster (1990), Wermuth (1992), and Muthen (1987) The Potential-outcome assault (1985-present) Causality is meaningful but, since SEM is a regression technique, it could not possibly have causal interpretation. Rubin (2004, 2010), Holland (1986) Imbens (2009), and Sobel (1996, 2008) 5

REGRESSION VS. STRUCTURAL EQUATIONS (THE CONFUSION OF THE CENTURY) Regression (claimless, nonfalsifiable):

Y = ax +

Y

Structural (empirical, falsifiable):

Y = bx + u Y

(

:

= assignment) Claim: (regardless of distributions):

E

(

Y | do

(

x

))

= E

(

Y | do

(

x

)

, do

(

z

)) =

bx

The mothers of all questions: Q. When would

b

equal

a

?

A. When (

u Y X

), read from the diagram Q. When is

b

a partial regression?

b

A. Shown in the diagram, Slide 36.

= 

YX • Z

6

THE POTENTIAL-OUTCOME ASSAULT (1985-PRESENT) • SEM = regressional strawman “I am speaking, of course, about the equation: {

y

What does it mean? The only meaning I have ever 

a

bx

  }.

determined for such an equation is that it is a shorthand way of describing the conditional distribution of {

y

} given {

x

} .” (Holland 1995) • “The use of complicated causal-modeling software [read SEM] rarely yields any results that have any interpretation as causal effects .” (Wilkinson and Task Force 1999 on Statistical Methods in Psychology Journals: Guidelines and Explanations) 7

THE POTENTIAL-OUTCOME ASSAULT (1985-PRESENT) (Cont) • In general (even in randomized studies), the structural and causal parameters are not equal, implying that the structural parameters should not be interpreted as effect .” (Sobel 2008) • “Using the observed outcome notation entangles the science...Bad! Yet this is exactly what regression approaches, path analyses, directed acyclic graphs, and so forth essentially compel one to do.” (Rubin 1010) 8

SEM REACTION TO THE STRUCTURE-PHOBIC ASSAULT

NONE TO SPEAK OF!

Galles, Pearl, Halpern (1998 - Logical equivalence) Heckman-Sobel Morgan Winship (1997) Gelman Blog 9

WHY SEM INTERPRETATION IS “SELF-CONTRADICTORY”

D. Freedman JES (1987), p. 114, Fig. 3

X Y

 

aZ cX

bW

U

( 7 .

1 ) 

dZ

V

( 7 .

2 ) "Now try the direct effect of

Z

on

Y

: We intervene by fixing

W

and

X

but increasing

Z

by one unit; this should increase

Y

by

d

units. However, this hypothetical intervention is self-contradictory , because fixing

W

and increasing

Z

causes an increase in

X

. The oversight: Fixing

X

DISABLES equation (7.1); 10

SEM REACTION TO FREEDMAN CRITICS

• • • • Total Surrender: It would be very healthy if more researchers thinking of and using terms such as cause and effect (Muthen, 1987). abandoned “Causal modeling” is an outdated misnomer (Kelloway, 1998).

Causality-free, politically-safe vocabulary: “covariance structure” “regression analysis,” or “simultaneous equations.” [Causal Modeling] may be somewhat dated, however, as it seems to appear less often in the literature nowadays” (Kline, 2004, p. 9) 11

THE RESURRECTION

Why non-parametric perspective?

Y

 

x

u

What are the parameters all about and why we labor to identify them?

Can we do without them?

Consider:

Y

f

(

x

,

u

) Only he who lost the parameters and needs to find substitutes can begin to ask: Do I really need them? What do they really mean? What role do the play? 12

THE LOGIC OF SEM

A

- CAUSAL ASSUMPTIONS

(M

A

) A* -

Logical implications of

A

Causal inference

Q

Queries of interest

Q(P)

- Identified estimands Data (

D

)

T(M

A

) -

Testable implications Statistical inference

Q

- Estimates of

Q(P)

Q

(

Q

|

D

, Conditional claims

A

)

g

(

T

) Model testing Goodness of fit 13

TRADITIONAL STATISTICAL INFERENCE PARADIGM

Data

P

Joint Distribution

Q

(

P

) (Aspects of

P

) Inference e.g., Infer whether customers who bought product

A

would also buy product

B

.

Q

=

P

(

B

|

A

) 14

FROM STATISTICAL TO CAUSAL ANALYSIS: 1. THE DIFFERENCES Probability and statistics deal with static relations Data

P

Joint Distribution change

P

 Joint Distribution

Q

(

P

 ) (Aspects of

P

 ) Inference What happens when

P

changes?

e.g., Infer whether customers who bought product

A

would still buy

A

if we were to double the price .

15

FROM STATISTICAL TO CAUSAL ANALYSIS: 1. THE DIFFERENCES What remains invariant when

P P

 (

price

=2)=1 changes say, to satisfy Data

P

Joint Distribution change

P

 Joint Distribution

Q

(

P

 ) (Aspects of

P

 ) Note:

P

 (

v

) Inference 

P

(

v

|

price =

2)

P

does not tell us how it ought to change Causal knowledge: what remains invariant 16

FROM STATISTICAL TO CAUSAL ANALYSIS: 1. THE DIFFERENCES (CONT) 1.

Causal and statistical concepts do not mix.

CAUSAL Spurious correlation Randomization / Intervention Confounding / Effect Instrumental variable Exogeneity / Ignorability Mediation STATISTICAL Regression Association / Independence “Controlling for” / Conditioning Odd and risk ratios Collapsibility / Granger causality Propensity score 2.

3.

4.

17

FROM STATISTICAL TO CAUSAL ANALYSIS: 2. MENTAL BARRIERS 4.

1.

2.

3.

Causal and statistical concepts do not mix.

CAUSAL STATISTICAL Spurious correlation Regression Randomization / Intervention Confounding / Effect Association / Independence “Controlling for” / Conditioning Instrumental variable Odd and risk ratios Exogeneity / Ignorability Collapsibility / Granger causality Mediation Propensity score No causes in – no causes out statistical assumptions + data causal assumptions (Cartwright, 1989) }  causal conclusions Causal assumptions cannot be expressed in the mathematical language of standard statistics.

18

FROM STATISTICAL TO CAUSAL ANALYSIS: 2. MENTAL BARRIERS 1.

2.

3.

4.

Causal and statistical concepts do not mix.

CAUSAL STATISTICAL Spurious correlation Regression Randomization / Intervention Confounding / Effect Association / Independence “Controlling for” / Conditioning Instrumental variable Odd and risk ratios Exogeneity / Ignorability Collapsibility / Granger causality Mediation Propensity score No causes in – no causes out statistical assumptions + data causal assumptions (Cartwright, 1989) }  causal conclusions Causal assumptions cannot be expressed in the mathematical language of standard statistics.

Non-standard mathematics: a) Structural equation models (Wright, 1920; Simon, 1960) b) Counterfactuals (Neyman-Rubin (

Y x

)

,

Lewis (

x Y

)) 19

THE STRUCTURAL MODEL PARADIGM

Data Joint Distribution Data Generating Model

M Q

(

M

Inference

M

– Invariant strategy (mechanism, recipe, law, protocol) by which Nature assigns values to variables in the analysis.

) (Aspects of

M

) 20

STRUCTURAL CAUSAL MODELS

• • • • Definition: A structural causal model 

V,U, F, P

(

u

)  , where

V U

= { = {

V U

1 1

,...,V n ,...,U

}

m

} is a 4-tuple are endogenous variables are background variables

F v i

= {

f

1

,..., f n

} =

f i

(

v

,

u

)

P

(

u

) are functions determining

V

, e.g.,

y

   is a distribution over

U

x

u Y P

(

u

) and

F

induce a distribution observable variables

P

(

v

) over 21

CAUSAL MODELS AND COUNTERFACTUALS

Definition: The sentence: “

Y

would be

y

(in unit

u

), had

X

been

x ,

” denoted

Y x

(u) = y

, means: The solution for

Y

in a mutilated model

M x

, (i.e., the equations for

X

replaced by

X

=

x

) with input

U=u ,

is equal to

y .

U U X

(

u

)

Y

(

u

)

X

=

x Y X

(

u

)

M M x

22

CAUSAL MODELS AND COUNTERFACTUALS

Definition: The sentence: “

Y

would be

y

(in unit

u

), had

X

been

x ,

” denoted

Y x

(u) = y

, means: The solution for

Y

in a mutilated model

M x

, (i.e., the equations for

X

replaced by

X

=

x

) with input

U=u ,

is equal to

y .

The Fundamental Equation of Counterfactuals:

Y x

(

u

) 

Y M x

(

u

) 23

CAUSAL MODELS AND COUNTERFACTUALS

Definition: The sentence: “

Y

would be

y

(in situation

u

), had

X

been

x ,

” denoted

Y x

(u) = y

, means: The solution for

Y

in a mutilated model

M x

, (i.e., the equations for

X

replaced by

X

=

x

) with input

U=u ,

is equal to

y .

• Joint probabilities of counterfactuals:

P

(

Y x

y

,

Z w

z

) 

u

:

Y x

(

u

)  In particular: 

y

,

Z w

(

u

) 

z P

(

u

)

P

(

PN y

|

do

(

x

) (

Y x

' 

y

|' )

x

,

P

(

Y x y

)  

y

) 

u

:

Y x

 (

u

) 

y P

(

u

) 

u

:

Y x

' (

u

) 

y

'

P

(

u

|

x

,

y

) 24

3-LEVEL HIERARCHY OF CAUSAL MODELS

1. Probabilistic Knowledge

P(y | x)

Bayesian networks, covariance structure 2. Interventional Knowledge

P(y | do(x))

Causal Bayesian Networks (CBN) (Agnostic graphs, manipulation graphs) 3. Counterfactual Knowledge

P(Y

x

Structural equation models, physics,

= y,Y

x

=y

)

functional graphs 25

TWO PARADIGMS FOR CAUSAL INFERENCE

Observed:

P

(

X, Y, Z,...

) Conclusions needed:

P

(

Y x =y

),

P

(

X y =x

|

Z=z

)...

How do we connect observables,

X,Y,Z

,… to counterfactuals

Y x , X z , Z y

,… ?

N-R

model Counterfactuals are primitives, new variables Super-distribution

P

* (

X

,

Y

,...,

Y x

,

X z

,...)

X

,

Y

,

Z

constrain

Y x

,

Z y

,...

Structural model Counterfactuals are derived quantities Subscripts modify the model and distribution

P

(

Y x

y

) 

P M x

(

Y

y

) 26

ARE THE TWO PARADIGMS EQUIVALENT?

• Yes (Galles and Pearl, 1998; Halpern 1998) • In the

N-R

paradigm,

Y x

consistency: is defined by

Y

xY

1  ( 1 

x

)

Y

0 • In SEM, consistency is a theorem.

• Moreover, a theorem in one approach is a theorem in the other.

27

THE FIVE NECESSARY STEPS OF CAUSAL ANALYSIS

Define: Express the target quantity

Q Q

(

M

) as a function that can be computed from any model

M

.

Assume: Formulate causal assumptions

A

using some formal language. Identify: Determine if

Q

is identifiable given

A

. Estimate: Estimate

Q

if it is not. if it is identifiable; approximate it, Test: Test the testable implications of

A

(if any). 28

THE FIVE NECESSARY STEPS FOR EFFECT ESTIMATION

Define: Assume: Express the target quantity

Q

as a function

Q

(

M

) that can be computed from any model

M

.

ATE 

E

(

Y

|

do

(

x

1 )) 

E

(

Y

|

do

( Formulate causal assumptions

A x

0 )) using some formal language. Identify: Determine if

Q

is identifiable given

A

. Estimate: Estimate

Q

if it is not. if it is identifiable; approximate it, Test: Test the testable implications of

A

(if any). 29

FORMULATING ASSUMPTIONS THREE LANGUAGES

1. English: Smoking (

X

), Cancer (

Y

), Tar (

Z

), Genotypes (

U

) 2. Counterfactuals:

Z x

(

u X y

(

u

) )  

Z yx

(

u

),

X zy

(

u

) 

Y z

(

u

)

Z x

  

Y zx

(

u

), {

Y z

,

X

} Not too friendly: consistent? complete? redundant? arguable?

X z

(

u

) 

X

(

u

), 3.

Structural:

X Z Y

30

IDENTIFYING CAUSAL EFFECTS IN POTENTIAL-OUTCOME FRAMEWORK

Define: Assume: Identify: Express the target quantity

Q

as a counterfactual formula, e.g.,

E

(

Y

(1) –

Y

(0)) Formulate causal assumptions using the distribution:

P

* 

P

(

X

|

Y

,

Z

,

Y

( 1 ),

Y

( 0 )) Determine if

Q

is identifiable using

P

* and

Y=x Y

(1) + (1 –

x

)

Y

(0).

Estimate: Estimate

Q

if it is not.

if it is identifiable; approximate it, 31

GRAPHICAL – COUNTERFACTUALS SYMBIOSIS

Every causal graph expresses counterfactuals assumptions, e.g.,

X

Y

Z

1. Missing arrows

Y

Z Y x

,

z

(

u

) 

Y x

(

u

) 2. Missing arcs

Y Z Y x

 

Z y

consistent, and readable from the graph.

• Express assumption in graphs • Derive estimands by graphical or algebraic methods 32

NON-PARAMETRIC IDENTIFIABILITY

Definition: Let

Q

(

M

) be any quantity defined on a causal model

M

,

and let

A

be a set of assumption.

Q

is identifiable relative to

A

iff

P

(

M

1 ) =

P

(

M

2 ) 

Q

(

M

1 )

= Q

(

M

2 ) for all

M

1 ,

M

2 , that satisfy

A

.

In other words,

Q

can be determined uniquely from the probability distribution

P

(

v

) of the endogenous variables,

V

, and assumptions

A

.

is displayed in graph

G

.

33

IDENTIFICATION IN SEM

Find the effect of

X

on

Y

,

P

(

y

|

do

(

x

)), given the causal assumptions shown in are auxiliary variables.

G

, where

Z

1 ,...,

Z k G Z 1 Z 2 Z 3 Z 4 Z 5 X Z 6 Y

Can

P

(

y

|

do

(

x

)) be estimated if only a subset, can be measured?

Z

, 34

NON-PARAMETRIC IDENTIFICATION USING THE BACK-DOOR CRITERION

P

(

y

|

do

(

x

)) is estimable if there is a set

Z

variables such that

Z d

-separates

X

of

from

Y

in

G x .

Z 1 G Z 2 Z 1 G x Z Z 2 Z 3 Z 3 Z 4 Z 5 Z 5 Z 4 X Z 6 Y X Z 6 Y

• Moreover,

P

(

y

|

do

(“adjusting” for

Z

) (

x

 ))  

P

(

y

|

x z

Ignorability ,

z

)

P

(

z

)  

z P

(

x

,

y

,

z

)

P

(

x

|

z

) 35

THE MOTHER OF ALL QUESTIONS Question: Under what conditions can a path coefficient  be estimated as a regression coefficient, and what variables should serve as the regressors?

Answer: It is all in the diagram (Pearl 2009, p. 180)

S X

Y

The single door criterion  

r YX

S

  

x E

(

Y

|

x

,

s

) i if 1. No member of

2. S S

is a descendant of

Y

, and blocks all paths from

X

to

Y

, except the direct path

X

Y.

36

SEARCHING FOR THE TESTABLE

37

FINDING THE TESTABLES BY INSPECTION

Z 1 Z 2 W 1 Z 3 W 2 X W 3 Y

Missing edges imply conditional independencies and vanishing regression coefficients.

Z

1

Z

1

Z

2

Z

2

Y X

| {

X

1 ,

Z

2 ,

Z

3 } | {

Z

1 ,

Z

3 }

Z

1

Z

1

Z

2 

aZ

2    

b

1

Y c

1

X

 

b

2

X c

3

Z

1  

b

3

Z

2

c

4

Z

3  

b

4

Z

3  '' • Software for automatic detection of all tests (Kyono, 2010)   ' 38

FINDING INSTRUMENTAL VARIABLES

Find an instrument for identifying

b

34 (Duncan, 1975) Find a variable

Z

that, conditioned on

S

, is independent of the disturbance

v

.

By inspection:

X

is

d

-separated from

V

Therefore,

X

1 is a valid instrument 39

FIND EQUIVALENT MODELS

• • Model-equivalence implies

d

-separation-equivalence

d

-separation tests should replace Lee & Herschberger rules

Z X W Z X W Y

(a)

Y

(b) Reason: conditioned on

Y

, Path

Z

Y

X



W

is open in (a) and blocked in (b) 40

CONFOUNDING EQUIVALENCE WHEN TWO MEASUREMENTS ARE EQUALLY VALUABLE

L W

1

Z

W

4

T

Z

T W

2

W

3

V

1

X

 

x E

(

Y

|

x

,

z

)  

x E

(

Y

|

x

,

t

)

V

2

Y

41

FROM IDENTIFICATION TO ESTIMATION

Define: Assume: Identify: Express the target quantity

Q

as a function

Q

(

M

) that can be computed from any model

M

.

Q

P

(

y

|

do

(

x

)) Formulate causal assumptions using ordinary scientific language and represent their structural part in graphical form. Determine if

Q

is identifiable.

Estimate: Estimate

Q

if it is not .

if it is identifiable; approximate it, 42

PROPENSITY SCORE ESTIMATOR (Rosenbaum & Rubin, 1983)

P

(

y

|

do

(

x

))

= ?

Z 3 Z 1 Z 4 Z 2 Z 5

L

X Z 6 Y

Theorem:

L

(

z

1 ,

z

2 ,

z

3 ,

z

4 ,

z

5 ) 

P

(

X

 1 |

z

1 ,

z

2 ,

z

3 ,

z

4 ,

z

5 ) 

P

(

y z

|

z

,

x

)

P

(

z

)  

P

(

y l

|

L

l

,

x

)

P

(

L

l

) Adjustment for

L

replaces Adjustment for

Z

43

WHAT PROPENSITY SCORE (PS) PRACTITIONERS NEED TO KNOW

L

(

z

) 

P

(

X

 1 |

Z

z

) 

z P

(

y

|

z

,

x

)

P

(

z

)  

P

(

y l

|

l

,

x

)

P

(

l

) 1. The asymptotic bias of PS is EQUAL to that of ordinary adjustment (for same

Z

).

2. Including an additional covariate in the analysis CAN SPOIL the bias-reduction potential of others.

3. In particular, instrumental variables tend to amplify bias.

4. Choosing sufficient set for PS, requires knowledge of the model.

44

RETROSPECTIVE COUNTERFACTUALS ETT – EFFECT OF TREATMENT ON THE TREATED 1. Regret: I took a pill to fall asleep. Perhaps I should not have?

2. Program evaluation: What would terminating a program do to those enrolled?

P

(

Y x

y

|

x

' ) 45

COUNTERFACTUAL DEFINITION EFFECT OF TREATMENT ON THE TREATED Define: Assume: Identify: Express the target quantity

Q

as a function

Q

(

M

) that can be computed from any model

M

.

ETT 

P

(

Y x

y

|

X

x

' ) Formulate causal assumptions using ordinary scientific language and represent their structural part in graphical form. Determine if

Q

is identifiable.

Estimate: Estimate

Q

if it is not.

if it is identifiable; approximate it, 46

ETT - IDENTIFICATION

Theorem (Shpitser-Pearl, 2009) ETT is identifiable in

G

iff

P

( y |

do

(

x

),

w

) is identifiable in

G

G

'

W X Y

Moreover, ETT 

P

( (

Y

Complete graphical criterion | |

x

' ' ) ) 

P

(

y

|

do

(

x

),

w

)

G

| '

w

| 

x

' 47

EFFECT DECOMPOSITION (direct vs. indirect effects)

1. Why decompose effects?

2. What is the definition of direct and indirect effects?

3. What are the policy implications of direct and indirect effects?

4. When can direct and indirect effect be estimated consistently from experimental and nonexperimental data?

48

WHY DECOMPOSE EFFECTS?

1. To understand how Nature works 2. To comply with legal requirements 3. To predict the effects of new type of interventions: Signal routing , rather than variable fixing 49

LEGAL IMPLICATIONS OF DIRECT EFFECT

Can data prove an employer guilty of hiring discrimination?

(Gender)

X Z

(Qualifications)

Y

(Hiring) What is the direct effect of

X

on

Y

?

E

(

Y

|

do

(

x

1 ),

do

(

z

)) 

E

(

Y

|

do

(

x

0 ),

do

(

z

)) (averaged over

z

) Adjust for

Z

? No! No!

50

NATURAL INTERPRETATION OF AVERAGE DIRECT EFFECTS

Robins and Greenland (1992) – “Pure”

X Z z = f

(

x, u

)

y = g

(

x, z, u

)

Y

Natural Direct Effect of

X

on

Y

:

DE

(

x

0 ,

x

1 ;

Y

The expected change in

Y

, when we change

X

) from

x

0

x

1 and, for each

u

, we keep attained before the change.

Z

to constant at whatever value it

E

[

Y x

1

Z x

0 

Y x

0 ] In linear models,

DE

= Controlled Direct Effect   (

x

1 

x

0 ) 51

DEFINITION OF INDIRECT EFFECTS

X Z z = f

(

x, u

)

y = g

(

x, z, u

)

Y

Indirect Effect of

X

on

Y

: The expected change in

Y

IE

(

x

0 ,

x

1 ;

Y

) when we keep

X

constant, say at

x

0 , and let

Z

attained had

X

change to whatever value it would have changed to

x

1 .

E

[

Y x

0

Z

Y x

0 ]

x

1 In linear models,

IE = TE – DE

, but not in general.

52

POLICY IMPLICATIONS OF DIRECT EFFECTS

What is the direct effect of

X

on

Y

?

Residual hiring disparity if skills inequality is eliminated GENDER

X

DISABLE

Z

QUALIFICATION

f Y

HIRING Deactivating a link – a new type of intervention 53

POLICY IMPLICATIONS OF INDIRECT EFFECTS

What is the indirect effect of

X

on

Y

?

Residual hiring disparity if employers’ prejudices are eliminated.

GENDER

X Z

QUALIFICATION IGNORE

f Y

HIRING Deactivating a link – a new type of intervention 54

WHY

TE

DE

IE Z TE

DE

IE

(

rev

)

m

1

m

2

X

In linear systems

TE

 

DE

 

IE

IE

(

rev

) 

m

1

m

2  

IE

m

1

m

2 

TE

DE Y

TE

TE - DE

IE

 Effect sustained by mediation alone

DE

TE

DE

 Effect prevented by disabling mediation Disabling mediation

IE

Disabling direct path 55

MEDIATION FORMULAS

1. The natural direct and indirect effects are identifiable in Markovian models (no confounding), 2. And are given by:

DE

IE

  [

E

(

Y z

E

(

Y z

| |

do

(

x

1 ,

z

)) 

E

(

Y do

(

x

0 ,

z

))[

P

(

z

| |

do

(

x

0 ,

z

))]

P

(

z do

(

x

1 )) 

P

(

z

| |

do

(

x

0 )).

do

(

x

0 ))] 3. Applicable to linear and non-linear models, continuous and discrete variables, regardless of distributional form.

56

MEDIATION FORMULAS IN UNCONFOUNDED MODELS

Z DE IE TE

 

z

[

E

(

Y

|

X x

1 ,

z

) 

E

(

Y Y

|

x

0 ,

z

)]

P

(

z

|

x

0 )  

z

[

E

(

Y

|

x

0 ,

z

)[

P

(

z

|

x

1 ) 

P

(

z

|

x

0 ) ] 

E

(

Y

|

x

1 ) 

E

(

Y

|

x

0 ) IE = TE – DE = Fraction of responses explained by Fraction of responses owed to mediation mediation 57

MODERATING MEDIATORS

X

m

1 

Z

c

12

m

2

Y z y

 

m

0

b

0  

m

1

x

x

m

2

z

c

12 What portion of the effect can be explained-by/attributed-to mediation?

DE

  

m

0

c

12 The Mediation Formula yields:

IE TE

 

m

1

m

2

DE

IE

m

1

c

12 Effect Explained: Effect Attributed:

IE TE

DE

 

m

1

m

2

m

1

m

2 

m

1

c

12 58

X n

1

n

2

n

3

n

4

n

5

n

6

n

7

n

8

Z Y

MEDIATION FORMULA FOR BINARY VARIABLES

Y

0 1 0 1 0 1 0 1 59

RAMIFICATION OF THE MEDIATION FORMULA

DE

should be averaged over mediator levels,

IE

should NOT be averaged over exposure levels.

TE-DE

need not equal

IE TE-DE

= proportion for whom mediation is necessary

IE

= proportion for whom mediation is sufficient •

TE-DE

informs interventions on indirect pathways

IE

informs intervention on direct pathways.

60

CONCLUSIONS

• Formal basis for causal and counterfactual

on an explicit causal structure that is defensible on scientific grounds.

and structural equation approaches • Friendly and formal solutions to century-old problems and confusions.

• No other method can do better (theorem) 61

QUESTIONS???

They will be answered

62