an introduction to basic multilevel modeling

30
An introduction to basic multilevel modeling

Upload: kalia-briggs

Post on 02-Jan-2016

59 views

Category:

Documents


4 download

DESCRIPTION

An introduction to basic multilevel modeling. Two introductory points:. 1. Hierarchical cross-national data structures are common within the field of sociology and political sciences, and ESS is a good example of such structures. - PowerPoint PPT Presentation

TRANSCRIPT

An introduction to basic multilevel modeling

Two introductory points:

1. Hierarchical cross-national data structures are common within the field of sociology and political sciences, and ESS is a good example of such structures. Hierarchical cross-national data structures mean that we have information on the level of individuals and on the level of countries. The problem is how to explain one individual-level characteristic (variable) using both (a) other individual-level characteristics (variables), and (b) country-level characteristics (variables).

.2. The old solution is to pool data from all countries and use country-level characteristics assigned to individuals as all other individual characteristics. There are three consequences of this (cross-sectional pooling) strategy:

(a) For effects of individual-level variables we increase N so that these effects become significant while in the reality they might be not significant.

(b) For effects of country-level variables we also increase N so that these effects become significant while in the reality they might be not significant.

[Both (a) and (b) lead to increasing the probability of “false positive” = Error Type I, the error of rejecting a "correct" null hypothesis].

(c) Case dependency. Units of analysis are not independent – what is required by regular regression.

.

Ignoring the statistical significance, for effects of individual-level variables, a pooling strategy can only be successful if countries are homogeneous. The extent that this condition really is met is indicated by the intra-class correlation. Usually this condition is poorly met.

.

The basic two-level regression model with a continuous outcome

An example taken from Stefan Dahlberg, Department of Political Science, Goteborg University (2007)

We have opinion survey data from J countries with a number of respondents, i.e. voters i, ni, in each country j. On the individual level we have an outcome variable (Y) measured as the absolute distance between a voters self-placement on a left-right scale and the position, as ascribed by the voter, for the party voted for on the same scale. This

dependent variable is called “proximity.”

.

Two-levels

We have two independent variables: X - education, and Z - the degree of proportionality of the electoral system.

We assume that more educated people tend to vote consistently (high proximity). More proportional systems are expected to induce proximity voting since voters tend to view the elections as an expression of preferences rather than a process of selecting government.

We have three variables on two different levels:

Y proximity, X education (individual level)

Z proportionality (country-level).

.

Eq. 1

The simplest way to analyze the data is to make separate regressions in each country such as:

(1.1.) Yij = α0j + β1jXij + eij

The difference compared to a more ordinary regression model is that we assume that intercepts and slopes are varying between electoral systems, i.e. each country has its own intercept and slope, denoted by α0j and β1j.

.

Interpretation

The error term, eij is expected to have a mean of zero and a variance to be estimated. Since the intercept and slope coefficients are assumed to vary across countries they are often referred to as random coefficients.

Across all countries the regression coefficients βj have a

distribution with some mean and variance.

.

Eq. 2-3

The next step in the hierarchical regression analysis is to introduce the country-level explanatory variable proportionality in an attempt to explain the variation of the regression coefficients α0j and β1j.

(1.2) α0j = γ00 + γ01Zj + u0j

and

(1.3) β1j = γ10 + γ11Zj + u1j

.

Interpretation

Equation 1.2 predicts the average degree of proximity voting in a country (the intercept α0j ) by the degree of proportionality in the electoral system (Z).

Equation 1.3 denotes that the relationship, as expressed by the coefficient β1j between proximity voting (Y) and level of education (X), is depending on the degree of proportionality within the electoral system (Z).

The degree of proportionality is here functioning as an interactive variable where the relationship between proximity voting and education varies according to the values on the second level variable proportionality.

.

The u-terms u0j and u1j in equation 1.2 and 1.3 are random residual error terms at the country level.

The residual errors uj are assumed to have a mean of zero and to be independent from the residual errors eij at the individual level.

The variance of the residual errors u0j is specified as σ2u0 and the variance of the residual errors u1j is specified as σ2u1.

.

In equation 1.2 and 1.3 the regression coefficients γ are not assumed to vary across countries since they have the denotation j in order to indicate to which country they belong. As they apply to all countries the coefficients are referred to as fixed effects.

All the between-country variation left in the β coefficients, after predicting these with the country variable proportionality (Zj), is then assumed to be the residual error variation. This is captured by the error term uj, which are denoted with j to indicate which country it belongs to.

.

Eq. 4

The equations 1.2 and 1.3 can then be substituted into equation 1.1 and by rearranging them we have:

(1.4) Yij = γ00 + γ10Xij + γ01Zj + γ11XijZj + u1jXij + u0j + eij

.

Interpretation

In terms of variable labels the equation 1.4 states:

proximityij = γ00

+ γ10 educationij

+ γ01 proportionalityj

+ γ11 educationij * proportionalityj

+ u1j educationij + u0j + eij

.

The first part of equation 1.4:

γ00 + γ10Xij + γ01Zj + γ11XijZj

is the fixed coefficients/effects.

The term XijZj is an interaction term that is included in the model as a consequence of the modeling of the varying regression slope β1j on individual level variable Xij with the country level Zj.

.

The last part of the equation 1.4

u1jXij + u0j + eij

is a part of the random coefficients.

The fact that the explanatory variable Xij is multiplied with the error term, u1jXij, the resulting total error will be different for different values of Xij.

.

The error terms are heteroscedastic instead of homoscedastic as assumed in ordinary regression models where the residual errors are assumed to be independent of the values of the explanatory variable.

[Random variables are heteroscedastic if they have different variances for the relevant subgrups. The complementary concept is called homoscedasticity. Note: The alternative spelling homo- or heteroskedasticity is equally correct and is also used frequently. The term means "differing variance" and comes from the Greek "hetero" (different) and "skedasis" (dispersion).]

.

Dealing with the problem of heteroscedasticy is one of the main reasons for why multilevel models are preferable over regular OLS models when analyzing hierarchical nested data used.

SUMMING UP:

Known in literature under a variety of names

Hierarchical linear model (HLM) Random coefficient model Variance component model Multilevel model Contextual analysis Mixed linear model Mixed effects model

Acknowldegements: Materials in the following are based on Prof. Joop J. Hox introductory lectures on HLM delivered at the 2011 QMSS2 Summer School in Leuven

Multi-level data structure

Groups at different levels may have different sizes

Response (outcome) variable at lowest level

Explanatory variables at all levels

The statistical model assumes sampling at all levels

Ex: education family longitudinal

level 3 schools classes

level 2 classes families

level 1 pupils members occasions (waves)

Problems with Hierarchical Data Structure I

Assume DV (response) variable on lowest level (level 1). We want to predict this, using explanatory variables at all available levels- How?- What is the proper sample size?- What if relations are not the same in different groups?

Traditional approaches:

Disaggregate all variables to the lowest levelDo standard analyses (anova, multiple regression)

Aggregate all variables to the highest levelDo standard analyses (anova, multiple regression)

Ancova with groups as factor

Some improvements:explanatory variables as deviations from their group meanhave both deviation score & disaggregated group mean as predictor (separates individual and group effects)

What is wrong with this?

Problems with Hierarchical Data Structure II

Multiple Regression assumes:

independent observations independent error terms equal variances of errors for all observations (assumption of

homoscedastic errors) normal distribution for errors

With hierarchical data observations are not independent errors are not independent different observations may have errors with different variances

(heteroscedastic errors)

Problems with Hierarchical Data Structure III

Observations in the same group are generally not independent

- they tend to be more similar than observations from different

groups

- selection, shared history, contextual group effects

The degree of similarity is indicated by the intraclass correlation rho: r

Kish: roh (rate of homogeneity)

Standard statistical tests are not at all robust against violation of the independence assumption

Graphical Pictureof 2-level regression Model

School level

Student level

Outcome variable on pupil level

- Explanatory variables at both levels: individual & group

- Residual error at individual level

- Plus residual error at school level

School size

Student gender

Grade error

error

Graphical Pictureof 2-level regression Model

School level

Student level

Essential points:

Explanatory variables characterize individuals and/or groups

Average value of individual variables may differ across groups

= most variables have both within-group and between-group variation

Groups do make a difference: the effect of an individual explanatory variable may be different in different groups

School size

Student gender

Grade error

error

Assumptions

Yij= [g00+ g10Xij + g01Zj + g11ZjXij] + [u1jXij+ u0j+ eij]

Individual level errors eij independent; normal distributionwith mean zero and same variance se² in all groups Group level errors u.j independent; multivariate normaldistribution with means zero and (co)variances su²

in covariance matrix W (omega)

Group level errors u.j independent from individual errors eij- plus usual assumptions of multiple regression analysis- linear relations, explanatory variables measured

without error

Types of HLM Models

Intercept-only Model

Intercept only model (null model, baseline model)Contains only intercept and corresponding error termsAt the lowest (individual) level we haveYij= b0j+ eijand at the second levelb0j= g00+ u0j

henceYij= g00+ u0j+ eij

Used to decompose the total variance and to compute the intraclass correlation r (rho)

r = (group level variance / total variance)r = expected correlation between 2 individuals in same group

Fixed Model:

Only fixed effects for level-1 explanatory variables:

slopes are assumed not to vary across groups

At the lowest (individual) level we have

Yij = b0j+ b1jXij+ eij

and at the second level

b0j = g00+ u0j and b1j= g10

hence Yij= g00+ g10Xij+ u0j+ eij

Intercepts vary across groups, slopes are the same

Similar to ANCOVA with random grouping factor

also called variance component model

Fixed effects for level-2 explanatory variables added next

Random Coefficient Model

Assumes intercept and slopes vary across groups

Yij= b0j+ b1jXij+ eij

and at the second level

b0j= g00+ u0j and b1j= g10+ u1j

hence Yij= g00+ g10Xij+ u1jXij+ u0j+ eij

The full multilevel model adds interactions to model the slope variation

Full Multilevel Regression Model

Explanatory variables at all levels

Higher level variables predict variation of lowest level

intercept and slopes

At the lowest (individual) level we have

Yij= b0j+ b1jXij+ eij

and at the second level

b0j= g00+ g01Zj+ u0j and b1j= g10+ g11Zj+ u1j

hence

Yij= g00+ g10Xij+ g01Zj+ g11ZjXij+ u1jXij+ u0j+ eij

Predicting the intercept implies a direct effect

Predicting slopes implies cross-level interactions