• Aucun résultat trouvé

Equivalence des mesures à travers différents groupes dans un sondage

N/A
N/A
Protected

Academic year: 2022

Partager "Equivalence des mesures à travers différents groupes dans un sondage"

Copied!
36
0
0

Texte intégral

(1)

Equivalence des mesures à travers différents groupes dans un sondage

Melanie Revilla RECSM, UPF

(2)

Measurement equivalence

“Measurement equivalence is conceptually defined as whether or not, under different conditions of observing and studying phenomena, measurement operations yield measures of the same attribute.”

Horn and McArdle, 1992: 117

“Measurement invariance or measurement equivalence is a statistical property of measurement that indicates that the same construct is being measured across some

specified groups.”

Wikipedia

(3)

Measurement equivalence

• Usually discussed in the frame of cross-national research (across countries and languages ):

– Issue of cultural differences and translation

• But equivalence can also be checked across:

– Data collection modes (e.g. face-to-face vs Web) – Data collection designs (unimode vs mixed-mode) – Socio-demographic groups (education levels)

– Time (longitudinal studies) – Etc

Idea: in different groups people can express

themselves in different ways

(4)

Measurement equivalence (ME)

• If there is measurement equivalence, 2 persons with the

same opinion will give the same answer, whatever their

group

(5)

Why does it matters?

• Classic assumption to be able to combine answers and compare them

– Equality of the response function

• Important because observed differences may result from:

– Non equivalent measures

– And/or from real substantive differences

• If measurement equivalence does not hold:

– cannot make direct comparison across groups

(6)

Important preliminary distinction

(7)

Different kinds of concepts

(Northrop, 1947; Blalock, 1990)

• Concepts-by-intuition (CI)

– simple concepts whose meaning is immediately obvious.

• Concepts-by-postulation (CP)

– less obvious concepts which require explicit definitions.

– also called constructs

• For CP, distinction between:

– concepts with reflective indicators – concepts with formative indicators

(8)

CP with formative indicators

• Concept defined by a combination of indicators

• These variables are not necessarily correlated

• The combination defines the CP

• Requirement: the definition should take into account all necessary components

CP

y1 y2

(9)

CP with reflective indicators

• CP is indicated by some similar possible indicators (consequences)

• These indicators will always be correlated because all are affected by the CP

• Usually 3 indicators are enough (identified)

CP

y1 y2

(10)

• Operationalized as a CI

How satisfied are you with your job?

Example of job satisfaction

Job satisfaction Verbal

report

e

(11)

• Operationalized as a CP with formative indicators

How satisfied are you with :

your salary?

promotion possibilities?

your contacts with colleagues

your spare time

...

• Should be exaustive!

Example of job satisfaction

Job satisfaction

Salary

Task

Hours

Spare time

Social contac ts

Verbal report

Verbal report

Verbal report

Verbal report

Verbal report

e1

e2

e3

e4

e5

(12)

• Operationalized as a CP with reflective indicators

How satisfied are you with your job?

Would you recommend your job to a friend?

Would you chose this job again?

The responses to all three questions are supposed to be affected by the satisfaction of the respondent with his/her job.

Job satisfaction

Satisfaction with job

Recommend to a friend

Choose again

Verbal report

Verbal report

Verbal report

e1

e2

e5 u1

u2

u3

Example of job satisfaction

(13)

Testing for measurement equivalence

• Classic procedure only exist to test for measurement equivalence in the case of CP with reflective

indicators

– We start with this case

(14)

Testing for measurement equivalence

CP / reflective indicators

(15)

Basic Confirmatory Factor Analysis model

CP1

Y1

Y2

Y3 λ11

λ21 λ31

e1 e2 e3 τ1

τ3 τ2

Yi = τi + λij CP1 + ei i = 1,2,3

Political trust

Answer Trust in the parliament

Answer Trust in the legal system

Answer Trust in the police

intercepts slopes error terms

≈ regression

equation

[

Dependent variable

Independent variable

(16)

Multiple Group CFA approach

Group1 Group 2

• Multiple group:

– possible to test for equality of the parameters in the different groups – constraints across groups

• Can be extended to more groups

(17)

Different levels of invariance (Meredith, 1993)

• C onfigural

– Same model holds in all groups

• M etric

– Configural holds + Slopes (λij) the same in all groups

– Sufficient for comparison of relationships

• Scalar

– Metric holds + Intercepts (τi) the same in all groups

– Sufficient for comparison of means

• More: error terms, etc…

Group 1

Group 2

?

=

(18)

How to do it in practice?

• Analyses can be done with standard SEM softwares

– LISREL/Mplus/R…

– Based on covariance matrices & means

– Recommended sample size: >200 in each group – 3-step procedure: configural, metric, scalar

– Syntax quite easy to get estimates

• More tricky but crucial step: testing

(19)

Testing the model

• Assessing global fit :

– Chi2 test

– Fit indices: RMSEA (<.05), CFI (>.9), etc…

• Limits:

– Depends on sample size

– Sensitive to deviations from normality – Sensitive to model complexity

• Saris, Satorra and van der Veld (2009)

– Show that we should test at the parameter level + take into account type II errors (H0 not rejected despite being false)

(20)

Testing the model

• Assessing local fit

– Use MI, EPC and Power

• JRule software

– Available for Lisrel (van der Veld, Saris, Satorra)and Mplus (Oberski) – For each parameter, it tells if it is misspecified

– /!\ The researcher should decide how large can a deviation be before considering it a misspecification

(21)

Testing the model

• Limits:

– If the model is large and/or you have many different groups:

possible to get many misspecifications

• With which one should you start?

• When should you stop?

• Always free parameters only one by one

• Always check if estimates are really different when you free a new parameter in your model

– Difference may be statistically significant but not substantially big enough to be meaningful

(22)

What if equivalence does not hold?

• Partial invariance

(Byrne, Shavelson, Muthén, 1989)

– If some indicators are equivalent but not all

– Consistent estimates of the means of the latent variables if at least 2 indicators are scalar invariant

• If you have many groups, often at least several are invariant:

– Report the results for these ones, and report the deviant cases separately

– Example of how to do it: Coromina, Saris and Oberski (2008)

(23)

Example

(Revilla, 2013)

• Comparison:

– ESS round 4 (F2f) / LISS panel (Web) – Same questions / Same fieldwork period – The Netherlands

• Configural, metric, scalar invariance achieved

• Possible compare relationships + means across LISS and ESS for that concept

Political trust

Trust parliament

Trust legal system

Trust police .79

1

[1.80]

[1.23]

.91

1.82 0 Mean: 5.68

[1.79]

.74

unstandardized Liss: [3.36]

ESS: [2.98]

(24)

Criticisms

(Saris & Gallhofer, 2007)

• Test of “measurement equivalence” too strict

– Need to separate cognitive and measurement processes – Cognitive process

• Understanding of the question

• Relationship between the higher order variables (CP) and the CI

• Consistency

– Measurement process

• Expressing the answer

• Relationship between the observed answers and the latent variables they attempt to measure

• Validity (+ reliability, method effects)

(25)

Proposition

(Saris & Gallhofer, 2007)

CP1

Y1

Y2

Y3 c1

c2 c3

e1 e2 e3 τ1

τ3 τ2 CI1

CI2

CI3

v11

v22

v33 u1

u2 u3

α1

α2

α3

• Measurement process: Yi = τi + vij CIi + ei i = 1,2,3 (Method effect could be included too: distinction true score-factor)

• Cognitive process: CIi = αi + ci CP1 + ui i = 1,2,3

(26)

CP1

Y1

Y2

Y3 c1

c2 c3

e1 e2 e3 τ1

τ3 τ2 CI1

CI2

CI3

v11

v22

v33 u1

u2 u3

α1

α2

α3

• Testing equivalence measurement process

• Testing equivalence for CI

Proposition

(Saris & Gallhofer, 2007)

(27)

Can we test for measurement

equivalence for CI or CP with formative

indicators?

(28)

Case of CI

Y1

e1 τ1 CI1 λ11

• Testing equivalence for CI

• Testing equivalence for single indicator

Yi = τi + λij CIi + ei

Y1

e1 τ1 CI1 λ11

Group 1 Group 2

(29)

Single  multiple indicators?

• Problem: model just presented not identified

• We need to get several indicators for the CI

– CI by definition can be asked using a single question

– But possible to use multiple formulations/methods/scales for this question

• In that way, we can come back to a CFA model with one

latent variable with several observed indicators

(30)

Can apply again MGCFA

Unimode / Mode 1 Mixed-mode / Mode 2

• Then, same procedure to get the estimates and test the model

Trust in the parliament

11 points

6 points

4 points

(31)

Alternative / solution for CP formative indicators

• Saris, Pirralha and Zavala (under review)

– Proposed to vary slightly the question itself

– For instance ask 3 direct questions about job satisfaction

• “How satisfied are you with your job?”

• “How much do you like your job?”

• ”How happy are you with your job?”

• For CP with formative indicators:

– Same procedure, we need to come back to something similar as for CP with reflective indicators

– For each of the formative indicators, we need to ask different questions/use different methods

(32)

More general model

CP1

Y31

Y22

Y13 c1

c2

c3

e11

e12

e33 τ11

τ33 τ12 CI1

CI2

CI3

v21

v22

v23 u1

u2

u3

α1

α2

α3

Y21

Y12

Y23 Y33 Y11

Y32

e21 τ21

e31 τ31

e23 τ23 e13 τ13 e22

τ22

e32 τ32 1

v31

v32

v33 1 1

(33)

Implication

• In order to separate the cognitive and the measurement process, we need to get several indicators for each CI

• Can be done by using different methods (e.g. different scales) for the same CI

– Same persons get the question several times using different methods

– But 20 minutes at least to avoid memory effects (Van Meurs &

Saris, 1990)

– We cannot do it for all questions of the questionnaire

(34)

In summary

• Measurement equivalence can be assessed using M

ultiple

G

roup

C

onfirmatory

F

actor

A

nalysis

• Classic procedure exists for CP with reflective indicators

– Testing is the most crucial step in this case – Very important to consider also local fit

– Limit: the classic procedure does not separate cognitive and measurement process

• For CI, for CP with formative indicators, and if we want to separate cognitive and measurement:

– We need to repeat questions

(35)

Thank you for your attention!

Contact: melanie.revilla@upf.edu

(36)

References

Blalock, H.M. (1990). Auxilary measurement theories revisited. In Hox J. J., and J. de Jong-Gierveld (eds.), Operationalization and Research Strategy. Amsterdam: Swets and Zeitlinger, 33─49.

Byrne, B.M., R.J. Shavelson, and B. Muthén (1989). “Testing for the equivalence of factor covariance and mean structures: the issue of partial measurement invariance.” Psychological Bulletin 105(3): 456-466.

Coromina, L., Saris, W.E. and D. Oberski (2008). The quality of the measurement of interest in the political issues presented in the media in the ESS. ASK, 2008, nr 17, ISSN 1234–9224. Available at: http://daob.nl/wp- content/uploads/2013/03/Coromina_Saris_Oberski_IPIMedia_2008.pdf

Horn, J.L., and J.J. McArdle (1992). A practical and theoretical guide to measurement invariance in aging research. Experimental Aging Research, 18(3):117-144.

Meredith, W. (1993). Measurement invariance, factor analysis and factorial invariance. Psychometrika, 58, 525-543.

Northrop, F.S.C. (1947). The Logic of the Sciences and the Humanities. New York: World Publishing Company.

Revilla, M. (2013). “Measurement invariance and quality of composite scores in a face-to-face and a web survey” Survey Research Methods 7(1): 17-28.

Saris, W.E., and I. Gallhofer (2007). Design, Evaluation, and Analysis of Questionnaires for Survey Research. New York: Wiley.

Saris, W.E., Pirralha, A., and D. Zavala Rojas (under review). Testing the comparability of different types of social indicators across groups.

Saris, W. E., Satorra, A., and W.M. Van der Veld (2009). Testing Structural Equation Models or Detection of Misspecifications?

Structural equation modeling: A multidisciplinary Journal, 16(4), 561-582.

Van Meurs, L., and W.E. Saris (1990). Memory Effects in MTMM Studies.’ Pp. 134-146 in Evaluation of Measurement Instruments by Meta-analysis of Multitrait-Multimethod Studies, edited by Willem E. Saris and Lex van Meurs. Amsterdam, the Netherlands:

North Holland.

Références

Documents relatifs

tique de pression, via une cellule de dégazage remplie d'eau désaérée, avec contrôle visuel continu au moyen d'un manomètre à cadran de 2,5 MPa de capacité ; - d'une pression

Brownian motion in ~d has the following property for d &amp;#x3E; 4: The random set W intersects sets of dimension greater than d-4 with positive probability. and

When forming their value proposition, which is needed for acquiring new customers, the company may focus one of the three main assets that are needed to run process instances:

Recall that these zeta functions are defined using an additive and multiplicative invariant (invariant means under Nash isomorphisms) of real algebraic sets, the virtual

Apart from the childless two-adult household, average equivalence scales in Cyprus are always higher than in Germany and France which means that income needs of children are higher

In particular, a framing effect of the original survey questionnaire coming from the possibility that respondents feel sympathetic towards the poor and they are framed by a tendency

First, we obtain a criterion similar to Theorem 1.1 for the Witt group of a quadratic field extension equipped with its nontrivial automorphism:..

In this paper, we make the following contributions: 1) we formalize the concept of equivalence between two learning algorithms and define two notions of algorithmic equivalence,