* with additional sources
Multivariate Analysis
Multivariate analysis (MVA) is classified a second generation statistical technique\(^*\) which extends from first generation statistical techniques\(^*\).Structural Equation Modeling (SEM)
SEM is a MVA method that models causal complex structures. The causal structure represents and investigate how multiple, observable, and measurable factors of a latent theoretical phenomenon are related to each other and their connection to the underlying theoretical phenomenon.
The objectives of causal analysis is (1) provide guidance on actions and treatment on causes (“variables that can be changed or manipulated”) and (2) provide “scientific explanation”.
With these objectives, SEM is a method that can be used to understand how to engage and intervene on complex mechanisms like social determinant of health (SDOH).
About SEM method:
SEM model:
Considerations of SEM:
what is involved to build appropriate model:
define covariation, correlation, collinearity
Principles of SEM:
Criticisms of SEM methodology
Piotr Tarka’s response to criticisms “if the hypothesized model is only empirically determined, at minimal theory input, then such a model is simply a convenient summary of a covariance matrix. The interpretation of this model’s equations express only the statistical relationships between variables in reference to the analyzed set of data. If, however, the tested SEM model is theoretically determined by a strong theory, then it receives the status of representing stable causal relationships in the empirical process of data processing”
Developments of SEM
Considerations of PLS-SEM:
Data Characteristics | Model Characteristics | Model Estimation | Model Evaluation |
---|---|---|---|
sample size | number of construct measures higher order construct estimation allowed |
objective of estimates | overall model |
distribution non-normal |
relationships between constructs and measures | algorithm efficiency | structural model |
scale of measurements | Model complexity and setup | definition of constructs | measurement model |
missing values | parameter estimates | construct scores calculation and use | |
Data Characteristics
Model Characteristics
Comparing Methods
create sidenote definingt composite based and common factor
PLS SEM | CB SEM |
---|---|
composite based | common factor |
combines indicators based on linear method | total variance divided into common shared part, unique variance of individual indicator, and error |
forms composite variables | only uses common shared variance to estimate construct |
accounts for measurement errors | estimates the commonality of sets indicators and its covariations |
using ordinary least squares to minimize error of constructs | states that indicators and covariations are manifestation of underlying construct |
composite variable is a valid proxie of constructs | explains covariations between associated indicators |
comprehensive representation of constructs | direct and precise measurement of constructs |
best for predicting and explaining target constructs | best for confirming theory manifestation |
PLS SEM | Sum scores | PLS regression |
---|---|---|
composite variables give individual weights to indicators | form composite variable but all indicators are weighted equally | linear relationships between multiple independent variables and dependent variable(s) |
weights inform importance of indicator in composite | equalize differences in indicator | no structural model |
must have structural and measurement models | simplification of PLS SEM |
PLS-SEM application procedures
Path Models
NEED TO ADD VISUAL OF PATH MODEL
Structural Model
Measurement Model
Data Exploration
INSERT NOTES FROM COMMON BELIEFS AND REALITY ABOUT PLS, PARTIAL LEAST SQUARES STRUCTURAUAL EQUALTION MODELING WITH R, REFELCATIONS ON PARTIAL LEAST SQUARES PATH MODELING
PLS-SEM objective
PLS-SEM algorithm
Algorithmic Process:
Stage 1:
during each iteration, weights are scaled to obtain unit variance for the LVS over the N cases in the sample size
initial:
inner approximation of weights: choice between 3 primary weighting schemes
weighting schemes specify how focal LV’s neighboring LVs are combined in order to estimate focal LVS
centroid weighting scheme:
factor weighting scheme:
path weighting scheme:
outside approximation - using inside estimation of LVS, outside weights are calculated for the LVS estimation
weight calculation dependent on indicator type (reflective or formative)
Mode A: reflective indicator weights
Mode B: formative indicator weights
Stage 2:
Stage 3:
Algorithmic Process Equations:
Inner model:
\(\eta = \beta_{0} + \beta\eta + \Gamma\xi+\zeta\)
\(\eta_{j} = \displaystyle\sum_{i}\beta_{ji}eta_{i} + \displaystyle\sum_{h} \gamma_{jh}\xi_{h} + \zeta_{j}\)
\(\beta_{ji}\) & \(\Gamma_{jh}\) - path coefficient the link predictors
inner model is subject to predictor specification \(E(\eta_{j}| \forall\eta_{i}, \xi_{h}) = \displaystyle\sum_{i}\beta_{ji}\eta_{i}+\displaystyle\sum_{h} \gamma_{jh}\xi_{h}\)
structural form
need path model graphic to correspond with weighting schemes
factor weighting scheme: \(\xi_{1} = \rho_{110}\cdot\eta_{1,out} + \rho_{120}\cdot\eta_{2,out}\) \(\eta_{1} = \rho_{110}\cdot\xi_{1,out} + \rho_{210}\cdot\xi_{2,out}+\rho_{012}\cdot\eta_{2,out}\)
path weighting scheme: \(\xi_{1} = \rho_{110}\cdot\eta_{1,out}+\rho_{120}\cdot\eta_{2,out}\) \(\xi_{2} = \rho_{210}\cdot\eta_{1,out}+\rho_{220}\cdot\eta_{2,out}\) \(\eta_{1} = \beta_{1}\cdot\xi_{1,out}+\beta_{2}\cdot\xi_2,out+\rho_{012}\cdot\eta_{2,out}\) \(\eta_{2} = \beta_{1}\cdot\xi_{1,out}+\beta_{2}\cdot\xi_2,out+\beta_{3}\cdot\eta_{1,out}\)
Outer model:
Mode A - reflective indicators
\(x =
\Lambda_{x}\xi_+\varepsilon_{x}\) \(y =
\Lambda_{y}\eta_+\varepsilon_{y}\)
\(x\) & \(y\) - indicators
\(\Lambda_{x}\) & \(\Lambda_{y}\) - simple regression coefficient (loading matrices)
\(\varepsilon_{x}\) & \(\varepsilon_{y}\) - residuals (measurement error/ noise)
predictor specification \(E(x|\xi) = \Lambda_{x}\xi\) \(E(y|\eta) = \Lambda_{y}\eta\)
Mode B - formative indicators
\(\xi =
\Pi_{\xi}\cdot x + \delta_\xi\) \(\eta
= \Pi_{\eta}\cdot y + \delta_\eta\)
\(x\) & \(y\) - indicators
\(\Pi_{\xi}\) & \(\Pi_{\eta}\) - multiple regression coefficients
\(\delta_{\xi}\) & \(\delta_{\eta}\) - residual from regression
predictor specification \(E(\xi|x)= \Pi_{\xi}\cdot x\) \(E(\eta|y)= \Pi_{\eta}\cdot y\)
formative specification - represent outer relationship between indicators and true LV
Weights relation: how LV estimates are formed
\(\hat\xi_{h}=\displaystyle\sum_{kh}w_{kh}x_{kh}\) \(\hat\eta_{i}=\displaystyle\sum_{ki}w_{ki}x_{ki}\)
Statistical properties
Algorithm Options and Parameter Setting