Learn R Programming

semTools (version 0.5-2)

reliability: Calculate reliability values of factors

Description

Calculate reliability values of factors by coefficient omega

Usage

reliability(object, omit.imps = c("no.conv", "no.se"))

Arguments

object

A '>lavaan or '>lavaan.mi object, expected to contain only exogenous common factors (i.e., a CFA model).

omit.imps

character vector specifying criteria for omitting imputations from pooled results. Can include any of c("no.conv", "no.se", "no.npd"), the first 2 of which are the default setting, which excludes any imputations that did not converge or for which standard errors could not be computed. The last option ("no.npd") would exclude any imputations which yielded a nonpositive definite covariance matrix for observed or latent variables, which would include any "improper solutions" such as Heywood cases. NPD solutions are not excluded by default because they are likely to occur due to sampling error, especially in small samples. However, gross model misspecification could also cause NPD solutions, users can compare pooled results with and without this setting as a sensitivity analysis to see whether some imputations warrant further investigation.

Value

Reliability values (coefficient alpha, coefficients omega, average variance extracted) of each factor in each group

Details

The coefficient alpha (Cronbach, 1951) can be calculated by

$$ \alpha = \frac{k}{k - 1}\left[ 1 - \frac{\sum^{k}_{i = 1} \sigma_{ii}}{\sum^{k}_{i = 1} \sigma_{ii} + 2\sum_{i < j} \sigma_{ij}} \right],$$

where \(k\) is the number of items in a factor, \(\sigma_{ii}\) is the item i observed variances, \(\sigma_{ij}\) is the observed covariance of items i and j.

The coefficient omega (Bollen, 1980; see also Raykov, 2001) can be calculated by

$$ \omega_1 =\frac{\left( \sum^{k}_{i = 1} \lambda_i \right)^{2} Var\left( \psi \right)}{\left( \sum^{k}_{i = 1} \lambda_i \right)^{2} Var\left( \psi \right) + \sum^{k}_{i = 1} \theta_{ii} + 2\sum_{i < j} \theta_{ij} }, $$

where \(\lambda_i\) is the factor loading of item i, \(\psi\) is the factor variance, \(\theta_{ii}\) is the variance of measurement errors of item i, and \(\theta_{ij}\) is the covariance of measurement errors from item i and j.

The second coefficient omega (Bentler, 1972, 2009) can be calculated by

$$ \omega_2 = \frac{\left( \sum^{k}_{i = 1} \lambda_i \right)^{2} Var\left( \psi \right)}{\bold{1}^\prime \hat{\Sigma} \bold{1}}, $$

where \(\hat{\Sigma}\) is the model-implied covariance matrix, and \(\bold{1}\) is the \(k\)-dimensional vector of 1. The first and the second coefficients omega will have the same value when the model has simple structure, but different values when there are (for example) cross-loadings or method factors. The first coefficient omega can be viewed as the reliability controlling for the other factors (like \(\eta^2_partial\) in ANOVA). The second coefficient omega can be viewed as the unconditional reliability (like \(\eta^2\) in ANOVA).

The third coefficient omega (McDonald, 1999), which is sometimes referred to hierarchical omega, can be calculated by

$$ \omega_3 =\frac{\left( \sum^{k}_{i = 1} \lambda_i \right)^{2} Var\left( \psi \right)}{\bold{1}^\prime \Sigma \bold{1}}, $$

where \(\Sigma\) is the observed covariance matrix. If the model fits the data well, the third coefficient omega will be similar to the \(\omega_2\). Note that if there is a directional effect in the model, all coefficients omega will use the total factor variances, which is calculated by lavInspect(object, "cov.lv").

In conclusion, \(\omega_1\), \(\omega_2\), and \(\omega_3\) are different in the denominator. The denominator of the first formula assumes that a model is congeneric factor model where measurement errors are not correlated. The second formula accounts for correlated measurement errors. However, these two formulas assume that the model-implied covariance matrix explains item relationships perfectly. The residuals are subject to sampling error. The third formula use observed covariance matrix instead of model-implied covariance matrix to calculate the observed total variance. This formula is the most conservative method in calculating coefficient omega.

The average variance extracted (AVE) can be calculated by

$$ AVE = \frac{\bold{1}^\prime \textrm{diag}\left(\Lambda\Psi\Lambda^\prime\right)\bold{1}}{\bold{1}^\prime \textrm{diag}\left(\hat{\Sigma}\right) \bold{1}}, $$

Note that this formula is modified from Fornell & Larcker (1981) in the case that factor variances are not 1. The proposed formula from Fornell & Larcker (1981) assumes that the factor variances are 1. Note that AVE will not be provided for factors consisting of items with dual loadings. AVE is the property of items but not the property of factors.

Regarding categorical indicators, coefficient alpha and AVE are calculated based on polychoric correlations. The coefficient alpha from this function may be not the same as the standard alpha calculation for categorical items. Researchers may check the alpha function in the psych package for the standard coefficient alpha calculation.

Item thresholds are not accounted for. Coefficient omega for categorical items, however, is calculated by accounting for both item covariances and item thresholds using Green and Yang's (2009, formula 21) approach. Three types of coefficient omega indicate different methods to calculate item total variances. The original formula from Green and Yang is equivalent to \(\omega_3\) in this function. Green and Yang did not propose a method for calculating reliability with a mixture of categorical and continuous indicators, and we are currently unaware of an appropriate method. Therefore, when reliability detects both categorical and continuous indicators in the model, an error is returned. If the categorical indicators load on a different factor(s) than continuous indicators, then reliability can be calculated separately for those scales by fitting separate models and submitting each to the reliability function.

References

Bollen, K. A. (1980). Issues in the comparative measurement of political democracy. American Sociological Review, 45(3), 370--390. doi:10.2307/2095172

Bentler, P. M. (1972). A lower-bound method for the dimension-free measurement of internal consistency. Social Science Research, 1(4), 343--357. doi:10.1016/0049-089X(72)90082-8

Bentler, P. M. (2009). Alpha, dimension-free, and model-based internal consistency reliability. Psychometrika, 74(1), 137--143. doi:10.1007/s11336-008-9100-1

Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, 16(3), 297--334. doi:10.1007/BF02310555

Fornell, C., & Larcker, D. F. (1981). Evaluating structural equation models with unobservable variables and measurement errors. Journal of Marketing Research, 18(1), 39--50. doi:10.2307/3151312

Green, S. B., & Yang, Y. (2009). Reliability of summed item scores using structural equation modeling: An alternative to coefficient alpha. Psychometrika, 74(1), 155--167. doi:10.1007/s11336-008-9099-3

McDonald, R. P. (1999). Test theory: A unified treatment. Mahwah, NJ: Erlbaum.

Raykov, T. (2001). Estimation of congeneric scale reliability using covariance structure analysis with nonlinear constraints British Journal of Mathematical and Statistical Psychology, 54(2), 315--323. doi:10.1348/000711001159582

See Also

reliabilityL2 for reliability value of a desired second-order factor, maximalRelia for the maximal reliability of weighted composite

Examples

Run this code
# NOT RUN {
library(lavaan)

HS.model <- ' visual  =~ x1 + x2 + x3
              textual =~ x4 + x5 + x6
              speed   =~ x7 + x8 + x9 '

fit <- cfa(HS.model, data = HolzingerSwineford1939)
reliability(fit)

# }

Run the code above in your browser using DataLab