I was recently tasked with estimating the reliability of scores composed of items on different scales. The lack of a common scale could be seen in widely varying item score means and variances. As discussed by Graham, it's not widely known in applied educational research that Cronbach's alpha assumes essential tau-equivalence and underestimates reliability when the assumption isn't met. A friend made a good suggestion to consider stratified alpha, which is commonly used to estimate internal consistency when scores are composed of multiple choice items (scored 0-1) and constructed response items (e.g., scored 0-4). However, the assessment with which I was concerned does not have clear item-type strata. I decided to estimate congeneric reliability because it makes few assumptions (e.g., unidimensionality) and doesn't require grouping items into essentially tau-equivalent strata.

With the help of Graham's paper and a LISREL tutorial by Raykov I wrote an program that estimates congeneric reliability. The program uses Fox's sem() library to conduct the confirmatory factor analysis with minimal constraints (variance of common true score fixed at 1 for identifiability). The estimated loadings and error variances are then summed to calculate reliability (i.e., the ratio of true score variance to observed score variance) as:

.

I obtained a congeneric reliability estimate of 0.80 and an internal consistency estimate of 0.58 for the scores I was analyzing. If the items had been essentially tau-equivalent, then the reliability estimates would have been the same. If I had assumed tau-equivalence, than I would have underestimated the reliability of the total scores (and overestimated standard errors of measurement). The example below replicates Graham's heuristic example.

> ############################################ > #Replicate results from Graham (2006) to check congeneric reliability code/calculations. > > library(sem) > library(psych) > library(stringr) > > #Variance/covariance matrix > S.graham <- readMoments(diag = T, names = paste("x", 1:7, sep = "")) 1: 4.98 2: 4.60 5.59 4: 4.45 4.42 6.30 7: 3.84 3.81 3.66 6.44 11: 5.71 5.67 5.52 4.91 11.86 16: 23.85 23.68 22.92 19.87 34.28 127.65 22: 46.53 46.20 44.67 38.57 62.30 244.36 471.95 29: Read 28 items > > ############################################ > #A function to estimate and compare congeneric reliability and internal consistency > funk.congeneric <- function(cfa.out) { + names.loadings <- str_detect(names(cfa.out$coeff), "loading") + names.errors <- str_detect(names(cfa.out$coeff), "error") + r.congeneric <- sum(cfa.out$coeff[names.loadings]) ^ 2 / + (sum(cfa.out$coeff[names.loadings]) ^ 2 + sum(cfa.out$coeff[names.errors])) + round(c("Congeneric" = r.congeneric, "Alpha" = alpha(cfa.out$S)$total$raw_alpha), 2) + } > > ############################################ > #Congeneric model; tau-equivalent items > model.graham <- specifyModel() 1: T -> x1, loading1 2: T -> x2, loading2 3: T -> x3, loading3 4: T -> x4, loading4 5: T -> x5, loading5 6: x1 <-> x1, error1 7: x2 <-> x2, error2 8: x3 <-> x3, error3 9: x4 <-> x4, error4 10: x5 <-> x5, error5 11: T <-> T, NA, 1 12: Read 11 records > cfa.out <- sem(model = model.graham, S = S.graham, N = 60) > summary(cfa.out) Model Chisquare = 0.11781 Df = 5 Pr(>Chisq) = 0.99976 Chisquare (null model) = 232.13 Df = 10 Goodness-of-fit index = 0.9992 Adjusted goodness-of-fit index = 0.99761 RMSEA index = 0 90% CI: (NA, NA) Bentler-Bonnett NFI = 0.99949 Tucker-Lewis NNFI = 1.044 Bentler CFI = 1 SRMR = 0.0049092 AIC = 20.118 AICc = 4.6076 BIC = 41.061 CAIC = -25.354 Normalized Residuals Min. 1st Qu. Median Mean 3rd Qu. Max. -0.038700 -0.008860 -0.000002 0.004700 0.002450 0.119000 R-square for Endogenous Variables x1 x2 x3 x4 x5 0.9286 0.8175 0.6790 0.4962 0.5968 Parameter Estimates Estimate Std Error z value Pr(>|z|) loading1 2.15045 0.21595 9.9580 2.3263e-23 x1 <--- T loading2 2.13776 0.24000 8.9073 5.2277e-19 x2 <--- T loading3 2.06828 0.26941 7.6770 1.6281e-14 x3 <--- T loading4 1.78754 0.29136 6.1352 8.5040e-10 x4 <--- T loading5 2.66040 0.38141 6.9752 3.0535e-12 x5 <--- T error1 0.35559 0.17469 2.0356 4.1793e-02 x1 <--> x1 error2 1.02000 0.25339 4.0255 5.6861e-05 x2 <--> x2 error3 2.02222 0.41688 4.8509 1.2293e-06 x3 <--> x3 error4 3.24471 0.62679 5.1767 2.2583e-07 x4 <--> x4 error5 4.78227 0.94911 5.0387 4.6871e-07 x5 <--> x5 Iterations = 21 > pathDiagram(cfa.out, edge.labels = "values", ignore.double = F, rank.direction = "TB")

> funk.congeneric(cfa.out) Congeneric Alpha 0.91 0.91 > > ############################################ > #Congeneric model; tau-inequivalent items > model.graham <- specifyModel() 1: T -> x1, loading1 2: T -> x2, loading2 3: T -> x3, loading3 4: T -> x4, loading4 5: T -> x7, loading7 6: x1 <-> x1, error1 7: x2 <-> x2, error2 8: x3 <-> x3, error3 9: x4 <-> x4, error4 10: x7 <-> x7, error7 11: T <-> T, NA, 1 12: Read 11 records > cfa.out <- sem(model = model.graham, S = S.graham, N = 60) > summary(cfa.out) Model Chisquare = 0.0072298 Df = 5 Pr(>Chisq) = 1 Chisquare (null model) = 353.42 Df = 10 Goodness-of-fit index = 0.99995 Adjusted goodness-of-fit index = 0.99985 RMSEA index = 0 90% CI: (NA, NA) Bentler-Bonnett NFI = 0.99998 Tucker-Lewis NNFI = 1.0291 Bentler CFI = 1 SRMR = 0.0010915 AIC = 20.007 AICc = 4.497 BIC = 40.951 CAIC = -25.464 Normalized Residuals Min. 1st Qu. Median Mean 3rd Qu. Max. -2.76e-02 -1.27e-04 -1.00e-07 -1.92e-03 1.96e-04 3.70e-03 R-square for Endogenous Variables x1 x2 x3 x4 x7 0.9303 0.8171 0.6778 0.4942 0.9902 Parameter Estimates Estimate Std Error z value Pr(>|z|) loading1 2.15247 0.212984 10.10624 5.1835e-24 x1 <--- T loading2 2.13719 0.237279 9.00711 2.1156e-19 x2 <--- T loading3 2.06646 0.266419 7.75646 8.7336e-15 x3 <--- T loading4 1.78392 0.287599 6.20279 5.5470e-10 x4 <--- T loading7 21.61720 2.015041 10.72792 7.5272e-27 x7 <--- T error1 0.34688 0.090227 3.84451 1.2080e-04 x1 <--> x1 error2 1.02240 0.200635 5.09584 3.4720e-07 x2 <--> x2 error3 2.02973 0.382466 5.30694 1.1148e-07 x3 <--> x3 error4 3.25764 0.605376 5.38118 7.3998e-08 x4 <--> x4 error7 4.64661 6.387765 0.72742 4.6697e-01 x7 <--> x7 Iterations = 38 > pathDiagram(cfa.out, edge.labels = "values", ignore.double = F, rank.direction = "TB")

> funk.congeneric(cfa.out) Congeneric Alpha 0.99 0.56