This tutorial assumes a fairly good working knowledge of generalizability theory and the terms used in its analysis. If the terms “facet”, “variance component”, “generalizability”, “dependability”, “G-Study” and “D-study” are unfamiliar, it is strongly advised that the reader review an introduction to generalizability procedures using balanced/non-missing data. There are many choices for this such as Briesch, Swaminathan, Welsh, and Chafouleas (2014).
While data sets with balanced facets and with no missing data points are relatively straightforward to analyze for generalizability, the same is not true for data sets not meeting these criteria. In fact many data sets (including those used in machine learning) that researchers wish to subject to generalizability analyses are either: 1) unbalanced (have different numbers of cases in each combination/nesting of variables); and/or 2) are missing data. There have been few freely available options to analyze such data – exceptions include the urGENOVA (Brennan, 2001b) and G-String_V (Bloch & Norman, 2012; 2023) programs. The urGENOVA program allows for both unbalanced data sets and missing data but has an esoteric user interface and does not calculate generalizability coefficients as part of the output. The G-String_V program has an intuitive user interface and produces generalizability and decision-study outputs but does not allow for missing data.
Generalizability analyses are dependent on variance components for their calculation. In balanced designs with no missing data, these are readily calculated using the Sums of Squares/Mean Squares generated in typical ANOVA analyses. When these criteria are not met, another approach (“analogous ANOVA”) has been used (Brennan, 2001a) and is based on the “Method 1” approach introduced by Henderson (1953) to generate variance components in data sets that are unbalanced and/or have missing data. It is mathematically simple - using cell frequency counts and squared/summed values of the data point values to calculate the variance components. The generalizability/dependability calculations can follow from there.
The first step in this document is to “walk” the reader through the “Method 1” approach to calculate variance components introduced by Henderson (1953) and the one utilized by Brennan (2001a) in creating the urGENOVA generalizability program. Note that calculating generalizability coefficients is NOT the end goal of Henderson - it is just to generate variance coefficients, which indicate how the variance in a set of data is distributed among its various components (or “facets” in the vernacular of generalizability theory). It will be simplest to show how these work with an actual data set.
Table 1 below shows the data from Henderson’s (1953) Table 1. Its cells contain the butterfat levels at first lactations of 4 cow herds inseminated by 3 different sires over 4 years. The entries in each cell are somewhat unique – they represent number of cows in each herd (first number) and butterfat content (second number) in the first lactation of the year. The entry “3 – 1414” represents the total butterfat (1414) summed over 3 different cows. Thus, cows are nested within Herd/Sire combinations (C: HS).
HERD | SIRE | YEAR | YEAR | YEAR | YEAR | TOTAL |
---|---|---|---|---|---|---|
HERD | SIRE | 1 | 2 | 3 | 4 | TOTAL |
1 | 1 | 3 - 1414 | 2 – 981 | 5 – 2395 | ||
1 | 2 | 4 – 1766 | 2 – 862 | 6 – 2628 | ||
1 | 3 | 5 - 1609 | 5 – 1609 | |||
2 | 1 | 1 - 405 | 3 – 1270 | 4 – 1674 | ||
2 | 2 | 5 – 2109 | 5 – 2109 | |||
2 | 3 | 4 – 1563 | 2 - 740 | 6 - 2303 | ||
3 | 1 | 3 – 1705 | 3 – 1705 | |||
3 | 2 | 4 – 2310 | 2 – 1134 | 6 – 3444 | ||
4 | 1 | 3 - 1113 | 5 - 1951 | 8 – 3064 | ||
4 | 3 | 3 - 1291 | 6 - 2457 | 9 – 3948 | ||
TOTAL | 7 - 2931 | 21-9983 | 16 - 6959 | 13 - 4806 | 57-24679 |
It is worth noting:
\[Y_{hijk} = \mu + A_h(\text{Year}) + H_i(\text{Herd}) + S_j(\text{Sire}) + HS_{ij}(\text{Herd} \times \text{Sire}) + \text{Error}_{hijk}\]
Assumptions:
Step 1: Obtain the “T” values (uncorrected sums of squares) for each facet by summing over the squared values for each level of the facet and divide by each level’s relevant sample size (“^” indicates exponential). Henderson uses summation notation in his article – however actual values will make the computations more concrete.
\[ A \text{ (year)} = \sum \frac{(2931^2)}{7} + \frac{(9983^2)}{21} + \frac{(6959^2)}{16} + \frac{(4806^2)}{13} = 10,776,451 \]
\[ H \text{ (herd)} = \sum \frac{(2395+2628+1609)^2}{(5+6+5)} + \frac{(1674+2109+2303)^2}{(4+5+6)} + \frac{(1705+3444)^2}{(3+6)} + \frac{(3064+3748)^2}{(8+9)} = 10,893,666 \]
\[S \text{ (sire)} = \sum \frac{(2395+1674+1705+3064)^2}{(5+4+3+8)} + \frac{(2628+2109+3444)^2}{(6+5+6)} + \frac{(1609+2303+3748)^2}{(5+6+9)} = 10,776,278\]
\[HS \text{ (herd)} \times \text{(sire)} = \sum \frac{2395^2}{5} + \frac{2628^2}{6} + \frac{1609^2}{5} + \frac{1674^2}{4} + \frac{2109^2}{5} + \frac{2303^2}{6} + \frac{1705^2}{3} + \frac{3444^2}{6} + \frac{3064^2}{8} + \frac{3748^2}{9} = 10,970,369\]
\[\text{Mean} \text{ (Henderson uses "CF" for this term)} = \sum \text{ over all values } \frac{24679^2}{57} = 10,685,141\]
Total = 11,124,007
This would normally be accomplished by squaring and summing each individual cell value (all 57 of them). In this data set this would mean we would have to have the individual butterfat values for EACH “cow”. Instead, Henderson provides the sum of the butterfat values across cows in the herd. For example, in the cell 3-1414, this means that 3 cows provided 1414 butterfat. We DO NOT know how the 1414 butterfat is distributed between the 3 cows. If we did, we would be able to calculate the nested effect of cows within herds (c:H), that we are not provided. Henderson later in the paper does provide a T-value for the total (Table 6, p. 233) of 11,124,007. I tried generating this value a couple of different ways:
Taking the overall total 24679 and dividing by 57 (433). Then squaring this value 57 times and summing. This gave me a T-value of 10,686,873.
Taking each herd value, dividing it by the number of cows in the herd, then “counting” that value for each cow for that herd. This was repeated 57 times. Then these values were squared then summed. This gave me a T-value of 10,973,517.
Taking each herd value, dividing it by the number of cows in the herd. Then averaging across those values (440). Then squaring this value 57 times and summing. This gave me a T-value of 11,035,200.
Clearly the actual values for each cow would need to be provided to get the actual total T value for this analysis.
Step 2: Obtain the estimated coefficients by which the variance components will be multiplied
So far, things have been straightforward. The next step is more complicated and tedious. The coefficients of μ², σ²(year), σ²(herd), σ²(sire), σ²(herdXsire), σ²(error) are estimated using the SAMPLE SIZES on which they are based that contribute to the calculated T-values. This is tedious because the data set is unbalanced and there are lots of missing data! Fortunately, Henderson provides a slow and steady summation routine that allows for the estimations. Many of the cells can be filled using the simple overall sample size of the data set (N) (these are described below). The others (noted with a “?” in Table 2) have to be estimated individually. Again, Henderson provides the summation notation for calculating these terms for those interested in seeing them.
Facet | T-Value | μ² | σ²(Year) | σ²(Herd) | σ²(Sire) | σ²(HS) | σ²(error) |
---|---|---|---|---|---|---|---|
A (year) | 10,776,451 | N | N | ? - 1 | ? - 2 | ? - 3 | Levels of A (4) |
H (herd) | 10,893,666 | N | ? - 4 | N | ? - 5 | ? - 6 | Levels of H (4) |
S (sire) | 10,776,278 | N | ? - 7 | ? - 8 | N | ? - 9 | Levels of S (3) |
HS (herdX sire) | 10,970,369 | N | ? - 10 | ? - 11 | ? - 12 | N | #Combinations of HS (10) |
Mean | 10,685,141 | N | ? - 13 | ? - 14 | ? - 15 | ? - 16 | 1 |
Total | 11,124,007* | N | N | N | N | N | N |
*See note above regarding where this value comes from
Let’s begin to “fill in” the 16 different “?” terms with “brute force” calculations….
\[\text{Year 1: } \frac{(3^2) + (1^2) + (0^2) + (3^2)}{7} +\]
\[\text{Year 2: } \frac{(2+4)^2 + (3^2) + ((3+4)^2) + (5^2)}{21} +\]
\[\text{Year 3: } \frac{(2^2) + ((5+4)^2) + (2^2) + (3^2)}{16} +\]
\[\text{Year 4: } \frac{(5^2) + (2^2) + (0^2) + (6^2)}{13}\]
\[= 19.51\]
\[\text{Year 1: } \frac{((3+1+0+3)^2) + ((0+0+0+0)^2) + ((0+0+0+0)^2))}{7} +\]
\[\text{Year 2: } \frac{((2+3+3+5)^2) + ((4+0+4+0)^2) + ((0+0+0+0)^2)}{21} +\]
\[\text{Year 3: } \frac{((0+0+0+0)^2) + ((2+5+2+0)^2) + ((0+4+0+3)^2))}{16} +\]
\[\text{Year 4: } \frac{((0+0+0+0)^2) + ((0+0+0+0)^2) + ((5+2+0+6)^2))}{13}\]
\[= 39.22\]
\[\text{Year 1: } \frac{((3^2) + (0^2) + (0^2) + (1^2) + (0^2) + (0^2) + (0^2) + (0^2) + (3^2)+ (0^2))}{7} +\]
\[\text{Year 2: } \frac{((2^2) + (4^2) + (0^2) + (3^2) +(0^2) + (0^2) + (3^2) + (4^2) + (5^2) +(0^2))}{21} +\]
\[\text{Year 3: } \frac{((0^2) + (2^2) +(0^2) + (0^2) + (5^2) + (4^2) + (0^2) + (2^2)+ (0^2) + (3^2))}{16} +\]
\[\text{Year 4: } \frac{((0^2) + (0^2) +(5^2) + (0^2) + (0^2) + (2^2) + (0^2) + (0^2) + (0^2) + (6^2))}{13}\]
\[= 15.10\]
\[\text{Herd 1: } \frac{((3^2) + ((2+4)^2) + (2^2) + (5^2))}{16} +\]
\[\text{Herd 2: } \frac{((1^2) + (3^2) + ((5+4)^2) + (2^2))}{15} +\]
\[\text{Herd 3: } \frac{((0^2) + ((3+4)^2) + (2^2) + (0^2))}{9} +\]
\[\text{Herd 4: } \frac{((3^2) + (5^2) + (3^2) + (6^2))}{17}\]
\[= 21.49\]
\[\text{Herd 1: } \frac{(((3+2)^2) + ((4+2)^2) + (5^2))}{16} +\]
\[\text{Herd 2: } \frac{((3+1)^2) + (5^2) + ((4+2)^2))}{15} +\]
\[\text{Herd 3: } \frac{((3^2) + ((4+2)^2) + (0^2))}{9} +\]
\[\text{Herd 4: } \frac{(((3+5)^2) + (0^2) + ((3+6)^2))}{17}\]
\[= 24.04\]
\[\text{Herd 1: } \frac{((5^2) + (6^2) + (5^2))}{16} +\]
\[\text{Herd 2: } \frac{((4^2) + (5^2) + (6^2))}{15} +\]
\[\text{Herd 3: } \frac{((3^2) + (6^2) + (0^2))}{9} +\]
\[\text{Herd 4: } \frac{((8^2) + (0^2) + (9^2))}{17}\]
\[= 24.04\]
\[\text{Sire 1: } \frac{((3+1+3)^2) + ((2+3+3+5)^2) + (0^2)+(0))}{20} +\]
\[\text{Sire 2: } \frac{((0^2) + ((4+4)^2) + ((2+5+2)^2) + ((0^2)))}{17} +\]
\[\text{Sire 3: } \frac{((0^2) + (0^2) + ((4+3)^2) + ((5+2+6)^2))}{20}\]
\[= 30.33\]
\[\text{Sire 1: } \frac{((3+2)^2) + ((1+3)^2) + (3^2)+((3+5)^2))}{20} +\]
\[\text{Sire 2: } \frac{((4+2)^2) + (5^2) + ((4+2)^2) + (0^2))}{17} +\]
\[\text{Sire 3: } \frac{(5^2) + ((4+2)^2) + (0^2) + ((6+3)^2))}{20}\]
\[= 18.51\]
Note that this is the SAME value as that for the \(\sigma^2(\text{Herd})\) effect on Sire. The difference is in how the numbers were arrived at. For the \(\sigma^2(\text{Herd})\) effect on Sire, we had to sum across the Years to get the number of cows. For the \(\sigma^2(\text{HS})\) effect we just have used the sum of cows across years (like having rolled the Years into a single column).
\[\text{Sire 1: } \frac{(5^2) + (4^2) + (3^2)+(8^2))}{20} +\]
\[\text{Sire 2: } \frac{(6^2) + (5^2) + (6^2) + (0^2))}{17} +\]
\[\text{Sire 3: } \frac{(5^2) + (6^2) + (0^2) + (9^2))}{20}\]
\[= 18.51\]
\[\text{Herd/Sire 1: } \frac{((3^2) + (2^2) + (0^2)+(0^2))}{5} +\]
\[\text{Herd/Sire 2: } \frac{((0^2) + (4^2) + (2^2)+(0^2))}{6} +\]
\[\text{Herd/Sire 3: } \frac{((0^2) + (0^2) + (0^2)+(5^2))}{5} +\]
\[\text{Herd/Sire 4: } \frac{((1^2) + (3^2) + (0^2)+(0^2))}{4} +\]
\[\text{Herd/Sire 5: } \frac{((0^2) + (0^2) + (5^2)+(0^2))}{5} +\]
\[\text{Herd/Sire 6: } \frac{((0^2) + (0^2) + (4^2)+(2^2))}{6} +\]
\[\text{Herd/Sire 7: } \frac{((0^2) + (3^2) + (0^2)+(0^2))}{3} +\]
\[\text{Herd/Sire 8: } \frac{((0^2) + (4^2) + (2^2)+(0^2))}{6} +\]
\[\text{Herd/Sire 9: } \frac{((5^2) + (3^2) + (0^2)+(0^2))}{8} +\]
\[\text{Herd/Sire 10: } \frac{((0^2) + (0^2) + (3^2)+(6^2))}{9}\]
\[= 37.35\]
\[= \frac{(57^2)}{57} = 57\]
\[= \frac{(57^2)}{57} = 57\]
\[= \frac{((7^2) + (21^2) + (16^2) + (13^2))}{57} = 16.05\]
\[= \frac{((16^2) + (15^2) + (9^2) + (17^2))}{57} = 14.93\]
\[= \frac{((20^2) + (17^2) + (20^2))}{57} = 19.11\]
\[= \frac{((5^2) + (6^2) + (5^2) +(4^2) + (5^2) + (6^2) + (3^2) + (6^2) + (8^2) +(9^2))}{57} = 6.19\]
All these calculated values can now be put into Table 2 and are presented in Table 3.
Facet | T-Values | μ² | ^2(year) | σ²(herd) | σ²(sire) | σ²(HS) | σ²(error) |
---|---|---|---|---|---|---|---|
A (year) | 10,776,451 | 57 | 57 | 19.51 | 39.22 | 15.1 | 4 |
H (herd) | 10,893,666 | 57 | 21.49 | 57 | 24.04 | 24.04 | 4 |
S (sire) | 10,776,278 | 57 | 30.33 | 18.51 | 57 | 18.51 | 3 |
HS (herdX sire) | 10,970,369 | 57 | 37.35 | 57 | 57 | 57 | 10 |
Mean | 10,685,141 | 57 | 16.05 | 14.93 | 19.11 | 6.19 | 1 |
Total | 11,124,007 | 57 | 57 | 57 | 57 | 57 | 57 |
Although at this point Henderson creates a “Table 7” from which equations he says can be solved, this step is not necessary. What is necessary is to “solve” for the equations that are set up in Table 3. Specifically, what values need to be multiplied with each estimated coefficient down the entire column for all 6 equations that will solve the following equations? These will be the variance components: A, B, C, D, E, and F.
\[ 10,776,451 = A(57) + B(57) + C(19.51) + D(39.22) + E(15.10) +F(4) \]
\[ 10,893,666 = A(57) + B(21.49) + C(57) + D(24.04) + E(24.04) +F(4) \]
\[ 10,776,278 = A(57) + B(30.33) + C(18.51) + D(57) + E(18.51) +F(3) \]
\[ 10,970,369 = A(57) + B(37.35) + C(57) + D(57) + E(57) +F(10) \]
\[ 10,685,141 = A(57) + B(16.05) + C(14.93) + D(19.11) + E(6.19) +F(1) \]
\[ 11,124,007 = A(57) + B(57) + C(57) + D(57) + E(57) +F(57) \]
Using the data in Table 3, solve these equations simultaneously via matrix procedures to solve a system of equations (solving by linear least squares method). Alternatively, they can be solved by regressing the 6 T-values on the 6 variance estimates in a direct regression (all predictors entered simultaneously), and requesting to include an intercept in the model). This produces the following output (from SPSS – although other programs will provide the same results). Note that there is no “Mean” variance component calculated, as μ² is a constant and has no variance so is eliminated from the analysis. Since there are 6 predictors and 6 outcomes the data will “fit perfectly” (in fact, Excel WILL NOT run the regression because there are equal numbers of predictors and cell entries in the outcome). However, this is not relevant, as we are interested solely in the unstandardized B-values associated with each effect, which are the variance components. We now have the variance components (see Table 4) for Henderson’s data.
Model | Unstandardized Coefficients | Standardized Coefficients | |
---|---|---|---|
B | Std. Error | Beta | |
(Constant) | 10572974.498 | .000 | |
Year | 763.154 | .000 | .084 |
Herd | 4531.304 | .000 | .615 |
Sire | 1587.278 | .000 | .174 |
HS | -164.329 | .000 | -.023 |
Error | 2949.830 | .000 | .402 |
Facet | Variance Component | Proportion of Variance |
---|---|---|
A (year) | 763 | .08 |
H (herd) | 4531 | .46 |
S (sire) | 1587 | .16 |
HS | -164 | .30 |
Error | 2950 | |
Total |
We can see from the results that most of the variance (46%) in the butterfat amount, is due to the “Herd” variable and next by the Herd X Sire interaction (30%). Not a lot of the variance is due to the Sire (16%) and almost none is due to Year (8%). As noted earlier, Henderson stops at this point.
Before leaving Henderson, there are a couple notes about his protocol.
Brennan’s (2001a) approach (he calls it analogous-ANOVA) is based on Henderson’s calculations of T-values. With unbalanced or missing data, “The primary theoretical problem presented by most unbalanced designs is that there are many possible estimators of random effects variance components and no unambiguously clear basis for determining which estimators are best. Among the practical problems with unbalanced designs are that some estimation methods require distributional form assumptions, which are often difficult to justify in generalizability analyses.” (Manual for urGENOVA Brennan, 2001b p. 1). Thus, he advises using the variance components generated from urGENOVA for generalizability estimates that are strictly “descriptive” of the specific data set and not inferential. The urGENOVA program does not allow for any computations for D-studies, nor are standard errors and confidence intervals of the variance components appropriate except for balanced designs under the assumptions of normality. A shortcoming of the urGENOVA program is that IT DOES NOT produce generalizability or dependability coefficients. These have to be calculated by hand by the end user.
The G-String_V program is more restrictive in input than urGENOVA. It does NOT allow for missing data. This simplifies the calculations (as we will see in a moment). The authors of this program do not adhere to Brennan’s (2001b) admonition of D-study, standard errors and confidence intervals of the variance components. In fact, it is relatively straightforward in the case of unbalanced, but no missing data, to estimate D-study generalizability coefficients using harmonic mean estimates of the unbalanced effects. However, these should be taken as estimates only, given the issues raised by Brennan. The G-String_V program DOES calculate generalizability and dependability coefficients for the facet of differentiation specified by the user.
With this as a backdrop, the next step in this tutorial is to use a data set from Narayanan et al. (2010) (Table 5) to calculate the variance components using Henderson’s Method 1. It will be noted at this point that Narayanan uses a completely different approach to calculating generalizability in the paper. It is based on the variance, not the variance components, which is the hallmark of traditional approaches to calculating Generalizability. These data are unbalanced, but no data points are missing. There are 3 different doctors (d); rated by 16 patients (p) - note that patients are nested in doctors, and this facet is unbalanced; 8 patients rate Doctor A, 5 rate Doctor B and 3 rate Doctor C); all use 5 items (i) to rate the doctor. The model then is: i X (p:d). The dependent variable is “rating on a scale of 1-5” for each item.
Doctor | Patient | item1 | item2 | item3 | item4 | item5 |
---|---|---|---|---|---|---|
A | 1 | 4 | 4 | 4 | 4 | 4 |
A | 2 | 4 | 4 | 3 | 3 | 4 |
A | 3 | 3 | 4 | 4 | 3 | 4 |
A | 4 | 3 | 3 | 3 | 3 | 3 |
A | 5 | 3 | 3 | 4 | 4 | 3 |
A | 6 | 4 | 4 | 3 | 3 | 3 |
A | 7 | 3 | 3 | 3 | 3 | 3 |
A | 8 | 4 | 4 | 3 | 3 | 3 |
B | 9 | 4 | 4 | 4 | 4 | 3 |
B | 10 | 4 | 4 | 4 | 4 | 4 |
B | 11 | 4 | 4 | 4 | 4 | 4 |
B | 12 | 4 | 3 | 3 | 3 | 3 |
B | 13 | 4 | 4 | 3 | 3 | 3 |
C | 14 | 4 | 4 | 3 | 3 | 3 |
C | 15 | 3 | 3 | 3 | 3 | 3 |
C | 16 | 4 | 4 | 4 | 4 | 3 |
We are going to need a number of sums and their squares for the next analyses (Table 6):
Doctor | Patient | item1 | item2 | item3 | item4 | item5 | Sum across Patient Ratings | Squared sum across Patient Ratings | Squared sum across patient ratings/#ratings each completed |
---|---|---|---|---|---|---|---|---|---|
A | 1 | 4 | 4 | 4 | 4 | 4 | 20 | 400 | 80 |
A | 2 | 4 | 4 | 3 | 3 | 4 | 18 | 324 | 64.8 |
A | 3 | 3 | 4 | 4 | 3 | 4 | 18 | 324 | 64.8 |
A | 4 | 3 | 3 | 3 | 3 | 3 | 15 | 225 | 45 |
A | 5 | 3 | 3 | 4 | 4 | 3 | 17 | 289 | 57.8 |
A | 6 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
A | 7 | 3 | 3 | 3 | 3 | 3 | 15 | 225 | 45 |
A | 8 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
B | 9 | 4 | 4 | 4 | 4 | 3 | 19 | 361 | 72.2 |
B | 10 | 4 | 4 | 4 | 4 | 4 | 20 | 400 | 80 |
B | 11 | 4 | 4 | 4 | 4 | 4 | 20 | 400 | 80 |
B | 12 | 4 | 3 | 3 | 3 | 3 | 16 | 256 | 51.2 |
B | 13 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
C | 14 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
C | 15 | 3 | 3 | 3 | 3 | 3 | 15 | 225 | 45 |
C | 16 | 4 | 4 | 4 | 4 | 3 | 19 | 361 | 72.2 |
Item Rating Sums | 59 | 59 | 55 | 54 | 53 | ||||
Item Rating sums squared | 3481 | 3481 | 3025 | 2916 | 2809 | ||||
Item sums squared/#ratings For each item | 217.562 | 217.562 | 189.062 | 182.25 | 175.562 |
Sum all ratings across each item for each Patient; Sum the individual patient ratings for each doctor; square these sums; divide by the number of rating counts within each doctor; add the quotients.
\[\text{Doctor A} = (20+18+18+15+17+17+15+17) = 137; \frac{137^2}{40} = \frac{18769}{40} = 469.225\]
\[\text{Doctor B} = (19+20+20+16+17) = 92; \frac{92^2}{25} = \frac{8464}{25} = 338.560\]
\[\text{Doctor C} = (17+15+19) = 51; \frac{51^2}{15} = \frac{2601}{15} = 173.400\]
Sum across all 3 quotients = \(469.225 + 338.560 + 173.400 = 981.185\)
Sum all ratings across each item for each Patient; Square these sums; divide by the number of rating counts within each patient; sum across these quotients.
\[\text{p:d 1: } \frac{20^2}{5} = \frac{400}{5} = 80\]
\[\text{p:d 2: } \frac{18^2}{5} = \frac{324}{5} = 64.8\]
[continuing for all patients]
\[\text{p:d 15: } \frac{15^2}{5} = \frac{225}{5} = 45\]
\[\text{p:d 16: } \frac{19^2}{5} = \frac{361}{5} = 72.2\]
Sum across all 16 quotients = 989.200
Sum all ratings down each item; Square these sums; divide by the number of rating counts within each item; sum across these quotients.
\[\text{Item 1: } \frac{59^2}{16} = \frac{3481}{16} = 217.5625\]
\[\text{Item 2: } \frac{59^2}{16} = \frac{3481}{16} = 217.5625\]
\[\text{Item 3: } \frac{55^2}{16} = \frac{3025}{16} = 189.0625\]
\[\text{Item 4: } \frac{54^2}{16} = \frac{2916}{16} = 182.250\]
\[\text{Item 5: } \frac{53^2}{16} = \frac{2809}{16} = 175.5625\]
Sum across all 5 quotients = \((217.5625+217.5625+189.0625+182.250+175.5625) = 982.000\)
Sum down each set of items for each doctor (there will be 15 combinations); square these sums; divide each sum by the number of ratings that go into each of the 15 combinations; sum the quotients.
\[\text{di1: } \frac{(4+4+3+3+3+4+3+4)^2}{8} = \frac{28^2}{8} = \frac{784}{8} = 98.000\]
\[\text{di2: } \frac{(4+4+4+3+3+4+3+4)^2}{8} = \frac{29^2}{8} = \frac{841}{8} = 105.125\]
[continuing for all doctor-item combinations]
\[\text{di14: } \frac{(3+3+4)^2}{3} = \frac{10^2}{3} = \frac{100}{3} = 33.333\]
\[\text{di15: } \frac{(3+3+3)^2}{3} = \frac{9^2}{3} = \frac{81}{3} = 27.000\]
Sum across all quotients = 983.808
Note that this is the total uncorrected sums of squares. Each individual rating is first squared. Then these are summed across all ratings.
\[4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 3^2 + 3^2 + 4^2 + ... + 3^2 + 3^2 + 3^2 + 3^2 + 3^2 + 4^2 + 4^2 + 4^2 + 4^2 + 3^2\]
Summing across all values = 1000.000
Sum FIRST across all values; square this sum; divide by the total number of ratings that went into the sum.
\[\frac{(4 + 4 + 4 + 4 + 4 + ... + 3 + 3 + 3 + 3 + 3 + 4 + 4 + 4 + 4 + 3)^2}{80} = \frac{280^2}{80} = \frac{78400}{80} = 980.000\]
We can now put our Facet T-values into Table 7. below.
Facet | T-Value |
---|---|
D (doctor) | 981.185 |
P:D (patient nested in doctor) | 989.2 |
I (item) | 982 |
DI (doctor X item) | 983.808 |
PI:D (patient X item nested in doctor) (Equivalent to Henderson’s “total”) | 1000 |
Mean | 980 |
As we know from the Henderson example, the next part is tedious. We have to get the sample sizes for the variance coefficients. Like Henderson, we will first put in the “easy” ones.
Brennan (2001a, equations 7.2-7.6 p. 219) discusses how each of the cell entries are generated. These equations are difficult to “unpack” so I will be using Henderson’s “brute force” approach that allows for every count to be taken into consideration when calculating the terms.
When the data set is not missing any datapoints, Brennan uses a notation system for the coefficients in the other cells that need to be calculated. These are mostly simple, such as n(p) is the number of levels of the facet. In this design the n(d) = 3; n(p) = 16; n(i) = 5. However, n(r) is a unique notation. It is based on the number of “counts(individual ratings)” within each DI combination (recall there are 15 of them); square these values; sum them; divide by the number of counts in the entire data set.
\[ \frac{(8^2) + (8^2) + (8^2) + (8^2) + (8^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (3^2) + (3^2) + (3^2) + (3^2) + (3^2)}{80} = 6.125 \]
The results from Brennan’s notation are shown in Table 8 for this design. However, as noted, the entries ONLY work in a balanced design. When there are missing data, we need to go back to the Henderson approach.
Facet | σ²(d) | σ²(p:d) | σ²(i) | σ²(di) | σ²(pi:d) | μ² |
---|---|---|---|---|---|---|
D | N | n(i)*n(d) | n(p) | n(p) | n(d) | N |
P:D | N | N | n(p) | n(p) | n(p) | N |
I | n(r)*n(i) | n(i) | N | n(r)*n(i) | n(i) | N |
DI | N | n(i)*n(d) | N | N | n(i)*n(d) | N |
PI:D (total for Henderson) | N | N | N | N | N | N |
Mean | n(r)*n(i) | n(i) | n(p) | n(r) | 1 | N |
Henderson’s approach to completing the “non-simple” cells in the coefficient matrix follows. It allows for missing data in the calculations. There are 16 unique values (highlighted) to calculate.
Facet | σ²(d) | σ²(p:d) | σ²(i) | σ²(di) | σ²(pi:d) | μ² |
---|---|---|---|---|---|---|
D | N | 1 | 2 | 3 | n(d) | N |
P:D | 4 | N | 5 | 6 | n(p) | N |
I | 7 | 8 | N | 9 | n(i) | N |
DI | 10 | 11 | 12 | N | n(i)*n(d) | N |
PI:D (total for Henderson) | N | N | N | N | N | N |
Mean | 13 | 14 | 15 | 16 | 1 | N |
The σ²(p:d) term on D (data are collapsed across Items). Sample size starts by calculating the squared sum of:
counts of each p:d term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[\text{D1: } \frac{(5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2)}{40} = \frac{200}{40} = 5 +\]
\[\text{D2: } \frac{(5^2) + (5^2) + (5^2) + (5^2) + (5^2)}{25} = \frac{125}{25} = 5 +\]
\[\text{D3: } \frac{(5^2) + (5^2) + (5^2)}{15} = \frac{75}{15} = 5\]
\[= 15\] 2. The σ²(i) term on D (data are collapsed across P:D). Sample size starts by calculating the squared sum of:
counts of each i term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[\text{D1: } \frac{(8^2) + (8^2) + (8^2) + (8^2) + (8^2)}{40} = \frac{320}{40} = 8 +\]
\[\text{D2: } \frac{(5^2) + (5^2) + (5^2) + (5^2) + (5^2)}{25} = \frac{125}{25} = 5 +\]
\[\text{D3: } \frac{(3^2) + (3^2) + (3^2) + (3^2) + (3^2)}{15} = \frac{45}{15} = 3\]
\[= 16\]
The σ²(di) term on D (data are collapsed across P:D). Sample size starts by calculating the squared sum of:
counts of each di term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[\text{D1: } \frac{(8^2) + (8^2) + (8^2) + (8^2) + (8^2)}{40} = \frac{320}{40} = 8 +\]
\[\text{D2: } \frac{(5^2) + (5^2) + (5^2) + (5^2) + (5^2)}{25} = \frac{125}{25} = 5 +\]
\[\text{D3: } \frac{(3^2) + (3^2) + (3^2) + (3^2) + (3^2)}{15} = \frac{45}{15} = 3\]
\[= 16\]
The σ²(d) term on P:D (data are collapsed across Items). Sample size starts by calculating the squared sum of:
counts of each d term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[\text{PD 1: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 2: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 3: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 4: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 5: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 6: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 7: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 8: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 9: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 10: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 11: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 12: } \frac{(5^2)}{5} = 5+\]
\[\text{PD 13: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 14: } \frac{(5^2)}{5} = 5+\]
\[\text{PD 15: } \frac{(5^2)}{5} = 5 +\]
\[\text{PD 16: } \frac{(5^2)}{5} = 5\]
\[=80\]
The σ²(i) term on P:D. Sample size starts by calculating the squared sum of:
counts of each i term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[\text{PD 1: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 2: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 3: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 4: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 5: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 6: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 7: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 8: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 9: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 10: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 11: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 12: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 13: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 14: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 15: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 16: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1\]
\[=16\]
The σ²(di) term on P:D. Sample size starts by calculating the squared sum of:
counts of each di term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[\text{PD 1: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 2: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 3: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 4: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 5: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 6: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 7: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 8: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 9: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 10: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 11: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 12: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 13: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 14: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 15: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1 +\]
\[\text{PD 16: } \frac{((1^2) + (1^2) + (1^2) + 1^2) + (1^2))}{5} = 1\]
\[=16\]
The σ²(d) term on I. Sample size starts by calculating the squared sum of:
counts of each d term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[\text{I 1: } \frac{((8^2) + (5^2) + (3^2))}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 2: } \frac{((8^2) + (5^2) + (3^2))}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 3: } \frac{((8^2) + (5^2) + (3^2))}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 4: } \frac{((8^2) + (5^2) + (3^2))}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 5: } \frac{((8^2) + (5^2) + (3^2))}{16} = \frac{98}{16} = 6.125\]
\[=30.625\]
The σ²(p:d) term on I. Sample size starts by calculating the squared sum of:
counts of each p:d term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[\text{I 1: } \frac{((1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+(1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2))}{16} = \frac{16}{16} =1+\]
\[\text{I 2: } \frac{((1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+(1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2))}{16} = \frac{16}{16} =1+\]
\[\text{I 3: } \frac{((1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+(1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2))}{16} = \frac{16}{16} =1+\]
\[\text{I 4: } \frac{((1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+(1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2))}{16} = \frac{16}{16} =1+\]
\[\text{I 5: } \frac{((1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+(1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2)+ (1^2)+(1^2)+(1^2))}{16} = \frac{16}{16} =1\]
\[=5\]
The σ²(di) term on I. Sample size starts by calculating the squared sum of:
counts of each di term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[\text{I 1: } \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 2: } \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 3: } \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 4: } \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 +\]
\[\text{I 5: } \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125\]
\[=30.625\]
The σ²(d) term on DI. Sample size starts by calculating the squared sum of:
counts of each d term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[\text{DI 1: } \frac{8^2}{8} = 8 +\]
\[\text{DI 2: } \frac{8^2}{8} = 8 +\]
\[\text{DI 3: } \frac{8^2}{8} = 8 +\]
\[\text{DI 4: } \frac{8^2}{8} = 8 +\]
\[\text{DI 5: } \frac{8^2}{8} = 8 +\]
\[\text{DI 6: } \frac{5^2}{5} = 5 +\]
\[\text{DI 7: } \frac{5^2}{5} = 5 +\]
\[\text{DI 8: } \frac{5^2}{5} = 5 +\]
\[\text{DI 9: } \frac{5^2}{5} = 5 +\]
\[\text{DI 10: } \frac{5^2}{5} = 5 +\]
\[\text{DI 11: } \frac{3^2}{3} = 3 +\]
\[\text{DI 12: } \frac{3^2}{3} = 3 +\]
\[\text{DI 13: } \frac{3^2}{3} = 3 +\]
\[\text{DI 14: } \frac{3^2}{3} = 3 +\]
\[\text{DI 15: } \frac{3^2}{3} = 3\]
\[=80\]
The σ²(p:d) term on DI. Sample size starts by calculating the squared sum of:
counts of each p:d term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[\text{DI 1: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 +\]
\[\text{DI 2: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 +\]
\[\text{DI 3: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 +\]
\[\text{DI 4: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 +\]
\[\text{DI 5: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 +\]
\[\text{DI 6: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 +\]
\[\text{DI 7: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 +\]
\[\text{DI 8: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 +\]
\[\text{DI 9: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 +\]
\[\text{DI 10: } \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 +\]
\[\text{DI 11: } \frac{(1^2) + (1^2) + (1^2)}{3} = 1 +\]
\[\text{DI 12: } \frac{(1^2) + (1^2) + (1^2)}{3} = 1 +\]
\[\text{DI 13: } \frac{(1^2) + (1^2) + (1^2)}{3} = 1 +\]
\[\text{DI 14: } \frac{(1^2) + (1^2) + (1^2)}{3} = 1 +\]
\[\text{DI 15: } \frac{(1^2) + (1^2) + (1^2)}{3} = 1\]
\[=15\]
The σ²(i) term on DI. Sample size starts by calculating the squared sum of:
counts of each i term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[\text{DI 1: } \frac{(8^2)}{8} = 8 +\]
\[\text{DI 2: } \frac{(8^2)}{8} = 8 +\]
\[\text{DI 3: } \frac{(8^2)}{8} = 8 +\]
\[\text{DI 4: } \frac{(8^2)}{8} = 8 +\]
\[\text{DI 5: } \frac{(8^2)}{8} = 8 +\]
\[\text{DI 6: } \frac{(5^2)}{5} = 5 +\]
\[\text{DI 7: } \frac{(5^2)}{5} = 5 +\]
\[\text{DI 8: } \frac{(5^2)}{5} = 5 +\]
\[\text{DI 9: } \frac{(5^2)}{5} = 5 +\]
\[\text{DI 10: } \frac{(5^2)}{5} = 5 +\]
\[\text{DI 11: } \frac{(3^2)}{3} = 3 +\]
\[\text{DI 12: } \frac{(3^2)}{3} = 3 +\]
\[\text{DI 13: } \frac{(3^2)}{3} = 3 +\]
\[\text{DI 14: } \frac{(3^2)}{3} = 3 +\]
\[\text{DI 15: } \frac{(3^2)}{3} = 3\]
\[=80\]
The σ²(d) term on mean. Sample size starts by calculating the squared sum of:
counts of each d term within each mean, divided by the total number of counts for the mean.
\[\text{Mean: } \frac{(40^2) + (25^2) + (15^2)}{80} = \frac{1600 + 625 + 225}{80} = \frac{2450}{80} = 30.625\]
The σ²(p:d) term on mean. Sample size starts by calculating the squared sum of:
counts of each p:d term within each mean, divided by the total number of counts for the mean.
\[ \text{Mean: } \frac{(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)+(5^2)}{80}\] \[ =\frac{25 \times 16}{80} = \frac{400}{80} = 5 \]
The σ²(i) term on mean. Sample size starts by calculating the squared sum of:
counts of each i term within each mean, divided by the total number of counts for the mean
\[\text{Mean: } \frac{(16^2) + (16^2) + (16^2) + (16^2) + (16^2)}{80} = \frac{256 \times 5}{80} = \frac{1280}{80} = 16\]
The σ²(di) term on mean. Sample size starts by calculating the squared sum of:
counts of each di term within each mean, divided by the total number of counts for the mean.
\[\text{Mean: } \frac{(8^2) + (5^2)+(3^2)+(8^2)+(5^2)+(3^2)+(8^2)+(5^2)+(3^2)+(8^2)+(5^2)+(3^2)+( 8^2)+(5^2)+(3^2)}{80}\]
\[= \frac{64 \times 5 + 25 \times 5 + 9 \times 5}{80}\]
\[= \frac{320 + 125 + 45}{80}\]
\[= \frac{490}{80} = 6.125\]
We insert the calculated values into their respective cells (Table 10) and bring the T-values down to complete the table.
Facet | σ²(d) | σ²(p:d) | σ²(i) | σ²(di) | σ²(pi:d) | μ² | T-Value |
---|---|---|---|---|---|---|---|
D | 80 | 15 | 16 | 16 | 3 | 80 | 981.185 |
P:D | 80 | 80 | 16 | 16 | 16 | 80 | 989.2 |
I | 30.625 | 5 | 80 | 30.625 | 5 | 80 | 982 |
DI | 80 | 15 | 80 | 80 | 15 | 80 | 983.808 |
PI:D (total for Henderson) | 80 | 80 | 80 | 80 | 80 | 80 | 1000 |
Mean | 30.625 | 5 | 16 | 6.125 | 1 | 80 | 980 |
Using the data in Table 10, (via matrix procedures or regressing the T-values on the coefficients) the variance components for the facets are calculated. Below is the output from SPSS. Because μ² is a constant and has no variance, it is eliminated from the regression analysis. However, the Constant (978.972) when divided by the sample size (80) is 12.237, which is the mean variance component if matrix procedures had been used.
Model | Model | Unstandardized Coefficients | Unstandardized Coefficients | Standardized Coefficients |
---|---|---|---|---|
Model | Model | B | Std. Error | Beta |
1 | (Constant) | 978.972 | .000 | |
1 | D | .002 | .000 | .008 |
1 | P_D | .092 | .000 | .442 |
1 | I | .028 | .000 | .128 |
1 | DI | -.016 | .000 | -.071 |
1 | PI_D_E | .157 | .000 | .625 |
Examination of the variance components of the facets and their relative contribution to variance away from the mean (Table 11), we can see that most of the variance (56%) in this data set is due to the PI:D facet. This is problematic from an interpretive standpoint, indicating it is not possible to really understand the variance in the data set. 33% of the variance is due to the P:D facet – within patient ratings of the doctors. In examining the raw data, we see that the doctors are all rated similarly by all patients (very little variance in the data set with which to work).
Facet | Variance Component | Proportion of Variance |
---|---|---|
D | .002 | 0.007 |
P:D | .092 | 0.33 |
I | .028 | 0.1 |
DI | -.016** | 0 |
PI:D | .157 | 0.563 |
Using the Variance Component (VC) values, Generalizability (G) and Dependability (D) coefficients can be calculated. It is most likely that we would want to know “Are the patients reliable in making their ratings of the doctors?” That is, “patient” would be the face of differentiation. We might also want to know how reliable the items are so we can also use item as a facet of differentiation. We can also specify doctor as a facet of differentiation, but looking at the data, there is so little variance on this facet that the results will not likely be reliable.
To calculate G and D, we need the variance components for each facet as well as the denominators for each of the facets that will serve in the error term (the facet(s) of generalization) when calculating the G and D for each facet of differentiation. Note that these denominators will differ depending on which facet is the facet of differentiation, and which facet/facets are being generalized across.
The calculations for the denominators are extensions on Brennan’s (2001a, p. 232) equation 7.28. The calculation modifies equation 7.28 for each unique value within the facet of differentiation and returns the harmonic mean (see Appendix A for a description of the use of harmonic mean) of those values to give the appropriate level coefficient. The general formula for determining levels used for G Theory Calculations can be determined as follows:
Let:
For each g, the level is determined by the sum of counts squared over the sum of square counts for v ∈ V: \[ Lg = \frac{\left(\sum_{v=1}^{V}C\left(g, v\right)\right)^{2}}{\sum_{v=1}^{V}C\left(g, v\right)^{2}} \]
When the data set is balanced and complete (i.e. no missing terms) this expression simplifies to , which is equal to the counts of unique combinations of facets in the variance in question for a given facet of differentiation.
It should be straightforward to the reader to see why this is the case. Let’s imagine briefly that for this design [items x (patients:doctors)] we had a balanced nesting of 5 items, 3 doctors, and 5 patients per doctor (instead of 8, 5, and 3 patients respectively). If the facet of differentiation was ‘doctors’ and we were trying to determine the appropriate level for ‘patients:doctors’ the calculation would be as follows: \[ Lg = \frac{\left(\sum_{v=1}^{5}5\right)^{2}}{\sum_{v=1}^{5}5^{2}} = \frac{\left(5*5\right)^{2}}{5*5^{2}} = 5 \]
Here each C(g,v) = 5 because there are 5 items for each combination of patient:doctor, and since there are 5 patients in the balanced example v ∈ [1, 5]. Thus, in the balanced case where there are 5 patients per doctor, it makes sense that we must correct the variance(patient:doctor) by a factor of 5 when we are considering doctors as the facet of differentiation!
Notice that thus far we’ve actually only calculated the level, L, for a unique facet of differentiation, g. When the data set is balanced, or we are dealing with an unnested variable (i.e. “items” is unnested in our current [items x (patients:doctors)] design) in an unbalanced design, each unique g will have the same level, L. For example, consider the unbalanced dataset in question. If we are considering ‘items’ as the facet of differentiation, and calculating the levels for ‘p:d’, it should be clear that each of the five items, g, has 16 unique combinations. However, this is not the case for facets involved with unbalanced or missing data. We must extend our definition of L to account for all g ∈ G by taking the harmonic mean of each Lg: \[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{\sum_{v=1}^{V}C\left(g, v\right)^{2}}{\left(\sum_{v=1}^{V}C\left(g, v\right)\right)^{2}}} \] The harmonic mean is necessary as we are accounting for the ratio between the counts of the variance in question and the chosen facet of differentiation. When data are balanced and complete this expression is unnecessary as gi = gk for all g ∈ G and thus L = Lg. However, this is not the case for unbalanced and missing data, so we must correct the variance by the harmonic mean of all Lg, L.
Table 12 shows the denominators to be used in our unbalanced data set with the complete worked examples following.
Variance Term | P:D Facet of Differentiation Denominator for G and D | I Facet of Differentiation Denominator for G and D | D Facet of Differentiation Denominator for G and D |
---|---|---|---|
σ²(d) | 1 | 2.612 | 1 |
σ²(p:d) | 1 | 16 | 4.557 |
σ²(i) | 5 | 1 | 5 |
σ²(i x d) | 5 | 2.612 | 5 |
σ²(i x (p:d)) | 5 | 16 | 22.785 |
Level of the σ²(d) term on facet of differentiation P:D. Since d is included in P:D, there is a single unique D for each P:D. L = 1
Level of the σ²(p:d) term on facet of differentiation P:D. Again, p:d is included in P:D, so there is a single unique P:D for each P:D. L = 1
Level of the σ²(i) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, i) = 1, v = 5 : 1 count of item for each unique combination of p:d and i with i ∈ [1, 5] for each p:d.
\[ L = Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
Level of the σ²(i x d) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, i x d) = 1, v = 5: Again there is 1 count of item x doctor for each unique combination of p:d and i x d (notice that d occurs in both) with i x d ∈ [1, 5] for each p:d.
\[ L = Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
Level of the σ²(i x (p:d)) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, i x (p:d)) = 1, v = 5: Again there is 1 count of i x (p:d)) for each unique combination of p:d and i x (p:d)) (notice both d and p occurs in both) with i x (p:d)) ∈ [1, 5] for each p:d.
\[ L = Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
Level of the σ²(d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, d) = \(\left|p_{v}\right|\) , for p = [8, 5, 3] with v ∈ [1, 3]; g ∈ [1, 5]. Since i is not involved with the unbalanced nesting, all levels of i, i ∈ [1, 5], will be the same.
\[ L = L1-5 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+3\right)^{2}}{(8^{2}+5^{2}+3^{2})} = \frac{256}{98} = 2.612 \]
Level of the σ²(p:d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, p:d) = 1, v ∈ [1, 16]. For each unique i, there will be 16 unique counts of p:d.
\[ L = L1-5 = \frac{\left(\sum_{v=1}^{16}1\right)^{2}}{\sum_{v=1}^{163}1^{2}} = 16 \]
Level of the σ²(i) term on facet of differentiation I. Since i is included in I, there is a single unique i for each I. L = 1
Level of the σ²(i x d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, i x d) = \(\left|p_{v}\right|\) , for p = [8, 5, 3] with v ∈ [1, 3]; g ∈ [1, 5]. Since i is not involved with the unbalanced nesting, all levels of i, i ∈ [1, 5], will be the same.
\[ L = L1-5 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+3\right)^{2}}{(8^{2}+5^{2}+3^{2})} = \frac{256}{98} = 2.612 \]
Level of the σ²(i x (p:d)) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, I x (p:d)) = 1, v ∈ [1, 16]. For each unique i, there will be 16 unique counts of i x(p:d).
\[ L = L1-5 = \frac{\left(\sum_{v=1}^{16}1\right)^{2}}{\sum_{v=1}^{163}1^{2}} = 16 \]
\[ L = 1 \]
Level of the σ²(p:d) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, p:d) = 5, for [g =A, v ∈ [1, 8]; g=B, v ∈ [1, 5]; g=C, v ∈ [1, 3]]. Thus we have different levels of patients:doctors for each unique doctor.
\[ L_A = \frac{\left(\sum_{v=1}^{8}5\right)^{2}}{\sum_{v=1}^{8}5^{2}} = 8 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}5\right)^{2}}{\sum_{v=1}^{5}5^{2}} = 5 \]
\[ L_C = \frac{\left(\sum_{v=1}^{3}5\right)^{2}}{\sum_{v=1}^{3}5^{2}} = 3 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{8}+\frac{1}{5}+\frac{1}{3}} = 4.557 \]
Level of the σ²(i) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, i) = \(\left|g\right|\) , for g=[8, 5, 3] and v = 5. There are g counts of each item [1, 5]. For example, there are 8 counts of items [1, 5] for doctor, d, A because there are 8 instances of 8 (by marginalizing over patients).
\[ L_A = \frac{\left(\sum_{v=1}^{5}8\right)^{2}}{\sum_{v=1}^{5}8^{2}} = 5 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}5\right)^{2}}{\sum_{v=1}^{5}5^{2}} = 5 \]
\[ L_C = \frac{\left(\sum_{v=1}^{5}3\right)^{2}}{\sum_{v=1}^{5}3^{2}} = 5 \]
\[ L = Lc = L_B = L_A = 5 \]
Level of the σ²(i x d) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, i x d) = \(\left|p\right|\) , for p=[8, 5, 3] and v = 5. There are p counts of each item, v, [1, 5]. This is identical to the uncrossed items.
\[ L_A = \frac{\left(\sum_{v=1}^{5}8\right)^{2}}{\sum_{v=1}^{5}8^{2}} = 5 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}5\right)^{2}}{\sum_{v=1}^{5}5^{2}} = 5 \]
\[ L_C = \frac{\left(\sum_{v=1}^{5}3\right)^{2}}{\sum_{v=1}^{5}3^{2}} = 5 \]
\[ L = Lc = L_B = L_A = 5 \]
Level of the σ²(i x (p:d)) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, i x (p:d)) = 1, for v=[pg*5] with p=[8, 5, 3]. For each doctor, there is a single unique count of items crossed with the nested patient:doctor. Thus, to obtain the level for each unique facet of differentiation, g, we must sum over v, which is patients * items or p*5 (since patients is unbalanced under doctors and items is constant 5).
\[ L_A = \frac{\left(\sum_{v=1}^{8*5}1\right)^{2}}{\sum_{v=1}^{8*5}1^{2}} = \frac{1600}{40} = 40 \]
\[ L_A = \frac{\left(\sum_{v=1}^{5*5}1\right)^{2}}{\sum_{v=1}^{5*5}1^{2}} = \frac{625}{25} = 25 \]
\[ L_A = \frac{\left(\sum_{v=1}^{3*5}1\right)^{2}}{\sum_{v=1}^{3*5}1^{2}} = \frac{225}{15} = 15 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{40}+\frac{1}{25}+\frac{1}{15}} = 22.785 \]
In Table 13 each of the generalizability coefficients (G and D; also referred to as \(E\rho^2\) and \(\Phi\)), are calculated as a ratio of: true variance/(true variance + error variance). “True variance” is signified by the term “tau” (\(\tau\)) (and in general is equal to the variance of the facet of differentiation, more below) and the denominator represents “\(\tau\)” + error variance. Error variances for generalizability and dependability coefficients are referred to as \(\delta\) and \(\Delta\) respectively. Notice that for the D-coefficients, more terms are added into the denominator that will result in lower values for D versus G coefficients. Error variances are comprised as indicated in the notes after Table 13.
For nested facets of differentiation, such as patients nested in doctors, it is imperative that the researcher understands the true variance that they seek to measure. According to Brennan’s (2001a) discussion on group means (5.3, page 157) the generalizability coefficient (G) of the facet of differentiation (p:d) can assess the generalizability of the patient ratings for “a randomly selected doctor” and would result in the formula \(G(p:d) = σ²_{p:d}/[σ²{p:d} + σ²{pi:d}/levels_{pi:d}]\). Where \(\tau\) represents the variance component of p:d (0.092) and the error variance is 0.157/5. If, however, we wanted the generalizability coefficient that represents patient ratings over “all doctors” then the \(\tau\) needs to include both the p:d and d variance components [\(\tau = σ²(p:d) + σ² (d)\), 0.092 and .002, respectively)]. We will assume, in this example exercise, that we want the latter interpretation of generalizability for patient ratings.
Facet of Differentiation = P:D | Facet of Differentiation = I | Facet of Differentiation = D |
---|---|---|
\(E\rho^2 = \tau/(\tau + \delta)\) | \(E\rho^2 = \tau/(\tau + \delta)\) | \(E\rho^2 = \tau/(\tau + \delta)\) |
\(\tau = \text{VC(P:D)} +
\text{VC(D)}\) \(\tau = 0.092 + 0.002 = 0.094\) |
\(\tau = \text{VC(I)} = 0.028\) | \(\tau = \text{VC(D)} = 0.002\) |
\(\delta =
\text{VC(PI:D)}/\text{Level}_{P:D}(\text{PI:D})\) \(\delta = 0.157/5 = 0.031\) |
\(\delta =
\text{VC(IxD)}/\text{Level}_I(\text{IxD}) +
\text{VC(PI:D)}/\text{Level}_I(\text{PI:D})\) \(\delta = 0/2.61 + 0.157/16 = 0.010\) |
\(\delta =
\text{VC(P:D)}/\text{Level}_D(\text{P:D}) +
\text{VC(IxD)}/\text{Level}_D(\text{IxD}) +
\text{VC(PI:D)}/\text{Level}_D(\text{PI:D})\) \(\delta = 0.092/4.557 + 0/5 + 0.157/22.785 = 0.027\) |
\(E\rho^2 = 0.094/(0.094+0.031) = \mathbf{0.752}\) | \(E\rho^2 = 0.028/(0.028+0.010) = \mathbf{0.737}\) | \(E\rho^2 = 0.002/(0.002+0.027) = \mathbf{0.069}\) |
Facet of Differentiation = P:D | Facet of Differentiation = I | Facet of Differentiation = D |
---|---|---|
\(\Phi = \tau/(\tau + \Delta)\) | \(\Phi = \tau/(\tau + \Delta)\) | \(\Phi = \tau/(\tau + \Delta)\) |
\(\Delta = (0.028/5) + (0/5) + (0.157/5) = 0.037\) | \(\Delta = (0.002/2.61) + (0.092/16) + (0/2.61) + (0.157/16) = 0.016\) | \(\Delta = (0.028/5) + (0.092/4.557) + (0/5) + (0.157/22.785) = 0.033\) |
\(\Phi = 0.094/(0.094 + 0.037) = \mathbf{0.718}\) | \(\Phi = 0.028/(0.028 + 0.016) = \mathbf{0.636}\) | \(\Phi = 0.002/(0.002 + 0.033) = \mathbf{0.057}\) |
* \(\delta\) = Respective Error Variance for G = sum of all VCs (except the VC of differentiation) that contain the facet of differentiation and at least one facet of generalization, each divided by the size of the product of the respective facets of generalization
** VC = Respective Variance Component for facet of differentiation
*** \(\Delta\) = Respective Error Variance for D = sum of all VCs (except the VC of differentiation), each divided by the size of the product of the respective facets of generalization
**** Because the DI variance component term is negative (-.016), we will substitute “0” for it when calculating the error variances for G and D, as suggested by Brennan (2001a).
It is worth pointing out at this stage when we finally obtain our Generalizability coefficient (G) for the P:D effect (0.752), that is it substantially different from that reported in the Narayanan et al. (2010) paper (0.91 for “unaggregated G” and 0.80 for “aggregated G”, p. 371). This is due to the completely different approach used. We will continue to observe the more traditional approach introduced by Henderson and made popular by Brennan (2001a).
Decision Studies utilize the variance components calculated from Generalizability Studies and manipulate the levels of facets to show how generalizability coefficients may be improved. For example, in this study we saw that the generalizability coefficient for p:d was below 0.80, and a researcher may wonder how the coefficient would change if there were more items for the same variance. With balanced and complete studies this is straightforward as the user can just input a dictionary of potential study combinations. For the items x (patients:doctors) this may look like:
D_Study_Dict: {items: [6, 7, 8], patients: [5], doctors: [3]}
Such a D-Study would calculate G Coefficients for the balanced scenarios where there are 3 unique doctors with 5 unique patients per doctor, each answering 6, 7, or 8 items per study respectively. Working out the 6-item scenario, we would have the following table of pseudo counts:
Doctor | Patient | Item |
---|---|---|
A | 1 | 1 |
A | 1 | 2 |
A | 1 | 3 |
A | 1 | 4 |
A | 1 | 5 |
A | 1 | 6 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
A | 5 | 1 |
A | 5 | 2 |
A | 5 | 3 |
A | 5 | 4 |
A | 5 | 5 |
A | 5 | 6 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
B | 10 | 1 |
B | 10 | 2 |
B | 10 | 3 |
B | 10 | 4 |
B | 10 | 5 |
B | 10 | 6 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
C | 15 | 1 |
C | 15 | 2 |
C | 15 | 3 |
C | 15 | 4 |
C | 15 | 5 |
C | 15 | 6 |
We then would have new levels coefficients to evaluate p:d (Table 15):
Variance Term | P:D Facet of Differentiation Denominator for G and D |
---|---|
σ²(d) | 1 |
σ²(p:d) | 1 |
σ²(i) | 6 |
σ²(i x d) | 6 |
σ²(i x (p:d)) | 6 |
With the adjusted generalizability coefficients (Table 16):
Generalizability Coefficient | Facet of Differentiation = P:D |
---|---|
Generalizability (G) | \(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(P:D)} + \text{VC(D)}\) \(\tau = 0.092 + 0.002 = 0.094\) \(\delta = \text{VC(PI:D)}/\text{Level}_{P:D}(\text{PI:D})\) \(\delta = 0.157/6 = 0.026\) \(E\rho^2 = 0.094/(0.094 + 0.026)\) \(E\rho^2 = \mathbf{0.78}\) |
Dependability (D) | \(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (0.028/6) + (0/6) + (0.157/6) = 0.031\) \(\Phi = 0.094/(0.094 + 0.031)\) \(\Phi = \mathbf{0.752}\) |
We see in this example that the G and D coefficients for P:D improve with the addition of another item (and assuming that there are 5 patients rating 3 doctors), indicating that for the same conditions this study would have been more generalizable had a greater number of items been included.
We see in this example that the G and D coefficients for P:D improve with the addition of another item (and assuming that there are 5 patients rating 3 doctors), indicating that for the same conditions this study would have been more generalizable had a greater number of items been included.
However, D-Studies need not be balanced. Let’s turn our attention to the generalizability coefficients associated with the “items” facet (which indicates the effect of questionnaire items in delineating doctor quality from patients) if some other aspect of the study was modified. In the G-Study, recall we had 5 items and the following unbalanced nesting of patients under doctors: {A: 8, B:5, C:3}. A researcher may have 16 patients but wonder: “If the number of patients nested under doctors A, B, and C were 6, 5, 5 respectively, how would this affect the generalizability of “items”? For demonstration let’s perform a D Study with the following characteristics and associated pseudo counts table (Table 17):
Doctor | Patient | Item |
---|---|---|
A | 1 | 1 |
A | 1 | 2 |
A | 1 | 3 |
A | 1 | 4 |
A | 1 | 5 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
A | 6 | 1 |
A | 6 | 2 |
A | 6 | 3 |
A | 6 | 4 |
A | 6 | 5 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
B | 11 | 1 |
B | 11 | 2 |
B | 11 | 3 |
B | 11 | 4 |
B | 11 | 5 |
… (table con’t)
Doctor | Patient | Item |
---|---|---|
C | 16 | 1 |
C | 16 | 2 |
C | 16 | 3 |
C | 16 | 4 |
C | 16 | 5 |
Using the method to calculate levels presented in Table 12, these new pseudo counts would result in the following levels (Table 18) for facet of differentiation, item.
Variance Term | I Facet of Differentiation Denominator for G and D |
---|---|
σ²(d) | 2.98 |
σ²(p:d) | 16 |
σ²(i) | 1 |
σ²(i x d) | 2.98 |
σ²(i x (p:d)) | 16 |
σ²(d)and σ²(i x (p:d)) remain the same because there are still 16 unique combinations of p and d despite being distributed differently. However, σ²(d) and σ²(i x d) change because of this distribution. Example calculation for σ²(d) shown below (which is identical to σ²(i x d)):
C(i, d) = |\(p_v\)|, for p = [6, 5, 5] with v ∈ [1, 3]; g ∈ [1, 5]. Since i is not involved with the unbalanced nesting, all levels of i, i ∈ [1, 5], will be the same.
\[ L = L1-5 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(6+5+5\right)^{2}}{(6^{2}+5^{2}+5^{2})} = \frac{256}{86} = 2.98 \]
As a small aside, the reader should notice that as we have nearly balanced data (6, 5, 5) the levels of σ²(d) on facet of differentiation, i, are nearly equivalent to the balanced case (L=3 for 3 doctors).
Coefficient | Facet of Differentiation = I |
---|---|
Generalizability (G) | \(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(I)} = 0.028\) \(\delta = \text{VC(IxD)}/\text{Level}_I(\text{IxD}) + \text{VC(PI:D)}/\text{Level}_I(\text{PI:D})\) \(\delta = 0/2.98 + 0.157/16 = 0.010\) \(E\rho^2 = 0.028/(0.028+0.010) = \mathbf{0.737}\) |
Dependability (D) | \(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (0.002/2.98) + (0.092/16) + (0/2.98) + (0.157/16)\) \(\Delta = 0.016\) \(\Phi = 0.028/(0.028 + 0.016) = \mathbf{0.636}\) |
Our generalizability coefficients, G and D remain the same as in the prior example! This is a direct result of the fact that the D and DI variances are so low that they hardly affect the calculation (in the first example or this one), thus for any reasonable choice of 16 p:d these coefficients will remain approximately the same. Of course, the numbers of P and D may be increased to reduce these variances and improve the generalizability coefficients, but those exercises will be left for the reader.
Our generalizability coefficients, G and D remain the same as in the prior example! This is a direct result of the fact that the D and DI variances are so low that they hardly affect the calculation (in the first example or this one), thus for any reasonable choice of 16 p:d these coefficients will remain approximately the same. Of course, the numbers of P and D may be increased to reduce these variances and improve the generalizability coefficients, but those exercises will be left for the reader.
urGENOVA and G-String_V programs also calculate the Estimated Mean Squares and from those the estimated Sums of Squares. However, neither of these are needed for calculating G and D coefficients. In addition, as Brennan (2001a) notes, because the Estimated Mean Square equations are complicated, expressions for estimators of the variance components in terms of mean square are also complicated. It is simpler to express the estimators of the variance components with respect to T-terms.
Another data set (the Narayana et al., 2010 with changed values for Doctor B (all ratings for all items increased by 1 point) and Doctor C (all ratings for all items lowered by 1 point) was generated (Table 20). This, not surprisingly, dramatically increases the variance on the “Doctor” facet, and the subsequent generalizability coefficients on which they are based.
Doctor | Patient | item1 | item2 | item3 | item4 | item5 |
---|---|---|---|---|---|---|
A | 1 | 4 | 4 | 4 | 4 | 4 |
A | 2 | 4 | 4 | 3 | 3 | 4 |
A | 3 | 3 | 4 | 4 | 3 | 4 |
A | 4 | 3 | 3 | 3 | 3 | 3 |
A | 5 | 3 | 3 | 4 | 4 | 3 |
A | 6 | 4 | 4 | 3 | 3 | 3 |
A | 7 | 3 | 3 | 3 | 3 | 3 |
A | 8 | 4 | 4 | 3 | 3 | 3 |
B | 9 | 5 | 5 | 5 | 5 | 4 |
B | 10 | 5 | 5 | 5 | 5 | 5 |
B | 11 | 5 | 5 | 5 | 5 | 5 |
B | 12 | 5 | 4 | 4 | 4 | 4 |
B | 13 | 5 | 5 | 4 | 4 | 4 |
C | 14 | 3 | 3 | 2 | 2 | 2 |
C | 15 | 2 | 2 | 2 | 2 | 2 |
C | 16 | 3 | 3 | 3 | 3 | 2 |
The resulting Table 21 shows that all the T-values change (as they would be expected to do), but not the “count”-based values for the coefficients, since these are only based on the count-relevant values for each coefficient.
Facet | σ²(d) | σ²(p:d) | σ²(i) | σ²(di) | σ²(pi:d) | μ² | T-Value |
---|---|---|---|---|---|---|---|
D | 80 | 15 | 16 | 16 | 3 | 80 | 1103.18 |
P:D | 80 | 80 | 16 | 16 | 16 | 80 | 1111.2 |
I | 30.625 | 5 | 80 | 30.625 | 5 | 80 | 1053.25 |
DI | 80 | 15 | 80 | 80 | 15 | 80 | 1105.81 |
PI:D (total for Henderson) | 80 | 80 | 80 | 80 | 80 | 80 | 1122 |
Mean | 30.625 | 5 | 16 | 6.125 | 1 | 80 | 1051.25 |
The resulting Table 22 of variance components demonstrates that while the D variance component is dramatically changed, the others remain exactly the same.
Facet | Variance Component | Proportion of Variance |
---|---|---|
D | 1.03 | 0.788 |
P:D | 0.092 | 0.07 |
I | 0.028 | 0.021 |
DI | -0.016 | 0 |
PI:D | 0.157 | 0.12 |
This is a much more easily-interpretable set of variance components: Most of the variance is due to different ratings of the doctors (79%). That is, the doctors are performing differently as rated by their patients. The P:D effect is 7%, indicating that the patients are very consistent (not much variance is ascribed to them) in rating the different doctors. The items are also internally consistent with only 2% of the variance in the data set ascribed to using different items. The resulting generalizability coefficients for the P:D and D facets goes up dramatically (Table 23):
Coefficient | Facet of Differentiation = P:D | Facet of Differentiation = I | Facet of Differentiation = D |
---|---|---|---|
Generalizability (G) | \(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(P:D)} + \text{VC(D)}\) \(\tau = 0.092 + 1.030 = 1.122\) \(\delta = \text{VC(PI:D)}/\text{Level}_{P:D}(\text{PI:D})\) \(\delta = 0.157/5 = 0.031\) \(E\rho^2 = 1.122/(1.122 + 0.031)\) \(E\rho^2 = \mathbf{0.973}\) |
\(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(I)}\) \(\tau = 0.028\) \(\delta = \text{VC(IxD)}/\text{Level}_I(\text{IxD}) + \text{VC(PI:D)}/\text{Level}_I(\text{PI:D})\) \(\delta = 0/2.61 + 0.157/16 = 0.010\) \(E\rho^2 = 0.028/(0.028 + 0.010)\) \(E\rho^2 = \mathbf{0.737}\) |
\(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(D)}\) \(\tau = 1.030\) \(\delta = \text{VC(P:D)}/\text{Level}_D(\text{P:D}) + \text{VC(IxD)}/\text{Level}_D(\text{IxD}) + \text{VC(PI:D)}/\text{Level}_D(\text{PI:D})\) \(\delta = 0.092/4.557 + 0/5 + 0.157/22.785 = 0.027\) \(E\rho^2 = 1.030/(1.030 + 0.027)\) \(E\rho^2 = \mathbf{0.974}\) |
Dependability (D) | \(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (0.028/5) + (0/5) + (0.157/5) = 0.037\) \(\Phi = 1.122/(1.122 + 0.037)\) \(\Phi = \mathbf{0.968}\) |
\(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (1.030/2.61) + (0.092/16) + (0/2.61) + (0.157/16) = 0.410\) \(\Phi = 0.028/(0.028 + 0.410)\) \(\Phi = \mathbf{0.064}\) |
\(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (0.028/5) + (0.092/4.557) + (0/5) + (0.157/22.785) = 0.033\) \(\Phi = 1.030/(1.030 + 0.033)\) \(\Phi = \mathbf{0.969}\) |
Up to this point, there have been no missing data points in the calculations. However, many data sets do have missing data (in fact Henderson’s 1953 data set is mostly missing data). The procedure for calculating the variance components and denominators for calculating G and D values is presented next. Note that NO Decision-study follow ups can be done with missing data sets. We will use Narayanan et al.’s original (2010) data set again to go through the example with the following changes: There is 1 data point missing for each “doctor” (highlighted in red) (see Table 24). Note that the “counts” of ratings that go into the denominator of the “Squared sums” VARIES as opposed to remaining constant across the facet.
Doctor | Patient | item1 | item2 | item3 | item4 | item5 | Sum across Patient Ratings | Squared sum across Patient Ratings | Squared sum across patient ratings/#ratings each completed |
---|---|---|---|---|---|---|---|---|---|
A | 1 | missing | 4 | 4 | 4 | 4 | 16 | 256 | 64 |
A | 2 | 4 | 4 | 3 | 3 | 4 | 18 | 324 | 64.8 |
A | 3 | 3 | 4 | 4 | 3 | 4 | 18 | 324 | 64.8 |
A | 4 | 3 | 3 | 3 | 3 | 3 | 15 | 225 | 45 |
A | 5 | 3 | 3 | 4 | 4 | 3 | 17 | 289 | 57.8 |
A | 6 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
A | 7 | 3 | 3 | 3 | 3 | 3 | 15 | 225 | 45 |
A | 8 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
B | 9 | 4 | 4 | 4 | 4 | 3 | 19 | 361 | 72.2 |
B | 10 | 4 | missing | 4 | 4 | 4 | 16 | 256 | 64 |
B | 11 | 4 | 4 | 4 | 4 | 4 | 20 | 400 | 80 |
B | 12 | 4 | 3 | 3 | 3 | 3 | 16 | 256 | 51.2 |
B | 13 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
C | 14 | 4 | 4 | 3 | 3 | 3 | 17 | 289 | 57.8 |
C | 15 | 3 | 3 | missing | 3 | 3 | 12 | 144 | 36 |
C | 16 | 4 | 4 | 4 | 4 | 3 | 19 | 361 | 72.2 |
Item Rating Sums | 55 | 55 | 52 | 54 | 53 | ||||
Item Rating sums squared | 3025 | 3025 | 2704 | 2916 | 2809 | ||||
Item sums squared/#ratings For each item | 201.667 | 201.6667 | 180.2667 | 182.25 | 175.562 |
Next we calculate the T-Values for each facet:
Sum all ratings across each item for each Patient; Sum the individual patient ratings for each doctor; square these sums; divide by the number of rating counts within each doctor; add the quotients.
\[\text{Doctor A} = \frac{(16+18+18+15+17+17+15+17)^2}{39} = 453.564\]
\[\text{Doctor B} = \frac{(19+16+20+16+17)^2}{24} = 322.667\]
\[\text{Doctor C} = \frac{(17+12+19)^2}{14} = 164.571\]
\[\text{Sum across all 3 quotients} = 940.802\]
Sum all ratings across each item for each Patient; Square these sums; divide by the number of rating counts within each patient; sum across these quotients.
\[\text{p:d 1} = \frac{16^2}{4} = \frac{256}{4} = 64\]
\[\text{p:d 2} = \frac{18^2}{5} = \frac{324}{5} = 64.8\]
\[\ldots\]
\[\text{p:d 15} = \frac{12^2}{4} = \frac{144}{4} = 36\]
\[\text{p:d 16} = \frac{19^2}{5} = \frac{361}{5} = 72.2\]
\[\text{Sum across all 16 quotients} = 948.2\]
Sum all ratings down each item; Square these sums; divide by the number of rating counts within each item; sum across these quotients.
\[\text{Item 1} = \frac{55^2}{15} = 201.667\]
\[\text{Item 2} = \frac{55^2}{15} = 201.667\]
\[\text{Item 3} = \frac{52^2}{15} = 180.267\]
\[\text{Item 4} = \frac{54^2}{16} = 182.250\]
\[\text{Item 5} = \frac{53^2}{16} = 175.563\]
\[\text{Sum across all 5 quotients} = 941.413\]
Sum down each set of items for each doctor (there will be 15 combinations); square these sums; divide each sum by the number of ratings that go into each of the 15 combinations; sum the quotients.
\[\text{di1} = (4+3+3+3+4+3+4) = 24; \frac{24^2}{7} = \frac{576}{7} = 82.286\]
\[\text{di2} = (4+4+4+3+3+4+3+4) = 29; \frac{29^2}{8} = \frac{841}{8} = 105.125\]
\[\ldots\]
\[\text{di6} = (4+4+4+4+4) = 20; \frac{20^2}{5} = \frac{400}{5} = 80.000\]
\[\text{di7} = (4+4+3+4) = 15; \frac{15^2}{4} = \frac{225}{4} = 56.250\]
\[\ldots\]
\[\text{di14} = (3+3+4) = 10; \frac{10^2}{3} = \frac{100}{3} = 33.333\]
\[\text{di15} = (3+3+3) = 9; \frac{9^2}{3} = \frac{81}{3} = 27.000\]
\[\text{Sum across all quotients} = 943.311\]
Note that this is the total uncorrected sums of squares. Each individual rating is first squared. Then these are summed across all ratings.
Going across each row:
\[0^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 4^2 + 3^2 + \ldots + 3^2 + 3^2 + 0^2 + 3^2 + 3^2 + 4^2 + 4^2 + 4^2 + 4^2 + 3^2\]
\[\text{Summing across all values} = 959.00\]
Sum FIRST across all values; square this sum; divide by the total number of ratings that went into the sum.
\[\frac{(0 + 4 + 4 + 4 + 4 + 4 + 4 + 4 + 4 + 3 + \ldots + 3 + 3 + 0 + 3 + 3 + 4 + 4 + 4 + 4 + 3)^2}{77} = \frac{269^2}{77} = 939.753\]
We can now put our Facet T-values into Table 25.
Facet | T-Value |
---|---|
D (doctor) | 940.802 |
P:D (patient nested in doctor) | 948.2 |
I (item) | 941.413 |
DI (doctor X item) | 943.311 |
PI:D (patient X item nested in doctor) | 959 |
Mean | 939.753 |
As always, obtaining the “counts” that go into the analysis of the variance components, is tedious. It becomes even more tedious when there are missing values. Let’s begin….
As before, we first put in the “easy” ones.
This leaves 16 “?” values to “fill in” (Table 26).
effect | counts D | counts P:D | counts I | counts DI | counts pi:d | counts mu | T-values |
---|---|---|---|---|---|---|---|
D | 77 | ? | ? | ? | 3 | 77 | 940.802 |
P:D | ? | 77 | ? | ? | 16 | 77 | 948.200 |
I | ? | ? | 77 | ? | 5 | 77 | 941.413 |
DI | ? | ? | ? | 77 | 15 | 77 | 943.311 |
pi:D,e | 77 | 77 | 77 | 77 | 77 | 77 | 959.000 |
Mean | ? | ? | ? | ? | 1 | 77 | 939.753 |
The σ²(p:d) term on D (data are collapsed across Items). Sample size starts by calculating the squared sum of:
counts of each p:d term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[ \text{D1}: \frac{(4^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2)}{39} = \frac{191}{39} = 4.897 \]
\[ \text{D2}: \frac{(5^2) + (4^2) + (5^2) + (5^2) + (5^2)}{24} = \frac{116}{24} = 4.833 \]
\[ \text{D3}: \frac{(5^2) + (4^2) + (5^2)}{14} = \frac{66}{14} = 4.714 \]
\[ \text{Sum} = 4.897 + 4.833 + 4.714 = 14.444 \]
The σ²(i) term on D (data are collapsed across P:D). Sample size starts by calculating the squared sum of:
counts of each i term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[ \text{D}_1: \frac{(7^2) + (8^2) + (8^2) + (8^2) + (8^2)}{39} = \frac{305}{39} = 7.820 \]
\[ \text{D}_2: \frac{(5^2) + (4^2) + (5^2) + (5^2) + (5^2)}{24} = \frac{116}{24} = 4.833 \]
\[ \text{D}_3: \frac{(3^2) + (3^2) + (2^2) + (3^2) + (3^2)}{14} = \frac{40}{14} = 2.857 \]
\[ \text{Sum} = 7.820 + 4.833 + 2.857 = 15.510 \]
The σ²(di) term on D (data are collapsed across P:D). Sample size starts by calculating the squared sum of:
counts of each di term within each D, divided by the total number of counts for that D. Then sum the quotients.
\[ \text{D}_1: \frac{(7^2) + (8^2) + (8^2) + (8^2) + (8^2)}{39} = \frac{305}{39} = 7.820 \]
\[ \text{D}_2: \frac{(5^2) + (4^2) + (5^2) + (5^2) + (5^2)}{24} = \frac{116}{24} = 4.833 \]
\[ \text{D}_3: \frac{(3^2) + (3^2) + (2^2) + (3^2) + (3^2)}{14} = \frac{40}{14} = 2.857 \]
\[ \text{Sum} = 7.820 + 4.833 + 2.857 = 15.510 \]
The σ²(d) term on P:D (data are collapsed across Items). Sample size starts by calculating the squared sum of:
counts of each d term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[ \text{PD}_1 = \frac{(4^2)}{4} = 4 \]
\[ \text{PD}_2 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_3 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_4 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_5 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_6 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_7 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_8 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_9 = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_{10} = \frac{(4^2)}{4} = 4 \]
\[ \text{PD}_{11} = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_{12} = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_{13} = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_{14} = \frac{(5^2)}{5} = 5 \]
\[ \text{PD}_{15} = \frac{(4^2)}{4} = 4 \]
\[ \text{PD}_{16} = \frac{(5^2)}{5} = 5 \]
\[ \text{Sum} = 77 \]
The σ²(i) term on P:D. Sample size starts by calculating the squared sum of:
counts of each i term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[ \text{PD}_1 = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_2 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_3 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_4 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_5 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_6 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_7 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_8 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_9 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{10} = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_{11} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{12} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{13} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{14} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{15} = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_{16} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{Sum} = 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 = 16 \]
The σ²(di) term on P:D. Sample size starts by calculating the squared sum of:
counts of each di term within each P:D, divided by the total number of counts for that P:D. Then sum the quotients.
\[ \text{PD}_1 = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_2 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_3 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_4 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_5 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_6 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_7 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_8 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_9 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{10} = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_{11} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{12} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{13} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{14} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{PD}_{15} = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{PD}_{16} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{Sum} = 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 = 16 \]
The σ²(d) term on I. Sample size starts by calculating the squared sum of:
counts of each d term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[ \text{I}_1 = \frac{(7^2) + (5^2) + (3^2)}{15} = \frac{83}{15} = 5.533 \]
\[ \text{I}_2 = \frac{(8^2) + (4^2) + (3^2)}{15} = \frac{89}{15} = 5.933 \]
\[ \text{I}_3 = \frac{(8^2) + (5^2) + (2^2)}{15} = \frac{93}{15} = 6.200 \]
\[ \text{I}_4 = \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 \]
\[ \text{I}_5 = \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 \]
\[ \text{Sum} = 5.533 + 5.933 + 6.200 + 6.125 + 6.125 = 29.916 \]
The σ²(p:d) term on I. Sample size starts by calculating the squared sum of:
counts of each p:d term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[ \text{I}_1 = \frac{\sum_{v=1}^{15}(1^2)}{15} = \frac{15}{15} = 1 \]
\[ \text{I}_2 = \frac{\sum_{v=1}^{15}(1^2)}{15} = \frac{15}{15} = 1 \]
\[ \text{I}_3 = \frac{\sum_{v=1}^{15}(1^2)}{15} = \frac{15}{15} = 1 \]
\[ \text{I}_4 = \frac{\sum_{v=1}^{16}(1^2)}{16} = \frac{16}{16} = 1 \]
\[ \text{I}_5 = \frac{\sum_{v=1}^{16}(1^2)}{16} = \frac{16}{16} = 1 \]
\[ \text{Sum} = 1 + 1 + 1 + 1 + 1 = 5 \]
The σ²(di) term on I. Sample size starts by calculating the squared sum of:
counts of each di term within each I, divided by the total number of counts for that I. Then sum the quotients.
\[ \text{I}_1 = \frac{(7^2) + (5^2) + (3^2)}{15} = \frac{83}{15} = 5.533 \]
\[ \text{I}_2 = \frac{(8^2) + (4^2) + (3^2)}{15} = \frac{89}{15} = 5.933 \]
\[ \text{I}_3 = \frac{(8^2) + (5^2) + (2^2)}{15} = \frac{93}{15} = 6.200 \]
\[ \text{I}_4 = \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 \]
\[ \text{I}_5 = \frac{(8^2) + (5^2) + (3^2)}{16} = \frac{98}{16} = 6.125 \]
\[ \text{Sum} = 5.533 + 5.933 + 6.200 + 6.125 + 6.125 = 29.916 \]
The σ²(d) term on DI. Sample size starts by calculating the squared sum of:
counts of each d term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[ \text{DI}_1 = \frac{7^2}{7} = 7 \]
\[ \text{DI}_2 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_3 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_4 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_5 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_6 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_7 = \frac{4^2}{4} = 4 \]
\[ \text{DI}_8 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_9 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_{10} = \frac{5^2}{5} = 5 \]
\[ \text{DI}_{11} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{12} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{13} = \frac{2^2}{2} = 2 \]
\[ \text{DI}_{14} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{15} = \frac{3^2}{3} = 3 \]
\[ \text{Sum} = 7 + 8 + 8 + 8 + 8 + 5 + 4 + 5 + 5 + 5 + 3 + 3 + 2 + 3 + 3 = 77 \]
The σ²(p:d) term on DI. Sample size starts by calculating the squared sum of:
counts of each p:d term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[ \text{DI}_1 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{7} = 1 \]
\[ \text{DI}_2 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 \]
\[ \text{DI}_3 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 \]
\[ \text{DI}_4 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 \]
\[ \text{DI}_5 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{8} = 1 \]
\[ \text{DI}_6 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{DI}_7 = \frac{(1^2) + (1^2) + (1^2) + (1^2)}{4} = 1 \]
\[ \text{DI}_8 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{DI}_9 = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{DI}_{10} = \frac{(1^2) + (1^2) + (1^2) + (1^2) + (1^2)}{5} = 1 \]
\[ \text{DI}_{11} = \frac{(1^2) + (1^2) + (1^2)}{3} = 1 \]
\[ \text{DI}_{12} = \frac{(1^2) + (1^2) + (1^2)}{3} = 1 \]
\[ \text{DI}_{13} = \frac{(1^2) + (1^2)}{2} = 1 \]
\[ \text{DI}_{14} = \frac{(1^2) + (1^2) + (1^2)}{3} = 1 \]
\[ \text{DI}_{15} = \frac{(1^2) + (1^2) + (1^2)}{3} = 1 \]
\[ \text{Sum} = 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 + 1 = 15 \]
The σ²(i) term on DI. Sample size starts by calculating the squared sum of:
counts of each i term within each DI, divided by the total number of counts for that DI. Then sum the quotients.
\[ \text{DI}_1 = \frac{7^2}{7} = 7 \]
\[ \text{DI}_2 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_3 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_4 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_5 = \frac{8^2}{8} = 8 \]
\[ \text{DI}_6 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_7 = \frac{4^2}{4} = 4 \]
\[ \text{DI}_8 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_9 = \frac{5^2}{5} = 5 \]
\[ \text{DI}_{10} = \frac{5^2}{5} = 5 \]
\[ \text{DI}_{11} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{12} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{13} = \frac{2^2}{2} = 2 \]
\[ \text{DI}_{14} = \frac{3^2}{3} = 3 \]
\[ \text{DI}_{15} = \frac{3^2}{3} = 3 \]
\[ \text{Sum} = 7 + 8 + 8 + 8 + 8 + 5 + 4 + 5 + 5 + 5 + 3 + 3 + 2 + 3 + 3 = 77 \]
The σ²(d) term on mean. Sample size starts by calculating the squared sum of:
counts of each d term within each mean, divided by the total number of counts for the mean.
\[ \text{Mean} = \frac{(39^2) + (24^2) + (14^2)}{77} = \frac{1521 + 576 + 196}{77} = \frac{2293}{77} = 29.779 \]
The σ²(p:d) term on mean. Sample size starts by calculating the squared sum of:
counts of each p:d term within each mean, divided by the total number of counts for the mean.
\[ \text{Mean} = \frac{(4^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (4^2) + (5^2) + (5^2) + (5^2) + (5^2) + (5^2) + (4^2) + (5^2)}{77} \]
\[ = \frac{(16 \times 3) + (25 \times 13)}{77} \]
\[ = \frac{48 + 325}{77} \]
\[ = \frac{373}{77} = 4.844 \]
The σ²(i) term on mean. Sample size starts by calculating the squared sum of:
counts of each i term within each mean, divided by the total number of counts for the mean
\[ \text{Mean} = \frac{(15^2) + (15^2) + (15^2) + (16^2) + (16^2)}{77} = \frac{(225 \times 3) + (256 \times 2)}{77} = \frac{675 + 512}{77} = \frac{1187}{77} = 15.416 \]
The σ²(di) term on mean. Sample size starts by calculating the squared sum of:
counts of each di term within each mean, divided by the total number of counts for the mean.
\[ \text{Mean} = \frac{(7^2) + (5^2) + (3^2) + (8^2) + (4^2) + (3^2) + (8^2) + (5^2) + (2^2) + (8^2) + (5^2) + (3^2) + (8^2) + (5^2) + (3^2)}{77} \]
\[ = \frac{(64 \times 4) + (25 \times 4) + (9 \times 4) + 49 + 16 + 4}{77} \]
\[ = \frac{256 + 100 + 36 + 69}{77} \]
\[ = \frac{461}{77} = 5.987 \]
We can now insert the calculated values into our “?” cells (Table 27) and using matrix equations or regression, solve for the Variance Components.
effect | counts D | counts P:D | counts I | counts DI | counts pi:d | counts mu | T-values | variance component |
---|---|---|---|---|---|---|---|---|
D | 77 | 14.444 | 15.510 | 15.510 | 3 | 77 | 940.802 | 0.001 |
P:D | 77 | 77 | 16 | 16 | 16 | 77 | 948.200 | 0.083 |
I | 29.916 | 5 | 77 | 29.916 | 5 | 77 | 941.413 | 0.021 |
DI | 77 | 15 | 77 | 77 | 15 | 77 | 943.311 | -0.014 |
PI:D,e | 77 | 77 | 77 | 77 | 77 | 77 | 959.000 | 0.170 |
Mean | 29.779 | 4.844 | 15.416 | 5.987 | 1 | 77 | 939.753 | 12.1936 |
We now need the denominators to use when calculating G and D coefficients. They are calculated using the same approach as that used for unbalanced designs with the results in Table 28.
Level of the σ²(d) term on facet of differentiation P:D. Since d is included in P:D, there is a single unique D for each P:D. L = 1
Level of the σ²(p:d) term on facet of differentiation P:D. Again p:d is included in P:D, there is a single unique P:D for each P:D. L = 1
Level of the σ²(i) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, i) = 1, v = \(\left|items_{g}\right|\) : 1 count of item for each unique combinations of p:d, however since there is missing data the length of items for each combination varies (for this data set, either 4 or 5). Thus two potential calculations exist:
\[ Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
\[ Lg’ = \frac{\left(\sum_{v=1}^{4}1\right)^{2}}{\sum_{v=1}^{4}1^{2}} = 4 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{16}{\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}} = 4.78 \]
Level of the σ²(i x d) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, ixd) = 1, v = \(\left|items_{g}\right|\) : Again, 1 count of item x d for each unique combinations of p:d, however since there is missing data the length of items for each combination varies (for this data set, either 4 or 5). Thus two potential calculations exist:
\[ Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
\[ Lg’ = \frac{\left(\sum_{v=1}^{4}1\right)^{2}}{\sum_{v=1}^{4}1^{2}} = 4 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{16}{\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}} = 4.78 \]
Level of the σ²(i x (p:d)) term on facet of differentiation P:D. The sum of counts squared and sum of squared counts can be determined as follows:
C(p:d, pi:d) = 1, v =[1, \(\left|items_{g}\right|\) ]: Similarly, 1 count of items x (p:d) for each unique combinations of p:d, however since there is missing data the length of items for each combination varies (for this data set, either 4 or 5). Thus two potential calculations exist:
\[ Lg = \frac{\left(\sum_{v=1}^{5}1\right)^{2}}{\sum_{v=1}^{5}1^{2}} = 5 \]
\[ Lg’ = \frac{\left(\sum_{v=1}^{4}1\right)^{2}}{\sum_{v=1}^{4}1^{2}} = 4 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{16}{\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{5}+\frac{1}{4}+\frac{1}{5}} = 4.78 \]
Level of the σ²(d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, d) = \(\left|p_{v}\right|\) , for p = [8, 5, 3*] with v ∈ [1, 3]; g ∈ [1, 5] and *indicating missing data. Since i is not involved with the unbalanced nesting, all levels of i, i ∈ [1, 5], will be the same.
\[ L = L1 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(7+5+3\right)^{2}}{(7^{2}+5^{2}+3^{2})} = \frac{225}{83} = 2.71 \]
\[ L = L2 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+4+3\right)^{2}}{(8^{2}+4^{2}+3^{2})} = \frac{225}{89} = 2.53 \]
\[ L = L3 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+2\right)^{2}}{(8^{2}+5^{2}+2^{2})} = \frac{225}{93} = 2.42 \]
\[ L4 = L5 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+3\right)^{2}}{(8^{2}+5^{2}+3^{2})} = \frac{256}{98} = 2.612 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{5}{\frac{1}{2.71}+\frac{1}{2.53}+\frac{1}{2.42}+\frac{1}{2.61}+\frac{1}{2.61}} = 2.57 \]
Level of the σ²(p:d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, p:d) = 1, v ∈ [1, \(\sum_{}^{}g\) ]. For each unique i, there will be 1 count up to the total counts for that i.
\[ L1 = L2 = L3 = \frac{\left(\sum_{v=1}^{15}1\right)^{2}}{\sum_{v=1}^{15}1^{2}} = \frac{\left(1+1+1+1+1+1+1+1+1+1+1+1+1+1+1\right)^{2}}{(1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2})} = \frac{225}{15} = 15 \]
\[ L4 = L5 = \frac{\left(\sum_{v=1}^{16}1\right)^{2}}{\sum_{v=1}^{16}1^{2}} = \frac{\left(1+1+1+1+1+1+1+1+1+1+1+1+1+1+1+1\right)^{2}}{(1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2})} = \frac{256}{16} = 16 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{5}{\frac{1}{15}+\frac{1}{15}+\frac{1}{15}+\frac{1}{16}+\frac{1}{16}} = 15.38 \]
Level of the σ²(i) term on facet of differentiation I. Since i is included in I, there is a single unique i for each I. L = 1
Level of the σ²(i x d) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, d) = \(\left|p_{v}\right|\) , for p = [8, 5, 3*] with v ∈ [1, 3]; g ∈ [1, 5] and *indicating missing data, similar to 6. Since i is not involved with the unbalanced nesting, all levels of i, i ∈ [1, 5], will be the same.
\[ L = L1 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(7+5+3\right)^{2}}{(7^{2}+5^{2}+3^{2})} = \frac{225}{83} = 2.71 \]
\[ L = L2 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+4+3\right)^{2}}{(8^{2}+4^{2}+3^{2})} = \frac{225}{89} = 2.53 \]
\[ L = L3 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+2\right)^{2}}{(8^{2}+5^{2}+2^{2})} = \frac{225}{93} = 2.42 \]
\[ L4 = L5 = \frac{\left(\sum_{v=1}^{3}p_{v}\right)^{2}}{\sum_{v=1}^{3}p_{v}^{2}} = \frac{\left(8+5+3\right)^{2}}{(8^{2}+5^{2}+3^{2})} = \frac{256}{98} = 2.612 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{5}{\frac{1}{2.71}+\frac{1}{2.53}+\frac{1}{2.42}+\frac{1}{2.61}+\frac{1}{2.61}} = 2.57 \]
Level of the σ²(i x (p:d)) term on facet of differentiation I. The sum of counts squared and sum of squared counts can be determined as follows:
C(i, p:d) = 1, v ∈ [1, \(\sum_{}^{}g\) ]. Similar to 7, For each unique i, there will be 1 count up to the total counts for that i.
\[ L1 = L2 = L3 = \frac{\left(\sum_{v=1}^{15}1\right)^{2}}{\sum_{v=1}^{15}1^{2}} = \frac{\left(1+1+1+1+1+1+1+1+1+1+1+1+1+1+1\right)^{2}}{(1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2})} = \frac{225}{15} = 15 \]
\[ L4 = L5 = \frac{\left(\sum_{v=1}^{16}1\right)^{2}}{\sum_{v=1}^{16}1^{2}} = \frac{\left(1+1+1+1+1+1+1+1+1+1+1+1+1+1+1+1\right)^{2}}{(1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2}+1^{2})} = \frac{256}{16} = 16 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{5}{\frac{1}{15}+\frac{1}{15}+\frac{1}{15}+\frac{1}{16}+\frac{1}{16}} = 15.38 \]
\[ L = 1 \]
Level of the σ²(p:d) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, p:d) = \(i_{v}\) , where = \(i_{v}\) is the number of items in v for [g =A, v ∈ [1, 8]; g=B, v ∈ [1, 5]; g=C, v ∈ [1, 3]]. Thus we have different levels of patients:doctors for each unique doctor.
\[ L_A = \frac{\left(\sum_{v=1}^{8}i_{v}, \right)^{2}}{\sum_{v=1}^{8}i_{v} ^{2}} = \frac{\left(4+5+5+5+5+5+5+5\right)^{2}}{(4^{2}+5^{2}+5^{2}+5^{2}+5^{2}+5^{2}+5^{2}+5^{2})} = 7.96 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}i_{v}\right)^{2}}{\sum_{v=1}^{5}i_{v}^{2}} = \frac{\left(4+5+5+5+5\right)^{2}}{(4^{2}+5^{2}+5^{2}+5^{2}+5^{2})} = 4.97 \]
\[ L_C = \frac{\left(\sum_{v=1}^{3}i_{v}\right)^{2}}{\sum_{v=1}^{3}i_{v}^{2}} = \frac{\left(4+5+5\right)^{2}}{(4^{2}+5^{2}+5^{2})} = 2.97 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{7.96}+\frac{1}{4.97}+\frac{1}{2.97}} = 4.52 \]
Level of the σ²(i) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, i) = \(\left|g_{v}\right|\) , for g=[8, 5, 3] and v = 5. There are g counts of each item [1, 5]. For example there are 7 counts of item [1] for doctor A because 7 patients returned surveys including item 1 for doctor A.
\[ L_A = \frac{\left(\sum_{v=1}^{5}g_{v} \right)^{2}}{\sum_{v=1}^{5}g_{v} ^{2}} = \frac{\left(7+8+8+8+8\right)^{2}}{(7^{2}+8^{2}+8^{2}+8^{2}+8^{2})} = 4.99 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}g_{v}\right)^{2}}{\sum_{v=1}^{5}g_{v}^{2}} = \frac{\left(5+4+5+5+5\right)^{2}}{(5^{2}+4^{2}+5^{2}+5^{2}+5^{2})} = 4.97 \]
\[ L_C = \frac{\left(\sum_{v=1}^{5}g_{v}\right)^{2}}{\sum_{v=1}^{5}g_{v}^{2}} = \frac{\left(3+3+2+3+3\right)^{2}}{(3^{2}+3^{2}+2^{2}+3^{2}+3^{2})} = 4.9 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{4.99}+\frac{1}{4.97}+\frac{1}{4.90}} = 4.95 \]
Level of the σ²(i x d) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, ixd) = \(\left|g_{v}\right|\) , for g=[8, 5, 3] and v = 5. There are g counts of each item [1, 5]. This is identical to the uncrossed items, C(d, i).
\[ L_A = \frac{\left(\sum_{v=1}^{5}g_{v} \right)^{2}}{\sum_{v=1}^{5}g_{v} ^{2}} = \frac{\left(7+8+8+8+8\right)^{2}}{(7^{2}+8^{2}+8^{2}+8^{2}+8^{2})} = 4.99 \]
\[ L_B = \frac{\left(\sum_{v=1}^{5}g_{v}\right)^{2}}{\sum_{v=1}^{5}g_{v}^{2}} = \frac{\left(5+4+5+5+5\right)^{2}}{(5^{2}+4^{2}+5^{2}+5^{2}+5^{2})} = 4.97 \]
\[ L_C = \frac{\left(\sum_{v=1}^{5}g_{v}\right)^{2}}{\sum_{v=1}^{5}g_{v}^{2}} = \frac{\left(3+3+2+3+3\right)^{2}}{(3^{2}+3^{2}+2^{2}+3^{2}+3^{2})} = 4.9 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{4.99}+\frac{1}{4.97}+\frac{1}{4.90}} = 4.95 \]
Level of the σ²(i x (p:d)) term on facet of differentiation D. The sum of counts squared and sum of squared counts can be determined as follows:
C(d, i x (p:d)) = 1, for v=[ \(\sum_{}^{}g\) ] with g=[A, B, C]. For each doctor, there is a single unique count of items crossed with the nested patient:doctor. Thus to obtain the level for each unique facet of differentiation, g, we must sum over v, which is the sum of counts for each doctor (marginalizing over doctor).
\[ L_A = \frac{\left(\sum_{v=1}^{39}1\right)^{2}}{\sum_{v=1}^{39}1^{2}} = \frac{1521}{39} = 39 \]
\[ L_A = \frac{\left(\sum_{v=1}^{24}1\right)^{2}}{\sum_{v=1}^{24}1^{2}} = \frac{576}{24} = 24 \]
\[ L_A = \frac{\left(\sum_{v=1}^{14}1\right)^{2}}{\sum_{v=1}^{14}1^{2}} = \frac{196}{14} = 14 \]
\[ L = \frac{\left|G\right|}{\sum_{g=1}^{G}\frac{1}{L_{g}}} = \frac{3}{\frac{1}{39}+\frac{1}{24}+\frac{1}{14}} = 21.62 \]
Variance Term | P:D Facet of Differentiation | I Facet of Differentiation | D Facet of Differentiation |
---|---|---|---|
σ²(d) | 1.000 | 2.573 | 1.000 |
σ²(p:d) | 1.000 | 15.385 | 4.520 |
σ²(i) | 4.776 | 1.000 | 4.951 |
σ²(i × d) | 4.776 | 2.573 | 4.951 |
σ²(i × (p:d)) | 4.776 | 15.385 | 21.624 |
These values can now be used to calculate the G and D coefficients. We will complete the one for P:D (Table 29). Note that the values for G and D are slightly lower (.700 and .675, respectively) than those with no missing data (.752 and .718, respectively), which is to be expected.
Generalizability Coefficient | Facet of Differentiation = P:D |
---|---|
Generalizability (G) | \(E\rho^2 = \tau/(\tau + \delta)\)
\(\tau = \text{VC(P:D)} + \text{VC(D)}\) \(\tau = 0.083 + 0.001 = 0.084\) \(\delta = \text{VC(PI:D)}/\text{Level}_{P:D}(\text{PI:D})\) \(\delta = 0.170/4.776 = 0.036\) \(E\rho^2 = 0.084/(0.084+0.036)\) \(E\rho^2 = \mathbf{0.700}\) |
Dependability (D) | \(\Phi = \tau/(\tau + \Delta)\)
\(\Delta = (0.021/4.776) + (0/4.776) + (0.170/4.776) = 0.040\) \(\Phi = 0.084/(0.084 + 0.040)\) \(\Phi = \mathbf{0.675}\) |
While the only design that has been examined in great detail in this tutorial is the \(i \times (p:d)\) design, all other designs (crossed, nested, etc. – perhaps using many more terms) are theoretically sound with tests passing for many different synthetic datasets (see GeneralizIT GitHub repository). The only limiting factor for the automated calculation of generalizability coefficients for any study design is the data analyst’s ability to properly specify the linear random effects relationships present in a given study. For example, in the \(i \times (p:d)\) design:
\[X = \mu + \sigma_{i}^{2} + \sigma_{d}^{2} + \sigma_{p:d}^{2} + \sigma_{id}^{2} + \sigma_{pi:d}^{2}\]
The utility in creating a generalizability program allows for: 1) missing data, 2) unbalanced data, 3) providing both generalizability and dependability coefficients, and 4) D-values for designs with no missing data is very useful addition to the toolbox of researchers. There is currently not a freely-available one that does all of these things.
References
Bloch, R. & Norman, G. (2023). G String V User Manual. Hamilton, Ontario, Canada.
Bloch, R. & Norman, G. (2012). Generalizability theory for the perplexed: A practical introduction and guide: AMEE Guide No. 68. Medical Teacher, 34 (11), 960-992. DOI: 10.3109/0142159X.2012.703791
Brennan, R. L. (2001a). Generalizability Theory. New York: Springer.
B Brennan, R. L. (2001b). Manual for urGENOVA (Version 2.1) (Iowa Testing Programs Occasional Paper Number 49). Iowa City, IA: Iowa Testing Programs, University of Iowa.
Riesch, A.M., Swaminathan, H., Welsh, M. & Chafouleas, S.M. (2014). Generalizability theory: A practical guide to study design, implementation, and interpretation, Journal of School Psychology, 52 (1), 13-35.
Henderson, C.R. (1953). Estimation of variance and covariance components. Biometrics, 9(2), 226-252.
Narayanan, A., Greco, M., & Campbell, J.L. (2010). Generalisability in unbalanced, uncrossed and fully nested studies. Medical Education, 44(4), 367-387.
Water Quality Division (2010). Procedures to Implement the Texas Surface-Water-Quality Standards: TCEQ RG–194, p. 81.
Harmonic means are an average. It is calculated by:
Mathematically equivalently: taking the sum of the number of values in the data series and dividing it by the sum of the reciprocals of each value in a data series.
Harmonic means are most useful when you are dealing with a set of values based on rates and ratios (i.e., working with reciprocal relationships). Ratios (such as speed (km/hr); Price/Earnings ratios; and Nested variables #items/form, #patients/doctor) are such instances. Harmonic means are thus useful in unbalanced data designs where there are non-equivalent numbers of items nested within another factor (e.g., i:h or p:h). Clearly, these are ratios.
These websites are useful to differentiate between arithmetic, geometric and harmonic means:
Using the simple example of: What is the average speed of a vehicle that drives out 60 km at 60/km/hr and back the same distance at 20 km/hr?
The arithmetic mean would be (60+20)/2hrs = 80/2 = 40 km/hr. However, this is NOT the average speed travelled over a fixed distance (i.e., “there and back”). The problem is that on the way back, the driver would have only covered only 1/3 of the distance back since they are going at 20/km/hr. To cover the while distance, they would have to drive 3 hours. We CAN mentally work this through and take the weighted arithmetic mean to get at the correct average speed: (60+20+20+20)/4 hrs = 120/4 = 30 km/hr is the average speed for the fixed distance when travelling 60 km/hr out and 20 km/hr back.
While this worked with a simple example, a much more elegant way to do this with many numbers that don’t evenly work out for weighting, is to take the reciprocal of the values (1/60 and 1/20), then average them:
\[\frac{1/60 + 1/20}{2} = \frac{0.016667 + 0.05}{2} = \frac{0.066667}{2} = 0.033333\]
Then to “get back” to our original units of km/hr, we take the reciprocal of this value:
\[\frac{1}{0.033333} = 29.99999 \approx 30\]
We could also skip a step and when we get to 0.066667/2, simply take the reciprocal of it:
\[\frac{2}{0.0666667} = 29.999999 \approx 30\]
In looking at the reciprocal values, you can see that the larger the value to start with, the smaller its reciprocal value. By taking the average of the reciprocals, you are equally weighting each value before averaging them. This dampens the effect of larger values in the data set that is exerted with the arithmetic mean.
For example: The arithmetic mean of Brennan’s (2001a) data set (Table 7.5, p. 224) with 3 h’s (let’s refer to them as Forms) and within each Form there are 2, 4, and 2 items. So, in effect we have three ratio values: 2items/1 form, 4items/1 form, 2items/1 form.
The arithmetic mean of these values is: \[\frac{2+4+2}{3} = \frac{8}{3} = 2.667\]
The harmonic mean of these values is: \[\frac{3}{(1/2 + 1/4 + 1/2)} = \frac{3}{(0.5+0.25+0.5)} = \frac{3}{1.25} = 2.4\]
Why would we want the arithmetic versus the harmonic means?
The arithmetic mean (2.667) tells us that there are, on average, 2.667 items for each form for this data set. There are 8 persons (p) who take all the items in all the forms. If we take 2.667 and multiply it by the number of persons (8) we get 21.336 “counts” estimated for each form. If we multiply this value by the number of forms (3), we get 64 – the number of counts in the entire data set FOR PERSONS. In addition, for each person, (person 1 has 2.667 items for Form 1, 2.667 items for Form 2 and 2.667 items for Form 3) returns 8 counts (2.667 × 3) per person. This symmetry is useful to know and works when there are no missing data.
The harmonic mean “weights” the forms equally so that the effect of forms “assumes” they all have the same number of items in them (are pseudo-balanced). For example, Form 2 has 4 items in it; if we weight Forms 1 and 3 by doubling their number of items, then we can calculate the arithmetic average of:
\[ Form_{h1}(2+2) + Form_{h2}(4) + Form_{h3}(2+2) = 12 \]
5 values in the data set
\[\frac{12}{5} = 2.4\]
This equally weights the contribution of each of the forms when using the term in the denominators of the Generalizability equations.