The methods `CM`

(Cousineau, 2005;
Morey, 2008) and `LM`

(Loftus
& Masson, 1994) can be unified when the transformations they
required are considered. In the Cousineau-Morey method, the raw data
must be subject-centered and bias-corrected.

Subject centering is obtained from

\[ Y_{ij} = X_{ij} - \bar{X}_{i\cdot} + \bar{\bar{X}} \]

in which \(i = 1..n\) and \(j=1..C\) where \(n\) is the number of participants and \(C\) is the number of repeated measures (sometimes noted with \(J\)).

Bias-correction is obtained from

\[ Z_{ij} = \sqrt{\frac{C}{C-1}} \left( Y_{ij} - \bar{Y}_{\cdot{}j} \right) + \bar{Y}_{\cdot{}j} \]

These two operations can be performed with two matrix
transformations. In comparison, the `LM`

method requires one
additional step, that is, pooling standard deviation, also achievable
with the following transformation.

\[ W_{ij} = \sqrt{\frac{S_p^2}{S_i^2}} \left( Z_{ij} - \bar{Z}_{\cdot{}j} \right) + \bar{Z}_{\cdot{}j} \]

in which \(S_j^2\) is the variance in measurement \(j\) and \(S_p^2\) is the pooled variance across all \(j\) mesurements.

With this approach, we can categorize all the proposals to repeated measure precision as requiring or not certain transformations. Table 1 shows these.

**Table 1**. Transformations required to implement one
of the repeated-measures method. Preprocessing must precede
post-processing

Method | preprocessing | postprocessing | |
---|---|---|---|

Stand-alone | - | - | |

Cousineau, 2005 | Subject-centering | - | - |

CM | Subject-centering | Bias-correction | |

NKM | Subject-centering | - | pool standard deviations |

LM | Subject-centering | Bias-correction | pool standard deviations |

From that point of view, we see that the Nathoo, Kilshaw and Masson NKM (Nathoo, Kilshaw, & Masson, 2018; but see Heck, 2019 ) method is missing a bias-correction transformation, which explains why these error bars are shorter. The original proposal found in Cousineau, 2005, is also missing the bias correction step, which led Morey (2008) to supplement this approach. With these four approaches, we have exhausted all the possible combinations regarding decorrelation methods based on subject-centering.

We added two arguments in superbPlot to handle this transformation
approach, the first is `preprocessfct`

and the second is
`postprocessfct`

.

Assuming a dataset dta with replicated measures stored in say columns
called `Score.1`

, `Score.2`

and
`Score.3`

, the command

```
pCM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = "biasCorrectionTransform",
plotStyle = "pointjitter",
errorbarParams = list(color="red", width= 0.1, position = position_nudge(-0.05) )
)
```

will reproduce the `CM`

error bars because it decorrelates
the data as per this method. With one additional transformation,

```
pLM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = c("biasCorrectionTransform","poolSDTransform"),
plotStyle = "line",
errorbarParams = list(color="orange", width= 0.1, position = position_nudge(-0.0) )
)
```

the `LM`

method is reproduced. Finally, if the
`biasCorrectionTransform`

is omitted, we get the NKM error
bars with:

```
pNKM <- superbPlot(dta, WSFactors = "moment(3)",
variables = c("Score.1","Score.2","Score.3"),
adjustments=list(decorrelation="none"),
preprocessfct = "subjectCenteringTransform",
postprocessfct = c("poolSDTransform"),
plotStyle = "line",
errorbarParams = list(color="blue", width= 0.1, position = position_nudge(+0.05) )
)
```

In what follow, I justapose the three plots to see the differences:

```
tlbl <- paste( "(red) Subject centering & Bias correction == CM\n",
"(orange) Subject centering, Bias correction & Pooling SDs == LM\n",
"(blue) Subject centering & Pooling SDs == NKM", sep="")
ornate <- list(
xlab("Group"),
ylab("Score"),
labs( title=tlbl),
coord_cartesian( ylim = c(12,18) ),
theme_light(base_size=10)
)
# the plots on top are made transparent
pCM2 <- ggplotGrob(pCM + ornate)
pLM2 <- ggplotGrob(pLM + ornate + makeTransparent() )
pNKM2 <- ggplotGrob(pNKM + ornate + makeTransparent() )
# put the grobs onto an empty ggplot
ggplot() +
annotation_custom(grob=pCM2) +
annotation_custom(grob=pLM2) +
annotation_custom(grob=pNKM2)
```

The method from Cousineau (2005) missing the bias-correction step is not shown as it should not be used.

All the decorrelation methods based on transformations have (probably) been explored. An alternative approach using correlation was proposed in Cousineau (2019). All these approaches requires sphericity of the data. Other approaches are required to overcome this sphericity limitations.

Cousineau, D. (2005). Confidence intervals in within-subject designs: A
simpler solution to Loftus and Masson’s
method. *Tutorials in Quantitative Methods for Psychology*,
*1*, 42–45. https://doi.org/10.20982/tqmp.01.1.p042

Cousineau, D. (2019). Correlation-adjusted standard errors and
confidence intervals for within-subject designs: A simple multiplicative
approach. *The Quantitative Methods for Psychology*, *15*,
226–241. https://doi.org/10.20982/tqmp.15.3.p226

Heck, D. W. (2019). Accounting for estimation uncertainty and shrinkage
in bayesian within-subject intervals: A comment on Nathoo,
Kilshaw, and Masson (2018). *Journal of
Mathematical Psychology*, *88*, 27–31. https://doi.org/10.1016/j.jmp.2018.11.002

Loftus, G. R., & Masson, M. E. J. (1994). Using confidence intervals
in within-subject designs. *Psychonomic Bulletin & Review*,
*1*, 476–490. https://doi.org/10.3758/BF03210951

Morey, R. D. (2008). Confidence intervals from normalized data: A
correction to Cousineau (2005). *Tutorials in
Quantitative Methods for Psychology*, *4*, 61–64. https://doi.org/10.20982/tqmp.04.2.p061

Nathoo, F. S., Kilshaw, R. E., & Masson, M. E. J. (2018). A better
(bayesian) interval estimate for within-subject designs. *Journal of
Mathematical Psychology*, *86*, 1–9. https://doi.org/10.1016/j.jmp.2018.07.005