Applied Univariate, Bivariate, and Multivariate Statistics. Daniel J. Denis

Чтение книги онлайн.

Читать онлайн книгу Applied Univariate, Bivariate, and Multivariate Statistics - Daniel J. Denis страница 47

Applied Univariate, Bivariate, and Multivariate Statistics - Daniel J. Denis

Скачать книгу

rel="nofollow" href="#ulink_fa2348e5-5396-519d-9597-3e0672c3f3c5">Table 2.9.

      Now, here is the trick to understanding advanced modeling, including a primary feature of mixed effects modeling. We know that we expect the covariance between treatments to be unequal to 0. This is analogous to what we expected in the simple matched-pairs design. It seems then that a reasonable assumption to make for the data in Table 2.9 is that the covariances between treatments are equal, or at minimum, follow some hypothesized correlational structure. In multilevel and hierarchical models, attempts are made to account for the correlation between treatment levels instead of assuming these correlations to equal 0 as is the case for classical between‐subjects designs. In Chapter 6, we elaborate on these ideas when we discuss randomized block and repeated measures models.

      In many statistical techniques, especially multivariate ones, statistical analyses take place not on individual variables, but rather on linear combinations of variables. A linear combination in linear algebra can be denoted simply as:

equation

      where a ' = (a1, a2, …, ap). These values are scalars, and serve to weight the respective values of y1 through yp, which are the variables.

      Just as we did for “ordinary” variables, we can compute a number of central tendency and dispersion statistics on linear combinations. For instance, we can compute the mean of a linear combination ℓi as

equation

      We can also compute the sample variance of a linear combination:

equation

      For two linear combinations,

equation

      and

equation

      we can obtain the sample covariance between such linear combinations as follows:

equation

      The correlation of these linear combinations (Rencher and Christensen, 2012, p. 76) is simply the standardized version of images:

equation

      As we will see later in the book, if images is the maximum correlation between linear combinations on the same variables, it is called the canonical correlation, discussed in Chapter 12. The correlation between linear combinations plays a central role in multivariate analysis. Substantively, and geometrically, linear combinations can be interpreted as “projections” of one or more variables onto new dimensions. For instance, in simple linear regression, the fitting of a least‐squares line is such a projection. It is the projection of points such that it guarantees that the sum of squared deviations from the given projected line or “surface” (in the case of higher dimensions) is kept to a minimum.

      If we can assume multivariate normality of a distribution, that is, YN[μ, ], then we know linear combinations of Y are also normally distributed, as well as a host of other useful statistical properties (see Timm, 2002, pp. 86–88). In multivariate methods especially, we regularly need to make assumptions about such linear combinations, and it helps to know that so long as we can assume multivariate normality, we have some idea of how such linear combinations will be distributed.

      As an example of how matrices will be used to develop more complete and general models, consider the multivariate general linear model in matrix form:

equation

      where yi = 1 to yi = n are observed measurements on some dependent variable, X is the model matrix containing a constant of 1 in the first column to represent the common intercept term (i.e., “common” implying there is one intercept that represents all observations in our data), xi = 1 to xi = n are observed values on a predictor

Скачать книгу