MLR_ANOVA - Regression Analysis of Variance (ANOVA)

Calculates the regression model analysis of variance (ANOVA) values.



MLR_ANOVA(X, Mask, Y, Intercept, Return_type)

X is the independent (explanatory) variables data matrix, such that each column represents one variable.

Mask is the boolean array to choose the explanatory variables in the model. If missing, all variables in X are included.

Y is the response or the dependent variable data array (one dimensional array of cells (e.g. rows or columns)).

Intercept is the constant or the intercept value to fix (e.g. zero). If missing, an intercept will not be fixed and is computed normally.

Return_type is a switch to select the output (1 = SSR (default), 2 = SSE, 3 = SST, 4 = MSR, 5 = MSE, 6 = F-Stat, 7 = P-Value).

Method Description
1 SSR (sum of squares of the regression)
2 SSE (sum of squares of the residuals)
3 SST (sum of squares of the dependent variable)
4 MSR (mean squares of the regression)
5 MSE (mean squares error or residuals)
6 F-Stat (test score)
7 Significance F (P-value of the test)


  1. The underlying model is described here.
  2. $$\mathbf{y} = \alpha + \beta_1 \times \mathbf{x}_1 + \dots + \beta_p \times \mathbf{x}_p$$
  3. The regression ANOVA table examines the following hypothesis:

    $$\mathbf{H}_o: \beta_1 = \beta_2 = \dots = \beta_p = 0 $$
    $$\mathbf{H}_1: \exists \beta_i \neq 0, i \in \left[1,0 \right ] $$
  4. In other words, the regression ANOVA examines the probability that regression does NOT explain the variation in $\mathbf{y}$, i.e. that any fit is due purely to chance.
  5. The MLR_ANOVA calculates the different values in the ANOVA tables as shown below:

    $$\mathbf{SST}=\sum_{i=1}^N \left(Y_i - \bar Y \right )^2 $$
    $$\mathbf{SSR}=\sum_{i=1}^N \left(\hat Y_i - \bar Y \right )^2 $$
    $$\mathbf{SSR}=\sum_{i=1}^N \left(Y_i - \hat Y_i \right )^2 $$
    • $N$ is the number of non-missing observations in the sample data.
    • $\bar Y$ is the empirical sample average for the dependent variable.
    • $\hat Y_i$ is the regression model estimate value for the i-th observation.
    • $\mathbf{SST}$ is the total sum of squares for the dependent variable.
    • $\mathbf{SSR}$ is the total sum of squares for the regression (i.e. $\hat y$) estimate.
    • $\mathbf{SSE}$ is the total sum of error (aka residuals $\epsilon$) terms for the regression (i.e. $\epsilon = y - \hat y$) estimate.
    • $\mathbf{SST} = \mathbf{SSR} + \mathbf{SSE}$
    $$\mathbf{MSR} = \frac{\mathbf{SSR} }{p} $$

    $$\mathbf{MSE} = \frac{ \mathbf{SSE} }{N-p-1}$$

    $$\mathbf{F-Stat} = \frac{\mathbf{MSR} }{\mathbf{MSE} }$$

    • $p$ is the number of explanatory (aka predictor) variables in the regression.
    • $\mathbf{MSR}$ is the mean squares of the regression.
    • $\mathbf{MSE}$ is the mean squares of the residuals.
    • $\textrm{F-Stat}$ is the test score of the hypothesis.

    • $\textrm{F-Stat} \sim \mathbf{F}\left(p,N-p-1 \right)$.
  6. The sample data may include missing values.
  7. Each column in the inputm atrix corresponds to a separate variable.
  8. Each row in the input matrix corresponds to an observation.
  9. Observations (i.e. rows) with missing values in X or Y are removed.
  10. The number of rows of the response variable (Y) must be equal to the number of rows of the explanatory variable (X).
  11. The MLR_ANOVA function is available starting with version 1.60 APACHE.

Files Examples


  • Hamilton, J .D.; Time Series Analysis , Princeton University Press (1994), ISBN 0-691-04289-6
  • Kenney, J. F. and Keeping, E. S. (1962) "Linear Regression and Correlation." Ch. 15 in Mathematics of Statistics, Pt. 1, 3rd ed. Princeton, NJ: Van Nostrand, pp. 252-285
Have more questions? Submit a request