Warning: foreach() argument must be of type array|object, bool given in /var/www/html/web/app/themes/studypress-core-theme/template-parts/header/mobile-offcanvas.php on line 20

Often in regression the mean of the random variable \(Y\) is a linear function of \(p\) -values \(x_{1}, x_{2}, \ldots, x_{p}\), say \(\beta_{1} x_{1}+\beta_{2} x_{2}+\cdots+\beta_{p} x_{p}\), where \(\boldsymbol{\beta}^{\prime}=\left(\beta_{1}, \beta_{2}, \ldots, \beta_{p}\right)\) are the regression coefficients. Suppose that \(n\) values, \(\boldsymbol{Y}^{\prime}=\left(Y_{1}, Y_{2}, \ldots, Y_{n}\right)\) are observed for the \(x\) -values in \(\boldsymbol{X}=\left[x_{i j}\right]\), where \(\boldsymbol{X}\) is an \(n \times p\) design matrix and its ith row is associated with \(Y_{i}, i=1,2, \ldots, n .\) Assume that \(Y\) is multivariate normal with mean \(\boldsymbol{X} \boldsymbol{\beta}\) and variance-covariance matrix \(\sigma^{2} \boldsymbol{I}\), where \(\boldsymbol{I}\) is the \(n \times n\) identity matrix. (a) Note that \(Y_{1}, Y_{2}, \ldots, Y_{n}\) are independent. Why? (b) Since \(\boldsymbol{Y}\) should approximately equal its mean \(\boldsymbol{X} \boldsymbol{\beta}\), we estimate \(\boldsymbol{\beta}\) by solving the normal equations \(\boldsymbol{X}^{\prime} \boldsymbol{Y}=\boldsymbol{X}^{\prime} \boldsymbol{X} \boldsymbol{\beta}\) for \(\boldsymbol{\beta}\). Assuming that \(\boldsymbol{X}^{\prime} \boldsymbol{X}\) is non- singular, solve the equations to get \(\hat{\boldsymbol{\beta}}=\left(\boldsymbol{X}^{\prime} \boldsymbol{X}\right)^{-1} \boldsymbol{X}^{\prime} \boldsymbol{Y}\). Show that \(\hat{\boldsymbol{\beta}}\) has a multivariate normal distribution with mean \(\boldsymbol{\beta}\) and variance-covariance matrix $$ \sigma^{2}\left(\boldsymbol{X}^{\prime} \boldsymbol{X}\right)^{-1} $$ (c) Show that $$ (\boldsymbol{Y}-\boldsymbol{X} \boldsymbol{\beta})^{\prime}(\boldsymbol{Y}-\boldsymbol{X} \boldsymbol{\beta})=(\hat{\boldsymbol{\beta}}-\boldsymbol{\beta})^{\prime}\left(\boldsymbol{X}^{\prime} \boldsymbol{X}\right)(\hat{\boldsymbol{\beta}}-\boldsymbol{\beta})+(\boldsymbol{Y}-\boldsymbol{X} \hat{\boldsymbol{\beta}})^{\prime}(\boldsymbol{Y}-\boldsymbol{X} \hat{\boldsymbol{\beta}}) $$ say \(Q=Q_{1}+Q_{2}\) for convenience. (d) Show that \(Q_{1} / \sigma^{2}\) is \(\chi^{2}(p)\). (e) Show that \(Q_{1}\) and \(Q_{2}\) are independent. (f) Argue that \(Q_{2} / \sigma^{2}\) is \(\chi^{2}(n-p)\). (g) Find \(c\) so that \(c Q_{1} / Q_{2}\) has an \(F\) -distribution. (h) The fact that a value \(d\) can be found so that \(P\left(c Q_{1} / Q_{2} \leq d\right)=1-\alpha\) could be used to find a \(100(1-\alpha)\) percent confidence ellipsoid for \(\beta\). Explain.

Short Answer

Expert verified
This exercise involves mathematical proofs related to several quantities derived from observations and predictors that are organized into a matrix in multiple linear regression. With normal distribution, identity covariance matrix and linear relationship between predictors and observations, we derived several important properties and relationships like normal distribution of estimates, independence of residuals, chi-square and F-distributions for squared quantities and statistical inference foundations for confidence intervals.

Step by step solution

01

Part (a)

Given that the variance-covariance matrix is \(\sigma^{2} \boldsymbol{I}\), where \(\boldsymbol{I}\) is the identity matrix, it implies all the off-diagonal elements representing the covariance of any two different \(Y\) values are zero. Thus, \(Y_{1}, Y_{2}, \ldots, Y_{n}\) are uncorrelated and since they are normal, they are also independent.
02

Part (b)

Given the normal equation \(\boldsymbol{X}^{\prime} \boldsymbol{Y}=\boldsymbol{X}^{\prime} \boldsymbol{X} \boldsymbol{\beta}\), we can solve for \(\boldsymbol{\beta}\) to get the least squares estimates \(\hat{\boldsymbol{\beta}}=\left(\boldsymbol{X}^{\prime} \boldsymbol{X}\right)^{-1} \boldsymbol{X}^{\prime} \boldsymbol{Y}\). Under normal distribution assumption, the vector \(\boldsymbol{\beta}\) also has a multivariate normal distribution with mean \(\boldsymbol{\beta}\) and variance-covariance matrix \(\sigma^{2}\left(\boldsymbol{X}^{\prime} \boldsymbol{X}\right)^{-1}\) from elements variance for multivariate normal distribution.
03

Part (c)

To prove the equation, expand both sides, as \((a-b)'(a-b) = a'a - 2 a'b + b'b\). After simplification, both sides will be equal.
04

Part (d)

The quantity \(Q_{1} / \sigma^{2} =(\hat{\boldsymbol{\beta}}-\boldsymbol{\beta})^{\prime}\left(\boldsymbol{X}^{\prime}\boldsymbol{X}\right)(\hat{\boldsymbol{\beta}}-\boldsymbol{\beta}) / \sigma^{2}\) follows a \(\chi^2\) distribution with \(p\) degrees of freedom by the definition of chi-square distribution and proved under (b)
05

Part (e)

The quantities \(Q_1\) and \(Q_2\) can be shown to be independent because they are functions of uncorrelated random variables.
06

Part (f)

The quantity \(Q_{2} / \sigma^{2} = (\boldsymbol{Y}-\boldsymbol{X}\hat{\boldsymbol{\beta}})^{\prime}(\boldsymbol{Y}-\boldsymbol{X}\hat{\boldsymbol{\beta}}) / \sigma^{2}\) follows a \(\chi^2\) distribution with \((n-p)\) degrees of freedom, by property of chi-square distribution where each of \(n-p\) residuals \((\boldsymbol{Y}-\boldsymbol{X} \hat{\boldsymbol{\beta}})\) squares contributes 1 degree.
07

Part (g)

To give \(c Q_{1} / Q_{2}\) an \(F\) -distribution, \(c\) must be the ratio of the degrees of freedom of \(Q_{1}\) and \(Q_{2}\), so \(c= p / (n-p)\) as they are the degrees of freedom from (d) and (f). An F-distribution describes ratio of two chi-square divided by their degrees of freedom.
08

Part (h)

Using \(Q_1\) and \(Q_2\) we can get an F distribution. We can also calculate a cut-off score \(d\) by defining our acceptable error to be \(\alpha\). Using the F-distribution table or function, we can find a value \(d\) that \[P\left(c Q_{1} / Q_{2} \leq d\right)=1-\alpha\]. This is the basis of statistical testing and establishing confidence intervals. We can obtain a \(100(1-\alpha)\%\) confidence ellipsoid for \(\beta\) by creating bounds using the relationship of \(\beta\) to the F-distribution.

Unlock Step-by-Step Solutions & Ace Your Exams!

  • Full Textbook Solutions

    Get detailed explanations and key concepts

  • Unlimited Al creation

    Al flashcards, explanations, exams and more...

  • Ads-free access

    To over 500 millions flashcards

  • Money-back guarantee

    We refund you if you fail your exam.

Over 30 million students worldwide already upgrade their learning with Vaia!

One App. One Place for Learning.

All the tools & learning materials you need for study success - in one app.

Get started for free

Most popular questions from this chapter

If \(A_{1}, A_{2}, \ldots, A_{k}\) are events, prove, by induction, Boole's inequality $$ P\left(A_{1} \cup A_{2} \cup \cdots \cup A_{k}\right) \leq \sum_{1}^{k} P\left(A_{i}\right) $$ Then show that $$ P\left(A_{1}^{c} \cap A_{2}^{c} \cap \cdots \cap A_{k}^{c}\right) \geq 1-\sum_{1}^{b} P\left(A_{i}\right) $$

In Exercise 9.2.1, show that the linear functions \(X_{i j}-X_{. j}\) and \(X_{. j}-X\).. are uncorrelated. Hint: Recall the definition of \(\bar{X}_{. j}\) and \(\bar{X}_{. .}\) and, without loss of generality, we can let \(E\left(X_{i j}\right)=0\) for all \(i, j\)

Let \(\mathbf{X}^{\prime}=\left[X_{1}, X_{2}\right]\) be bivariate normal with matrix of means \(\boldsymbol{\mu}^{\prime}=\left[\mu_{1}, \mu_{2}\right]\) and positive definite covariance matrix \(\mathbf{\Sigma}\). Let $$ Q_{1}=\frac{X_{1}^{2}}{\sigma_{1}^{2}\left(1-\rho^{2}\right)}-2 \rho \frac{X_{1} X_{2}}{\sigma_{1} \sigma_{2}\left(1-\rho^{2}\right)}+\frac{X_{2}^{2}}{\sigma_{2}^{2}\left(1-\rho^{2}\right)} $$ Show that \(Q_{1}\) is \(\chi^{2}(r, \theta)\) and find \(r\) and \(\theta\). When and only when does \(Q_{1}\) have a central chi-square distribution?

A random sample of size \(n=6\) from a bivariate normal distribution yields a value of the correlation coefficient of \(0.89 .\) Would we accept or reject, at the 5 percent significance level, the hypothesis that \(\rho=0\).

With the background of the two-way classification with \(c>1\) observations per cell, show that the maximum likelihood estimators of the parameters are $$ \begin{aligned} \hat{\alpha}_{i} &=\bar{X}_{i . .}-\bar{X}_{\ldots} \\ \hat{\beta}_{j} &=\bar{X}_{. j .}-\bar{X}_{\cdots} \\ \hat{\gamma}_{i j} &=\bar{X}_{i j .}-\bar{X}_{i .}-\bar{X}_{. j}+\bar{X}_{\ldots} \\ \hat{\mu} &=\bar{X}_{\ldots} \end{aligned} $$ Show that these are unbiased estimators of the respective parameters. Compute the variance of each estimator.

See all solutions

Recommended explanations on Math Textbooks

View all explanations

What do you think about this solution?

We value your feedback to improve our textbook solutions.

Study anywhere. Anytime. Across all devices.

Sign-up for free