Chapter 9: Problem 13
Fit by the method of least squares the plane \(z=a+b x+c y\) to the five points \((x, y, z):(-1,-2,5),(0,-2,4),(0,0,4),(1,0,2),(2,1,0)\).
Chapter 9: Problem 13
Fit by the method of least squares the plane \(z=a+b x+c y\) to the five points \((x, y, z):(-1,-2,5),(0,-2,4),(0,0,4),(1,0,2),(2,1,0)\).
All the tools & learning materials you need for study success - in one app.
Get started for freeUsing the background of the two-way classification with one observation per cell, show that the maximum likelihood estimator of \(\alpha_{i}, \beta_{j}\), and \(\mu\) are \(\hat{\alpha}_{i}=\bar{X}_{i .}-\bar{X}_{. .}\) \(\hat{\beta}_{j}=\bar{X}_{. j}-\bar{X}_{. .}\), and \(\hat{\mu}=\bar{X}_{. .}\), respectively. Show that these are unbiased estimators of their respective parameters and compute \(\operatorname{var}\left(\hat{\alpha}_{i}\right), \operatorname{var}\left(\hat{\beta}_{j}\right)\), and \(\operatorname{var}(\hat{\mu})\).
Suppose \(\mathbf{A}\) is a real symmetric matrix. If the eigenvalues of \(\mathbf{A}\) are only 0 's and 1 's then prove that \(\mathbf{A}\) is idempotent.
Let the independent random variables \(Y_{1}, \ldots, Y_{n}\) have the joint pdf. $$ L\left(\alpha, \beta, \sigma^{2}\right)=\left(\frac{1}{2 \pi \sigma^{2}}\right)^{n / 2} \exp \left\\{-\frac{1}{2 \sigma^{2}} \sum_{1}^{n}\left[y_{i}-\alpha-\beta\left(x_{i}-\bar{x}\right)\right]^{2}\right\\} $$ where the given numbers \(x_{1}, x_{2}, \ldots, x_{n}\) are not all equal. Let \(H_{0}: \beta=0(\alpha\) and \(\sigma^{2}\) unspecified). It is desired to use a likelihood ratio test to test \(H_{0}\) against all possible alternatives. Find \(\Lambda\) and see whether the test can be based on a familiar statistic. Hint: In the notation of this section show that $$ \sum_{1}^{n}\left(Y_{i}-\hat{\alpha}\right)^{2}=Q_{3}+\widehat{\beta}^{2} \sum_{1}^{n}\left(x_{i}-\bar{x}\right)^{2} $$
Let \(X_{1}, X_{2}, X_{3}, X_{4}\) denote a random sample of size 4 from a distribution which is \(N\left(0, \sigma^{2}\right) .\) Let \(Y=\sum_{1}^{4} a_{i} X_{i}\), where \(a_{1}, a_{2}, a_{3}\), and \(a_{4}\) are real constants. If \(Y^{2}\) and \(Q=X_{1} X_{2}-X_{3} X_{4}\) are independent, determine \(a_{1}, a_{2}, a_{3}\), and \(a_{4}\).
(Bonferroni Multiple Comparison Procedure). In the notation of this section, let \(\left(k_{i 1}, k_{i 2}, \ldots, k_{i b}\right), i=1,2, \ldots, m\), represent a finite number of \(b\) -tuples. The problem is to find simultaneous confidence intervals for \(\sum_{j=1}^{b} k_{i j} \mu_{j}, i=1,2, \ldots, m\), by a method different from that of Scheffé. Define the random variable \(T_{i}\) by $$ \left(\sum_{j=1}^{b} k_{i j} \bar{X}_{. j}-\sum_{j=1}^{b} k_{i j} \mu_{j}\right) / \sqrt{\left(\sum_{j=1}^{b} k_{i j}^{2}\right) V / a}, \quad i=1,2, \ldots, m $$ (a) Let the event \(A_{i}^{c}\) be given by \(-c_{i} \leq T_{i} \leq c_{i}, i=1,2, \ldots, m\). Find the random variables \(U_{i}\) and \(W_{i}\) such that \(U_{i} \leq \sum_{1}^{b} k_{i j} \mu_{j} \leq W_{j}\) is equivalent to \(A_{i}^{c}\) (b) Select \(c_{i}\) such that \(P\left(A_{i}^{c}\right)=1-\alpha / m ;\) that is, \(P\left(A_{i}\right)=\alpha / m .\) Use Exercise 9.4.1 to determine a lower bound on the probability that simultaneously the random intervals \(\left(U_{1}, W_{1}\right), \ldots,\left(U_{m}, W_{m}\right)\) include \(\sum_{j=1}^{b} k_{1 j} \mu_{j}, \ldots, \sum_{j=1}^{b} k_{m j} \mu_{j}\) respectively. (c) Let \(a=3, b=6\), and \(\alpha=0.05\). Consider the linear functions \(\mu_{1}-\mu_{2}, \mu_{2}-\mu_{3}\), \(\mu_{3}-\mu_{4}, \mu_{4}-\left(\mu_{5}+\mu_{6}\right) / 2\), and \(\left(\mu_{1}+\mu_{2}+\cdots+\mu_{6}\right) / 6 .\) Here \(m=5 .\) Show that the lengths of the confidence intervals given by the results of Part (b) are shorter than the corresponding ones given by the method of Scheffé as described in the text. If \(m\) becomes sufficiently large, however, this is not the case.
What do you think about this solution?
We value your feedback to improve our textbook solutions.