Warning: foreach() argument must be of type array|object, bool given in /var/www/html/web/app/themes/studypress-core-theme/template-parts/header/mobile-offcanvas.php on line 20

In Example 12.6.7, let \(\left( {{X^*},{Y^*}} \right)\) be a random draw from the sample distribution\({F_n}\) . Prove that the correlation \({X^*} and {Y^*}\) is R in Eq. (12.6.2).

Short Answer

Expert verified

\(R = \frac{{\sum\limits_{i = 1}^n {\left( {{X_i} - \bar X} \right)} \left( {{Y_i} - \bar Y} \right)}}{{\sqrt {\sum\limits_{i = 1}^n {{{\left( {{X_i} - \bar X} \right)}^2}} \sum\limits_{i = 1}^n {{{\left( {{Y_i} - \bar Y} \right)}^2}} } }}\)

Step by step solution

01

To Prove that the correlation between \({X^*} and {Y^*}\) is R 

The correlation coefficient of X and Y is

\({\rho _{X,Y}} = \frac{{Cov(X,Y)}}{{{\sigma _X} \times {\sigma _Y}}}\)

Notice first that for such a random sample from the sample distribution,\({F_n}\)the expected values are

\(\begin{aligned}{l}E\left( {{X^*}} \right) &= \bar X\\E\left( {{Y^*}} \right) &= \bar Y\end{aligned}\)

and the variances are given by

\(\begin{aligned}{l}V\left( {{X^*}} \right) &= \frac{1}{n}\sum\limits_{i = 1}^n {{{\left( {{X_i} - \bar X} \right)}^2}} \\\\V\left( {{Y^*}} \right) &= \frac{1}{n}\sum\limits_{i = 1}^n {{{\left( {{Y_i} - \bar Y} \right)}^2}} \end{aligned}\)

Finally, to get the sample correlation coefficient of

\(R = \frac{{\sum\limits_{i = 1}^n {\left( {{X_i} - \bar X} \right)} \left( {{Y_i} - \bar Y} \right)}}{{\sqrt {\sum\limits_{i = 1}^n {{{\left( {{X_i} - \bar X} \right)}^2}} \sum\limits_{i = 1}^n {{{\left( {{Y_i} - \bar Y} \right)}^2}} } }}\)

it notices that the covariance\(is Cov(X,Y)\)is

\(\begin{aligned}{l}Cov(X,Y) &= E\left( {\left( {{X^*} - \bar X} \right)\left( {{Y^*} - \bar Y} \right)} \right)\\\\ &= \sum\limits_{i = 1}^n {\left( {{X_i} - \bar X} \right)} \left( {{Y_i} - \bar Y} \right)\end{aligned}\)

Because the sample distribution\({F_n}\)is discrete. Finally, by substituting all values and using the fact that\({\sigma _{{X^*}}} = \sqrt {V\left( {{X^*}} \right)} \)it is true that

\(\begin{aligned}{l}{\rho _{X,Y}} &= \frac{{Cov(X,Y)}}{{{\sigma _X} \times {\sigma _Y}}}\\\\ &= \frac{{\sum\limits_{i = 1}^n {\left( {{X_i} - \bar X} \right)} \left( {{Y_i} - \bar Y} \right)}}{{\sqrt {\sum\limits_{i = 1}^n {{{\left( {{X_i} - \bar X} \right)}^2}} \sum\limits_{i = 1}^n {{{\left( {{Y_i} - \bar Y} \right)}^2}} } }}\quad \quad \end{aligned}\)

Unlock Step-by-Step Solutions & Ace Your Exams!

  • Full Textbook Solutions

    Get detailed explanations and key concepts

  • Unlimited Al creation

    Al flashcards, explanations, exams and more...

  • Ads-free access

    To over 500 millions flashcards

  • Money-back guarantee

    We refund you if you fail your exam.

Over 30 million students worldwide already upgrade their learning with Vaia!

One App. One Place for Learning.

All the tools & learning materials you need for study success - in one app.

Get started for free

Most popular questions from this chapter

Use the data in the Table \({\bf{11}}{\bf{.5}}\) on page \({\bf{699}}\) suppose that \({{\bf{y}}_{\bf{i}}}\) is the logarithm of pressure \({x_i}\)and is the boiling point for the I the observation \({\bf{i = 1,}}...{\bf{,17}}{\bf{.}}\) Use the robust regression scheme described in Exercise \({\bf{8}}\) to \({\bf{a = 5, b = 0}}{\bf{.1}}\,\,{\bf{and f = 0}}{\bf{.1}}{\bf{.}}\) Estimate the posterior means and standard deviations of the parameter \({{\bf{\beta }}_{\bf{0}}}{\bf{,}}{{\bf{\beta }}_{\bf{1}}}\,\) and n.

Let \({{\bf{X}}_{\bf{1}}},...,{{\bf{X}}_n}\) be i.i.d. with the normal distribution having mean \(\mu \) and precision \(\tau \).Gibbs sampling allows one to use a prior distribution for \(\left( {\mu ,\tau } \right)\) in which \(\mu \) and\(\tau \) are independent. With mean \({\mu _0}\) and variance, \({\gamma _0}\) Let the prior distribution of \(\tau \)being the gamma distribution with parameters \({\alpha _0}\) and \({\beta _0}\) .

a. Show that the Table \({\bf{12}}{\bf{.8}}\) specifies the appropriate conditional distribution for each parameter given the other.

b. Use the new Mexico nursing home data(Examples \({\bf{12}}{\bf{.5}}{\bf{.2}}\,{\bf{and}}\,{\bf{12}}{\bf{.5}}{\bf{.3}}\) ). Let the prior hyperparameters be \({{\bf{\alpha }}_{\bf{0}}}{\bf{ = 2,}}{{\bf{\beta }}_{\bf{0}}}{\bf{ = 6300,}}{{\bf{\mu }}_{\bf{0}}}{\bf{ = 200}}\), and \({\gamma _0} = 6.35 \times {10^{ - 4}}.\) Implement a Gibbs sampler to find the posterior distribution \(\left( {\mu ,\tau } \right).\,\) . In particular, calculate an interval containing \(95\) percent of the posterior distribution of \(\mu \)

In Sec. 10.2, we discussed \({\chi ^2}\) goodness-of-fit tests for composite hypotheses. These tests required computing M.L.E.'s based on the numbers of observations that fell into the different intervals used for the test. Suppose instead that we use the M.L.E.'s based on the original observations. In this case, we claimed that the asymptotic distribution of the \({x^2}\) test statistic was somewhere between two different \({\chi ^2}\) distributions. We can use simulation to better approximate the distribution of the test statistic. In this exercise, assume that we are trying to test the same hypotheses as in Example 10.2.5, although the methods will apply in all such cases.

a. Simulate \(v = 1000\) samples of size \(n = 23\) from each of 10 different normal distributions. Let the normal distributions have means of \(3.8,3.9,4.0,4.1,\) and \(4.2\) Let the distributions have variances of 0.25 and 0.8. Use all 10 combinations of mean and variance. For each simulated sample, compute the \({\chi ^2}\) statistic Q using the usual M.L.E.'s of \(\mu \) , and \({\sigma ^2}.\) For each of the 10 normal distributions, estimate the 0.9,0.95, and 0.99 quantiles of the distribution of Q.

b. Do the quantiles change much as the distribution of the data changes?

c. Consider the test that rejects the null hypothesis if \(Q \ge 5.2.\) Use simulation to estimate the power function of this test at the following alternative: For each \(i,\left( {{X_i} - 3.912} \right)/0.5\) has the t distribution with five degrees of freedom.

Let \({\bf{Y}}\) be a random variable with some distribution. Suppose that you have available as many pseudo-random variables as you want with the same distribution as \({\bf{Y}}\). Describe a simulation method for estimating the skewness of the distribution of \({\bf{Y}}\). (See Definition 4.4.1.)

Show how to simulate Cauchy random variables using the probability integral transformation.

See all solutions

Recommended explanations on Math Textbooks

View all explanations

What do you think about this solution?

We value your feedback to improve our textbook solutions.

Study anywhere. Anytime. Across all devices.

Sign-up for free