Warning: foreach() argument must be of type array|object, bool given in /var/www/html/web/app/themes/studypress-core-theme/template-parts/header/mobile-offcanvas.php on line 20

An experiment consists of a large, but unknown, number \(n(\gg 1)\) of trials in each of which the probability of success \(p\) is the same, but also unkown. In the \(i\) th trial, \(i=1,2, \ldots, N\), the total number of successes is \(x_{i}(\gg 1)\). Determine the log-likelihood function. Using Stirling's approximation to \(\ln (n-x)\), show that $$ \frac{d \ln (n-x)}{d n} \approx \frac{1}{2(n-x)}+\ln (n-x) $$ and hence evaluate \(\partial\left({ }^{n} C_{x}\right) / \partial n\) By finding the (coupled) equations determining the ML estimators \(\hat{p}\) and \(\hat{n}\), show that, to order \(N^{-1}\), they must satisfy the simultaneous 'arithmetic' and 'geometric' mean constraints $$ \hat{n} \hat{p}=\frac{1}{N} \sum_{i=1}^{N} x_{i} \quad \text { and } \quad(1-\hat{p})^{N}=\prod_{i=1}^{N}\left(1-\frac{x_{i}}{\hat{n}}\right). $$

Short Answer

Expert verified
The ML estimators \(\hat{n}\) and \(\hat{p}\) satisfy \(\hat{n} \hat{p} = \frac{1}{N} \sum_{i=1}^N x_i\) and \((1-\hat{p})^N = \prod_{i=1}^{N}\left(1-\frac{x_i}{\hat{n}}\right)\).

Step by step solution

01

Log-likelihood function

The probability of obtaining exactly \(x_i\) successes in any given trial follows the binomial distribution given by \({n \brace x} p^{x} (1-p)^{n-x}\). The full log-likelihood function is therefore the sum over all \(N\) trials: \[ \ln L(n, p) = \sum_{i=1}^{N} \left[ x_i \ln(p) + (n-x_i) \ln(1-p) + \ln \left(\frac{n!}{x_i!(n-x_i!)}\right) \right] \].
02

Applying Stirling's approximation

Stirling's approximation states that \(\ln(k!) \approx k\ln(k) - k\). This gives \[ \ln \left(\frac{n!}{x_i!(n-x)!}\right) \approx n \ln(n) - n - x_i \ln(x_i) + x_i - (n-x_i)\ln(n-x_i) + n - x_i \].
03

Simplifying log-likelihood function

Simplifying using Stirling’s approximation, we get \[ \ln L(n, p) = \sum_{i=1}^{N} \left[ x_i \ln(p) + (n-x_i) \ln(1-p) + n \ln(n) - x_i \ln(x_i) - (n-x_i) \ln(n-x_i) \right]. \]
04

Derivative of log-likelihood

First, find the derivative of the log-likelihood with respect to n. Applying Stirling's approximation, we approximate \(\frac{d \ln(n-x)}{d n} \approx \frac{1}{2(n-x)} + \ln(n-x)\). Partial derivative is then: \[ \frac{\partial}{\partial n} \ln L(n, p) \approx \sum_{i=1}^{N} \left[ -\ln(1-p) - x_i \left( \frac{1}{2(n-x_i)} + \ln(n-x_i)\right) \right]. \]
05

Evaluate \(\partial\left(\binom{n}{x}\right) / \partial n\)

We have the term \(\binom{n}{x}\) involved in log-likelihood. The derivative with respect to \(n\) can now be evaluated considering the approximation: \[ \frac{\partial}{\partial n}\binom{n}{x} \approx \sum_{i=1}^{N} \left[ \binom{x_i}{n} \left( \frac{1}{2(n-x_i)} + \ln(n-x_i)\right) \right] \]
06

Setup coupled equations

To find ML estimators \(\hat{p}\) and \(\hat{n}\), solve the equations: \[ \hat{n} \hat{p}=\frac{1}{N} \sum_{i=1}^{N} x_{i} \text{ and } (1-\hat{p})^{N} = \prod_{i=1}^{N}\left(1-\frac{x_{i}}{\hat{n}}\right). \]
07

Validate equations to order \(N^{-1}\)

Lastly, ensure the equations meet the simultaneous 'arithmetic' (\(\hat{n} \hat{p}\)) and 'geometric' ((1-\hat{p})^{N}) mean constraints to order \(N^{-1}\). Hence, verify: \[ \hat{n} \hat{p}=\frac{1}{N} \sum_{i=1}^{N} x_{i} \text{ and } (1-\hat{p})^{N} = \prod_{i=1}^{N}\left(1-\frac{x_{i}}{\hat{n}}\right).\]

Unlock Step-by-Step Solutions & Ace Your Exams!

  • Full Textbook Solutions

    Get detailed explanations and key concepts

  • Unlimited Al creation

    Al flashcards, explanations, exams and more...

  • Ads-free access

    To over 500 millions flashcards

  • Money-back guarantee

    We refund you if you fail your exam.

Over 30 million students worldwide already upgrade their learning with Vaia!

Key Concepts

These are the key concepts you need to understand to accurately answer the question.

Maximum Likelihood Estimation (MLE)
Maximum Likelihood Estimation (MLE) is a method used for estimating the parameters of a statistical model. In the context of the binomial distribution, we are trying to find the values of the parameters that make the observed data most probable. Here, the parameters are the number of trials \(n\) and the probability of success \(p\).
To perform MLE, we start by defining a likelihood function, which expresses the probability of observing the given set of data as a function of the parameters to be estimated. We then find the parameter values that maximize this likelihood function. This is often done by taking the natural logarithm of the likelihood function (resulting in the log-likelihood function) for easier differentiation. We solve the partial derivatives of the log-likelihood function with respect to each parameter and set them to zero. This gives us the estimators for the parameters.

The log-likelihood function for the binomial distribution is particularly useful because its additive properties make it easier to differentiate.
Stirling's Approximation
Stirling's Approximation is a mathematical formula used to approximate the factorial of a large number. It is especially handy in probability and statistics, particularly when dealing with the log-likelihood function of the binomial distribution. According to Stirling's Approximation, \(\ln(k!) \approx k\ln(k) - k\).
This simplifies complex factorial expressions that appear in the binomial coefficients. For example, the term \(\frac{n!}{x!(n-x)!}\) in the binomial distribution's probability mass function can be cumbersome for large values of \(n\) and \(x\). By applying Stirling's Approximation, this term becomes manageable:
\[ \ln(\frac{n!}{x!(n-x)!}) \approx n \ln(n) - n - x \ln(x) + x - (n-x) \ln(n-x) + (n-x). \]
This approximation eases the differentiation process in the next step of finding the MLE.
Binomial Distribution
The binomial distribution is a discrete probability distribution that models the number of successes in a fixed number of independent and identically distributed Bernoulli trials. Each trial results in a success with probability \(p\) and a failure with probability \(1-p\). The binomial distribution is denoted by \(B(n, p)\), where \(n\) is the number of trials, and \(p\) is the probability of success.
The probability mass function for the binomial distribution is:
\[ P(X = x) = {n \choose x} p^x (1-p)^{n-x}, \]
where \({n \choose x}\) is the binomial coefficient, representing the number of ways to choose \(x\) successes out of \(n\) trials.
In MLE, our goal is to find the values of \(n\) and \(p\) that make our observed data most probable, by maximizing the likelihood of observing a specific sequence of successes and failures.
Log-likelihood Function
The log-likelihood function is the natural logarithm of the likelihood function. It is used in maximum likelihood estimation for easier differentiation. When dealing with the binomial distribution, the log-likelihood function is:
\[ \ln L(n, p) = \sum_{i=1}^{N} \left[ x_i \ln(p) + (n - x_i) \ln(1-p) + \ln \left(\frac{n!}{x_i!(n-x_i)!}\right) \right]. \]
Applying Stirling's Approximation, the log-likelihood function simplifies to:
\[ \ln L(n, p) = \sum_{i=1}^{N} \left[ x_i \ln(p) + (n-x_i) \ln(1-p) + n\ln(n) - x_i\ln(x_i) - (n-x_i)\ln(n-x_i) \right]. \]
The log-likelihood function helps us find the MLE by simplifying the process of taking derivatives and solving for the parameters.
Partial Derivatives
Partial derivatives are used in MLE to find the points where the log-likelihood function reaches its maximum, which correspond to the estimates of the parameters. For our problem, we compute the partial derivatives of the log-likelihood function \(\ln L(n, p)\) with respect to \(n\) and \(p\):
- The partial derivative with respect to \(p\) is:
\[ \frac{\partial \ln L(n, p)}{\partial p} = \sum_{i=1}^{N} \left( \frac{x_i}{p} - \frac{n - x_i}{1 - p} \right). \]
- The partial derivative with respect to \(n\) involves approximations and can be expressed as:
\[ \frac{\partial \ln L(n, p)}{\partial n} \approx \sum_{i=1}^{N} \left[ -\ln(1-p) - x_i \left( \frac{1}{2(n - x_i)} + \ln(n - x_i) \right) \right]. \]
We set these partial derivatives to zero and solve the resulting system of equations to find the MLE for \(n\) and \(p\).

One App. One Place for Learning.

All the tools & learning materials you need for study success - in one app.

Get started for free

Most popular questions from this chapter

It is claimed that the two following sets of values were obtained (a) by randomly drawing from a normal distribution that is \(N(0,1)\) and then (b) randomly assigning each reading to one of the two sets \(\mathrm{A}\) and \(\mathrm{B}\). \(\begin{array}{lrrrrrrr}\text { Set A } & -0.314 & 0.603 & -0.551 & -0.537 & -0.160 & -1.635 & 0.719 \\ & 0.610 & 0.482 & -1.757 & 0.058 & & & \\ \text { Set B } & -0.691 & 1.515 & -1.642 & -1.736 & 1.224 & 1.423 & 1.165\end{array}\) Make tests, including \(t\) - and \(F\)-tests, to establish whether there is any evidence that either claims is, or both claims are, false.

On a certain (testing) steeplechase course there are 12 fences to be jumped and any horse that falls is not allowed to continue in the race. In a season of racing a total of 500 horses started the course and the following numbers fell at each fence: \(\begin{array}{lrrrrrrrrrrrr}\text { Fence: } & 1 & 2 & 3 & 4 & 5 & 6 & 7 & 8 & 9 & 10 & 11 & 12 \\ \text { Falls: } & 62 & 75 & 49 & 29 & 33 & 25 & 30 & 17 & 19 & 11 & 15 & 12\end{array}\) Use this data to determine the overall probability of a horse falling at a fence, and test the hypothesis that it is the same for all horses and fences as follows. (a) draw up a table of the expected number of falls at each fence on the basis of the hypothesis; (b) consider for each fence \(i\) the standardised variable $$ z_{i}=\frac{\text { estimated falls }-\text { actual falls }}{\text { standard deviation of estimated falls }} $$ and use it in an appropriate \(\chi^{2}\) test; (c) show that the data indicates that the odds against all fences being equally testing are about 40 to \(1 .\) Identify the fences that are significantly easier or harder than the average.

The function \(y(x)\) is known to be a quadratic function of \(x\). The following table gives the measured values and uncorrelated standard errors of \(y\) measured at various values of \(x\) (in which there is negligible error): $$ \begin{array}{lccccc} x & 1 & 2 & 3 & 4 & 5 \\ y(x) & 3.5 \pm 0.5 & 2.0 \pm 0.5 & 3.0 \pm 0.5 & 6.5 \pm 1.0 & 10.5 \pm 1.0 \end{array} $$ Construct the response matrix \(R\) using as basis functions \(1, x, x^{2} .\) Calculate the matrix \(\mathrm{R}^{\mathrm{T}} \mathrm{N}^{-1} \mathrm{R}\) and show that its inverse, the covariance matrix \(\mathrm{V}\), has the form $$ \mathrm{V}=\frac{1}{9184}\left(\begin{array}{ccc} 12592 & -9708 & 1580 \\ -9708 & 8413 & -1461 \\ 1580 & -1461 & 269 \end{array}\right). $$ Use this matrix to find the best values, and their uncertainties, for the coefficients of the quadratic form for \(y(x)\).

The following are the values obtained by a class of 14 students when measuring a physical quantity \(x: 53.8,53.1,56.9,54.7,58.2,54.1,56.4,54.8,57.3,51.0,55.1\) \(55.0,54.2,56.6\) (a) Display these results as a histogram and state what you would give as the best value for \(x\). (b) Without calculation estimate how much reliance could be placed upon your answer to (a). (c) Databooks give the value of \(x\) as \(53.6\) with negligible error. Are the data obtained by the students in conflict with this?

Two physical quantities \(x\) and \(y\) are connected by the equation $$ y^{1 / 2}=\frac{x}{a x^{1 / 2}+b} $$ and measured pairs of values for \(x\) and \(y\) are as follows: \(\begin{array}{rrrrr}x: & 10 & 12 & 16 & 20 \\ y: & 409 & 196 & 114 & 94\end{array}\) Determine the best values for \(a\) and \(b\) by graphical means and (either by hand or by using a built-in calculator routine) by a least squares fit to an appropriate straight line,

See all solutions

Recommended explanations on Combined Science Textbooks

View all explanations

What do you think about this solution?

We value your feedback to improve our textbook solutions.

Study anywhere. Anytime. Across all devices.

Sign-up for free