فترة الثقة لنسبة ثنائية الحدود

In statistics, a binomial proportion confidence interval is a confidence interval for the probability of success calculated from the outcome of a series of success–failure experiments (Bernoulli trials). In other words, a binomial proportion confidence interval is an interval estimate of a success probability p when only the number of experiments n and the number of successes nS are known.

There are several formulas for a binomial confidence interval, but all of them rely on the assumption of a binomial distribution. In general, a binomial distribution applies when an experiment is repeated a fixed number of times, each trial of the experiment has two possible outcomes (success and failure), the probability of success is the same for each trial, and the trials are statistically independent. Because the binomial distribution is a discrete probability distribution (i.e., not continuous) and difficult to calculate for large numbers of trials, a variety of approximations are used to calculate this confidence interval, all with their own tradeoffs in accuracy and computational intensity.

A simple example of a binomial distribution is the set of various possible outcomes, and their probabilities, for the number of heads observed when a coin is flipped ten times. The observed binomial proportion is the fraction of the flips that turn out to be heads. Given this observed proportion, the confidence interval for the true probability of the coin landing on heads is a range of possible proportions, which may or may not contain the true proportion. A 95% confidence interval for the proportion, for instance, will contain the true proportion 95% of the times that the procedure for constructing the confidence interval is employed.[1]

. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .

Normal approximation interval

A commonly used formula for a binomial confidence interval relies on approximating the distribution of error about a binomially-distributed observation, , with a normal distribution.[2] This approximation is based on the central limit theorem and is unreliable when the sample size is small or the success probability is close to 0 or 1.[3]

Using the normal approximation, the success probability p is estimated as

or the equivalent

where is the proportion of successes in a Bernoulli trial process, measured with trials yielding successes and failures, and is the quantile of a standard normal distribution (i.e., the probit) corresponding to the target error rate . For a 95% confidence level, the error , so and .

An important theoretical derivation of this confidence interval involves the inversion of a hypothesis test. Under this formulation, the confidence interval represents those values of the population parameter that would have large p-values if they were tested as a hypothesized population proportion. The collection of values, , for which the normal approximation is valid can be represented as

where is the quantile of a standard normal distribution. Since the test in the middle of the inequality is a Wald test, the normal approximation interval is sometimes called the Wald interval, but it was first described by Pierre-Simon Laplace in 1812.[4]


فترة نتيجة ولسون

The Wilson score interval is an improvement over the normal approximation interval in that the actual coverage probability is closer to the nominal value. It was developed by Edwin Bidwell Wilson (1927).[5]

Wilson started with the normal approximation to the binomial:

with the analytic formula for the sample standard deviation given by

.

Combining the two, and squaring out the radical, gives an equation that is quadratic in p:

Transforming the relation into a standard-form quadratic equation for p, treating and n as known values from the sample (see prior section), and using the value of z that corresponds to the desired confidence for the estimate of p gives this:

,

where all of the values in parentheses are known quantities. The solution for p estimates the upper and lower limits of the confidence interval for p. Hence the probability of success p is estimated by

or the equivalent

The practical observation from using this interval, is that it has good properties even for a small number of trials and / or an extreme probability.

Intuitively, the center value of this interval is the weighted average of and , with receiving greater weight as the sample size increases. Formally, the center value corresponds to using a pseudocount of 1/2 z², the number of standard deviations of the confidence interval: add this number to both the count of successes and of failures to yield the estimate of the ratio. For the common two standard deviations in each direction interval (approximately 95% coverage, which itself is approximately 1.96 standard deviations), this yields the estimate , which is known as the "plus four rule".

Although the quadratic can be solved explicitly, in most cases Wilson's equations can also be solved numerically using the fixed-point iteration

with .

The Wilson interval can be derived from Pearson's chi-squared test with two categories. The resulting interval,

can then be solved for to produce the Wilson score interval. The test in the middle of the inequality is a score test.


See also

References

  1. ^ Sullivan, Lisa (2017-10-27). "Confidence Intervals". Boston University School of Public Health.
  2. ^ Wallis, Sean A. (2013). "Binomial confidence intervals and contingency tests: mathematical fundamentals and the evaluation of alternative methods" (PDF). Journal of Quantitative Linguistics. 20 (3): 178–208. doi:10.1080/09296174.2013.799918. S2CID 16741749.
  3. ^ Brown, Lawrence D.; Cai, T. Tony; DasGupta, Anirban (2001). "Interval Estimation for a Binomial Proportion". Statistical Science. 16 (2): 101–133. CiteSeerX 10.1.1.50.3025. doi:10.1214/ss/1009213286. MR 1861069. Zbl 1059.62533.
  4. ^ Laplace, Pierre Simon (1812). Théorie analytique des probabilités (in الفرنسية). Ve. Courcier. p. 283.
  5. ^ Wilson, E. B. (1927). "Probable inference, the law of succession, and statistical inference". Journal of the American Statistical Association. 22 (158): 209–212. doi:10.1080/01621459.1927.10502953. JSTOR 2276774.

خطأ استشهاد: الوسم <ref> ذو الاسم "New" المُعرّف في <references> غير مستخدم في النص السابق.
خطأ استشهاد: الوسم <ref> ذو الاسم "Rei" المُعرّف في <references> غير مستخدم في النص السابق.
خطأ استشهاد: الوسم <ref> ذو الاسم "SL" المُعرّف في <references> غير مستخدم في النص السابق.

خطأ استشهاد: الوسم <ref> ذو الاسم "Ross" المُعرّف في <references> غير مستخدم في النص السابق.