Statistics/Interval Estimation
Introduction
Previously, we have discussed Template:Colored em, which gives us an estimator for the value of an unknown parameter . Now, suppose we want to know the size of Template:Colored em of the point estimator , i.e. the difference between and the unknown parameter . Of course, we can make use of the value of the Template:Colored em of , , or other things.
However, what if we only know about one specific Template:Colored em? We cannot calculate the mean squared error of its corresponding point estimator with Template:Colored em this point estimates, right? So, how do we know the possible size of error of this Template:Colored em? Indeed, it is impossible to tell, since we are only given a particular estimated value of parameter , but of course we do know the value of the unknown parameter , thus the difference between this point estimate and is also unknown.
To illustrate this, consider the following example: suppose we take a random sample of 10 students from one particular course in university to estimate the mean score of the students in the final exam in that course, denoted by , (assume the score is normally distributed), and the observed value of the sample mean is . Then, what is the difference between this point estimate and the true unknown parameter ? Can we be "confident" that this sample mean is close to , say ?
It is possible that is, say 90, and somehow the students in the sample are the one with very poor performance. On the other hand, it is also possible that is, say 30, and somehow the students in the sample are the one who perform well (relatively). Of course, it is also possible that the is quite close to 60, say 59. From this example, we can see that a particular value does not tell us the possible size of error: the error can be very large, and also can be very small.
In this chapter, we will introduce Template:Colored em where we use Template:Colored em that can describe the size of error through providing the probability for the random interval (i.e. interval with at least one of its bounds to be a random variable) given by the interval estimator to contain the unknown parameter , which measures the "accuracy" of the interval estimator of , and hence the size of error.
As suggested by the name Template:Colored em, the estimator involves some sort of Template:Colored em. Also, as one may expect, Template:Colored em is also based on Template:Colored em: Template:Colored definition Of course, we would like the probability for the unknown parameter to lie in the interval to be close to 1, so that the interval estimator is very accurate. However, a very accurate interval estimator may have a very bad "precision", i.e. the interval covers "too many" plausible values of an unknown parameter, and therefore even if we know that is very likely to be one of such values, there are too many different possibilities. Hence, such interval estimator is not very "useful". To illustrate this, suppose the interval concerned is , which is the parameter space of . Then, of course (so the "confidence" is high) since must lie in its parameter space. However, such interval has basically "zero precision", and is quite "useless", since the "plausible values" of in the intervals are essentially all possible values of .
From this, we can observe the need of the "precision" of the interval, that is, we also want the Template:Colored em of the interval to be small, so that we can have some ideas about the "location" of . However, as the interval becomes smaller, it is more likely that such interval misses , i.e. does not cover the actual value of , and therefore the probability for to lie in that interval becomes smaller, i.e. the interval becomes less "accurate". To illustrate this, let us consider the extreme case: the interval is so small that it becomes an interval containing a single point (the two end-points of the interval coincide). Then, the "interval estimator" basically becomes a "point estimator" in some sense, and we know that it is very unlikely that the true value of equals the value of the point estimator ( lies in that "interval" is equivalent to in this case). Indeed, if the distribution of is Template:Colored em, then .
As we can see from above, although we want the interval to have a very high "confidence" and also "very precise" (i.e. the interval is very narrow), we cannot have both of them, since an increase in confidence causes a decrease in "precision", and an increase in "precision" causes a decrease in confidence. Therefore, we need to make some compromises between them, and pick an interval that gives a sufficiently high confidence, and also is quite precise. In other words, we would like to have a Template:Colored em interval that will cover with a Template:Colored em.
Terminologies
Now, let us formally define some terminologies related to Template:Colored em. Template:Colored definition Template:Colored remark Template:Colored definition Template:Colored example Template:Colored definition Template:Colored remark Template:Colored example Template:Colored remark Template:Colored example Template:Colored exercise
Construction of confidence intervals
After understanding what confidence interval is, we would like to know how to construct one naturally. A main way for such construction is using the Template:Colored em, which is defined below. Template:Colored definition Template:Colored remark After having such pivotal quantity , we can construct a confidence interval for by the following steps:
- For that value of , find such that [1] ( does not involve since is a pivotal quantity).
- After that, we can transform to since the expression of involves , as we have assumed (the resulting inequalities should be Template:Colored em to the original inequalities, that is, , so that ).
Template:Colored example Template:Colored exercise Template:Colored example Template:Colored example
Confidence intervals for means of normal distributions
In the following, we will use the concept of pivotal quantity to construct confidence intervals for means and variances of Template:Colored em distributions. After that, because of the central limit theorem, we can construct Template:Colored em confidence intervals for means and variances of other types of distributions that are not normal.
Mean of a normal distribution
Before discussing this confidence interval, let us first introduce a notation:
- is the upper percentile of at level , i.e. it satisfies where .
We can find (or calculate) the values of for different from Template:Colored em. Template:Colored theorem Template:Colored remark
Proof. Let . Since is a pivotal quantity (its distribution is independent from ), we set where is a constant (and does not involve ). Then, we have The result follows.
The following graph illustrates :
|
*-|-*
/##|##\
/###|###\ <----- area 1-a
/####|####\
/#####|#####\
/######|######\
/|######|######|\
area --*.|######|######|.*--
a/2 --> ....|######|######|.... <--- area a/2
------------*---------------
-z_{a/2} z_{a/2}
Template:Colored example Template:Colored example We have previously discussed a way to construct confidence interval for the mean when the variance is Template:Colored em. However, this is not always the case in practice. We may not know the variance, right? Then, we cannot use the in the confidence interval from the previous theorem.
Intuitively, one may think that we can use the Template:Colored em to "replace" the , according to the weak law of large number. Then, we can simply replace the unknown in the confidence interval by the known (or its realization for interval estimate). However, the flaw in this argument is that the sample size may not be large enough to apply the weak law of large number for approximation. Template:Colored remark So, you may now ask that when the sample size is large enough, can we do such "replacement" for approximation. The answer is Template:Colored em, and we will discuss in the last section about approximated confidence intervals.
Before that section, the confidence intervals discussed is Template:Colored em in the sense that no approximation is used to construct them. Therefore, the confidence intervals constructed "work" for Template:Colored em sample size, no matter how large or how small it is (it works even if the sample size is 1, although such confidence interval constructed may not be very "nice", in the sense that the width of the interval may be quite large).
Before discussing how to construct an confidence interval for the mean when the variance is unknown, we first give some results that are useful for deriving such confidence interval.
Proof.
(i) One may use Basu's theorem to prove this, but the details about Basu's theorem and the proof are omitted here, since they are a bit complicated.
(ii) We will use the following definition of chi-squared distribution : where are independent. Also, we will use the fact that the mgf of is .
Now, first let which follows since are independent. Then, we write as Applying the definition of chi-squared distribution, we have .
By (i), and are independent. Thus, (a function of ) is independent from (a function of ). Now, let and . Since and are independent, and also we have from above derivation, the mgf Since and , we can further write which implies that the mgf of is , which is exactly the mgf of . Hence, .
(iii) We will use the following definition of -distribution : where , , and and are independent.
After using this definition, it is easy to prove (iii) with (ii), as follows: By (ii), . Also, we know that and are independent since and are independent by (i). Then, it follows by the above definition that .
Using this proposition, we can prove the following theorem. Again, before discussing this confidence interval, let us introduce a notation:
- is the upper percentile of at level , i.e. it satisfies where .
Template:Colored theorem Template:Colored remark
Proof. By (iii) in the previous proposition, we have . Since is independent from , is a pivotal quantity of . Hence, we set where is a constant (-distribution is symmetric (about ), so we have ). It follows that The result follows.
Template:Colored example Template:Colored example
Difference in means of two normal distributions
Sometimes, apart from estimating mean of a Template:Colored em normal distribution, we would like to estimate the Template:Colored em in means of Template:Colored em normal distributions for making comparison. For example, apart from estimating the mean amount of time (lifetime) for a bulb until it burns out, we are often interested in estimating the Template:Colored em between life of two different bulbs, so that we know which of the bulbs will last longer in average, and then we know that bulb has a higher "quality".
First, let us discuss the case where the two normal distributions are Template:Colored em.
Now, the problem is that how should we construct a confidence interval for the Template:Colored em in two means. It seems that we can just construct two confidence intervals for each of the two means respectively. Then, the confidence interval for is . However, this is indeed incorrect since when we have and , it does Template:Colored em mean that (there are no results in probability that justify this).
On the other hand, it seems that since and are independent (since the normal distributions we are considering are independent), then we have Then, when and , we have so which means is a confidence interval.
However, this is actually also incorrect. The flaw is that "when and , we have " only means (we do not have the reverse subset inclusion in general). This in turn means So, is actually Template:Colored em a confidence interval (in general).
So, the above two "methods" to construct confidence intervals for difference in means of two independent normal distributions actually do not work. Indeed, we do Template:Colored em use the confidence interval for each of the two means, which is constructed previously, to construct a confidence interval for difference in the two means. Instead, we consider a Template:Colored em of the difference in the two means, which is a standard way for constructing confidence intervals.
Template:Colored theorem Template:Colored remark Template:Colored exercise Now, we will prove the above theorem based on the result shown in the previous exercise:
Proof. Let (from the previous exercise). Then, is a pivotal quantity of . Hence, we have
Template:Colored example Now, we will consider the case where the variances are Template:Colored em. In this case, the construction of the confidence interval for the difference in means is more complicated, and even more complicated when . Thus, we will only discuss the case where is unknown. As you may expect, we will also use some results mentioned previously for constructing confidence interval for when is unknown in this case. Template:Colored theorem Template:Colored remark
Proof. Let (the reason for this to follow is shown in a previous exercise). From a previous result, we know that and . Then, we know that the mgf of is and the mgf of is . Since the distributions and are independent, the mgf of is Hence, .
By the independence of sample mean and sample variance ( and are independent, and are independent), we can deduce that and are independent. Thus, by the definition of -distribution, follows . Therefore, is a pivotal quantity of . Hence, we have The result follows.
Now, what if the two normal distributions concerned are Template:Colored em? Clearly, we cannot use the above results anymore, and we need to develop a new method to construct a confidence interval for the difference of means in this case. In this case, we need to consider the notion of Template:Colored em.
Template:Colored proposition Template:Colored remark
Proof.
1. Independence of :
Since are independent, and are independent, it follows that are independent, which is what we want to show.
2. :
- To show that still follow normal distribution, we can consider the pdf of for each . The pdf can be obtained using the transformation of random variables formula: e.g., let and where and . Then, the pdf of obtained, which is the pdf of , should be in the form of normal distribution.
- However, since the actual derivation process is somewhat complicated, it is omitted here.
- Of course, the mean and variance of can be observed from the pdf of determined previously. Alternatively, before determining the pdf of , we can also know that the mean of is (we use the linearity of expectation here, which does not require independence assumption), and the variance of is ( and ).
Template:Colored corollary Template:Colored remark
Proof. From the previous proposition, we know that is a random sample from . Since is known, it follows from a previous theorem that a confidence interval for is
Template:Colored corollary Template:Colored remark Template:Colored exercise Template:Colored example
Confidence intervals for variances of normal distributions
Variance of a normal distribution
After discussing the confidence intervals for means of normal distributions, let us consider the confidence intervals for Template:Colored em of normal distributions. Similarly, we need to consider a pivotal quantity of . Can you suggest a pivotal quantity of , based on a previous result discussed?
Recall that we have , and is independent from with some suitable assumptions. Thus, this result gives us a pivotal quantity of , namely . Before discussing the theorem for constructing a confidence interval for . Let us introduce a notation:
- is the upper percentile of at level , i.e. it satisfies where .
Some values of can be found in the chi-squared table.
- To find the value of , locate the row for degrees of freedom and the column for "probability content" .
Template:Colored theorem Template:Colored remark
Proof. Since , set [2] Then, we have The result follows.
Template:Colored example Template:Colored remark
Ratio of variances of two independent normal distributions
Similar to the case for means, we would also sometimes like to compare the variances of two normal distributions. One may naturally expect that we should construct a confidence interval for Template:Colored em in variances, similar to the case for means. However, there are simple ways to do this, since we do not have some results that help with this construction. Therefore, we need to consider an alternative way to Template:Colored em the variances, without using the Template:Colored em in variances. Can you suggest a way?
Recall the definition of Template:Colored em in point estimation. Template:Colored em gives us a nice way to compare two variances without considering their difference, where the Template:Colored em of two variances is considered. Fortunately, we have some results that help us to construct a confidence interval for the Template:Colored em of two variances.
Recall that the definition of -distribution: if and are independent, then follows the -distribution with and degrees of freedom, denoted by . From the definition of -distribution, we can see that it involves a ratio of two independent chi-squared random variables. How can it be linked to the ratio of two variances?
Recall that we have with some suitable assumptions. This connects the variance with the chi-squared random variable, and thus we can use this property together with the definition of -distribution to construct a pivotal quantity, and hence a confidence interval.
Let us introduce a notation before discussing the construction of confidence interval:
- is the upper percentile of at level , i.e. it satisfies .
Some values of can be found in -tables (there is different -tables for different values of , and the row and column of each table indicates the first and second degrees of freedom respectively). Also, using the property that , we can obtain some more values of which are not included in the -tables. Template:Colored theorem Template:Colored remark
Proof. By the assumptions, we have Thus, by the definition of -distribution, we have which is a pivotal quantity of . Hence, we have as desired.
Apart from using this confidence interval to Template:Colored em variances (or standard deviations), it can also be useful to Template:Colored em some assumptions about variances. Let us illustrate these two usages in the following examples. Template:Colored example Template:Colored example Template:Colored remark
Approximated confidence intervals for means
Previously, the distributions for the population are assumed to be normal, but the distributions are often Template:Colored em normal in reality. So, does it mean our previous discussions are meaningless in reality? No. The discussions are indeed still quite meaningful in reality, since we can use the Template:Colored em to "connect" the distributions in reality (which are usually not normal) to normal distribution. Through this, we can construct Template:Colored em confidence intervals, since we use central limit theorem for approximation.
To be more precise, recall that the Template:Colored em suggests that with some suitable assumptions. Therefore, if the sample size is large enough (a rule of thumb: at least 30), then follows Template:Colored em standard normal distribution. Hence, is a pivotal quantity (approximately). Recall from the property of normal distribution that if is a random sample from , then we have Template:Colored em (not approximately), and we have used this for the pivotal quantity for the confidence interval for mean when variance is known, and also the confidence interval for when is known. Therefore, we can just use basically the same confidence interval in these cases, but we need to notice that such confidence intervals are approximated, but not exact since we have used the central limit theorem for constructing the pivot quantity.
Now, how about the other confidence intervals where the pivotal quantity is "not in this form"? In the confidence interval for difference in means when variance is unknown, the pivotal quantity is similar in some sense: (see the corresponding theorem for the meaning of the notations involved). Can we use the central limit theorem to conclude that when the distributions involved are not normal (but are still independent), and the sample sizes and are both large enough, then Template:Colored em? The answer is yes. For the proof, see the following exercise. Template:Colored exercise As a result, we know that we can again just use basically the same confidence interval in this case, but of course such confidence interval is approximated.
There are still some confidence intervals that are not considered yet. Let us first consider the confidence interval for mean when the variance is Template:Colored em.
Recall that we have mentioned that we can simply replace the "" by "" according to the weak law of large number, which is quite intuitive. But why can we do this? Consider the following theorem. Template:Colored theorem Template:Colored remark
Proof. Under the assumption that the random sample has finite mean and variance, applying weak law of large number gives (we have shown that , then we can just apply continuous mapping theorem to get this). Hence, () by property of convergence in probability.
By central limit theorem, we have . Thus, by Slutsky's theorem.
Therefore, is a pivotal quantity, which follows approximately. Notice that its approximated distribution, , is the same as that of pivotal quantity for confidence interval for when is known, namely . As a result, we can use similar steps to obtain the approximated confidence interval, where "" is replaced by "".
So far, we have not discussed how to construct an approximated confidence interval for when is unknown, as well as approximated confidence intervals of variances. Since the pivotal quantities used are constructed according to some results that are exclusive to normal distributions, they all do Template:Colored em work when the distributions involved are not normal. Therefore, there are no simple ways to perform such constructions.
The following table summarizes the approximated confidence intervals in different cases: Template:Colored remark Template:Colored example Let us consider an application of the approximated confidence intervals. Template:Colored proposition Template:Colored remark
Proof. Since , by the fundamental bridge between probability and expectation, we have
Applying the result for constructing an approximated confidence interval for mean when variance is unknown (the variance of is since follows the Bernoulli distribution actually), an approximated confidence interval for is
Template:Colored example Template:Colored example Template:Nav Template:BookCat
- β Usually, we choose and such that and because of convenience (if the pdf of is symmetric about , then we know that ).
- β
We need to do this since chi-squared distribution is Template:Colored em symmetric about .
Graphically, it looks like
| area: 1-a | # | | #....# v | # .......# | # |..........# |# |..........| # *---*----------*------ chi^2 1-a/2 chi^2 a/2