Statistics/Interval Estimation

From testwiki
Jump to navigation Jump to search

Template:Nav

Introduction

Previously, we have discussed Template:Colored em, which gives us an estimator θ^ for the value of an unknown parameter θ. Now, suppose we want to know the size of Template:Colored em of the point estimator θ^, i.e. the difference between θ^ and the unknown parameter θ. Of course, we can make use of the value of the Template:Colored em of θ^, 𝔼[(θ^θ)2], or other things.

However, what if we only know about one specific Template:Colored em? We cannot calculate the mean squared error of its corresponding point estimator with Template:Colored em this point estimates, right? So, how do we know the possible size of error of this Template:Colored em? Indeed, it is impossible to tell, since we are only given a particular estimated value of parameter θ, but of course we do know the value of the unknown parameter θ, thus the difference between this point estimate and θ is also unknown.

To illustrate this, consider the following example: suppose we take a random sample of 10 students from one particular course in university to estimate the mean score of the students in the final exam in that course, denoted by μ, (assume the score is normally distributed), and the observed value of the sample mean is x=60. Then, what is the difference between this point estimate and the true unknown parameter μ? Can we be "confident" that this sample mean is close to μ, say μ[x5,x+5]=[55,65]?

It is possible that μ is, say 90, and somehow the students in the sample are the one with very poor performance. On the other hand, it is also possible that μ is, say 30, and somehow the students in the sample are the one who perform well (relatively). Of course, it is also possible that the μ is quite close to 60, say 59. From this example, we can see that a particular value x=60 does not tell us the possible size of error: the error can be very large, and also can be very small.

In this chapter, we will introduce Template:Colored em where we use Template:Colored em that can describe the size of error through providing the probability for the random interval (i.e. interval with at least one of its bounds to be a random variable) given by the interval estimator to contain the unknown parameter θ, which measures the "accuracy" of the interval estimator of θ, and hence the size of error.

As suggested by the name Template:Colored em, the estimator involves some sort of Template:Colored em. Also, as one may expect, Template:Colored em is also based on Template:Colored em: Template:Colored definition Of course, we would like the probability for the unknown parameter θ to lie in the interval to be close to 1, so that the interval estimator is very accurate. However, a very accurate interval estimator may have a very bad "precision", i.e. the interval covers "too many" plausible values of an unknown parameter, and therefore even if we know that θ is very likely to be one of such values, there are too many different possibilities. Hence, such interval estimator is not very "useful". To illustrate this, suppose the interval concerned is ℝ, which is the parameter space of θ. Then, of course β„™(θℝ)=1 (so the "confidence" is high) since θ must lie in its parameter space. However, such interval has basically "zero precision", and is quite "useless", since the "plausible values" of θ in the intervals are essentially all possible values of θ.

From this, we can observe the need of the "precision" of the interval, that is, we also want the Template:Colored em of the interval to be small, so that we can have some ideas about the "location" of θ. However, as the interval becomes smaller, it is more likely that such interval misses θ, i.e. does not cover the actual value of θ, and therefore the probability for θ to lie in that interval becomes smaller, i.e. the interval becomes less "accurate". To illustrate this, let us consider the extreme case: the interval is so small that it becomes an interval containing a single point (the two end-points of the interval coincide). Then, the "interval estimator" basically becomes a "point estimator" in some sense, and we know that it is very unlikely that the true value of θ equals the value of the point estimator θ^ (θ lies in that "interval" is equivalent to θ=θ^ in this case). Indeed, if the distribution of θ^ is Template:Colored em, then β„™(θ^=θ)=0.

As we can see from above, although we want the interval to have a very high "confidence" and also "very precise" (i.e. the interval is very narrow), we cannot have both of them, since an increase in confidence causes a decrease in "precision", and an increase in "precision" causes a decrease in confidence. Therefore, we need to make some compromises between them, and pick an interval that gives a sufficiently high confidence, and also is quite precise. In other words, we would like to have a Template:Colored em interval that will cover θ with a Template:Colored em.

Terminologies

Now, let us formally define some terminologies related to Template:Colored em. Template:Colored definition Template:Colored remark Template:Colored definition Template:Colored example Template:Colored definition Template:Colored remark Template:Colored example Template:Colored remark Template:Colored example Template:Colored exercise

Construction of confidence intervals

After understanding what confidence interval is, we would like to know how to construct one naturally. A main way for such construction is using the Template:Colored em, which is defined below. Template:Colored definition Template:Colored remark After having such pivotal quantity Q(𝐗,θ), we can construct a 1α confidence interval for θ by the following steps:

  1. For that value of α, find a,b such that β„™(aQ(𝐗,θ)b)=1α [1] (a,b does not involve θ since Q(𝐗,θ) is a pivotal quantity).
  2. After that, we can transform aQ(𝐗,θ)b to L(𝐗)θU(𝐗) since the expression of Q(𝐗,θ) involves θ, as we have assumed (the resulting inequalities should be Template:Colored em to the original inequalities, that is, aQ(𝐗,θ)bL(𝐗)θU(𝐗), so that β„™(L(𝐗)θU(𝐗))=β„™(aQ(𝐗,θ)b)).

Template:Colored example Template:Colored exercise Template:Colored example Template:Colored example

Confidence intervals for means of normal distributions

In the following, we will use the concept of pivotal quantity to construct confidence intervals for means and variances of Template:Colored em distributions. After that, because of the central limit theorem, we can construct Template:Colored em confidence intervals for means and variances of other types of distributions that are not normal.

Mean of a normal distribution

Before discussing this confidence interval, let us first introduce a notation:

  • zα is the upper percentile of 𝒩(0,1) at level α, i.e. it satisfies β„™(Zzα)=α where Z𝒩(0,1).

We can find (or calculate) the values of zα for different α from Template:Colored em. Template:Colored theorem Template:Colored remark

Proof. Let Z=Xμσ/n𝒩(0,1). Since Z is a pivotal quantity (its distribution is independent from μ), we set 1α=1β„™(Zzα/2)β„™(Zzα/2)=β„™(zα/2<Z<zα/2)=β„™(zα/2Zzα/2), where zα/2 is a constant (and does not involve μ). Then, we have 1α=β„™(zα/2Zzα/2)=β„™(zα/2Xμσ/nzα/2)=β„™(zα/2σnXμzα/2σn)=β„™(zα/2σnμXzα/2σn)=β„™(zα/2σnμXzα/2σn)(rewrite)=β„™(Xzα/2σnμX+zα/2σn). The result follows.

The following graph illustrates β„™(zα/2Zzα/2)=1α:

                    |
                  *-|-*
                 /##|##\   
                /###|###\  <----- area 1-a
               /####|####\
              /#####|#####\
             /######|######\
            /|######|######|\
 area    --*.|######|######|.*-- 
 a/2 --> ....|######|######|....  <---  area a/2
        ------------*---------------
           -z_{a/2}       z_{a/2}

Template:Colored example Template:Colored example We have previously discussed a way to construct confidence interval for the mean when the variance is Template:Colored em. However, this is not always the case in practice. We may not know the variance, right? Then, we cannot use the σ in the confidence interval from the previous theorem.

Intuitively, one may think that we can use the Template:Colored em S2 to "replace" the σ2, according to the weak law of large number. Then, we can simply replace the unknown σ in the confidence interval by the known S (or its realization s for interval estimate). However, the flaw in this argument is that the sample size may not be large enough to apply the weak law of large number for approximation. Template:Colored remark So, you may now ask that when the sample size is large enough, can we do such "replacement" for approximation. The answer is Template:Colored em, and we will discuss in the last section about approximated confidence intervals.

Before that section, the confidence intervals discussed is Template:Colored em in the sense that no approximation is used to construct them. Therefore, the confidence intervals constructed "work" for Template:Colored em sample size, no matter how large or how small it is (it works even if the sample size is 1, although such confidence interval constructed may not be very "nice", in the sense that the width of the interval may be quite large).

Before discussing how to construct an confidence interval for the mean when the variance is unknown, we first give some results that are useful for deriving such confidence interval.

Template:Colored proposition

Proof.

(i) One may use Basu's theorem to prove this, but the details about Basu's theorem and the proof are omitted here, since they are a bit complicated.

(ii) We will use the following definition of chi-squared distribution χk2 : i=1kZi2χk2 where Z1,Z2,,Zk𝒩(0,1) are independent. Also, we will use the fact that the mgf of χk2 is M(t)=(12t)k/2,t<12.

Now, first let W=i=1n(Xiμσ)2 which follows χn2 since X1μσ,,Xnμσ𝒩(0,1) are independent. Then, we write W as W=i=1n(Xiμσ)2=i=1n(XiXσ+Xμσ)2=i=1n(XiXσ)2+i=1n(Xμσ)2+0(2i=1n(Xμ)(XiX)σ2=2(Xμ)σ2(i=1nXi=nXi=1nXconstant wrt i=nX)=0)=1σ2i=1n(XiX)2=nS2+(n(Xμ)σ𝒩(0,1) by property)2=nS2σ2+Z2(Z=n(Xμ)σ𝒩(0,1)) Applying the definition of chi-squared distribution, we have Z2χ12.

By (i), X and S2 are independent. Thus, nS2σ2 (a function of S2) is independent from Z2 (a function of X). Now, let U=nS2σ2 and V=Z2. Since U and V are independent, and also we have W=U+V from above derivation, the mgf MW(t)=MU+V(t)=MU(t)MV(t). Since Wχn2 and Vχ12, we can further write (12t)n/2=MU(t)(12t)1/2,t<12, which implies that the mgf of U is MU(t)=(12t)(n1)/2,t<12, which is exactly the mgf of χn12. Hence, U=nS2σ2χn12.

(iii) We will use the following definition of t-distribution tk: ZY/ktk where Z𝒩(0,1), Yχk2, and Z and Y are independent.

After using this definition, it is easy to prove (iii) with (ii), as follows: XμS/n1=n(Xμ)/σnS/(σn1)=n(Xμ)/σnS2σ2/n1. By (ii), nS2σ2χn12. Also, we know that n(Xμ)σ and nS2σ2 are independent since X and S2 are independent by (i). Then, it follows by the above definition that XμS/n1tn1.

Using this proposition, we can prove the following theorem. Again, before discussing this confidence interval, let us introduce a notation:

  • tα,ν is the upper percentile of tν at level α, i.e. it satisfies β„™(Ttα,ν)=α where Ttν.

Template:Colored theorem Template:Colored remark

Proof. By (iii) in the previous proposition, we have T=XμS/n1tn1. Since tn1 is independent from μ, T is a pivotal quantity of μ. Hence, we set 1α=1β„™(Ttα/2,n1))β„™(Ttα/2,n1)=β„™(tα/2,n1Ttα/2,n1) where tα/2,n1 is a constant (t-distribution is symmetric (about x=0), so we have β„™(Ttα/2,n1)=α/2). It follows that 1α=β„™(tα/2,n1Ttα/2,n1)=β„™(tα/2,n1XμS/n1tα/2,n1)=β„™(tα/2,n1Sn1Xμtα/2,n1Sn1)=β„™(tα/2,n1Sn1μXtα/2,n1Sn1)=β„™(tα/2,n1Sn1μXtα/2,n1Sn1)(rewrite)=β„™(Xtα/2,n1Sn1μX+tα/2,n1Sn1). The result follows.

Template:Colored example Template:Colored example

Difference in means of two normal distributions

Sometimes, apart from estimating mean of a Template:Colored em normal distribution, we would like to estimate the Template:Colored em in means of Template:Colored em normal distributions for making comparison. For example, apart from estimating the mean amount of time (lifetime) for a bulb until it burns out, we are often interested in estimating the Template:Colored em between life of two different bulbs, so that we know which of the bulbs will last longer in average, and then we know that bulb has a higher "quality".

First, let us discuss the case where the two normal distributions are Template:Colored em.

Now, the problem is that how should we construct a confidence interval for the Template:Colored em in two means. It seems that we can just construct two 1α confidence intervals [L(𝐗),U(𝐗)],[L(𝐘),U(𝐘)] for each of the two means μX,μY respectively. Then, the 1α confidence interval for μXμY is [L(𝐗)L(𝐘),U(𝐗)U(𝐘)]. However, this is indeed incorrect since when we have β„™(L(𝐗)μXU(𝐗))=1α and β„™(L(𝐘)μYU(𝐘))=1α, it does Template:Colored em mean that β„™(L(𝐗)L(𝐘)μXμYU(𝐗)U(𝐘))=1α (there are no results in probability that justify this).

On the other hand, it seems that since {L(𝐗)μXU(𝐗)} and {L(𝐘)μYU(𝐘)} are independent (since the normal distributions we are considering are independent), then we have β„™(L(𝐗)μXU(𝐗) and L(𝐘)μYU(𝐘))=(1α)2. Then, when L(𝐗)μXU(𝐗) and L(𝐘)μYU(𝐘), we have L(𝐗)U(𝐘)μXμYU(𝐗)L(𝐘), so β„™(L(𝐗)U(𝐘)μXμYU(𝐗)L(𝐘))=(1α)2, which means [L(𝐗)U(𝐘),U(𝐗)L(𝐘)] is a (1α)2 confidence interval.

However, this is actually also incorrect. The flaw is that "when L(𝐗)μXU(𝐗) and L(𝐘)μYU(𝐘), we have L(𝐗)U(𝐘)μXμYU(𝐗)L(𝐘)" only means {L(𝐗)μXU(𝐗) and L(𝐘)μYU(𝐘)}{L(𝐗)U(𝐘)μXμYU(𝐗)L(𝐘)} (we do not have the reverse subset inclusion in general). This in turn means (1α)2=β„™(L(𝐗)μXU(𝐗) and L(𝐘)μYU(𝐘))β„™(L(𝐗)U(𝐘)μXμYU(𝐗)L(𝐘)). So, [L(𝐗)U(𝐘),U(𝐗)L(𝐘)] is actually Template:Colored em a (1α)2 confidence interval (in general).

So, the above two "methods" to construct confidence intervals for difference in means of two independent normal distributions actually do not work. Indeed, we do Template:Colored em use the confidence interval for each of the two means, which is constructed previously, to construct a confidence interval for difference in the two means. Instead, we consider a Template:Colored em of the difference in the two means, which is a standard way for constructing confidence intervals.

Template:Colored theorem Template:Colored remark Template:Colored exercise Now, we will prove the above theorem based on the result shown in the previous exercise:

Proof. Let Z=(XY)(μXμY)σX2/n+σY2/m𝒩(0,1) (from the previous exercise). Then, Z is a pivotal quantity of μXμY. Hence, we have 1α=β„™(zα/2Zzα/2)=β„™(zα/2(XY)(μXμY)σX2/n+σY2/mzα/2)=β„™(zα/2σX2/n+σY2/m(XY)(μXμY)zα/2σX2/n+σY2/n)=β„™(zα/2σX2n+σY2m(μXμY)(XY)zα/2σX2n+σY2m)=β„™((XY)zα/2σX2n+σY2mμXμY(XY)+zα/2σX2n+σY2m).

Template:Colored example Now, we will consider the case where the variances are Template:Colored em. In this case, the construction of the confidence interval for the difference in means is more complicated, and even more complicated when σX2σY2. Thus, we will only discuss the case where σX2=σY2 is unknown. As you may expect, we will also use some results mentioned previously for constructing confidence interval for μ when σ2 is unknown in this case. Template:Colored theorem Template:Colored remark

Proof. Let Z=(XY)(μXμY)σ2/n+σ2/m𝒩(0,1) (the reason for this to follow 𝒩(0,1) is shown in a previous exercise). From a previous result, we know that V=nSX2σ2χn12 and W=mSY2σ2χm12. Then, we know that the mgf of V is MV(t)=(12t)(n1)/2 and the mgf of W is MW(t)=(12t)(m1)/2. Since the distributions 𝒩(μX,σ2) and 𝒩(μY,σ2) are independent, the mgf of U=V+W is MU(t)=MV+W(t)=MV(t)MW(t)=(12t)(n1)/2(m1)/2=(12t)(n+m2)/2. Hence, Wχn+m22.

By the independence of sample mean and sample variance (X and SX2 are independent, Y and SY2 are independent), we can deduce that Z and U are independent. Thus, by the definition of t-distribution, T=ZU/(n+m2)=((XY)(μXμY))/σ2/n+σ2/m(nSX2+mSY2)/(σ2(n+m2))=((XY)(μXμY))/(σ1/n+1/m)(1/σ)(nSX2+mSY2)/(n+m2)(this step is not possible without the equal variance assumption)=(XY)(μXμY)1/n+1/m(nSX2+mSY2)/(n+m2)=((XY)(μXμY))/nSX2+mSY2n+m2(1n+1m)=(XY)(μXμY)R(R=nSX2+mSY2n+m2(1n+1m)) follows tn+m2. Therefore, Ttn+m2 is a pivotal quantity of μXμY. Hence, we have 1α=β„™(tα/2,n+m2Ttα/2,n+m2)=β„™(tα/2,n+m2(XY)(μXμY)Rtα/2,n+m2)=β„™(tα/2,n+m2R(XY)(μXμY)tα/2,n+m2R)=β„™((XY)+tα/2,n+m2R(μXμY)(XY)tα/2,n+m2R)=β„™((XY)tα/2,n+m2R(μXμY)(XY)+tα/2,n+m2R)=β„™((XY)tα/2,n+m2nSX2+mSY2n+m2(1n+1m)(μXμY)(XY)+tα/2,n+m2nSX2+mSY2n+m2(1n+1m)). The result follows.

Template:Colored example

Now, what if the two normal distributions concerned are Template:Colored em? Clearly, we cannot use the above results anymore, and we need to develop a new method to construct a confidence interval for the difference of means in this case. In this case, we need to consider the notion of Template:Colored em.

Template:Colored proposition Template:Colored remark

Proof.

1. Independence of D1,,Dn:

Since X1,,Xn are independent, and Y1,,Yn are independent, it follows that X1Y1,,XnYn are independent, which is what we want to show.

2. D1,,Dn𝒩(μXμY,σX2+σY2+2Cov(X,Y)):

  • To show that D1,,Dn still follow normal distribution, we can consider the pdf of Di for each i{1,,n}. The pdf can be obtained using the transformation of random variables formula: e.g., let U=XY and V=Y where X𝒩(μX,σX2) and 𝒩(μY,σY2). Then, the pdf of U obtained, which is the pdf of Di, should be in the form of normal distribution.
  • However, since the actual derivation process is somewhat complicated, it is omitted here.
  • Of course, the mean and variance of Di can be observed from the pdf of Di determined previously. Alternatively, before determining the pdf of Di, we can also know that the mean of Di is 𝔼[Di]=𝔼[Xi]𝔼[Yi]=μXμY (we use the linearity of expectation here, which does not require independence assumption), and the variance of Di is Var(Di)=Var(Xi)+(1)2Var(Yi)+2Cov(Xi,Yi)=σX2+σY2+2(1)Cov(Xi,Yi)=σX2+σY22Cov(X,Y) (X𝒩(μX,σX2) and Y𝒩(μY,σY2)).

Template:Colored corollary Template:Colored remark

Proof. From the previous proposition, we know that D1,,Dn is a random sample from 𝒩(μD,σD2). Since σD2 is known, it follows from a previous theorem that a 1α confidence interval for μD=μXμY is [Dzα/2σDn,D+zα/2σDn].

Template:Colored corollary Template:Colored remark Template:Colored exercise Template:Colored example

Confidence intervals for variances of normal distributions

Variance of a normal distribution

After discussing the confidence intervals for means of normal distributions, let us consider the confidence intervals for Template:Colored em of normal distributions. Similarly, we need to consider a pivotal quantity of σ2. Can you suggest a pivotal quantity of σ2, based on a previous result discussed?

Recall that we have nS2σ2χn12, and χn12 is independent from σ2 with some suitable assumptions. Thus, this result gives us a pivotal quantity of σ2, namely nS2σ2. Before discussing the theorem for constructing a confidence interval for σ2. Let us introduce a notation:

  • χα,ν2 is the upper percentile of χν2 at level α, i.e. it satisfies β„™(Xχα,ν2)=α where Xχν2.

Some values of χα,ν2 can be found in the chi-squared table.

  • To find the value of χα,ν2, locate the row for ν degrees of freedom and the column for "probability content" α.

Template:Colored theorem Template:Colored remark

Proof. Since Y=nS2σ2χn12, set 1α=(1α/2)α/2=β„™(Yχ1α/2,n12)β„™(Yχα/2,n12)=β„™(χ1α/2,n12Y<χα/2,n12)=β„™(χ1α/2,n12Y<χα/2,n12). [2] Then, we have 1α=β„™(χ1α/2,n12nS2σ2χα/2,n12)=β„™(χ1α/2,n12nS21σ2χα/2,n12nS2)=β„™(nS2χ1α/2,n12σ2nS2χα/2,n12)=β„™(nS2χα/2,n12σ2nS2χ1α/2,n12). The result follows.

Template:Colored example Template:Colored remark

Ratio of variances of two independent normal distributions

Similar to the case for means, we would also sometimes like to compare the variances of two normal distributions. One may naturally expect that we should construct a confidence interval for Template:Colored em in variances, similar to the case for means. However, there are simple ways to do this, since we do not have some results that help with this construction. Therefore, we need to consider an alternative way to Template:Colored em the variances, without using the Template:Colored em in variances. Can you suggest a way?

Recall the definition of Template:Colored em in point estimation. Template:Colored em gives us a nice way to compare two variances without considering their difference, where the Template:Colored em of two variances is considered. Fortunately, we have some results that help us to construct a confidence interval for the Template:Colored em of two variances.

Recall that the definition of F-distribution: if Uχν12 and Vχν22 are independent, then U/ν1V/ν2 follows the F-distribution with ν1 and ν2 degrees of freedom, denoted by Fν1,ν2. From the definition of F-distribution, we can see that it involves a ratio of two independent chi-squared random variables. How can it be linked to the ratio of two variances?

Recall that we have nS2σ2χn12 with some suitable assumptions. This connects the variance with the chi-squared random variable, and thus we can use this property together with the definition of F-distribution to construct a pivotal quantity, and hence a confidence interval.

Let us introduce a notation before discussing the construction of confidence interval:

  • Fα,ν1,ν2 is the upper percentile of Fν1,ν2 at level α, i.e. it satisfies β„™(XFα,ν1,ν2)=α.

Some values of Fα,ν1,ν2 can be found in F-tables (there is different F-tables for different values of α, and the row and column of each table indicates the first and second degrees of freedom respectively). Also, using the property that Fα,ν1,ν2=1F1α,ν2,ν1, we can obtain some more values of Fα,ν1,ν2 which are not included in the F-tables. Template:Colored theorem Template:Colored remark

Proof. By the assumptions, we have nSX2σX2χn12 and mSY2σY2χm12. Thus, by the definition of F-distribution, we have mSY2σY2(m1)/nSX2σX2(n1)=m(n1)SY2σX2n(m1)SX2σY2Fm1,n1, which is a pivotal quantity of σX2/σY2. Hence, we have 1α=(1α/2)α/2=β„™(F1α/2,m1,n1m(n1)SY2σX2n(m1)SY2σX2Fα/2,m1,n1)=β„™(n(m1)SX2m(n1)SY2F1α/2,m1,n1σX2σY2n(m1)SX2m(n1)SY2Fα/2,m1,n1), as desired.

Apart from using this confidence interval to Template:Colored em variances (or standard deviations), it can also be useful to Template:Colored em some assumptions about variances. Let us illustrate these two usages in the following examples. Template:Colored example Template:Colored example Template:Colored remark

Approximated confidence intervals for means

Previously, the distributions for the population are assumed to be normal, but the distributions are often Template:Colored em normal in reality. So, does it mean our previous discussions are meaningless in reality? No. The discussions are indeed still quite meaningful in reality, since we can use the Template:Colored em to "connect" the distributions in reality (which are usually not normal) to normal distribution. Through this, we can construct Template:Colored em confidence intervals, since we use central limit theorem for approximation.

To be more precise, recall that the Template:Colored em suggests that Xμσ/ndZ𝒩(0,1) with some suitable assumptions. Therefore, if the sample size n is large enough (a rule of thumb: at least 30), then Xμσ/n follows Template:Colored em standard normal distribution. Hence, Xμσ/n is a pivotal quantity (approximately). Recall from the property of normal distribution that if X1,,Xn is a random sample from 𝒩(μ,σ2), then we have Xμσ/n𝒩(0,1) Template:Colored em (not approximately), and we have used this for the pivotal quantity for the confidence interval for mean when variance is known, and also the confidence interval for μXμY when σD2 is known. Therefore, we can just use basically the same confidence interval in these cases, but we need to notice that such confidence intervals are approximated, but not exact since we have used the central limit theorem for constructing the pivot quantity.

Now, how about the other confidence intervals where the pivotal quantity is "not in this form"? In the confidence interval for difference in means when variance is unknown, the pivotal quantity is similar in some sense: (XY)(μXμY)σX2/n+σY2/m𝒩(0,1) (see the corresponding theorem for the meaning of the notations involved). Can we use the central limit theorem to conclude that when the distributions involved are not normal (but are still independent), and the sample sizes n and m are both large enough, then (XY)(μXμY)σX2/n+σY2/m𝒩(0,1) Template:Colored em? The answer is yes. For the proof, see the following exercise. Template:Colored exercise As a result, we know that we can again just use basically the same confidence interval in this case, but of course such confidence interval is approximated.

There are still some confidence intervals that are not considered yet. Let us first consider the confidence interval for mean when the variance is Template:Colored em.

Recall that we have mentioned that we can simply replace the "σ" by "S" according to the weak law of large number, which is quite intuitive. But why can we do this? Consider the following theorem. Template:Colored theorem Template:Colored remark

Proof. Under the assumption that the random sample has finite mean and variance, applying weak law of large number gives Spσ (we have shown that S2pσ2, then we can just apply continuous mapping theorem to get this). Hence, σSpσσ=1 (σ>0) by property of convergence in probability.

By central limit theorem, we have Xμσ/ndZ𝒩(0,1). Thus, XμS/n=Xμσ/nσSdZ𝒩(0,1) by Slutsky's theorem.

Therefore, XμS/n is a pivotal quantity, which follows 𝒩(0,1) approximately. Notice that its approximated distribution, 𝒩(0,1), is the same as that of pivotal quantity for confidence interval for μ when σ2 is known, namely Xμσ/n. As a result, we can use similar steps to obtain the approximated confidence interval, where "σ" is replaced by "S".

So far, we have not discussed how to construct an approximated confidence interval for μXμY when σX2=σY2=σ2 is unknown, as well as approximated confidence intervals of variances. Since the pivotal quantities used are constructed according to some results that are exclusive to normal distributions, they all do Template:Colored em work when the distributions involved are not normal. Therefore, there are no simple ways to perform such constructions.

The following table summarizes the approximated 1α confidence intervals in different cases: (approximated (1α) confidence intervals)meandifference in meansknown variance[Xzα/2σn,X+zα/2σn][(XY)zα/2σX2n+σY2m,(XY)+zα/2σX2n+σY2m] OR [Dzα/2σDn,D+zα/2σDn](paired samples)unknown variance[Xzα/2Sn,X+zα/2Sn][Dzα/2SDn,D+zα/2SDn](paired samples) Template:Colored remark Template:Colored example Let us consider an application of the approximated confidence intervals. Template:Colored proposition Template:Colored remark

Proof. Since ξ=𝟏{XS}, by the fundamental bridge between probability and expectation, we have 𝔼[ξ]=𝔼[𝟏{XS}]=β„™(XS)=p.

Applying the result for constructing an approximated confidence interval for mean when variance is unknown (the variance of ξ is p(1p) since ξ follows the Bernoulli distribution Ber(p) actually), an approximated 1α confidence interval for p is [ξzα/2Sξn,ξ+zα/2Sξn].

Template:Colored example Template:Colored example Template:Nav Template:BookCat

  1. ↑ Usually, we choose a and b such that β„™(Q(𝐗,θ)<a)=α/2 and β„™(Q(𝐗,θ)>b)=α/2 because of convenience (if the pdf of Q(𝐗,θ) is symmetric about x=0, then we know that b=a).
  2. ↑ We need to do this since chi-squared distribution is Template:Colored em symmetric about x=0. Graphically, it looks like
    |         area: 1-a
    |     #    |
    |   #....# v
    |  # .......#   
    | # |..........#
    |#  |..........|  #
    *---*----------*------
    chi^2 1-a/2  chi^2 a/2