Probability/Joint Distributions and Independence

From testwiki
Revision as of 06:07, 28 September 2022 by imported>LeoChiukl (β†’Independence)
(diff) ← Older revision | Latest revision (diff) | Newer revision β†’ (diff)
Jump to navigation Jump to search

Template:Nav

Template:TOC Template:Clear

Motivation

Suppose we are given a pmf of a discrete random variable X and a pmf of a discrete random variable Y. For example, fX(x)=(𝟏{x=0}+𝟏{x=1})/2andfY(y)=(𝟏{y=0}+𝟏{y=2})/2 We cannot tell the relationship between X and Y with only such information. They may be related or not related.

For example, the random variable X may be defined as X=1 if head comes up and X=0 otherwise from tossing a fair coin, and the random variable Y may be defined as Y=2 if head comes up and Y=0 otherwise from tossing the coin another time. In this case, X and Y are unrelated.

Another possibility is that the random variable Y is defined as Y=2X if head comes up in the first coin tossing, and Y=0 otherwise. In this case, X and Y are related.

Yet, in the above two examples, the pmf of X and Y are exactly the same.

Therefore, to tell the Template:Colored em between X and Y, we define the Template:Colored em cumulative distribution function, or joint cdf.

Joint distributions

Template:Colored definition Sometimes, we may want to know the random behaviour in one of the random variables involved in a joint cdf. We can do this by computing the marginal cdf from joint cdf. The definition of marginal cdf is as follows: Template:Colored definition Template:Colored remark Template:Colored proposition

Proof. When we set the arguments other than i-th argument to be , e.g. X1limxX1x , the joint cdf becomes β„™(X1Xi1XixXi+1Xn)=β„™(X1Xi1)1β„™(Xix)β„™(Xi+1Xn)1by independence=β„™(Xix)=FXi(x)

Template:Colored remark Template:Colored example Similar to the one-variable case, we have joint pmf and joint pdf. Also, analogously, we have marginal pmf and marginal pdf.

Template:Colored definition Template:Colored definition Template:Colored proposition

Proof. Consider the case in which there are only two random variables, say X and Y. Then, we have yf(x,y)=yβ„™(X=xY=y)=β„™(X=x)by law of total probability. Similarly, in general case, we have unf(u1,,ui1,x,ui+1,,un)=unβ„™(X1u1Xi1ui1XixXi+1ui+1Xn1un1Xnun)=β„™(X1u1Xi1ui1XixXi+1ui+1Xn1un1)by law of total probability. Then, we perform similar process on each of the other variables (n2 left), with one extra summation sign added for each process. Thus, in total we will have n1 summation sign, and we will finally get the desired result.

Template:Colored remark Template:Colored example Template:Colored exercise Template:Colored exercise Template:Hide For Template:Colored em continuous random variables, the definition is generalized version of the one for continuous random variables (univariate case). Template:Colored definition Template:Colored remark Template:Colored definition Template:Colored proposition

Proof. Recall the proposition about obtaining marginal cdf from joint cdf. We have FXi(x)=F(,,,xi-th position,,,)xfXi(u)du=xf(u1,,un)dunduidu1by definitionsddxxfXi(u)du=ddxxf(u1,,un)dunduidu1fXi(x)=n1integrationsf(u1,,ui1,x,ui+1,,un)du1dui1dui+1dunby fundamental theorem of calculus

Template:Colored proposition

Proof. It follows from using fundamental theorem of calculus n times.

Template:Colored example Template:Colored exercise

Independence

Recall that multiple events are independent if the probability for the intersection of them equals the product of probabilities of each event, by definition. Since {XA} is also an event, we have a natural definition of independence for Template:Colored em as follows: Template:Colored definition Template:Colored remark Template:Colored theorem

Proof. Partial:

Only if part: If random variables X1,X2,,Xn are independent, β„™(X1A1XnAn)=β„™(X1A1)β„™(XnAn) for each n and for each subset A1,A2,,Anℝ. Setting A1=(,x1),,An=(,xn), and we have β„™(X1x1Xnxn)=β„™(X1x1)β„™(Xnxn)F(x1,,xn)=FX1(x1)FXn(xn). Thus, we obtain the result for the joint cdf part.

For the joint pdf part, F(x1,,xn)=FX1(x1)FXn(xn)nx1xnF(x1,,xn)=nx1xn(FX1(x1)FXn(xn))f(x1,,xn)=fXn(xn)nx1xn1(FX1(x1)FXn1(xn1))=fXn(xn)fXn1(xn1)nx1xn2(FX1(x1)FXn2(xn2))==fX1(x1)fXn(xn)

Template:Colored remark Template:Colored example Template:Colored exercise Template:Colored proposition Template:Colored example Template:Colored exercise

Sum of independent random variables (optional)

In general, we use joint cdf, pdf or pmf to determine the distribution of sum of independent random variables by first principle. In particular, there are some interesting results related to the distribution of Template:Colored em of Template:Colored em random variables. Template:Collapse top Template:Colored proposition

Proof.

  • Continuous case:
  • cdf: FX+Y(z)=β„™(X+Yz)by definition=x+yzfX(x)fY(y)dxdyby definition and independence=zyfX(x)fY(y)dxdyby Fubini's theorem=(zyfX(x)dx)fY(y)dy=FX(zy)fY(y)dyby definition.
/\                                     
//\ y                                
///\|
////*
////|\
////|/\
////|//\ x+y=z <=> x=z-y
////|///\
////|////\
----*-----*--------------- x 
////|//////\
////|///////\

-->: -infty to z-y
^
|: -infty to infty
 
*--*
|//| : x+y <= z
*--*
  • pdf: fX+Y(z)=ddzFX(zy)fY(y)dy=ddzFX(zy)fY(y)dyby fundamental theorem of calculus=fX(zy)fY(y)dy.

Template:Colored remark Template:Colored example Template:Colored proposition

Proof.

  • Let Ei={X=i}{Y=ni}.
  • For each nonnegative integer n,

{X+Y=n}=E0E1En.

  • Since EiEj= for each ij, Ei's are pairwise disjoint.
  • Hence, by extended P3 and independence of X and Y, β„™(X+Y=n)=β„™(X=0)β„™(Y=n)+β„™(X=1)β„™(Y=n1)++β„™(X=n)β„™(Y=0).
  • The result follows by definition.

Template:Colored example Template:Colored proposition

Proof.

  • The pmf of X1+X2 is

fX1+X2(a)=k=0neλ1λ1kk!eλ2λ2nk(nk)!by the proposition about convolution of pmf's=eλ1λ2k=0nλ1kλ2nkk!(nk)!=e(λ1+λ2)n!k=0nn!k!(nk)!λ1kλ2nk=(λ1+λ2)n by binomial theorem.

  • This pmf as the pmf of Pois(λ1+λ2), and so X1+X2Pois(λ1+λ2).
  • We can extend this result to n Poisson r.v.'s by induction.

Template:Colored example Template:Collapse bottom

Order statistics

Template:Colored definition Template:Colored proposition

Proof.

  • Consider the event {X(k)x}.
                          Possible positions of x
                      |<--------------------->
    *---*----...------*----*------...--------*
X  (1)  (2)          (k)  (k+1)             (n)
                      |----------------------> when x moves RHS like this, >=k X_i are at the LHS of x
  • We can see from the above figure that {X(k)x}={at least k of the Xi's are x}.
  • Let no. of Xi's that are less than or equal to x be N.
  • Since NBinom(n,β„™(Xix))= def Binom(n,F(x)) (because for each Xi, we can treat Xix and Xi>x be the two outcomes in a Bernoulli trial),
  • The cdf is

β„™(X(k)x)=β„™(Nk)=j=kn(nj)(F(x))j(1F(x)).

Template:Colored example

Poisson process

Template:Colored definition There are several important properties for Poisson process. Template:Colored proposition

Proof.

  • The time to n-th event is X1++Xn, with each following Exp(λ).
  • It suffices to prove that X1+X2Gamma(2,λ), and then the desired result follows by induction.
  • fX1+X2(z)=λ2𝟏{zx0xz}𝟏{x0}eλ(zx)eλxdxby proposition about convolution of pdf's=λ20zeλ(zx)λxdx=λ20zeλzdx=λ2zeλz=λ2zeλzΓ(2)since Γ(2)=1!=1,
which is the pdf of Γ(2,λ), as desired.

Template:Colored remark Template:Colored proposition

Proof. For each nonnegative integer n, let V be the interarrival time between the n-th and n+1-th arrival, and W be the time to nth event, starting from the beginning of the fixed time interval (we can treat the start to be time zero because of the memoryless property). The joint pdf of (V,W) is f(v,w)=fV(v)fW(w)by independence=(λeλv)pdf ofExp(λ)(λnwn1eλw(n1)!)pdf ofGamma(n,λ). Let N the number of arrivals within the fixed time interval. The pmf of N is β„™(N=n)=β„™(WtV+W>tV>tW)=0ttwf(v,w)joint pdf of(V,W)dvdw=0ttw(λeλv)(λnwn1eλw(n1)!)dvdw=0tλnwn1eλw(n1)!twλeλvdvdw=λn(n1)!0twn1eλw(0(eλ(tw)))dw=λneλt(n1)!0twn1dw=λneλt(n1)!(tnn0)=eλt(λt)nn! which is the pmf of Pois(λt). The result follows.

Template:Colored proposition

Proof. For each t>0, β„™(T>t)=β„™(T1>tTn>t)=β„™(T1>t)β„™(Tn>t)by independence=[1(1eλ1tcdf ofExp(λ1))][1(1eλntcdf ofExp(λn))]=et(λ1++λn)β„™(Tt)=1et(λ1++λn)TExp(λ1+λ2++λn)

Template:Colored example Template:Colored exercise

Template:Nav