/ , / Marginal and Conditional Probability Distributions

Marginal and Conditional Probability Distributions

Marginal and Conditional Probability Distributions

With each two-dimensional random variable (X, Y) we associate two onedimensional
random variables, namely X and Y, individually. That is, we may be
interested in the probability distribution of X or the probability distribution of Y.

EXAMPLE 1
Let us again consider .Let us also compute the "marginal" totals, that is, the sum of the 6 columns and 4 rows of the table. (See Table 1)
          The probabilities appearing in the row and column margins represent the probability distribution of Y and X, respectively. For instance, P( Y = l) = 0.26, P(X = 3) = 0.21, etc. Because of the appearance of Table 1 we refer, quite generally, to the marginal distribution of X or the marginal distribution of Y, whenever we have a two-dimensional random variable (X, Y), whether discrete or continuous.

X\Y012345Sum
000.010.030.050.070.090.25
10.010.020.040.050.060.080.26
20.010.030.050.050.050.060.25
30.010.020.040.060.060.050.24
Sum0.030.080.160.210.240.281.00

In the discrete case we proceed as follows: Since X = xi must occur with Y = yj for some j and can occur with Y = yj for only one j, we have
p(xi) = P(X = xi) = P(X = xi, Y = y1 or X = xi, Y = y2 or···)
\[\sum\limits_{j=1}^{\infty }{p\left( {{x}_{i}},{{y}_{j}} \right)}\] The function p defined for x1, x2, • • . , represents the marginal probability distribution of X Analogously we define q(yj) = P( Y yj) = \[\sum\limits_{j=1}^{\infty }{p\left( {{x}_{i}},{{y}_{j}} \right)}\]  as the marginal probability distribution of Y. In the continuous case we proceed as follows: Let f be the joint pdf of the continuous two-dimensional random variable ( X, Y). We define g and h, the marginal probability density functions of X and Y, respectively, as follows: \[g\left( x \right)=\int_{-\infty }^{+\infty }{f\left( x,y \right)}dy;\] \[h\left( y \right)=\int_{-\infty }^{+\infty }{f\left( x,y \right)}dx;\]

These pdrs correspond to the basic pdrs of the one-dimensional random variables X and Y, respectively.
For example
P( c ≤ X ≤  d) = P [ c ≤ X ≤  d, - ∞ < Y < ∞]

\[\int_{c}^{d}{\int_{-\infty }^{+\infty }{f\left( x,y \right)dydx=\int_{c}^{d}{g\left( x \right)dx}}}\]

EXAMPLE 2
Two characteristics of a rocket engine's performance are thrust X and mixture ratio Y. Suppose that (X, Y) is a two-dimensional continuous random variable with joint pdf:

f(x, y) = 2(x + y - 2xy),              0 ≤ x ≤  1, 0 ≤  y ≤  1,
           = 0,                                    elsewhere.

(The units have been adjusted in order to use values between 0 and 1.) The marginal
pdf of X is given by

\[g\left( x \right)=\int_{0}^{1}{2\left( x+y-2xy \right)}dy=2\left( xy+\frac{{{y}^{2}}}{2}-x{{y}^{2}} \right)|_{0}^{1}\]

  =1 ,               0 ≤  x  ≤  1

That is, X is uniformly distributed over [O, 1 ].
The marginal pdf of Y is given by


\[h\left( y \right)=\int_{0}^{1}{2\left( x+y-2xy \right)}dx=2\left( xy+\frac{{{x}^{2}}}{2}-y{{x}^{2}} \right)|_{0}^{1}\]

  =1 ,               0 ≤  y  ≤  1

Hence Y is also uniformly distributed over [0, 1 ].

Definition. 
We say that the two-dimensional continuous random variable is uniformly distributed over a region R in the Euclidean plane if
f(x, y) = const                   for (x, y) ϵR,
           = 0,                          elsewhere.

Because of the requirement the above implies  \[\int_{-\infty }^{+\infty }{\int_{-\infty }^{+\infty }{f\left( x,y \right)dydx=1}}\] that the constant equals I/area (R). We are assuming that R is a region
with finite, nonzero area.
Note: This definition represents the two-dimensional analog to the one-dimensional
uniformly distributed random variable.
EXAMPLE 3
. Suppose that the two-dimensional random variable (X, Y) is uniformly distributed over the shaded region R
 f(x, y) = 1/area(R)'                    (x, y)  ϵ R.
We find that
\[area\left( R\right)=\int_{0}^{1}{\left( x-{{x}^{2}} \right)}dx=\frac{1}{6}\]

Therefore the pdf is given by
f(x,y) = 6,                        (x,y) R
          = 0,                        (x, y) . R.

In the following equations we find the marginal pdf's of X and Y.
\[g\left( x \right)=\int_{-\infty }^{+\infty }{f\left( x,y \right)dy=\int_{{{x}^{2}}}^{x}{6dy=6\left( x-{{x}^{2}} \right)}},\]                                              
0 ≤  x  ≤  1
\[h\left( y \right)=\int_{-\infty }^{+\infty }{f\left( x,y \right)dx=\int_{y}^{\sqrt{y}}{6dy=6\left( \sqrt{y}-y \right)}},\]
                                       0 ≤  y  ≤  1

The concept of conditional probability may be introduced in a very natural way.

EXAMPLE 4
Suppose that we want to evaluate the conditional probability P(X = 2 | Y = 2). According to the definition
of conditional probability we have

P(X = 2 | y = 2) ={P(X= 2 , y = 2)} / P(Y = 2)=0.05/ 0.25= 0 20

We can carry out such a computation quite generally for the discrete case. We have
p(xi | yj) = P(X = xi |Y = yj)
             =p(xi ,yj) /q(yi)                        if  q(yj ) > 0

q(yj | xi ) = P( y = yj | x = xi)
                 =p(x, yj) / p(xi)                 if p(xi) > 0.

Note: For given j, p(xi | yj) satisfies all the conditions for a probability distribution.
We have 0 ≤  p(xi | yj) and also

\[\sum\limits_{i=1}^{\infty }{p\left( {{x}_{i}}|{{y}_{j}} \right)=\sum\limits_{i=1}^{\infty }{\frac{p\left( {{x}_{i}}|{{y}_{j}} \right)}{q\left( {{y}_{j}} \right)}}}=\frac{q\left( {{y}_{j}} \right)}{q\left( {{y}_{j}} \right)}=1\]

In the continuous case the formulation of conditional probability presents some
difficulty since for any given x, y0 we have P(X = x) = P( Y =y) = 0. We
make the following formal definitions.

Definition. 
Let (X, Y) be a continuous two-dimensional random variable with joint pdf f Let g and h be the marginal pdrs of X and Y, respectively. The conditional pdf of X for given Y = y is defined by
g(x | y ) = f(x, y)' / h(y)                      h(y) > 0.
The conditional pdf of Y for given X = x is defined by
h(y | x) = f(x,y)' / g(x)                       g(x) > 0.

Notes:
(a) The above conditional pdf's satisfy all the requirements for a one-dimensional
pdf. Thus, for fixed y, we have g(x | y)  0 and

\[\int_{-\infty }^{+\infty }{g\left( x|y \right)=\int_{-\infty }^{+\infty }{\frac{f\left( x,y \right)}{h\left( y \right)}}}dx=\frac{1}{h\left( y \right)}\int_{-\infty }^{+\infty }{f\left( x,y \right)}dx=\frac{h\left( y \right)}{h\left( y \right)}=1\]

An analogous computation may be carried out for h(y | x).
(b) An intuitive interpretation of g(x | y) is obtained if we consider slicing the surface represented by the joint pdf /with the plane y = c, say. The intersection of the plane with the surface z = f(x, y) will result in a one-Dimensional pdf, namely the pdf of X for Y = c. This will be precisely g(x | c).
(c) Suppose that· (X, Y) represents the height and weight of a person, respectively. Let /be the joint pdf of (X, Y) and let g be the marginal pdf of X (irrespective of Y).

Hence  ∫65.8 g(x) dx would represent the probability of the event {5.8 ≤ X ≤  6} irrespective
of the weight Y. And ∫65.8g(x  | 150) dx would be interpreted as P(5.8 ≤ X 6 | Y = 150). Strictly speaking, this conditional probability is not defined in view of
our previous convention with conditional probability, since P(Y = 150) = 0. However,
we simply use the above integral to define this probability. Certainly on intuitive grounds
this ought to be the meaning of this number.

about author

Blogger Sens it website about blogger templates and blogger widgets you can find us on social media
Previous Post :Go to tne previous Post
Next Post:Go to tne Next Post

No comments:

Post a Comment