Page 164 - Elements of Distribution Theory
P. 164

P1: JZP
            052184472Xc05  CUNY148/Severini  May 24, 2005  17:53





                            150                   Parametric Families of Distributions

                              Suppose that λ is distributed according to a normal distribution with mean µ and standard
                            deviation τ. Note that X 1 has characteristic function
                                                                                 1  2 2

                                      E[exp(it X 1 )] = E{E[exp(it X 1 |λ)]}= E exp itλ − σ t
                                                                                 2
                                                                                         ;

                                                             1  2    2  2
                                                 = exp itµ − (τ + σ )t
                                                             2
                            it follows that X 1 is distributed according to a normal distribution with mean µ and variance
                             2
                                 2
                            τ + σ . Clearly, X 2 has the same marginal distribution as X 1 .
                              However, X 1 and X 2 are no longer independent. Since
                                                                             2
                                                                       2
                                                                                 2
                                             E(X 1 X 2 ) = E[E(X 1 X 2 |λ)] = E(λ ) = τ + µ ,
                            it follows that
                                                        Cov(X 1 , X 2 ) = τ 2
                                                                 2
                                                                         2
                                                                     2
                            and, hence, that the correlation of X 1 and X 2 is τ /(τ + σ ). Hence, although, conditionally
                            on λ, X 1 and X 2 are independent, marginally they are dependent random variables.
                              The distribution of (X 1 , X 2 )is called the bivariate normal distribution; its properties
                            will be considered in detail in Chapter 8.


                                                     5.5 Regression Models
                                                            d
                            Consider a parametric model on Y ⊂ R , P ={P(·; λ): λ ∈  }. Suppose that Y 1 ,..., Y n
                            are independent random variables such that, for each j = 1,..., n, the distribution of Y j is
                            the element of P corresponding to a parameter value λ j . Hence, Y 1 ,..., Y n are independent,
                            but are not necessarily identically distributed.
                              Let x 1 , x 2 ,..., x n denote a known sequence of nonrandom vectors such that, for each
                            j = 1,..., n, there exists a function h such that

                                                          λ j = h(x j ; θ)
                            for some θ in a set  . Thus, the distribution of Y j depends on the value of x j , along with the
                            value of θ and the function h. The vectors x 1 ,..., x n are known as covariates or explana-
                            tory variables; the random variables Y 1 ,..., Y n are called the response variables. The
                            response variables and covariates are sometimes called dependent and independent vari-
                            ables, respectively; however, those terms will not be used here, in order to avoid confusion
                            with the concept of independence, as discussed in Section 2.2.
                              In a regression model, the function h is known, while θ is an unknown parameter. Interest
                            generally centers on the relationship between Y j and x j , j = 1,..., n,asexpressed through
                            the function h.Regression models are very widely used in statistics; the goal of this section
                            is to present a few examples illustrating some of the regression models commonly used.


                            Example 5.20 (Additive error models). Suppose that each distribution in P has a finite
                            mean. Let

                                                        µ(x j ,θ) = E(Y j ; θ).
   159   160   161   162   163   164   165   166   167   168   169