Page 356 - Fundamentals of Probability and Statistics for Engineers
P. 356
Linear Models and Linear Regression 339
^
at ^ and . Elementary calculations show that
2
q Q
2n > 0;
q^ 2
and
n
X 2
D 4n
x i x > 0
i1
The proof of this theorem is thus complete. Note that D would be zero if all
x i take the same value. Hence, at least two distinct x i values are needed for the
^
determination of ^ and .
It is instructive at this point to restate the foregoing results by using a more
compact vector–matrix notation. As we will see, results in vector–matrix form
facilitate calculations. Also, they permit easy generalizations when we consider
more general regression models.
In terms of observed sample values (x 1 , y 1 ), (x 2 , y 2 ), ..., (x n , y n ), we have a
system of observed regression equations
y i x i e i ; i 1; 2; ... ; n:
11:11
Let
2 3 2 3 2 3
1 x 1 y 1 e 1
6 1 7 6 y 2 7 6 7
. . 7; . 7; . 7;
6 x 2 7 6 7 6 e 2 7
.
C 6 y 6 e 6
4 . . . 5 4 . . 5 4 . . 5
1 x n y n e n
and let
q :
Equations (11.11) can be represented by the vector–matrix equation
y Cq e:
11:12
The sum of squared residuals given by Equation (11.6) is now
T
T
Q e e
y Cq
y Cq:
11:13
TLFeBOOK