Books in black and white
 Main menu Share a book About us Home
 Books Biology Business Chemistry Computers Culture Economics Fiction Games Guide History Management Mathematical Medicine Mental Fitnes Physics Psychology Scince Sport Technics

# Detection, Estimation modulation theory part 1 - Vantress H.

Vantress H. Detection, Estimation modulation theory part 1 - Wiley & sons , 2001. - 710 p.
ISBN 0-471-09517-6 Previous << 1 .. 64 65 66 67 68 69 < 70 > 71 72 73 74 75 76 .. 254 >> Next It is clear that there may not be more than one solution. For example,
KJj, u) = af2 f{t) f{u), 0 < t,u <T. (49)
has only one nonzero eigenvalue and one normalized eigenfunction.
2. By looking at (46) we see that if <?;(t) is a solution then ñô,(ã) is also a solution. Therefore we can always normalize the eigenfunctions.
t Here we follow Davenport and Root , p. 373.
Properties of Integral Equations 181
3. If ô&) and ô2{t) are eigenfunctions associated with the same eigenvalue Ë, then c^x(t) + ñ2ô2{0 is also an eigenfunction associated with Ë.
4. The eigenfunctions corresponding to different eigenvalues are orthogonal.
5. There is at most a countably infinite set of eigenvalues and all are bounded.
6. For any particular A there is at most a finite number of linearly independent eigenfunctions. [Observe that we mean algebraic linear independence;/^) is linearly independent of the set ôõ(t), ³ = 1, 2Ê, if it cannot be written as a weighted sum of the &(/)•] These can always be orthonormalized (e.g., by the Gram-Schmidt procedure; see Problem 4.2.7 in Chapter 4).
7. Because Kx(t, u) is nonnegative definite, the kernel Kx(t, è) can be expanded in the series
00
Kx{t, «) = 2 Ô³(è)’ î < t, è < T, (50)
² = 1
where the convergence is uniform for 0 < t, è < T. (This is called Mercer’s theorem.)
8. If Kx(t, u) is positive definite, the eigenfunctions form a complete orthonormal set. From our results in Section 3.2 this implies that we can expand any deterministic function with finite energy in terms of the eigenfunctions.
9. If Kx(t, u) is not positive definite, the eigenfunctions cannot form a complete orthonormal set. [This follows directly from (35) and (40).] Frequently, we augment the eigenfunctions with enough additional orthogonal functions to obtain a complete set. We occasionally refer to these additional functions as eigenfunctions with zero eigenvalues.
10. The sum of the eigenvalues is the expected value of the energy of the process in the interval (0, Ã), that is,
E [Jo *] = Jo ^ ^ ^ 2 Ai*
(Recall that x(t) is assumed to be zero-mean.)
These properties guarantee that we can find a set of ô&) that leads to uncorrelated coefficients. It remains to verify the assumption that we made in (42). We denote the expected value of the error if x(f) is approximated by the first N terms as ?N(t):
(51)
182 3.3 Random Process Characterization
Evaluating the expectation, we have
ø = Kx(t, t) - 2e [x(o 2 xt m] + E [ 2 2 XiX< mo
‘_1 r æ (52)
Ø = Kx(t, t) - 2E^x(t) 2 (JJ x{u)Uu)du} ^(o] + 2A<-M0-M0,
*7 7"1 (53)
mo = Kx(t, o-22 (J*o Kx(t, è)^³(ì)j^,(0 + 2 êøø,
f_1 i_1 (54)
i = 1 y=l
JV
ûî-àäî- %\¹Ø-
(55)
i = l
Property 7 guarantees that the sum will converge uniformly to f) as TV-> oo. Therefore
which is the desired result. (Observe that the convergence in Property 7 implies that for any ˆ > 0 there exists an Nx independent of t such that Û0 < « for all N > N±).
The series expansion we have developed in this section is generally referred to as the Karhunen-Loeve expansion. (Karhunen , Loeve , p. 478, and .) It provides a second-moment characterization in terms of uncorrelated random variables. This property, by itself, is not too important. In the next section we shall find that for a particular process of interest, the Gaussian random process, the coefficients in the expansion are statistically independent Gaussian random variables. It is in this case that the expansion finds its most important application.
3.3.3 Gaussian Processes
We now return to the question of a suitable complete characterization of a random process. We shall confine our attention to Gaussian random processes. To find a suitable definition let us recall how we defined jointly Gaussian random variables in Section 2.6. We said that the random variables jci, x2,..xN were jointly Gaussian if
lim (N(t) = 0, 0 < t < T,
(56)
N
(57)
was a Gaussian random variable for any set of gt. In 2.6 N was finite and we required the gt to be finite. If N is countably infinite, we require the gt to be such that E[y2] < oo. In the random process, instead of a linear
Gaussian Processes 183
transformation on a set of random variables, we are interested in a linear functional of a random function. This suggests the following definition:
Definition. Let jc(0 be a random process defined over some interval [Ta, T0] with a mean-value mx{t) and covariance function Kx(t, w). If every linear functional of ë:(´) is a Gaussian random variable, then x(t) is a Gaussian random process. In other words, if Previous << 1 .. 64 65 66 67 68 69 < 70 > 71 72 73 74 75 76 .. 254 >> Next 