<< Chapter < Page
  Signal theory   Page 1 / 1
Chapter >> Page >
Introduces the concept of the eigendecomposition of a linear operator, with properties and examples.

Definition 1 A scalar λ is an eigenvalue of A B ( X , X ) if there exists a vector e (dubbed the eigenvector for λ ) such that A e = λ e .

Multiples of eigenvectors are also eigenvectors, as shown below:

A ( c e ) = c ( A e ) = c λ e = λ ( c e )

Definition 2 The eigenspace of A corresponding to λ is defined by ϵ λ = { e X : A e = λ e } .

For example, if a given eigenvalue λ has two eigenvectors, the eigenspace is given by [ { e 1 , e 2 } ] where A e 1 = λ e 1 and A e 2 = λ e 2 .

Definition 3 An operator A B ( X , X ) is said to be self-adjoint if A * = A , i.e., A x , y = x , A y for all x , y X .

If X = R N , a self-adjoint operator corresponds to a symmetric matrix.

Theorem 1 All eigenvalues of a self-adjoint operator are real.

Let λ be a complex eigenvalue of a self-adjoint operator A . Then A e = λ e for some e . For such an e , we have

λ e 2 = λ e , e = λ e , e = A e , e = e , A * e = e , A e = e , λ e = λ ¯ e , e .

Therefore we know that lambda is the same as its complex conjugate ( λ = λ ¯ ). The only way for this to be possible is if the imaginary part of λ is zero, and therefore λ R .

Theorem 2 If λ 1 , λ 2 are distinct eigenvalues of a self-adjoint operator A B ( X , X ) , then ϵ λ 1 ϵ λ 2 .

Assume we pick some arbitrary e 1 ϵ λ 1 and e 2 ϵ λ 2 . Then,

λ 1 e 1 , e 2 = λ 1 e 1 , e 2 = A e 1 , e 2 = e 1 , A e 2 = e 1 , λ 2 e 2 = λ 2 e 1 , e 2 .

Therefore we know that λ 1 e 1 , e 2 = λ 2 e 1 , e 2 . Since we chose two distinct eigenvalues, we know that λ 1 λ 2 . Therefore we must have e 1 , e 2 = 0 . This implies e 1 e 2 . Since e 1 and e 2 were chosen arbitrarily, this implies ϵ λ 1 ϵ λ 2 .

Theorem 3 (Schur's Lemma) Let X be an N-dimensional Hilbert space and A B ( X , X ) . Then there exists an orthonormal basis { φ 1 , φ 2 , ... , φ N } for X and a set of coefficients { A i j } i , j = 1 N such that A φ j = i = 1 j A i j φ i for j = 1 , 2 , . . . N .

An important note: since { φ j } makes an orthonormal basis for the domain X and range X , there exist coefficients a i such that A φ j = i = 1 N a i φ i for j = 1 , 2 , . . . . N .

Example 1 Recall the matrix representation of an operator: given A B ( X , X ) and an orthonormal basis { φ i } i = 1 N for X , we can write the matrix representation A ˜ of the operator A with entries

A ˜ i j = A φ j , φ i = k = 1 j A k j φ k , φ i = k = 1 j A k j φ k , φ i = A i j , if i j , 0 , if i > j .

We can represent A ˜ as an upper-triangular matrix (where x represents a non-zero entry in the matrix):

A ˜ = x x x x 0 x x x 0 0 x x 0 0 0 x

This shows that there exists an orthonormal basis φ for which the matrix representation of A is an upper-triangular matrix.

Theorem 4 If X is an N -dimensional space and A B ( X , X ) , then there exists an orthonormal basis { φ i } i = 1 N such that A φ i = λ i φ i for a set of scalars λ 1 , λ 2 , . . . λ N . The matrix representation A ˜ is a diagonal matrix whose entries are the eigenvalues; that is, A ˜ = diag ( { λ i } ) .

Note that, according to the theorem, we can fully represent the operator A by the aforementioned orthonormal basis { φ i } and the diagonal matrix A ˜ .

Pick the orthonormal basis { φ 1 , φ 2 , . . . φ N } specified by Schur's Lemma. For i < j , we have:

A i j = A φ j , φ i = k = 1 j A k j φ k , φ i = φ j , A φ i = φ j , k = 1 i A k i φ k = k = 1 i A k i ˜ φ j , φ k .

For k < j , the term in the sum is equal to zero. We then have:

A φ j = k = 1 j A k j φ k = A j j φ j .

Thus, the only non-zero entries of the representation matrix A ˜ are the diagonal entries. Furthermore, these entries are eigenvalues of A .

Recall that to compute A x using its matrix representation A ˜ , there are three steps:

  1. Represent x using the orthonormal basis { φ i } and collect the coefficients into a vector c .
  2. Perform the matrix-vector product d = A ˜ c .
  3. Then obtain A x = b = i = 1 N d i φ i .

A matrix-vector product with a dense matrix is computationally intensive. If we can diagonalize A , we can find the matrix-vector product A x by performing the operation A ˜ c , where A ˜ is a diagonal matrix, making the operation more efficient.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Signal theory. OpenStax CNX. Oct 18, 2013 Download for free at http://legacy.cnx.org/content/col11542/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Signal theory' conversation and receive update notifications?

Ask