Appendix C - Vectors and Linear Algebra

From Qunet
Revision as of 19:05, 23 February 2010 by Mbyrd (talk | contribs) (More Dirac Notation)
Jump to: navigation, search

Introduction

This appendix introduces some aspects of linear algebra and complex algebra which will be helpful for the course. In addition, Dirac notation is introduced and explained.

Vectors

Here we review some facts about real vectors before discussing the representation and complex analogues used in quantum mechanics.

Real Vectors

The simple definition of a vector - an object which has magnitude and direction - is helpful to keep in mind even when dealing with complex and/or abstract vectors as we will here. In three dimensional space, a vector is often written as

where the hat () denotes a unit vector. The components , are just numbers. These unit vectors are also called basis vectors because any vector (in real three-dimensional space) can be written in terms of them. In some sense they are basic components of any vector. Other basis vectors could be used though. Some other common choices are those the are used for spherical coordinates and those used for cylindrical coordinates. When dealing with more abstract and/or complex vectors, it is often helpful to ask what one would do for an ordinary three-dimensional vector. For example, properties of unit vectors, dot products, etc. in three-dimensions are similar to the analogous constructions in more dimensions.

The inner product, or dot product for two real three-dimensional vectors

can be computed as follows

For the inner product of with itself, we get the square of the magnitude of denoted ,

If we want a unit vector in the direction of , then we divide by its magnitude to get a unit vector,

Now, of course, as can easily be checked.

There are several ways to represent a vector. The ones we will use most often are column and row vector notations. So, for example, we could write the vector above as

In this case, our unit vectors are represented by the following

We next turn to the subject of complex vectors along with the relevant notation. We will see how to compute the inner product later, but some other definitions will be required.


Complex Vectors

For complex vectors in quantum mechanics, Dirac notation is most often used. This notation uses a , called a ket, for a vector, so our vector would be

For qubits, i.e. two-state quantum systems, will often be written as complex vectors


(C.1)

where

are the basis vectors. The two numbers and are complex numbers, so the vector is said to be a complex vector.


Linear Algebra: Matrices

There are many aspects of linear algebra that are quite useful in quantum mechanics. We will discuss (briefly) several of these here, but first we will provide some definitions and properties which will be useful as well as fixing notation. Some familiarity with matrices will be assumed, but many basic defintions are also included.


Let us denote some matrix by . The set of all matrices with real entries is . Such matrices are said to be real since they have all real entries. Similarly, the set of complex matrices is . For the set of set of square complex matrices, we simply write .


We will also refer to the set of matrix elements, where the first index ( in this case) labels the row and the second labels the column. Thus the element is the element in the second row and third column. A comma is inserted if there is some ambiguity. For example, in a large matrix the element in the 2nd row and 12th column is written as to distinguish between the 21st row and 2nd column.


Complex Conjugate

The complex conjugate of a matrix is the matrix with each element replaced by its complex conjugate. In other words, to take the complex conjugate of a matrix, one takes the complex conjugate of each entry in the matrix. We denote the complex conjugate with a ``star, e.g. . For example,


(C.2)

(Notice that the notation for a matrix is a capital letter, whereas the entries are numbers, so they are represented by lower case letters.)

Transpose

The transpose of a matrix is the same set of elements but the first row becomes the first column, the second row becomes the second column, etc. Thus the rows and columns are interchanged. For example, for a square matrix, the transpose is given by


(C.3)


Hermitian Conjugate

The complex conjugate and transpose is called the Hermitian conjugate, or simply the dagger of a matrix. It is called the dagger, because the symbol used to denote it is a dagger (), viz.


(C.4)

For our example,

If a matrix is its own Hermitian conjugate, i.e. , the we call it a Hermitian matrix. (Clearly this is only possible for square matrices.) Hermitian matrices are very important in quantum mechanics since their eigenvalues are real. (See Sec.(Eigenvalues and Eigenvectors).)

Index Notation

Very often we write the product of two matrices and simply as and let . However, it is also quite useful to write this in component form. In this case, if these are matrices

This says that the element in the row and column of the matrix is the sum . The transpose of has elements

Now if we were to transpose and as well, this would read

This gives us a way of seeing the general rule that

It follows that



The Trace

The trace of a matrix is the sum of the diagonal elements and is denoted . So for example, the trace of an matrix is

Some useful properties of the trace are the following:

Using the first of these results,

This relation is used so often that we state it here explicitly.


The Determinant

For a square matrix, the determinant is quite a useful thing. For example, an matrix is invertible if and only if its determinant is not zero. So let us define the determinant and give some properties and examples.


The determinant of a matrix


(C.5)

is given by


(C.6)

Higher-order determinants can be written in terms of smaller ones in the standard way.


The determinant of a matrix can be also be written in terms of its components as


(C.7)

where the symbol


(C.8)

Let us consider the example of the matrix given above. The determinant can be calculated by

where, explicitly,


(C.9)

so that


(C.10)

Now given the values of in Eq.~(\ref{eq:3depsilon}), this is

The determinant has several properties which are useful to know. A few are listed here.

  1. The determinant of the transpose of a matrix is the same as the determinant of the matrix itself:
  2. The determinant of a product is the product of determinants:

From this last property, another specific property can be derived. Suppose we take the determinant of the product of a matrix and its inverse we find

since the determinant of the identity is one. This implies that

The Inverse of a Matrix

The inverse of a square matrix is another matrix, denoted such that

where is the identity matrix consisting of zeroes everywhere except the diagonal which has ones. For example the identity matrix is

It is important to note that a matrix is invertible if and only if its determinant is nonzero. Thus one only needsd to calculate the determinant to see if a matrix has an inverse or not.


Unitary Matrices

A unitary matrix is one whose inverse is also its Hermitian conjugate, so that

If the unitary matrix also has determinant one, it is said to be a special unitary matrix. The set of unitary matrices is denoted U and the set of special unitary matrices is denoted SU.

Unitary matrices are particularly important in quantum mechanics because they describe the evolution, or change, of quantum states. They are able to do this because unitary matrices have the property that rows and columns, viewed as vectors, are orthonormal. (To see this, an example is provided below.) This means that when they act on a basis vector of the form (one 1, in say the th spot, and zeroes everywhere else)


(C.11)

the result is a normalized complex vector. Acting on a set of orthonormal vectors of the form given in Eq.(C.11), will produce another orthonormal set.

Let us consider the example of a unitary matrix,


(C.12)

The inverse of this matrix is the Hermitian conjugate, so the inverse is given by


(C.13)

provided that the matrix satisfies the constraints


(C.14)

and


(C.15)

Looking at each row as a vector, the constraints in Eq.(C.14) are the orthonormality conditions for the vectors forming the rows. Similarly, the constraints in Eq.(C.15) are the orthonormality conditions for the vectors forming the columns when viewed as vectors.

More Dirac Notation

Now that we have a definition of Hermitian conjugate, we consider the case for a matrix, i.e. a vector. In Dirac notation, we had

So the Hermitian conjugate comes up so often that we use the following notation for vectors,

This is a row vector. Let us consider a second complex vector

The inner product between and is computed as follows:


(C16)

If these two vectors are orthogonal, then their inner product is zero, . The inner product of with itself is

If this vector is normalized then .

More generally, we will consider vectors in dimensions. In this case we write the vector in terms of a set of basis vectors , where . This is an ordered set of vectors which are just labeled by integers. If the set is orthogonal, then

and if they are normalized, then

If both of these are true, i.e., the entire set is orthonormal, we can write,

where the symbol is called the Kronecker delta and is defined by


(C.17)

Now consider -dimensional vectors by letting two such vectors be expressed in the same basis as

and

Then the inner product is


(C.18)

where we have used the fact that the delta function is zero unless to get the last equality. For the inner product of a vector with itself, we get

This immediately gives us a very important property of the inner product. It tells us that in general,

(Just in case you don't know, the symbol means "if and only iff" sometimes written as "iff.")

We could also expand a vector in a different basis. Let us suppose that the set is an orthonormal basis () which is different from the one considered earlier. We could expand our vector in terms of our new basis by expanding our new basis in terms of our old basis. Let us first expand the in terms of the :


(C.19)

so that


(C.20)

where


(C.21)

Notice that the insertion of didn't do anything to our original vector. It is the same vector, just in a different basis. Therefore, this is effectively the identity operator

This is an important and quite useful relation. Now, to interpret Eq.(C.19), we can draw a close analogy with three-dimensional real vectors. The inner product can be interpreted as the projection of one vector onto another. This provides the part of along .

Transformations

Suppose we have two different orthogonal bases, , . The numbers for all the different and are often referred to as matrix elements since the set forms a matrix with labelling the rows, and labelling the columns. Therefore, we can write the transformation from one basis to another with a matrix transformation. Let be the matrix with elements . Then the transformation from one basis to another, written in terms of the coefficients of , is


(C.22)

where

This sort of transformation is a change of basis. However, most often when one vector is transformed to another, the transformation is represented by a matrix. Such transformations can either be represented by the matrix equation, like Eq.~(\ref{eq:matrixeq}), or the components


(C.23)

For a general transformation matrix , acting on a vector, the matrix elements in a particular basis are

just as elements of a vector can be found using

A similarity transformation of an matrix by an invertible matrix is . There are (at least) two important things to note about similarity transformations,

  1. Similarity transformations leave determinants unchanged. (We say the determinant is invariant under similarity transformations.) This is because
  2. Simultaneous similarity transformations of matrices in an equation will leave the equation unchanged. Let , , and . If , then since . The two matrices and are said to be similar.

Eigenvalues and Eigenvectors

A matrix can always be diagonalized. By this, it is meant that for every complex matrix there is a diagonal matrix such that


(C.24)

where and are unitary matrices. The entries of the diagonal matrix are called the singular values of the matrix . However, the singular values are not always easy to find.

For the special case that the matrix is Hermitian (), the matrix can be written as


(C.25)

where is unitary (). In this case the elements of the matrix are called eigenvalues. Very often eigenvalues are introduced as solutions to the equation

where a vector called an eigenvector.

To find the eigenvalues and eigenvectors of a matrix , we follow a standard procedure which is to calculate the following


(C.26)

and solve for . The different solutions for is the set of eigenvalues and this set is called the spectrum. Let the different eigenvalues be denoted by , fo an vector. If two eigenvalues are equal, we say the spectrum is degenerate. To find the eigenvectors, which correspond to different eigenvalues, the equation

must be solved for each value of . Notice that this equations holds even if we multiply both sides by some complex number. This implies that an eigenvector can always be scaled. Usually they are normalized to obtain an orthonormal set. As we will see by example, degenerate eigenvalues require some care.

Examples

Consider a Hermitian matrix


(C.27)

To find the eigenvalues of this, we follow a standard procedure which is to calculate the following


(C.28)

and solve for . The eigenvalues of this matrix are given by

which implies the eigenvalues are

and the eigenvectors are

These expressions are useful for calculating properties of qubit states as will be seen in the text.

Now consider a matrix

First we calculate

This implies that the eigenvalues \index{eigenvalues} are

Let , , and . To find eigenvectors, we calculate


(2.29)

for each . For we get the following equations:


(2.30)

so , , and is any non-zero number, but we choose it to normalize the vector. For ,


(2.31)

and for ,


(2.32)

so that . Therefore, our three eigenvectors are

The matrix

is the matrix that diagonalizes in the following way,

where

or, we may write this as

This is sometimes called the eigenvalue decompostion of the matrix and is also written as,


(C.33)

(ONE MORE EXAMPLE WITH DEGENERATE E-VALUES)

Tensor Products

The tensor product (also called the Kronecker product) is used extensively in quantum mechanics and throughout the course. It is commonly denoted with a symbol, but this symbol is also often left out. In fact the following are commonly found in the literature as notation for the tensor product of two vectors and


(C.34)

Each of these has its advantages and we will use all of them in different circumstances.

The tensor product is also often used for operators. So several examples will be given, one which explicitly calculates the tensor product for two vectors and one which calculates it for two matrices which could represent operators. However, these are not different in the sense that a vector is a or an matrix. It is also noteworthy that the two objects in the tensor product need not be of the same type. In general a tensor product of an object (array) with a object will produce an object.


In general, the tensor product of two objects is computed as follows. Let be an and be a array


(C.35)

and similarly for . Then


(C.36)

Let us now consider two examples. First let and be as before,

Then


(2.37)

Also


(2.38)

Now consider two matrices

Then


(2.39)

Properties of Tensor Products

Some properties of tensor products which are useful are the following (with , , , any type):

(See Ref.~(\cite{Horn/JohnsonII}), Chapter 4.)