Introduction
This appendix introduces some aspects of linear algebra and complex
algebra which will be helpful for the course. In addition, Dirac
notation is introduced and explained.
Vectors
Here we review some facts about real vectors before discussing the representation and complex analogues used in quantum mechanics.
Real Vectors
The simple definition of a vector - an object which has magnitude and
direction - is helpful to keep in mind even when dealing with complex
and/or abstract vectors as we will here. In three dimensional space,
a vector is often written as
where the hat () denotes a unit vector, and the components
, are just numbers. The unit vectors, denoted by the hats () are also
known as basis vectors.
This is because any vector
(in real three-dimensional space) can be written in terms of these unit/basis vectors. In
some sense they are basic components of any vector. Other basis
vectors could be used though. Some other common choices are those that
are used for spherical and cylindrical
coordinates. When dealing with more abstract and/or complex vectors,
it is often helpful to ask what one would do for an ordinary
three-dimensional vector. For example, properties of unit vectors,
dot products, etc. in three-dimensions are similar to the analogous
constructions in more dimensions.
The inner product, or dot product for two real three-dimensional vectors
can be computed as follows
For the inner product of with itself, we get the square of
the magnitude of denoted ,
If we want a unit vector in the direction of , then we divide
by its magnitude to get a unit vector,
Now, of course, as can easily be checked.
There are several ways to represent a vector. The ones we will use
most often are column and row vector notations. So, for example, we
could write the vector above as
In this case, our unit vectors are represented by the following
We next turn to the subject of complex vectors along with the relevant
notation.
We will see how to compute the inner product later, but some other
definitions will be required.
Complex Vectors
For complex vectors in quantum mechanics, Dirac notation is most often
used. This notation uses a ,
called a ket, for a vector, so our vector would be
For qubits, i.e. two-state quantum systems, will often be written as
complex vectors
|
|
|
(C.1)
|
where
are the basis vectors. The two numbers and
are complex numbers, so the vector is said to
be a complex vector.
Linear Algebra: Matrices
There are many aspects of linear algebra that are quite useful in
quantum mechanics. We will discuss (briefly) several of these here,
but first we will provide some definitions and properties which will
be useful as well as fixing notation. Some familiarity with matrices
will be assumed, but many basic defintions are also included.
Let us denote some matrix by . The set of all matrices with real entries is . Such matrices
are said to be real since they have all real entries. Similarly, the
set of complex matrices is . For the
set of set of square complex matrices, we simply write
.
We will also refer to the set of matrix elements, where the
first index ( in this case) labels the row and the second
labels the column. Thus the element is the element in the
second row and third column. A comma is inserted if there is some
ambiguity. For example, in a large matrix the element in the
2nd row and 12th
column is written as to distinguish between the
21st row and 2nd column.
Complex Conjugate
The complex conjugate of a matrix
is the matrix with each element replaced by its complex conjugate. In
other words, to take the complex conjugate of a matrix, one takes the
complex conjugate of each entry in the matrix. We denote the complex
conjugate with a ``star, e.g. . For example,
|
|
|
(C.2)
|
(Notice that the notation for a matrix is a capital letter, whereas
the entries are numbers, so they are represented by lower case
letters.)
Transpose
The transpose of a matrix is the same set of
elements but the first row becomes the first column, the second row
becomes the second column, etc. Thus the rows and columns are
interchanged. For example, for a square matrix, the
transpose is given by
|
|
|
(C.3)
|
Hermitian Conjugate
The complex conjugate and transpose is called the Hermitian conjugate, or simply the dagger of a matrix. It is called the dagger, because the symbol used to denote it is a dagger
(), viz.
|
|
|
(C.4)
|
For our example,
If a matrix is its own Hermitian conjugate, i.e. , the
we call it a Hermitian matrix.
(Clearly this is only possible for square matrices.) Hermitian
matrices are very important in quantum mechanics since their
eigenvalues are real. (See Sec.(Eigenvalues and Eigenvectors).)
Index Notation
Very often we write the product of two matrices and simply as
and let . However, it is also quite useful to write this
in component form. In this case, if these are matrices
This says that the element in the row and
column of the matrix is the sum . The transpose of has elements
Now if we were to transpose and as well, this would read
This gives us a way of seeing the general rule that
It follows that
The Trace
The trace of a matrix is the sum of the diagonal
elements and is denoted . So for example, the trace of an
matrix is
Some useful properties of the trace are the following:
Using the first of these results,
This relation is used so often that we state it here explicitly.
The Determinant
For a square matrix, the determinant is quite a useful thing. For
example, an matrix is invertible if and only if its
determinant is not zero. So let us define the determinant and give
some properties and examples.
The determinant of a matrix
|
|
|
(C.5)
|
is given by
|
|
|
(C.6)
|
Higher-order determinants can be written in terms of smaller ones in
the standard way.
The determinant of a matrix can be
also be written in terms of its components as
|
|
|
(C.7)
|
where the symbol
|
|
|
(C.8)
|
Let us consider the example of the matrix given
above. The determinant can be calculated by
where, explicitly,
|
|
|
(C.9)
|
so that
|
|
|
(C.10)
|
Now given the values of in Eq.~(\ref{eq:3depsilon}),
this is
The determinant has several properties which are useful to know. A
few are listed here.
- The determinant of the transpose of a matrix is the same as the determinant of the matrix itself:
- The determinant of a product is the product of determinants:
From this last property, another specific property can be derived.
Suppose we take the determinant of the product of a matrix and its
inverse we find
since the determinant of the identity is one. This implies that
The Inverse of a Matrix
The inverse of a square matrix is another matrix,
denoted such that
where is the identity matrix consisting of zeroes everywhere
except the diagonal which has ones. For example the
identity matrix is
It is important to note that a matrix is invertible if and only if its determinant is nonzero. Thus one only needsd to calculate the
determinant to see if a matrix has an inverse or not.
Unitary Matrices
A unitary matrix is one whose
inverse is also its Hermitian conjugate, so that
If the unitary matrix also has determinant one, it is said to be a special unitary matrix. The set of
unitary matrices is denoted
U and the set of special unitary matrices is denoted SU.
Unitary matrices are particularly important in quantum mechanics
because they describe the evolution, or change, of quantum states.
They are able to do this because unitary matrices have the property that rows and
columns, viewed as vectors, are orthonormal. (To see this, an example
is provided below.) This means that when
they act on a basis vector of the form (one 1, in say the th spot, and zeroes everywhere else)
|
|
|
(C.11)
|
the result is a normalized complex vector. Acting on a set of
orthonormal vectors of the form given in Eq.(C.11),
will produce another orthonormal set.
Let us consider the example of a unitary matrix,
|
|
|
(C.12)
|
The inverse of this matrix is the Hermitian conjugate, so the inverse
is given by
|
|
|
(C.13)
|
provided that the matrix satisfies the constraints
|
|
|
(C.14)
|
and
|
|
|
(C.15)
|
Looking at each row as a vector, the constraints in
Eq.(C.14) are the orthonormality conditions for the
vectors forming the rows. Similarly, the constraints in
Eq.(C.15) are the orthonormality conditions for the
vectors forming the columns when viewed as vectors.
More Dirac Notation
Now that we have a definition of Hermitian conjugate, we consider the
case for a matrix, i.e. a vector. In Dirac notation, we
had
So the Hermitian conjugate comes up so often that we use the following
notation for vectors,
This is a row vector. Let us consider a second complex vector
The inner product between and
is computed as follows:
|
|
|
(C16)
|
If these two vectors are orthogonal,
then their inner product is zero, .
The inner product of with itself is
If this vector is normalized then .
More generally, we will consider vectors in dimensions. In this
case we write the vector in terms of a set of basis vectors
, where . This is an ordered set of
vectors which are just labeled by integers. If the set is orthogonal,
then
and if they are normalized, then
If both of these are true, i.e., the entire set is orthonormal, we can
write,
where the symbol is called the Kronecker delta and is defined by
|
|
|
(C.17)
|
Now consider -dimensional vectors by letting two such vectors
be expressed in the same basis as
and
Then the inner product is
|
|
|
(C.18)
|
where we have used the fact that the delta function is zero unless
to get the last equality. For the inner product of a vector
with itself, we get
This immediately gives us a very important property of the inner
product. It tells us that in general,
(Just in case you don't know, the symbol means "if and only if" sometimes written as "iff.")
We could also expand a vector in a different basis. Let us suppose
that the set is an orthonormal basis which is different from the one considered earlier. We
could expand our vector in terms of our new basis by
expanding our new basis in terms of our old basis. Let us first
expand the in terms of the :
|
|
|
(C.19)
|
so that
|
|
|
(C.20)
|
where
|
|
|
(C.21)
|
Notice that the insertion of didn't do anything to our original vector. It is the same vector, just in a
different basis. Therefore, this is effectively the identity operator
This is an important and quite useful relation.
Now, to interpret Eq.(C.19), we can draw a close
analogy with three-dimensional real vectors. The inner product
can be interpreted as the projection of one vector onto
another. This provides the part of along .
Transformations
Suppose we have two different orthogonal bases, , .
The numbers for all the different and are
often referred to as matrix elements since the set forms a matrix with
labelling the rows, and labelling the columns. Therefore, we
can write the transformation from one basis to another with a matrix
transformation. Let be the matrix with elements . Then the transformation from one basis to another,
written in terms of the coefficients of , is
|
|
|
(C.22)
|
where
This sort of transformation is a change of basis. However, most
often when one vector is transformed to another, the transformation is
represented by a matrix. Such transformations can either be
represented by the matrix equation, like Eq.(C.22), or the components
|
|
|
(C.23)
|
For a general transformation matrix , acting on a vector,
the matrix elements in a particular basis are
just as elements of a vector can be found using
A similarity transformation
of an matrix by an invertible matrix is .
There are (at least) two important things to note about similarity
transformations,
- Similarity transformations leave determinants unchanged. (We say the determinant is invariant under similarity transformations.) This is because
- Simultaneous similarity transformations of matrices in an equation will leave the equation unchanged. Let , , and . If , then since . The two matrices and are said to be similar.
Eigenvalues and Eigenvectors
A matrix can always be diagonalized. By this, it is meant that for
every complex matrix there is a diagonal matrix such that
|
|
|
(C.24)
|
where and are unitary matrices. This form is called a singular value decomposition of the matrix and the entries of the diagonal matrix are called the singular values
of the matrix . However, the singular values are not always easy to find.
For the special case that the matrix is Hermitian ,
the matrix can be written as
|
|
|
(C.25)
|
where is unitary . In this case the elements
of the matrix are called eigenvalues.
Very often eigenvalues are introduced as solutions to the equation
where a vector called an eigenvector.
To find the eigenvalues and eigenvectors of a matrix , we follow a
standard procedure which is to calculate the following
|
|
|
(C.26)
|
and solve for . The different solutions for is the
set of eigenvalues and this set is called the spectrum. Let the different eigenvalues be denoted by , fo an vector. If two
eigenvalues are equal, we say the spectrum is
degenerate. To find the
eigenvectors, which correspond to different eigenvalues, the equation
must be solved for each value of . Notice that this equations
holds even if we multiply both sides by some complex number. This
implies that an eigenvector can always be scaled. Usually they are
normalized to obtain an orthonormal set. As we will see by example,
degenerate eigenvalues require some care.
Examples
Consider a Hermitian matrix
|
|
|
(C.27)
|
To find the eigenvalues
of this, we follow a standard procedure which
is to calculate the following
|
|
|
(C.28)
|
and solve for . The eigenvalues of this matrix are given by
which implies the eigenvalues are
and the eigenvectors are
These expressions are useful for calculating properties of qubit
states as will be seen in the text.
Now consider a matrix
First we calculate
This implies that the eigenvalues \index{eigenvalues} are
Let , , and .
To find eigenvectors, we calculate
|
|
|
(C.29)
|
for each .
For we get the following equations:
|
|
|
(C.30)
|
so , , and is any non-zero number, but we choose
it to normalize the vector. For ,
|
|
|
(C.31)
|
and for ,
|
|
|
(C.32)
|
so that .
Therefore, our three eigenvectors are
The matrix
is the matrix that diagonalizes in the following way,
where
or, we may write this as
This is sometimes called the eigenvalue decompostion of the matrix and is also written as,
|
|
|
(C.33)
|
(ONE MORE EXAMPLE WITH DEGENERATE E-VALUES)
Tensor Products
The tensor product
(also called the Kronecker product)
is used extensively in quantum mechanics and
throughout the course. It is commonly denoted with a
symbol, but this symbol is also often left out. In fact the following
are commonly found in the literature as notation for the tensor
product of two vectors and
|
|
|
(C.34)
|
Each of these has its advantages and we will use all of them in
different circumstances.
The tensor product is also often used for operators. So several
examples
will be given, one which explicitly calculates the tensor product for
two vectors and one which calculates it for two matrices which could
represent operators. However, these are not different in the sense
that a vector is a or an matrix. It is also
noteworthy that the two objects in the tensor product need not be of
the same type. In general a tensor product of an object
(array) with a object will produce an
object.
In general, the tensor product of two objects is computed as follows.
Let be an and be a array
|
|
|
(C.35)
|
and similarly for . Then
|
|
|
(C.36)
|
Let us now consider two examples. First let and
be as before,
Then
|
|
|
(C.37)
|
Also
|
|
|
(C.38)
|
Now consider two matrices
Then
|
|
|
(C.39)
|
Properties of Tensor Products
Some properties of tensor products which are useful are the following
(with , , , any type):
(See Ref.~(\cite{Horn/JohnsonII}), Chapter 4.)