Lecture 2: Linear Operators: IIT Kanpur
Lecture 2: Linear Operators: IIT Kanpur
Lecture 2: Linear Operators: IIT Kanpur
Rajat Mittal
IIT Kanpur
The mathematical formulation of Quantum computing requires vector spaces and linear operators. So, we
need to be comfortable with linear algebra to study quantum computing. These notes will assume that the
reader is familiar with the concept of vector space, basis, linear independence and tensor product. Strang’s
book, Linear Algebra and its applications, is a good source to brush up these concepts.
This exposition will focus on vectors, matrices and properties of matrices. Dirac’s notation is used widely
in quantum computing, because it simplifies the understanding of quantum mechanical concepts. We will
switch between the standard vector notation and Dirac notation in these notes.
Exercise 1. Read about vector space, basis, linear independence and tensor product if you are not comfortable
with these words.
One of the most fundamental concept in linear algebra is that of a vector space. We will mostly concern
ourselves with the vector space Cn , the vector space of dimension n over the field of complex numbers. This
means that the scalars used in these vector spaces are complex numbers.
A vector is the most basic unit of a vector space. Using Dirac’s notation, a column vector will be denoted
by |ψi. Suppose |v1 i, |v2 i, · · · , |vn i is the basis of the vector space, then any vector |vi can be written as,
For a vector space with dimension n, the standard basis is denoted by |0i, |1i, · · · , |n − 1i. Here you can
think of |ii as the vector with 1 at the (i − 1)th position and 0 otherwise. For example, a qubit lives in a
2-dimensional space with basis |0i, |1i.
Note 1. There is a difference between vector |0i and vector 0, the vector with all entries 0. First one is a
basis vector with the first entry 1 and rest 0.
The notation hψ| denotes the row vector whose entries are complex conjugate of the entries of the vector
|ψi. The space Cn is equipped with the natural inner product,
n
X
((x1 , x2 , · · · , xn ), (y1 , y2 , · · · , yn )) = x∗i yi .
i=1
The inner product (dot product) between two vectors |ψi, |φi is denoted by hψ|φi. The vector for the
tensor product space, |ψi⊗|φi will be simply written as |ψi|φi. In Dirac’s notation, the expression A = |ψihφ|
is a matrix which takes |vi to hφ|vi|ψi. The analog of this expression in the simple vector notation would
be, A = ψ(φ)T .
1 Operators
Given two vector spaces, V and W over C, a linear operator M : V → W is defined as an operator satisfying
the following properties.
– M (x + y) = M (x) + M (y).
– M (αx) = αM (x), ∀α ∈ C.
These conditions imply that the zero of the vector space V is mapped to the zero of the vector space W .
Also,
M (α1 x1 + · · · + αk xk ) = α1 M (x1 ) + · · · + αk M (xk )
Where x1 , · · · , xk are elements of V and αi ’s are in C. Because of this linearity, it is enough to specify the
value of a linear operator on any basis of the vector space V . In other words, a linear operator is uniquely
defined by the values it takes on any particular basis of V .
Let us define the addition of two linear operators as (M + N )(u) = M (u) + N (u). Similarly, αM (scalar
multiplication) is defined to be the operator (αM )(u) = αM (u). The space of all linear operators from V to
W (denoted L(V, W )) is a vector space in itself. The space of linear operators from V to V will be denoted
by L(V ).
Exercise 2. Given the dimension of V and W , what is the dimension of the vector spaces L(V, W )?
Exercise 3. What is the lower-bound/ upper-bound on the dimension of the vector space spanned by
{w1 , w2 , · · · , wn }?
The easiest way to see the above fact is: notice that the matrix Mf and the operator f act exactly the
same on the basis elements of V . Since both the operations are linear, they are exactly the same operation.
This proves that any linear operation can be specified by a matrix.
The previous discussion does not depend upon the chosen basis. We can pick our favorite basis, and the
linear operator can similarly be written in the new basis as a matrix (The columns of this matrix are images
of the basis elements). In other words, given bases of V and W and a linear operator f , it has a unique
matrix representation.
To compute the action of a linear operator, express v ∈ V in the preferred basis and multiply it with
the matrix representation. The output will be in the chosen basis of W . We will use the two terms, linear
operator and matrix, interchangeably in future (the bases will be clear from the context).
For a matrix A, AT denotes the transpose of the matrix and A∗ denotes the adjoint of the matrix (take
complex conjugate and then transpose).
Let us look at some simple matrices which will be used later.
– Zero matrix: The matrix with all the entries 0. It acts trivially on every element and takes them to the
0 vector.
– Identity matrix: The matrix with 1’s on the diagonal and 0 otherwise. It takes v ∈ V to v itself.
– All 1’s matrix (J): All the entries of this matrix are 1.
2
1.2 Kernel, image and rank
For a linear operator/matrix (from V to W ), the kernel is defined to be the set of vectors which map to 0.
ker(M ) = {x ∈ V : M x = 0}
The image is the set of vectors which can be obtained through the action of the matrix on some element
of the vector space V .
img(M ) = {x ∈ W : ∃y ∈ V, x = M y}
Notice that ker(M ) is a subset of V , but img(M ) is a subset of W . The dimension of img(M ) is known
as the rank of M (rank(M )). The dimension of ker(M ) is known as the nullity of M (nullity(M )). For a
matrix M ∈ L(V, W ), by the famous rank-nullity theorem,
Proof. Suppose u1 , · · · , uk is the basis for ker(M ). We can extend it to the basis of V , u1 , · · · , uk , vk+1 , · · · , vn .
We need to prove that the dimension of img(M ) is n − k. It can be proved by showing that the set
{M vk+1 , · · · , M vn } forms a basis of img(M ).
Exercise 9. Prove that any vector in the image of M can be expressed as linear combination of M vk+1 , · · · , M vn .
Also any linear combination of M vk+1 , · · · , M vn can’t be zero vector.
Exercise 10. Prove the rank of a matrix is equal to the dimension of the vector space spanned by its columns
(column-space).
The dimension of the column space is sometimes referred as the column-rank. We can similarly define
the row-rank, the dimension of the space spanned by the rows of the matrix. Luckily, row-rank turns out to
be equal to column-rank and we will call both of them as the rank of the matrix. This can be proved easily
using Gaussian elimination. We will give a visual proof of the theorem.
Proof. Given an m × n matrix M , say {c1 , c2 , · · · , ck } span the column space of M . Suppose, C be the m × k
matrix with columns {c1 , c2 , · · · , ck }. Then, there exist an k × n matrix R, s.t., CR = M . If {d1 , d2 , · · · , dk }
are the columns of R, then the equation CR = M can be viewed as,
.. .. .. .. .. .. .. .. .. .. .. .. .. .. ..
. . . . . . . . . . . . . . .
c1 c2 · · ck d1 d2 · · dk = Cd1 Cd2 · · Cdk
.. .. .. .. .. .. .. .. .. .. .. .. .. .. ..
. . .. . . . .. . . . .. .
Another way to view the same equation is,
3
P
··· r1 ··· · · · Pi C1i ri · · ·
··· ···
r2 ··· i C2i ri · · ·
···
C · ··· = ··· · ···
··· · ··· ··· P · ···
··· rk ··· ··· i Cki ri · · ·
This shows that the k columns of R span the row-space of M . Hence, column-rank is smaller than the
row-rank.
Exercise 11. Show that row-rank is less than column-rank by a similar argument.
Note 2. The column-rank is equal to row-rank. It does not mean that the row-space is same as the column-
space.
Using these characterizations of rank, it can be proved easily that rank(A) = rank(A∗ ) and rank(A +
B) ≤ rank(A) + rank(B).
– Trace: The trace of a matrix is the sum of all the diagonal elements.
X
tr(A) = A[i, i]
i
– Entry-wise multiplication: The entry-wise multiplication of two matrices is known as Hadamard product
and only makes sense when both of them have same number of rows and columns. The Hadamard product
of two matrices A, B is
(A ◦ B)[i, j] = A[i, j]B[i, j].
The related operation is when you add up the entries of this Hadamard product.
X
(A • B) = A[i, j]B[i, j]
i,j
Exercise 13. Given a matrix, express • operation in terms of multiplication and trace operation.
– Inverse: Inverse of a matrix M is the matrix M −1 , s.t., M M −1 = M −1 M = I. The inverse only exists if
the matrix has full rank ( the columns of M span the whole space).
Exercise 16. Show that the inverse of a matrix exists iff it has full rank.
4
2 Eigenvalues and eigenvectors
A matrix M ∈ L(V, W ) is square if dim(V ) = dim(W ). In particular, a matrix M ∈ L(V ) is always square.
Consider a matrix M ∈ L(V ), any vector v ∈ V satisfying,
M v = λv for some λ ∈ C,
is called the eigenvector of matrix M with eigenvalue λ.
Exercise 17. Given two eigenvectors v, w, when is their linear combination an eigenvector itself?
The previous exercise can be used to show that all the eigenvectors corresponding to a particular eigen-
value form a subspace. This subspace is called the eigenspace of the corresponding eigenvalue.
An eigenvalue λ of an n × n matrix M satisfies the equation
Det(λI − M ) = 0,
where Det(M ) denotes the determinant of the matrix M . The polynomial Det(λI − M ) = 0, in λ, is called
the characteristic polynomial of M . The characteristic polynomial has degree n and will have n roots in the
field of complex numbers. Though, these roots might not be real.
Exercise 18. Give an example of a matrix with no real eigenvalue.
Theorem 1. Given a matrix P of full rank, matrix M and matrix P −1 M P have the same set of eigenvalues.
Proof. Suppose λ is an eigenvalue of P −1 M P , we need to show that it is an eigenvalue for M too. Say λ is
an eigenvalue with eigenvector v. Then,
P −1 M P v = λv ⇒ M (P v) = λP v.
Hence P v is an eigenvector with eigenvalue λ.
The opposite direction follows similarly. Given an eigenvector v of M , it can be shown that P −1 v is an
eigenvector of P −1 M P .
P −1 M P (P −1 v) = P −1 M v = λP −1 v
Hence proved.
Exercise 19. Where did we use the fact that P is a full rank matrix?
3 Spectral decomposition
Exercise 20. Let v1 , v2 be two eigenvectors of a matrix M with distinct eigenvalues. Show that these two
eigenvectors are linearly independent.
Given an n × n matrix M , it need not have n linearly independent eigenvectors. The matrix M is called
diagonalizable iff the set of eigenvectors of M span the complete space Cn .
For a diagonalizable matrix, the basis of eigenvectors need not be an orthogonal basis. We will be
interested in matrices which have an orthonormal basis of eigenvectors.
A normal matrix is defined to be a matrix M , s.t., M M ∗ = M ∗ M . Spectral theorem shows that we can
form an orthonormal basis of Cn using the eigenvectors of a normal matrix.
Theorem 2. Spectral theorem: For a normal matrix M ∈ L(V ), there exists an orthonormal basis |x1 i, · · · , |xk i
of V , s.t.,
n
X
M= λi |xi ihxi |.
i=1
Here ∀i : λi ∈ C.
5
Note 3. It means that any normal matrix M = U ∗ DU for some unitary U and diagonal matrix D.
Exercise 21. Show that |xi i is an eigenvector of M with eigenvalue λi .
Note 4. hy|xi is a scalar, but |yihx| is a matrix.
Note 5. The λi ’s need not be different. If we collect all the |xi i’s corresponding to a particular eigenvalue λ,
the space spanned by those |xi i’s is the eigenspace of λ.
Proof idea. The proof of spectral theorem essentially hinges on the following lemma.
Lemma 1. Given an eigenspace S (of eigenvalue λ) of matrix M , the matrix M acts on the space S and
S ⊥ separately. In other words, M |vi ∈ S if |vi ∈ S and M |vi ∈ S ⊥ if |vi ∈ S ⊥ .
Proof of lemma. Since S is an eigenspace, M |vi ∈ S if |vi ∈ S. For a vector |vi ∈ S,
M M ∗ |vi = M ∗ M |vi = λM ∗ |vi.
This shows that M ∗ preserves the subspace S. Suppose |v1 i ∈ S ⊥ , |v2 i ∈ S, then M ∗ |v2 i ∈ S. So,
0 = hv1 |M ∗ |v2 i = hM v1 |v2 i.
Hence M |v1 i ∈ S ⊥ . Hence, matrix M acts separately on S and S ⊥ .
The lemma implies that M is a linear operator on S ⊥ , i.e., it moves every element of S ⊥ to an element
in S ⊥ linearly. It can be easily shown that this linear operator (the action of M on S ⊥ ) is also normal. The
proof of spectral theorem follows by using induction and is given below.
From the fundamental theorem of Algebra, there is at least one root λ0 of det(λI − M ) = 0. Start with
the eigenspace of the eigenvalue λ0 . Using Lem. 1, we can restrict the matrix to orthogonal subspace (which
is of smaller dimension). We can divide the entire space into orthogonal eigenspaces by induction.
Exercise 22. Show that if we take the orthonormal basis of all these eigenspaces, then we get the required
decomposition.
Exercise 23. Given the spectral decomposition of M , what is the spectral decomposition of M ∗ ?
Exercise 24. If M is normal, prove that the rank of M is the sum of the dimension of the non-zero eigenspaces.
It is easy to show that any matrix with orthonormal set of eigenvectors is a normal matrix. Hence, spectral
decomposition provides another characterization of normal matrices.
Clearly the spectral decomposition is not unique (essentially because of the multiplicity of eigenvalues).
But the eigenspaces corresponding to each eigenvalue are fixed. So there is a unique decomposition in terms
of eigenspaces and then any orthonormal basis of these eigenspaces can be chosen.
Note 6. It is also true that if an eigenvalue is a root of characteristic polynomial with multiplicity k, then
its eigenspace is of dimension k.
The eigenvalues and eigenvectors have more structure if we look at specific classes of normal matrices.
We will take a look at these special classes of normal matrices below.
6
3.2 Unitary matrix
Unitary matrices can be viewed as matrices which implement a change of basis. Hence they preserve the
angle (inner product) between the vectors. So for unitary M ,
If two matrices A, B are related by A = M −1 BM , where M is unitary, then they are unitarily equivalent.
If two matrices are unitarily equivalent then they are similar.
Spectral theorem can be stated as the fact that normal matrices are unitarily equivalent to a diagonal
matrix. The diagonal of a diagonal matrix contains its eigenvalues.
A matrix M is positive semidefinite if it is Hermitian and all its eigenvalues are non-negative. If all eigenvalues
are strictly positive then it is called a positive definite matrix.
Proof. 1 ⇒ 2: Say λ is an eigenvalue of M . Then there exist an eigenvector |vi ∈ V , s.t., M |vi = λ|vi. So
0 ≤ hv|M |vi = λhv|vi. Since hv|vi is positive for all |vi, implies λ is non-negative.
2 ⇒ 3: Since the matrix M is Hermitian, it has a spectral decomposition.
X
M= λi |xi ihxi |
i
√
Define |yi i = λi |xi i. This definition is possible because λi ’s are non-negative. Then,
X
M= |yi ihyi |.
i
Define B ∗ to be the matrix whose columns are yi . Then it is clear that B ∗ B = M . From this construction,
B’s columns are orthogonal.
Note 7. In general, any matrix of the form B ∗ B is positive semi-definite. The matrix B need not have
orthogonal columns (it can even be rectangular).
But this representation is not unique and there always exists a matrix B with orthogonal columns for
M , s.t., B ∗ B = M . This decomposition is unique if B is positive semidefinite. The positive semidefinite B,
s.t., B ∗ B = M , is called the square root of M .
7
Hint: Use the spectral decomposition to find one of the square root. Suppose A is any square root of M .
Then use the spectral decomposition of A and show the square root is unique (remember the decomposition
to eigenspaces is unique) .
3 ⇒ 1: We are given a matrix B, s.t., B ∗ B = M . Then,
P
Note 8. A matrix M of the form M = i |xi ihxi | is positive semidefinite (Exercise: Prove it), even if xi ’s
are not orthogonal to each other.
Note 9. A matrix of the form |yihx| is a rank one matrix. It is rank one because all columns are scalar
multiples of |yi. Similarly, all rank one matrices can be expressed in this form.
Singular value decomposition is one of the most important factorizations of a matrix. The statement says,
Theorem 4. Given a linear operator M in L(V, W ). There exists a decomposition of the form:
r
X
M= si |yi ihxi |
i=1
Where |x1 i, · · · , |xr i (called right singular vectors) and |y1 i, · · · , |yr i (called left singular vectors) are or-
thonormal basis of V and W respectively. The numbers s1 , · · · , sr (called singular values) are positive real
numbers and r itself is the rank of the matrix M .
The statement of the theorem can also be written as M = A∆B ∗ , where A ∈ L(W ), B ∈ L(V ) are unitary
matrices and ∆ is the diagonal matrix of singular values. With this interpretation, any linear operation can
be viewed as rotation in subspace V then scaling the standard basis and then another rotation in W subspace.
The statement of singular value decomposition is easy to prove if we don’t need any condition on |yi i’s.
Any basis of V will be sufficient to construct such a decomposition (why?). We can even choose all singular
values to be 1 in that case. But it turns out that with the singular values we can make the yi ’s to be
orthonormal.
The proof of singular value decomposition follows by applying spectral decomposition on matrices M M ∗
and M ∗ M .
Exercise 31. Prove that M ∗ M and M M ∗ have the same set of eigenvalues.
M xi
Suppose the eigenvectors of M ∗ M are |xi i’s for eigenvalues λi , then eigenvectors of M M ∗ are kM xi k =
yi ’s.
Exercise 32. Prove the above statement.
From the assignment rank(A) = rank(A∗ A) and hence it is enough to specify the action of M on xi ’s.
So,
r
X
M= kM xi k|yi ihxi |
i=1
√
Exercise 33. Prove that kM xi k = λi .
8
This implies the singular value decomposition.
r p
X r
X
M= λi |yi ihxi | = si |yi ihxi |.
i=1 i=1
The eigenvectors of M M ∗ are left singular vectors and eigenvectors of M ∗ M are right singular vectors
of M . The eigenvalues of M M ∗ or M ∗ M are the squares of the singular values of M .
We will restrict our attention to normal matrices for the rest of this lecture note.
6 Operator functions
The first notion is of applying a function on a linear operator. We will assume that the linear operators given
to us belong to the set of normal operators or some subset of it.
Suppose we have a function, f : C → C, from complex numbers to complex numbers. It can naturally
extended to be a function on a normal linear operator in L(Cn ). By definition of operator function, we apply
the function on all the eigenvalues of the operator. So, if
A = λ1 x1 x∗1 + · · · + λn xn x∗n .
then
f (A) = f (λ1 )x1 x∗1 + · · · + f (λn )xn x∗n .
In particular, we can now define the square-root, exponential and logarithm of an operator.
Exercise 35. Check that this definition of square-root agrees with the definition of square root defined in the
previous lecture notes.
Note 10. This means that we define square root only for positive semi-definite operators.
Pauli matrices are used widely in quantum computing. They are defined as,
01 0 −i 1 0
X= Y = Z= .
10 i 0 0 −1
X is the quantum NOT gate and Z is known as the phase gate.
9
Exercise 36. Show that the Pauli matrices are Hermitian as well as Unitary by calculating their eigenvalue.
Exercise 37. Show that the Pauli matrices (with identity) form a basis of all Hermitian 2 × 2 operators.
P Another very important function on operators introduced before is trace. We defined trace to be T r(A) =
i Aii . At this point, it is a function on matrices and not linear operators.
For a linear operator, trace might be different for different bases. In other words, there is no guarantee
that it is independent of the basis (from the definition given above).
Exercise 40. Show that the trace is cyclic, i.e., tr(AB) = tr(BA).
∗
This exercise implies that tr(UP AU ) = tr(A). Hence, trace is independent of the representation.
We also know that hv|A|wi = ij Aij vi∗ wj .
Exercise 41. Show that Aij = hi|A|ji, where matrix A is represented in the standard basis |1i, · · · , |ni.
P
From the previous exercise, tr(A) = i hi|A|ii. In fact, for any orthonormal basis v1 , · · · , vn ,
X
tr(A) = hvi |A|vi i.
i
7 Tensor product
We have talked about the concept of tensor product before, it was used to combine the two systems.
Suppose there is a ball which can be colored blue or red. The state of a “quantum” ball is a vector in
two dimensions,
|vi = α|ri + β|bi.
Where |ri, |bi represent the classical states, the ball being red or blue, the coefficients α, β follow the
usual law.
How about if there are two different balls. The classical states possible are |rri, |rbi, |bri, |bbi, i.e., we take
the set multiplication of possible states of individual system.
What are the possible states if this system is quantum?
This idea motivates the definition of tensor product. Given two vector spaces V, W equipped with an
inner product and spanned by the orthonormal basis v1 , v2 , · · · , vn and w1 , w2 , · · · , wm , the tensor product
V ⊗ W is the space spanned by the mn vectors (v1 ⊗ w1 ), · · · , (v1 ⊗ wn ), (v2 ⊗ w1 ), · · · , (vn ⊗ wm ).
10
– Linearity in the first component: for v1 , v2 ∈ V and w ∈ W ,
(v1 + v2 ) ⊗ w = v1 ⊗ w + v2 ⊗ w.
v ⊗ (w1 + w2 ) = v ⊗ w1 + v ⊗ w2 .
We can define the tensor product of two vectors in a canonical way for the vector spaces Cn and Cm .
The tensor product of two vectors a = (a1 , · · · , an ) ∈ V and b = (b1 , · · · , bm ) is the vector a ⊗ b ∈ V ⊗ W ,
a1 b1
a1 b2
a⊗b= .
..
an bm
We can define the inner product on the tensor product space in the natural way,
ha ⊗ b|c ⊗ di = ha|cihb|di.
Given two linear operators A ∈ L(V ) and B ∈ L(W ), their tensor product is in space L(V ⊗ W ). The
action is specified by,
(A ⊗ B)(a ⊗ b) = Aa ⊗ Bb.
Extend this by linearity to define the action on the complete space V ⊗ W .
Exercise 44. Given the matrix representation of A, B, come up with matrix representation of A ⊗ B.
Exercise 45. Write out the matrix representation of H ⊗2 = H ⊗ H where H is the Hadamard Matrix.
Partial trace is a linear operator on the tensor product of the operators. Given two vector spaces, V and
W , let A ∈ L(V ) and B ∈ L(W ) be two linear operators. The partial trace T rW is a linear operator from
L(V ⊗ W ) to L(V ), s.t.,
T rW (A ⊗ B) = AT r(B).
We can similarly define T rV .
Exercise 46. Show that the partial trace defined above is a linear operator.
Partial trace is mostly used in quantum computing to understand the state/operation on a part of a
composite system.
11
8 Direct sum
A closely related concept is called direct sum. Notice the difference in the motivation and definition.
Suppose you have a ball which can be colored red or blue. Then the “quantum state” of that ball could
be any superposition of these classical states,
Where |ri, |bi represent the classical states, the ball being red or blue, the coefficients α, β follow the usual
law. One day you discover that ball can also have a yellow color. Then the basis states become |ri, |bi, |yi
and the state of the ball can be,
|vi = α|ri + β|bi + γ|yi.
This idea gives rise to the concept of direct sum of two vector spaces. Given two vector spaces V, W
spanned by basis v1 , v2 , · · · , vn and w1 , w2 , · · · , wm , the direct sum V ⊕W is the space spanned by the vectors
(v1 , 0), · · · , (vn , 0), (0, w1 ), · · · , (0, wm ). Here the zeroes in the first n entries is the vector 0 of dimension m.
In the last m entries it is the 0 vector with dimension n. Similar to tensor product, direct sum of two vector
spaces should satisfy the following two conditions,
– Scalar multiplication:
α(v, w) = (αv, αw).
– Linearity:
(v1 , w1 ) + (v2 , w2 ) = (v1 + v2 , w1 + w2 ).
Exercise 47. What is the dimension of this vector space dim(V ) ⊕ dim(W ).
Exercise 48. Why did we define direct sum of matrices this way?
The description of both direct sum as well as tensor product is given in a very simplified manner in terms
of basis vectors, sufficient for use in our course. Readers are encouraged to check out the formal definitions.
9 Assignment
Exercise 49. Show that the matrix M and M ∗ have the same singular values.
Exercise 50. Prove that the eigenvalues of a Hermitian matrix are real.
Exercise 51. Prove that the absolute value of the eigenvalues of an unitary matrix is 1. Is the converse true.
What condition do we need to get the converse?
Exercise 52. Show that a matrix M is positive semi-definite if it is the Gram matrix of vectors |u1 i, · · · , |un i.
That is,
Mij = hui |uj i.
Exercise 53. Read about polar decomposition and how to get singular decomposition using polar decompo-
sition.
Exercise 54. Prove that a matrix M is Hermitian iff hv|M |vi is real for all |vi.
Exercise 55. Show that if two matrices are simultaneously diagonalizable then they commute.
12
Exercise 56. Show that the set of Hermitian matrices of a fixed dimension form a vector space (over which
field?). What is the dimension of this vector space?
Exercise 57. Read about polar decomposition and prove it using singular value decomposition.
Exercise 59. Prove that rank(A) = rank(A∗ A) without using singular or spectral decomposition.
Hint: rank(A) ≥ rank(A∗ A) is easy. For the other direction, reduce A to its reduced row echelon form.
Exercise 60. What are the eigenvalues of A ⊗ B, where A, B are normal matrices?
Exercise 61. What are the eigenvalues of A ⊕ B, where A, B are normal matrices?
Exercise 62. Give a characterization of the linear operators over V ⊗ W in terms of linear operators over V
and W . Remember that they form a vector space.
2 2 2
Exercise 64. Let σ = α1 X + α2 Y + α3 Z, where |α1 | + |α2 | + |α3 | = 1. Show that,
References
1. M. A. Nielsen and I. L. Chuang. Quantum computation and quantum information. Cambridge, 2010.
2. S. Arora and B. Barak. Computational Complexity: A modern approach. Cambridge, 2009.
3. R. Lidl and H. Niederreiter. Finite Fields. Cambridge University Press, 1997.
4. B. Kleinberg. Course notes: Introduction to algorithms. http://www.cs.cornell.edu/courses/cs4820/2010sp/handouts/MillerRabin.pd
2010.
5. D. R. Simon. On the power of quantum computation. Foundations of Computer Science, 1994 Proceedings., 35th
Annual Symposium on: 116123, 1994.
6. A. Childs. Course notes: Quantum algorithms. https://cs.umd.edu/ amchilds/teaching/w13/qic823.html, 2013.
13