Using an "educated guess" one observes that
. With this, it is easy to see that
,
,
and
.
, as all the coefficients divisible by 7 reduce to 0.
Let us denote the matrices (appearing in the same order as in the book) by
. We need to check the following:
is a group with matrix addition and
as the identity. We get the following addition table
|
0
|
1
|
A
|
B
|
0
|
0
|
1
|
A
|
B
|
1
|
1
|
0
|
B
|
A
|
A
|
A
|
B
|
0
|
1
|
B
|
B
|
A
|
1
|
0
|
So we see that the elements with addition form an abelian group with
as the identity.
is a group with
as the identity. Again we have the multiplication table
|
1
|
A
|
B
|
1
|
1
|
A
|
B
|
A
|
A
|
B
|
1
|
B
|
B
|
1
|
A
|
Again, we see that the elements with matrix multiplication form an abelian group with
as the identity.
- The distributive law follows from the distributive law for matrices in general.
Writing out a product and sum of two elements from the given set, and noticing that the coefficients of both elements, and thus their sums and products are in
, implies that the sum and product are in the set. To see that each non-zero element has an inverse in the
-operation is trivial. To see the same for the product operation, write the equations coming from the condition
, where
is a known element from the set and
a candidate for its inverse with unknown coefficients as a linear system. By Corollary 3.2.8 this system has a solution. The distribution law is immediate.
a) The space of symmetric matrices is a vector space, since the sum of two symmetric matrices is a symmetric matrix, and a scaling of a symmetric matrix by a scalar is symmetric.
b) The space of invertible matrices is not a vector space, since it does not contain the zero matrix.
c) The space of upper triangular matrices is also a vector space by similar reasoning as used in part a).
One possible basis for the space of symmetric matrices is for example the matrices
for
that have zeros everywhere but in the
and
entry. There are
such matrices, and they are linearly independent, since no two such matrices have ones in the same entry. Furthermore, the matrices
are symmetric and clearly any symmetric matrix can be written as a linear combination of
.
Let
be coefficients such that
. (1)
The matrix
has as the
th column the vector
where
is the
th element of the vector
. Denote
, so that the (1) implies together with the condition that the vectors
form a basis that
for all
. So then we must have
for all
. This implies that
for all
, but since the vectors
form a basis, we must have
for all
.
Let
be the matrix with the vectors
as column vectors. Let
. Then
is equivalent to saying that
is a linear combination of the vectors
. By Theorem 1.2.21,
has a unique solution
if and only if
is invertible.
In particular, this means that also
has a unique solution
. This shows that 1)
span the space
and 2)
are linearly independent.
a)
.
b)
.
c)
or
.
The given operations correspond to row operations on matrices. By Theorem 1.2.16, any matrix that is invertible can be reduced to the identity using such operations. In Exercise 3.8 we proved that the columns of a matrix form a basis if and only if the matrix is invertible.
a) Any basis in
corresponds to a matrix that is invertible, i.e., an element of
. On the other hand the column vectors of any element from
form basis vectors for
.
b) For
we have that there are in total
matrices in
of which we have to count the ones that are not invertible. Considering the columns of a matrix in
, we have
first columns that are not the
column vector, and
scalings of the first column with a value other than zero.
- If the first column is
, the second column can be chosen in
ways such that it is not also the
vector.
- If the second column is
, the first column can also be chosen in
ways such that it is not the
vector.
- There is exactly one matrix with both columns
.
Combining these facts, we get that there are
invertible matrices in
.
For
we want to compute the number of matrices in
with determinant equal to 1. In
there are equally many elements with determinant 1, 2, 3, etc.Therefore, the number of elements in
is the number of elements with determinant 1 times
. From the previous calculation we thus get that the number of elements in
is
.
a) The key for finding the number of subspaces is to find the number of linearly independent vectors in
.
- Subspaces of dimension 0: 1.
- Subspaces of dimension 1: Each subspace is spanned by a nonzero vector of the form
with
. There are
such vectors. For any such given vector, there are
nonzero scalings with a scalar in
. Hence, the number of linearly independent vectors is
. Each such vector spans a subspace that is different from the subspaces spanned by the others.
- Subspaces of dimension 2: Let
be some maximal collection of linearly independent vectors of
. We know that
, and any two vectors from
span a two-dimensional subspace of
. We can choose two vectors from
in
ways, but this is not the number of two-dimensional subspaces of
. Indeed, say we choose
such that
. Then
is a subspace of
containing
points and
linearly independent vectors. As
, this means
contains some vector
. The number of pairs of linearly independent vectors in
is
, and hence the number of two-dimensional subspaces of
is
. Another way to arriving the same conclusion is as follows: Let
be a subspace of
of dimension 2. Then,
is spanned by two linearly independent vectors, and there is a vector
such that
. In other words, the vectors in
are linearly independent of
. We know that there are
linearly independent vectors in
, so whenever we choose one of such vectors, we are left with a subspace of dimension 2 that does not contain the chosen vector (but contains all the others). Hence, there are also
subspaces of dimension 2.
- Subspaces of dimension 3: 1.
b) The case of
can be generalised from the previous case:
- Subspaces of dimeansion 0: 1.
- Subspaces of dimension 1: The number of linearly independent vectors can be calculated similarly as in a), and we get
.
- Subspaces of dimension 2: Similarly as in the case of
, we have that the number of two-dimensional subspaces is
.
- Subspaces of dimension 3: Similarly as in the case of
, we have that for each three-dimensional subspace we have a one-dimensional subspace "left over". Therefore the number of three-dimensional subspaces is
.
- Subspaces of dimension 4: 1.
Let
be the space of symmetric and
the space of skew-symmetric matrices. It is clear that
and
and that
contains only the zero matrix, so the spaces are independent. By Proposition 3.6.4 b), we have
and so by Proposition 3.4.23,
.
The condition
introduces a linear dependency between the elements of the matrix. Therefore, we have
, and thus any one-dimensional subspace of
that is independent of
suffices. For example we can take as
the span of the matrix for which the top-left corner element is 1 and the rest are 0. Then
.
The given vectors span the set of sequences that are constant apart from a finite set of indices.
a) Let
and
, and
. Then we have also
. The coefficients
are linear in the coefficients
, so
we can solve as follows
Setting each
to zero yields a system of equations
.
By Corollary 1.2.14 this system has a solution where at least one of the coefficients
is non-zero, so there is a polynomial
that is not identically zero, but
for every
.
b) We can solve for example
using similar approach to part a).
c) Let
be a polynomial of degree
and
polynomial of degree
, so that
and
. Let
be a polynomial with unknown coefficients
. In order to have this polynomial vanish at
, we have to solve the equations that set the coefficient of
to 0 for each
in the polynomial
. These equations are linear in
, and there are
of them. On the other hand, there are
variables
, so by Corollary 1.2.14 the linear system has a non-zero solution. Note that in part a) we restricted the degree of the polynomial
to 2, and thus did not end up with as many equations as in this proof.