We've reviewed vector spaces, but the operations available to us are a little too basic to be useful yet. In this article, we'll introduce the concept of linear algebra.
In linear algebra, we'll be using the vectors introduced in the last article and adding some extra pieces called linear operators. Linear operators will be represented by capital letters like , , etc. Linear operators will act on vectors to produce new vectors. We'll notate it like this:
In English this says: When the linear operator acts on the vector , it produces the vector . Linear operators can also act across different vector spaces. i.e. in the example above, it might be the case that and are from different vector spaces.
We define the vector spaces a linear operator acts on as:
This says: the linear operator takes in vectors from the vector space and produces vectors in the vector space .
It's worth at this point mentioning a word that might be on the mind of some readers: Matrices. If you've studied matrices before, you will notice that linear operators appear to be very similar. They both involve operations on vectors. Matrices are one possible representation of linear operators. But, as we've said before, we're more interested in studying abstract objects and their properties rather than specific representations. So, we'll be focusing on the abstract properties of linear operators rather than concrete representations.
Linear operators only follow 2 very simple rules:
From these simple rules, we can derive a very important property of linear operators. Remember from the previous article we inroduced the idea of a basis. One of the properties we had for a basis is that it must be spanning. i.e. any vector in the vector space can be written as a linear combination of the basis vectors.
So, say we have a vector in a vector space . If we have a basis in this vector space, then we can write as for some set of scalars .
Now look what happens if we apply a linear operator to :
This shows us something very interesting: To calculate how acts on any vector we only need to know how it acts on the basis vectors of the vector space. This is a very powerful property as it means, for example in a 2D vector space, we only need to know how acts on two basis vectors instead of every possible vector in the space.
Let's look for example at the identity operator acting on a 2D vector space. The identity operator is essentially the "do nothing" operator as it takes every vector to itself. We define on our 2 basis vectors as:
And can therefore verify that behaves as we expect:
So it works!
The next concept we'll need to introduce is an inner-product. An inner product is an operation that takes two vectors and produces a scalar. We'll denote it with a dot () like this:
Note that we've reused the symbol that we use to multiply scalars together: . This is just convention and usually won't be confusing in context as we denote scalars and vectors differently.
Here are the rules that an inner-product must follow:
This shows us that the inner product distributes over addition and scalar multiplication
Where denotes the complex conjugate operator. Of course this is only relevant for vector spaces over the field of complex numbers. If we were working with real numbers we'd have:
We specify this complex conjugate explicitly here since it will be useful later in the course.
This means that the inner-product of a vector with itself is always greater than or equal to and is only equal to if the vector is .
Remember that is the additive identity of the field and is the additive identity of the vector space.
Some of you might be aware of something called the dot product, which is a specific inner-product in euclidean vector spaces but, as we've said before, we're more interested in the abstract properties of inner-products rather than any specific representations here.
With this inner-product, we will now define two new concepts: orthogonality and normality.
We say two vectors and are orthogonal if.
We say a single vector is normal if . This is sometimes called a unit vector.
If some vectors are orthogonal and normal we will describe them as orthonormal.
Something we will be interested in are orthonormal bases. That is, a set of basis vectors that are not only linearly independent and spanning, but also orthonormal:
Why is this useful? Well, consider taking the inner product of two vectors and . We can represent these vectors in terms of the orthonormal basis vectors: and . Then the inner product of these 2 vectors is:
But from our definition earlier, we know that the inner product of any 2 basis vectors is 0 unless . So we can simplify it:
And since our basis vectors are normal, when we take the inner product of them with themselves we get 1:
So by representating our vectors in an orthonormal basis we managed to find a nice way to calculate it. This also shows us that no matter what orthonormal basis we chose, we would always get the same result which tells us that the inner product is independent of the basis we chose.
Show that the inner product is conjugate linear in the first argument. That is:
In the space of 2D Euclidian vectors, prove or disprove that the following vectors are:
By saying "In the space of 2D Euclidian vectors" here, you may assume the following holds: