# Introduction to eigenvalues and eigenvectors

## Video transcript

For any transformation that maps from Rn to Rn, we've done it implicitly, but it's been interesting for us to find the vectors that essentially just get scaled up by the transformations. So the vectors that have the form-- the transformation of my vector is just equal to some scaled-up version of a vector. And if this doesn't look familiar, I can jog your memory a little bit. When we were looking for basis vectors for the transformation-- let me draw it. This was from R2 to R2. So let me draw R2 right here. And let's say I had the vector v1 was equal to the vector 1, 2. And we had the lines spanned by that vector. We did this problem several videos ago. And I had the transformation that flipped across this line. So if we call that line l, T was the transformation from R2 to R2 that flipped vectors across this line. So it flipped vectors across l. So if you remember that transformation, if I had some random vector that looked like that, let's say that's x, that's vector x, then the transformation of x looks something like this. It's just flipped across that line. That was the transformation of x. And if you remember that video, we were looking for a change of basis that would allow us to at least figure out the matrix for the transformation, at least in an alternate basis. And then we could figure out the matrix for the transformation in the standard basis. And the basis we picked were basis vectors that didn't get changed much by the transformation, or ones that only got scaled by the transformation. For example, when I took the transformation of v1, it just equaled v1. Or we could say that the transformation of v1 just equaled 1 times v1. So if you just follow this little format that I set up here, lambda, in this case, would be 1. And of course, the vector in this case is v1. The transformation just scaled up v1 by 1. In that same problem, we had the other vector that we also looked at. It was the vector minus-- let's say it's the vector v2, which is-- let's say it's 2, minus 1. And then if you take the transformation of it, since it was orthogonal to the line, it just got flipped over like that. And that was a pretty interesting vector force as well, because the transformation of v2 in this situation is equal to what? Just minus v2. It's equal to minus v2. Or you could say that the transformation of v2 is equal to minus 1 times v2. And these were interesting vectors for us because when we defined a new basis with these guys as the basis vector, it was very easy to figure out our transformation matrix. And actually, that basis was very easy to compute with. And we'll explore that a little bit more in the future. But hopefully you realize that these are interesting vectors. There was also the cases where we had the planes spanned by some vectors. And then we had another vector that was popping out of the plane like that. And we were transforming things by taking the mirror image across this and we're like, well in that transformation, these red vectors don't change at all and this guy gets flipped over. So maybe those would make for good bases. Or those would make for good basis vectors. And they did. So in general, we're always interested with the vectors that just get scaled up by a transformation. It's not going to be all vectors, right? This vector that I drew here, this vector x, it doesn't just get scaled up, it actually gets changed, this direction gets changed. The vectors that get scaled up might switch direct-- might go from this direction to that direction, or maybe they go from that. Maybe that's x and then the transformation of x might be a scaled up version of x. Maybe it's that. The actual, I guess, line that they span will not change. And so that's what we're going to concern ourselves with. These have a special name. And they have a special name and I want to make this very clear because they're useful. It's not just some mathematical game we're playing, although sometimes we do fall into that trap. But they're actually useful. They're useful for defining bases because in those bases it's easier to find transformation matrices. They're more natural coordinate systems. And oftentimes, the transformation matrices in those bases are easier to compute with. And so these have special names. Any vector that satisfies this right here is called an eigenvector for the transformation T. And the lambda, the multiple that it becomes-- this is the eigenvalue associated with that eigenvector. So in the example I just gave where the transformation is flipping around this line, v1, the vector 1, 2 is an eigenvector of our transformation. So 1, 2 is an eigenvector. And it's corresponding eigenvalue is 1. This guy is also an eigenvector-- the vector 2, minus 1. He's also an eigenvector. A very fancy word, but all it means is a vector that's just scaled up by a transformation. It doesn't get changed in any more meaningful way than just the scaling factor. And it's corresponding eigenvalue is minus 1. If this transformation-- I don't know what its transformation matrix is. I forgot what it was. We actually figured it out a while ago. If this transformation matrix can be represented as a matrix vector product-- and it should be; it's a linear transformation-- then any v that satisfies the transformation of-- I'll say transformation of v is equal to lambda v, which also would be-- you know, the transformation of [? v ?] would just be A times v. These are also called eigenvectors of A, because A is just really the matrix representation of the transformation. So in this case, this would be an eigenvector of A, and this would be the eigenvalue associated with the eigenvector. So if you give me a matrix that represents some linear transformation. You can also figure these things out. Now the next video we're actually going to figure out a way to figure these things out. But what I want you to appreciate in this video is that it's easy to say, oh, the vectors that don't get changed much. But I want you to understand what that means. It literally just gets scaled up or maybe they get reversed. Their direction or the lines they span fundamentally don't change. And the reason why they're interesting for us is, well, one of the reasons why they're interesting for us is that they make for interesting basis vectors-- basis vectors whose transformation matrices are maybe computationally more simpler, or ones that make for better coordinate systems.