Main content

# Introduction to linear independence

## Video transcript

Let's say I had the set of vectors-- I don't want to do it that thick. Let's say one of the vectors is the vector 2, 3, and then the other vector is the vector 4, 6. And I just want to answer the question: what is the span of these vectors? And let's assume that these are position vectors. What are all of the vectors that these two vectors can represent? Well, if you just look at it, and remember, the span is just all of the vectors that can be represented by linear combinations of these. So it's the set of all the vectors that if I have some constant times 2 times that vector plus some other constant times this vector, it's all the possibilities that I can represent when I just put a bunch of different real numbers for c1 and c2. Now, the first thing you might realize is that, look, this vector 2, this is just the same thing as 2 times this vector. So I could just rewrite it like this. I could just rewrite it as c1 times the vector 2, 3 plus c2 times the vector-- and here, instead of writing the vector 4, 6, I'm going to write 2 times the vector 2, 3, because this vector is just a multiple of that vector. So I could write c2 times 2 times 2, 3. I think you see that this is equivalent to the 4, 6. 2 times 2 is 4. 2 times 3 is 6. Well, then we can simplify this a little bit. We can rewrite this as just c1 plus 2c2, all of that, times 2, 3, times our vector 2, 3. And this is just some arbitrary constant. It's some arbitrary constant plus 2 times some other arbitrary constant. So we can just call this c3 times my vector 2, 3. So in this situation, even though we started with two vectors, and I said, well, you know, the span of these two vectors is equal to all of the vectors that can be constructed with some linear combination of these, any linear combination of these, if I just use this substitution right here, can be reduced to just a scalar multiple of my first vector. And I could have gone the other way around. I could have substituted this vector as being 1/2 times this, and just made any combination of scalar multiple of the second vector. But the fact is, that instead of talking about linear combinations of two vectors, I can reduce this to just a scalar combination of one vector. And we've seen in R2 a scalar combination of one vector, especially if they're position vectors. For example, this vector 2, 3. It's 2, 3. It looks like this. All the scalar combinations of that vector are just going to lie along this line. So 2, 3, it's going to be right there. They're all just going to lie along that line right there, so along this line going in both directions forever. And if I take negative values of 2, 3, I'm going to go down here. If I take positive values, I'm going to go here. If I get really large positive values, it's going to go up here. But I can just represent the vectors, and when you put them in standard form, their arrows essentially would trace out this line. So you could say that the span of my set of vectors-- let me put it over here. The span of the set of vectors 2, 3 and 4, 6 is just this line here. Even though we have two vectors, they're essentially collinear. They're multiples of each other. I mean, if this is 2, 3, 4, 6 is just this right here. It's just that longer one right there. They're collinear. These two things are collinear. Now, in this case, when we have two collinear vectors in R2, essentially their span just reduces to that line. You can't represent some vector like-- let me do a new color. You can't represent this vector right there with some combination of those two vectors. There's no way to kind of break out of this line. So there's no way that you can represent everything in R2. So the span is just that line there. Now, a related idea to this, and notice, you had two vectors, but it kind of reduced to one vector when you took its linear combinations. The related idea here is that we call this set-- we call it linearly dependent. Let me write that down: linearly dependent. This is a linearly dependent set. And linearly dependent just means that one of the vectors in the set can be represented by some combination of the other vectors in the set. A way to think about it is whichever vector you pick that can be represented by the others, it's not adding any new directionality or any new information, right? In this case, we already had a vector that went in this direction, and when you throw this 4, 6 on there, you're going in the same direction, just scaled up. So it's not giving us any new dimension, letting us break out of this line, right? And you can imagine in three space, if you have one vector that looks like this and another vector that looks like this, two vectors that aren't collinear, they're going to define a kind of two-dimensional space. They can define a two-dimensional space. Let's say that this is the plane defined by those two vectors. In order to define R3, a third vector in that set can't be coplanar with those two, right? If this third vector is coplanar with these, it's not adding any more directionality. So this set of three vectors will also be linearly dependent. And another way to think about it is that these two purple vectors span this plane, span the plane that they define, essentially, right? Anything in this plane going in any direction can be-- any vector in this plane, when we say span it, that means that any vector can be represented by a linear combination of this vector and this vector, which means that if this vector is on that plane, it can be represented as a linear combination of that vector and that vector. So this green vector I added isn't going to add anything to the span of our set of vectors and that's because this is a linearly dependent set. This one can be represented by a sum of that one and that one because this one and this one span this plane. In order for the span of these three vectors to kind of get more dimensionality or start representing R3, the third vector will have to break out of that plane. It would have to break out of that plane. And if a vector is breaking out of that plane, that means it's a vector that can't be represented anywhere on that plane, so it's outside of the span of those two vectors. Where it's outside, it can't be represented by a linear combination of this one and this one. So if you had a vector of this one, this one, and this one, and just those three, none of these other things that I drew, that would be linearly independent. Let me draw a couple more examples for you. That one might have been a little too abstract. So, for example, if I have the vectors 2, 3 and I have the vector 7, 2, and I have the vector 9, 5, and I were to ask you, are these linearly dependent or independent? So at first you say, well, you know, it's not trivial. Let's see, this isn't a scalar multiple of that. That doesn't look like a scalar multiple of either of the other two. Maybe they're linearly independent. But then, if you kind of inspect them, you kind of see that v, if we call this v1, vector 1, plus vector 2, if we call this vector 2, is equal to vector 3. So vector 3 is a linear combination of these other two vectors. So this is a linearly dependent set. And if we were to show it, draw it in kind of two space, and it's just a general idea that-- well, let me see. Let me draw it in R2. There's a general idea that if you have three two-dimensional vectors, one of them is going to be redundant. Well, one of them definitely will be redundant. For example, if we do 2, 3, if we do the vector 2, 3, that's the first one right there. I draw it in the standard position. And I draw the vector 7, 2 right there, I could show you that any point in R2 can be represented by some linear combination of these two vectors. We can even do a kind of a graphical representation. I've done that in the previous video, so I could write that the span of v1 and v2 is equal to R2. That means that every vector, every position here can be represented by some linear combination of these two guys. Now, the vector 9, 5, it is in R2. It is in R2, right? Clearly. I just graphed it on this plane. It's in our two-dimensional, real number space. Or I guess we could call it a space or in our set R2. It's there. It's right there. So we just said that anything in R2 can be represented by a linear combination of those two guys. So clearly, this is in R2, so it can be represented as a linear combination. So hopefully, you're starting to see the relationship between span and linear independence or linear dependence. Let me do another example. Let's say I have the vectors-- let me do a new color. Let's say I have the vector-- and this one will be a little bit obvious-- 7, 0, so that's my v1, and then I have my second vector, which is 0, minus 1. That's v2. Now, is this set linearly independent? Is it linearly independent? Well, can I represent either of these as a combination of the other? And really when I say as a combination, you'd have to scale up one to get the other, because there's only two vectors here. If I am trying to add up to this vector, the only thing I have to deal with is this one, so all I can do is scale it up. Well, there's nothing I can do. No matter what I multiply this vector by, you know, some constant and add it to itself or scale it up, this term right here is always going to be zero. It's always going to be zero. So nothing I can multiply this by is going to get me to this vector. Likewise, no matter what I multiply this vector by, the top term is always going to be zero. So there's no way I could get to this vector. So both of these vectors, there's no way that you can represent one as a combination of the other. So these two are linearly independent. And you can even see it if we graph it. One is 7, 0, which is like that. Let me do it in a non-yellow color. 7, 0. And one is 0, minus 1. And I think you can clearly see that if you take a linear combination of any of these two, you can represent anything in R2. So the span of these, just to kind of get used to our notion of span of v1 and v2, is equal to R2. Now, this is another interesting point to make. I said the span of v1 and v2 is R2. Now what is the span of v1, v2, and v3 in this example up here? I already told you. I already showed you that this third vector can be represented as a linear combination of these two. It's actually just these two summed up. I can even draw it right here. It's just those two vectors summed up. So it clearly can be represented as a linear combination of those two. So what's its span? Well, the fact that this is redundant means that it doesn't change its span. It doesn't change all of the possible linear combinations. So its span is also going to be R2. It's just that this was more vectors than you needed to span R2. R2 is a two-dimensional space, and you needed two vectors. So this was kind of a more efficient way of providing a basis, and I haven't defined basis formally, yet, but I just want to use it a little conversationally, and then it'll make sense to you when I define it formally. This provides a better basis, or this provides a basis, kind of a non-redundant set of vectors that can represent R2. While this one, right here, is redundant. So it's not a good basis for R2. Let me give you one more example in three dimensions. And then in the next video, I'm going to make a more formal definition of linear dependence or independence. So let's say that I had the vector 2, 0, 0. Let me make a similar argument that I made up there: the vector 2, 0, 0, the vector 0, 1, 0, and the vector 0, 0, 7. We are now in R3, right? Each of these are three-dimensional vectors. Now, are these linear dependent or linearly independent? Sorry, are they linearly dependent or independent? Well, there's no way with some combination of these two vectors that I can end up with a non-zero term right here to make this third vector, right? Because no matter what I multiply this one by and this one by, this last term is going to be zero. So this is kind of adding a new direction to our set of vectors. Likewise, there's nothing I can do-- there's no combination of this guy and this guy that I can get a non-zero term here. And finally, no combination of this guy and this guy that I can get a non-zero term here. So this set is linearly independent. And if you were to graph these in three dimensions, you would see that none of these-- these three do not lie on the same plane. Obviously, any two of them lie on the same plane, but if you were to actually graph it, you get 2, 0. Let me say that that's x-axis. That's 2, 0, 0. Then you have this, 0, 1, 0. Maybe that's the y-axis. And then you have 0, 0, 7. It would look something like this. So it almost looks like, your three-dimensional axes, it almost looks like the vectors i, j, k. They're just scaled up a little bit. But you can always correct that by just scaling them down, right? Because we care about any linear combination of these. So the span of these three vectors right here, because they're all adding new directionality, is R3. Anyway, I thought I would leave you there in this video. I realize I've been making longer and longer videos, and I want to get back in the habit of making shorter ones. In the next video, I'm going to make a more formal definition of linear dependence, and we'll do a bunch more examples.