3
$\begingroup$

My book, professor, and friends make this theorem look very significant:

If A is an m by n matrix, then the following are either all true or all false:

  • Each vector b in R^m is a linear combination of the columns of A.

  • For each b in R^m, the equation Ax = b has a solution.

  • The columns of A span R^m.

  • A has a pivot position in every row.

My question is: isn't that a fairly obvious tautology? I mean, the definition of matrix multiplication simply expands the second equation into a linear combination, so why do people get so excited about this?

Sometimes I get the feeling that, in linear algebra, we're just finding fifty ways to state the same thing in different words, and getting excited even though they all stem from the same definition. :\

(Apologies in advance about the formatting, I'm not sure how the math formatting works here.)

  • 2
    The equivalence of those two statements establishes a connection between something you want (knowing when you can solve certain equations) and a certain property those equations may have. It is quite not earth shattering---and I suspect the *other* statements in that theorem were more interesting. In any case, you should not subestimate the value of being able to view the same truth in many, many different forms: I would venture the claim that most of the power and usefulness of linear algebra stems from the fact that it is found all over the place, in many different forms and shapes.2011-01-31
  • 1
    @Mariano: I was just too lazy to write the other parts, but I've edited my question to include them. :) The issue I'm having is, those are equivalent simply because of how we *defined* matrices and matrix multiplication... if you define things to turn out equal, then of course they will turn out equal. So it makes me wonder if we're actually discovering anything new here, or just wasting time defining new definitions and setting them equal to what we already had...2011-01-31
  • 11
    Well, you are ignoring one little fact: you arrived quite late to the party! Those definitions are the result of a very long, arduous process of trying to capture what is important to prove certain things that, in specific situations, were proved originally *by hand* and *concretely*. That linear algebra has reached a point where basic facts seem to be obvious is not a hint of its underlying triviality but a testament of the amazing work of giants in picking a good formalization!2011-01-31
  • 1
    @Mariano: But I don't understand: the *definition* of "The columns of *A* span R^m" is literally that you can get any vector **b** in R^m with some linear combination of the columns of *A*, isn't it? If that's the definition, then what two different facts are being proven equal? It's like saying "I am taller than my sister if my height is more than hers" and "My sister is shorter than me if her height is less than mine", and then "proving" this theorem: "I am taller than my sister if and only if she is shorter than me", and then getting excited at this **remarkable** fact!2011-01-31
  • 1
    The most interesting thing in that theorem (the equivalenxe of the first 3 statements being pretty obvious) is that it shows that to decide solvability of a certain family of equations it is sufficient to apply a specific algorithm and see it it comes with an answer: if it does, well, there is a solution; if it doesn't, then in fact there was no solution to begin with. (More or less...)2011-01-31
  • 1
    Wait, but the theorem says nothing about the algorithm... Gaussian elimination isn't mentioned at all. The only thing it says is, if we know the pivot positions, then we know the other things -- but it doesn't say anything about how to find the pivot positions. :\2011-01-31
  • 1
    That the matrix has pivots means exactly that elimination with pivots succeeds. (finding the pivots is trivial: when you perform elimination, they jump at you, assuming you can tell a zero value from a nonzero value: the theorem says that the equations are solvable iff they exist)2011-01-31
  • 1
    Hm, I don't know... this all seems like a bit of recursive logic to me, with all these definitions that lead back to each other, I'm not sure why... I just don't feel like I'm having the same "aha!" moment with this as when I learned calculus, even though the concept of adding up the parts to calculate the whole is relatively obvious too. :\2011-01-31
  • 4
    Oh well... Faith will come.2011-01-31
  • 1
    What is more surprising is that the dimension of the row space is equal to the dimension of the column space.2011-01-31
  • 2
    @Mehrdad: Yes, this is a fairly trivial theorem (at least the first three items), since the proof just consists of "unraveling the definitions". I wish all students understood this as easily as you do. Good for you, now you can move on to more interesting things! :-)2011-01-31
  • 1
    @Hans: Haha thanks... yeah, I'm eagerly waiting to get to other topics like eigenvalues, since they definitely sound more interesting. :)2011-01-31
  • 5
    One last comment: the hallmark of a well-developed theory in mathematics is often that the definitions are sufficiently refined so that what _ought_ to be completely obvious _is in fact_ completely obvious. In view of that you may want to take the apparent triviality of that theorem to be the _a posteriori_ justification for the precise phrasing of the definition.2011-01-31
  • 0
    Perhaps another point to make is that the equivalence of the first two statements tell you that using matrices is not merely a convenient book-keeping device, as it might seem at first (based on the definition of "coefficient matrix" and "augmented matrix" of a system of linear equations), but actually gives you an algebraic structure to play with.2011-01-31

1 Answers 1

4

I have seen many, many students (and you have probably also seen them yourself) who would answer a definitive 'No' to your first question.

Of course it is obvious to most of us, and if it is obvious to you as well, then fantastic. As Hans said, you can move on to bigger and better things. If you are looking for more difficult and deeper results in linear algebra (I notice that you are worried linear algebra is a study in the invariants under change of notation) then perhaps you can look in to eigenvalue problems, and if you have studied some calculus, how they relate to the solutions of ODEs.

Linear algebra is one of those fields which crops up all the time. For example, a standard method of solving a very non-linear system of equations involves 'linearising' the system, whereupon certain nice theorems from linear theory apply, and then returning to the non-linear situation. While I'm not saying that linear algebra is all one needs to solve linear PDEs (there is a fair bit of analysis after all!), I am trying to say that linear algebra plays a role, and thus attempt to convey the far-reaching nature of linear algebra as a discipline.

I do not know what you are looking for as an answer to this question after the excellent comments you have received already. But I think that one thing to keep in mind is what Willie already mentioned: mathematics is an ever-changing discipline, where 'good' definitions are those which allow one to prove 'obvious' facts are true. If you find this hard to believe, I can give one model example, again from PDEs. A large class of solutions to second order parabolic equations remain positive, if initially so. However for fourth (and higher) order equations this is not true. The unsatisfactory point is the extent to which it is not true. The negative part of the solution is very 'small' when compared with the magnitude of the positive part. In certain situations one can make this precise, by speaking about the 'average positivity' or 'eventual positivity'. Both of these are variations on the definition of 'positive' which allow us to quantify the behaviour we see from these equations. Before discovery, it was very hard to show this. (In fact, the opposite was a conjecture of Hadamard.) But now, working backward from a modern definition of 'almost positive' or 'eventually positive', the proof appears quite simple---almost elementary!

I hope this answer complements the other comments and instills a little bit of that faith Mariano mentioned ;).