Advertisement
If you have a new account but are having problems posting or verifying your account, please email us on hello@boards.ie for help. Thanks :)
Hello all! Please ensure that you are posting a new thread or question in the appropriate forum. The Feedback forum is overwhelmed with questions that are having to be moved elsewhere. If you need help to verify your account contact hello@boards.ie
Hi there,
There is an issue with role permissions that is being worked on at the moment.
If you are having trouble with access or permissions on regional forums please post here to get access: https://www.boards.ie/discussion/2058365403/you-do-not-have-permission-for-that#latest

More Linear Algebra Fun!

  • 27-11-2010 8:39pm
    #1
    Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭


    I'm so angry :mad: You know basic matrix multiplication? Every book I've
    looked in, & I've spent all frickin' day on googlebooks & amazon checking
    this out
    , defines matrix multiplication in the shorthand summation notation
    or else as the standard row-column algorithm you should parrot off.
    Every forum I've read defines matrix multiplication as these things -
    I mean it's definition, who am I to question it? - or else says
    '
    it's because it works, it's just a convenient way to define it this way'
    or else uses some linear transformation explanation that I haven't studied
    yet (but will in 1 chapter!!) but this linear transformation thing doesn't look
    convincing to me from what I understand of it. Basically the only person
    who went to the trouble of explaining this properly was Sal of
    khanacademy :cool: here Did you know about this (I'll explain in a minute)?
    Where &/or in what book did you learn about it?
    Well, not every book defines it as these things. My book by Lang gives
    a slightly better explanation in terms of dot products but it wasn't
    satisfying enough. It was enough of a hint at the right way to do this
    but he didn't explain it properly unfortunately.

    Basically I am partly posting this to find out more about a specific operation
    known as the transpose. I'm actually a little confused because in one
    book, page 4, he defines a vector in two ways:

    X = (x,y,z)

    or

    ___|x|
    X= |y|
    ___|z|
    (Obviously the ___ are just to get the | | things to form a column shape :p)

    which are equivalent but then in the video I linked to above Sal calls
    the vector:
    ___|x|
    X= |y|
    ___|z|
    as if it's normal but calls X = (x,y,z) the transpose, Xᵀ = (x,y,z).

    I remember from my old study of linear algebra (which I hated & quit
    because it made no sense memorizing algorithms and faking my way through proofs)
    that
    the transpose is different somehow and is used in inverting a matrix
    I think but is a vector and it's transpose the same thing or something?

    Anyway, using this idea of a transpose it makes the whole concept of
    matrix multiplication 100% completely, lovingly, passionately, painfully,
    hatefully, relievingly intelligible. I think the picture is clear,

    blaea.jpg

    In part 3 you just take the transpose of each row of the 2x3 matrix &
    dot it with the 3x1 matrix. I just wrote part 4 in as well because
    in that book around page 4 he defines both modes of dot product as
    being the same, which they are.
    It seems like a trick the way Ive decomposed the matrix though,
    I mean I could use these techniques to multiply matrices regardless
    of their dimensions.
    I've just written down a method using these to multiply two matrices
    of dimensions 2x3, i.e. (2x3)•(2x3) and gotten a logical answer.

    If we copy the exact algorithm I've used in the picture then multiplying
    two matrices of equal size is indeed meaningless as you take the dot
    product of two differently sized vectors but if I play with the techniques
    used to decompose a big matrix I can swing it so that I get a logical
    answer.

    [1,2,3][a,b,c] = [1,2,3][a] [1,2,3] [1,2,3][c]
    [4,5,6][d,e,g] = [4,5,6][d] [4,5,6][e] [4,5,6][g]

    (This is the same as I do in the picture, then instead of transposing
    straight away I just use more of this decomposition only this time I
    decompose the left matrix instead of the right one):

    [1,2,3][a] [1,2,3] [1,2,3][c] = [1][a] [2][a] [1] [2] [3] ...
    [4,5,6][d] [4,5,6][e] [4,5,6][g] = [4][d] [5][d] [6][d] [4][e] [5][e] ...

    I can view this as a dot product:

    [1]•[a] [2]•[a] [1]• [2]• [3]• ...
    [4]•[d] [5]•[d] [6]•[d] [4]•[e] [5]•[e] ...

    (Obviously I just wrote 2 •'s @ each row vector to keep it neat ;))

    and I end up with some ridiculously crazy matrix essentially being
    meaningless but still following the "rules" I came up with something.

    This is important, my little knowledge of Hamilton is that he just defined
    ijk = -1 because it worked. Maybe this is true, and from what I know
    using this kind of algebra is useful in special relativity but I think it's
    literally a cheat, there is no explanation other than "it works".
    Hopefully I'm wrong! But, with my crazy matrix up here, why is it wrong?
    Is it really just that "it works" when we do it the way described in
    the picture but it doesn't work (i.e. it doesn't describe physical reality)
    when I do it the way I did here? What does this say about mathematics
    being independent from reality when we ignore things like this, my
    ridiculous matrix, and focus on the ones that describe reality?
    I know it's stupid but I don't knw why :P

    I'm also worried because just magically defining these things seems to
    be common, looking at differential forms I think, & this is because I
    haven't studied them properly, that you literally invoke this witchcraft
    when doing algebra with the dx's and dy's.

    I seriously hope that there are reasons behind all of this, thinking
    about the cross product post I made there was a perfect reason why
    things like ixj = k & j x i = -k make sense but here I'm worried.
    In the cross product example the use of a determinant, an abuse of
    notation, is a clear sign we're invoking magic spells to get the right
    answers but with matrix multiplication I haven't even located the
    source of the sourcery yet & it's driving me crazy :o
    Honestly, tell me now if I've got more of this to expect with
    differential forms or will I get a solid answer??? :pac:

    TL;DR - The method in the picture of multiplying matrices seems to me
    to be the most logical explanation of matrix multiplication, but why is
    it done that particular way & not the way I described in this part:
    [1,2,3][a,b,c] = [1,2,3][a] [....
    of the post? Also, with differential forms when you multiply differential's
    dx's and dy's etc... you are using magic sorcery adding minuses yada
    yada yada by definition, how come? Is there a beautiful reason for
    all of this like that described in the cross product post? Oh, and what's
    the deal with transposes? Transposing vectors is the reason why I can
    use this method in the picture, but I mean I could stupidly take the matrix
    [1,2,3]
    [4,5,6]
    as being either:
    (1,2,3) transposed from it's column vector form, or
    (1,4) & (2,5) & (3,6) as being the vectors, it's so weird...
    Also, I could have taken part 2 of the picture differently, multiplying
    the Y matrix by 3 1x2 X vectors, again it's so weird.... :o

    /pent_up_rant...


Comments

  • Registered Users, Registered Users 2 Posts: 360 ✭✭CJC86


    Ok, I'll give this one a shot, but I'll preface it by saying that I won't ever reply to a rant like this again. If you want help, then you should try to make your post readable.

    If you understand what a matrix actually is, then the way the multiplication is done is natural and is not an "algorithm" as you call it.

    A mxn matrix is a representation of a linear transformation from an m-dimensional vector space to an n-dimensional vector space. This means that if we take an m-dimensional vector, then we can form an n-dimensional vector from linear combinations of its m "components" (this is not really fully correct, but I'm trying to give the basic idea). A matrix is just a way to visually express these linear combinations.

    Now, if we have a linear transformation from l-dimensions to m-dimensions, and then one from m-dimensions to n-dimensions, then it would make sense that we can compose them to get one transformation from l to n dimensions, this is where matrix multiplication comes from, and why we can only multiply lxm matrices by mxn matrices.

    I was going to actually fill out the details for you, but it would be much better if you read what I've written above and compose the linear transformations, and see what you get. You will see that matrix multiplication is completely natural as long as you know what a matrix really is.

    (By the way, I have not seen a vector in row form since secondary school. Vectors are always assumed to be in column form, and their transposes are rows.)


  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    CJC86 wrote: »
    Ok, I'll give this one a shot, but I'll preface it by saying that I won't ever reply to a rant like this again. If you want help, then you should try to make your post readable.

    :eek:

    Seriously? First of all I wasn't ranting & I think it's obvious I was joking
    when I included it at the end of my post. Second my post is readable &
    I think it's obvious that I went to considerable effort by giving two
    versions of my post, a tl;dr one & the main one to make sure I got my
    point across. I know people on this forum would actually take the time to
    read it so the tl;dr wasn't included for lazy people it was just to sum up
    the main points to be extra careful. Third, if my post comes across as
    confused it is because I was confused. If I could have made it any better
    I wouldn't have felt the need to post. If I have to follow some personal
    rules of yours to get your help I think I'll be fine without it, most likely I
    am to expect further qualifications with every post & I think that passive
    aggressive rudeness is best kept out of a maths post.

    While composition of linear mappings does give a good explanation I
    think it's fair to say that the question I am asking is analogous to
    the situation with the dot product. We can view the dot product as
    a linear mapping from Rⁿ to R defined by L:X ↦ A•X we can also
    view the dot product as just being A•X = a₁x₁ + a₂x₂ + ... a_nx_n.
    If I asked what the dot product meant & the first explanation I got
    was that it was a special case of a linear mapping I'd be confused
    because I'd have no geometric or analytic insight. Qualifications that
    the dot product satisfies further properties like L(u + v) = L(u) + L(v) &
    L(cv) = cL(v) wouldn't shed light on it for me. For me, the further
    abstraction from dot product to linear mapping makes sense, not the
    other way round. I think that matrices as I've shown in the picture
    have some novel explanation akin to the secret behind the cross
    product I mentioned & I was just curious since I seen that video by Sal
    I linked to. I' just curious if there is more to that rabbit hole as I
    can't find anything else on it. It seems intuitive for me to think about
    matrices like that in the picture but it also raises a lot of questions & I
    just wanted to see if anyone who has come across this knows, if not
    then it's fine. I just can't ignore the feeling that these explanations
    are simply an abstraction & formalism of rules that were already in
    place with matrices. It seems I'm right because the history of matrices
    goes back to ancient times...


  • Registered Users, Registered Users 2 Posts: 1,595 ✭✭✭MathsManiac


    I strongly suspect that CJC86 was not relying on your miniscule finishing tag in order to describe your post as a rant.

    I came to the same conclusion on the basis of the duck test.
    ;)


  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    I strongly suspect that CJC86 was not relying on your miniscule finishing tag in order to describe your post as a rant.

    I came to the same conclusion on the basis of the duck test.
    ;)
    A rant is a speech or text that does not present a calm argument;
    rather, it is typically an enthusiastic speech or talk or lecture on an idea,
    a person or an institution. Compare with a dialectic.
    Rants can be based on partial fact or may be entirely factual but written
    in a comedic/satirical form.
    Rants can also be used in the defense of an individual, idea or organization.
    Rants of this type generally occur after the subject has been attacked by
    another individual or group."
    http://en.wikipedia.org/wiki/Rant

    So asking questions about ideas, giving your reasons for those questions
    & highlighting your concerns is nothing but a rant, that's good to know.
    If you guys don't know anything about the idea I'm asking about that's
    fine but if you're going to try to insult my questions at least insult me
    using the right words.


  • Registered Users, Registered Users 2 Posts: 360 ✭✭CJC86


    Ok, I guess that was a bit rude, but I did find your post extremely difficult to read, and since it was there for more than a day without a response I'd guess that many others thought the same. I did use the word rant because you had used it yourself, by the way, and I found it to be a fitting description.

    I read your post fully, and the abstraction from dot product to matrix multiplication seemed in no way "natural", whereas getting the dot product from linear transformations is completely natural, whichever came first in the history of mathematics. Taking transpose of single rows inside the matrix and then taking the dot product with the columns, while it gives the correct answer and clearly works, in my mind is a very convoluted and complicated way of explaining something that is simple viewed another way.

    That is what I was trying to get across in my post. While your way works and gives the right answer, it is terribly complicated to my eyes (why would you take the transpose of rows inside a matrix, unless you already knew that you had to multiply the row of the matrix on the left by the columns on the right?).

    Having watched the Khan academy video you linked to, I can see where you got it from, but it seems needlessly complicated to me. It seems he likes to teach people with something he has already used, which is fair enough, but I don't think the "transpose rows and take the dot product" method here helps with understanding why matrix multiplication is how it is.

    Again, apologies for being a bit rude, but I was stating a fact, that if I have to make a serious effort to read a post, it's very unlikely I'll take time to decipher it and reply. The reason I replied to you at all is because before this all of your posts I have read have been clear and making a serious effort to understand the maths you are coming across.

    Edit: Looking at your definition of a rant in your last post, I would say that your post was an enthusiastic post on an idea, which was not an entirely calm argument (in my opinion).


  • Advertisement
  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    I'm sorry if my post is a bit difficult to read but there is no mandate on
    anyone to respond nor do I want someone who would prefer not to
    respond to do so. The only idea I'm arguing for is that there has to be a
    good explanation of this way of looking at matrices akin to the idea behind
    the cross product (that I still haven't found mentioned in any book
    describing it even though, to me, it's the most logical way of understanding
    it), the rest is just questions about the ideas & of course it's not calm
    because I wasn't calm I was furious writing it. So if I was expounding on
    the idea that all the textbooks are wrong because they don't explain
    matrices the way I want it done I'd be ranting but I wasn't I
    was just asking about an idea I have many questions about & saying
    that I'd spent the whole day browsing books looking for it. I think I'll
    contact Sal & ask for a source...

    I'm just wondering about the specific idea in the video, I'm perfectly
    happy with thinking of matrices in terms of linear transformations &
    all that but I was just curious about this specific idea seeing as it's
    another way to look at things & personally it just seems like a nice
    way to look at it & I'd like to know about it.


  • Closed Accounts Posts: 6,081 ✭✭✭LeixlipRed


    Ladies, let's keep it civilised please.


  • Registered Users, Registered Users 2 Posts: 360 ✭✭CJC86


    In my opinion, the only reason he does it this way in the video is to use something that he has already defined to show people how to perform the calculation. It does not seem to be a natural explanation, rather a manipulation of the algebra to avoid saying "multiply the rows on the left by the columns on the right", which can be confusing at first.

    He doesn't necessarily want to go into the detail of composing the transformations at first, and he has a very natural looking multiplication at hand in the dot product. What I take issue with is that what he is doing in the middle seems quite unnatural, and he doesn't give a good explanation for doing so. At the end of the video he finally explains what a matrix is, and why you multiply them that way, but I personally don't like this.

    I think it's better to try explain the theory before doing a calculation, but he seems to think it's better to know how to do the calculation rather than have the theory nailed down. This is, in my opinion, a very bad way to be teaching maths, and why people end up learning algorithms rather than understanding what they are doing.


  • Registered Users, Registered Users 2 Posts: 2,481 ✭✭✭Fremen


    Sorry sponsoredwalk, I haven't looked through your whole post. It is a bit of an eyeful. ;)

    It's probably best to begin by thinking about points in the plane. In that setting, matrices represent very natural geometric transforms: dilations, rotations, projections and reflections off the top of my head, though there may be others.

    It's natural to be interested in compositions of transforms. A moment's thought will show you that a rotation followed by a projection is not necessarily the same as that same projection followed by the same rotation: composition of these transforms is noncommutative!

    It turns out that all these transforms can be represented by 2x2 matrices, and matrix multiplication is exactly what you need to represent the composition given the two transforms.

    Moreover, because of the linearity property, you know exactly what a transform will do to every vector in the plane just by knowing what it does to the vectors (1,0) and (0,1), since any vector (x,y) = x(1,0) + y(0,1).

    I know that's not quite what you asked, but for me it's by far the most natural way to think of a linear transformation.


  • Registered Users, Registered Users 2 Posts: 1,595 ✭✭✭MathsManiac


    I'll add my apologies too. But I was surprised that you were surprised at the reaction. I too found the post difficult to read and I think that the tenor of your opening paragraph was not conducive to putting us all in a happy mood and eager to read on.

    Anyway, setting all that aside, I'll try to offer a helpful perspective on the substance of the post.

    I agree with CJC that the method in the video and repeated in your post is not especially helpful. (However, I appreciate that the fact that you found it enlightening indicates that it is clearly helpful to some). The effort to link the operation of matrix multiplication to the dot product is a bit contrived, in my opinion.

    I think that there's a more general issue here. It strikes me that perhaps your dissatisfaction arises from the way that mathematics in general proceeds into new territory. Of necessity, when we are introducing a new structure or operation, we do so by defining it and its properties in the abstract. These can be viewed to be quite arbitrary in one sense, but we lay out the definition clearly and precisely so that everyone can be clear about what is involved. Usually, we choose to define things in the way we do because, by doing so, the structures we are studying and the properties they possess turn out to model some aspect of reality (or some other branch of mathematics) in a useful way.

    This happens from the basic level all the way through to the most advanced. For example, building axiomatically, we define the integers to be a set and some operations with certain properties. The particular rules we choose might seem quite arbitrary. However, by taking the definitions we do, we note a few crucial things: the (previously defined) natural numbers can be seen to be naturally embedded in this new structure, so that it is in a sense an extension of them, and secondly, this new structure happens to be a good way of modelling the reality of some things: a handy way to count the floors above and below ground in a building, for example, and how can move up and down through them, or credit and debit in a bank account, and so on. To the pure mathematician, however, this reflection of reality is almost coincidental, as the defined structure has its own internal logic and consistency that stans in its own right.

    Similarly, when defining addition and multiplication of rational numbers, why do we pick (a,b)+(c,d) := (ad+bc,bd) and (a,b)*(c,d) := (ac,bd), respectively? There are lots of other completely different operations that we could define instead, many of which could be made to still satisfy laws of commutivity, distributivity, etc, and which might admit an embedded subset equivalent to the integers. We choose these operations because our abstract structure then models, in a useful way, the business of chopping up pizzas and apples and sharing them between two and a quarter friends, etc.

    [Aside, this modelling step is non-trivial. e.g. How do you answer the primary school student who asks "Why is a half times a half equal to a quarter?" You might, by chopping up pizzas, convince them that a half[B] of [/B]a half is a quarter, but the tricky bit is: why should "of" be represented by multiplication? Not impossible to explain, but not easy either. To the mathematician, of course, the only reason that a half by a half is a quarter is because this follows immediately from the definition of multiplication in the rationals.]

    If we refer, in advance of gving our definitions, to the reality we hope to model, then this is generally presented only by way of motivation. For the whole structure of mathematics to hang together logically, we have to take the abstract definition given as the fundamental way we proceed.

    Bringing all this back to the point: it seems to me that you are dissatisfied because the definition of matrix multiplication appears to have been plucked out of the air. But of course, this is how all such business happens. There are lots of other ways to define what one might call a "multiplication" of matrices. For example, here's one: for two matrices of the same shape, let the product be given by multiplying the corresponding entries. This is a very well behaved and intuitive operation: it's commutative (assuming we're talking about matrices whose entries are in R or C), it's associative, it's distributive over matrix addition, etc. So, why is this not the one that everyone agrees to refer to as "matrix multiplication"? The answer is that it's not especially useful for anything, whereas the one we do use is.

    Matrix multiplication is defined the way it is pretty much for the sole reason that, by doing so, it happens to correspond with composition of transformations.

    In cases where matrices happen to nicely model something else, and when some other operation on matrices happens to correspond with an important aspect of that other thing, then the relevant operation would be so defined and used. It would not be called multiplication, because mathematicians (usually) have enough sense to avoid using the same term in the same context to mean two completely different things.


  • Advertisement
  • Registered Users, Registered Users 2 Posts: 5,083 ✭✭✭RoundTower


    I just can't ignore the feeling that these explanations
    are simply an abstraction & formalism of rules that were already in
    place with matrices. It seems I'm right because the history of matrices
    goes back to ancient times...

    this is the only thing that doesn't make sense to me. Is it really true that matrices, in their current form, go back to "ancient times"? Arranging numbers in a rectangular array is one thing (like the Chinese puzzle on the back of the tortoise which I forget right now) but imo you can only call it a matrix once you have, at a minimum, matrix multiplication as defined now. I don't believe this was around before Gauss or so.

    Assuming I am right, I would agree with MathsManiac: multiplication is defined that way because it represents something useful once you start to think of a matrix as a representation of a linear transformation.


  • Registered Users, Registered Users 2 Posts: 1,595 ✭✭✭MathsManiac


    According to Struk, the Chinese Han Dynasty text Nine Chapters on the Mathematical Art has systems of linear equations written as a matrix of the coefficients, and shows their solution by matrix transformations (presumably elementary row operations).

    The Wikipedia entry on Matrices has brief history section, indicating that it was reall in the mid-nineteenth century that they were developed in the way they are considered today.

    It would be interesting to find out when exactly the concept of matrix multiplication as we know it today was invented.


  • Registered Users, Registered Users 2 Posts: 144 ✭✭gentillabdulla




  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    Holy jesus I've cracked it!!!

    I started revising my linear algebra, it's fun - each time you come back
    the level of abstraction goes up & with it understanding.

    I went back and read some of Arthur Cayley's "A Memoir on the Theory of
    Matrices"
    (.pdf) because I was confronted with the issue of matrix multiplication
    again. Needless to say the idea of linear mappings never satisfied me in the
    way what I'm about to write does.

    If you read Cayley's paper he denotes his matrices as follows:

    (a, b, c)
    |α, β, γ|
    |θ, λ, μ|

    He includes the "( )" in the top line of his matrix, why? Well I thought
    it was some notational hazard seeing as he's writing in 1858, basically it
    just put me on the alert for typo's relating to age.

    Then, he wrote something very novel, he tells us that a more convenient
    way to represent the system of linear equations:

    T = ax + by + cz
    U = αx + βy + γz
    V = θx + λy + μz

    is as follows:

    (T,U,V) = (a, b, c)(x,y,z)
    ________|α, β, γ|
    ________|θ, λ, μ|


    Notice the shape of the (T,U,V) & (x,y,z)! It just looks far more natural
    this way. At first I thought it was a notational hazard, (1858!), but then
    I thought no because looking at the paper he certainly took advantage
    of Gutenberg so I thought about it & jeesh is this notation far clearer!
    I want to stress this point because if we write the above system as
    a linear combination in the way the notation clearly suggests it is
    extremely clear what's going on (first of all) and later on it is pivotal,
    so I'll do it:

    (T,U,V) = (a, b, c)(x,y,z) = _|a| __ |b| __ |c| = ax + by + cz
    ________|α, β, γ| _____ = x|α| + y|β| + z|γ| = αx + βy + γz
    ________|θ, λ, μ| _____ = _|θ| __ |λ| __ |μ| = θx + λy + μz


    He goes on to explain that

    (T,U,V) = (a, b, c)(x,y,z)
    ________|α, β, γ|
    ________|θ, λ, μ|

    represents the set of linear functions:

    ((a, b, c)(x,y,z),(α, β, γ)(x,y,z),(θ, λ, μ)(x,y,z))

    I think it's clear that

    (T,U,V) = ((a, b, c)(x,y,z),(α, β, γ)(x,y,z),(θ, λ, μ)(x,y,z))

    So we see that

    T = (a, b, c)(x,y,z) = ax + by + cz
    U = (α, β, γ)(x,y,z) = αx + βy + γz
    V = (θ, λ, μ)(x,y,z) = θx + λy + μz

    If you compare this notation to modern notation:

    |T| = |a, b, c||x| = |(ax + by + cz)|
    |U| = |α, β, γ||y|=_|(αx + βy + γz)|
    |V| = |θ, λ, μ||z| = |(θx + λy + μz)|

    multiplication of matrices is extremely clear when multiplying an (A)mxn
    matrix by an (X)nx1 matrix. It just follows from unfurling a system.
    We're going to use this idea when multiplying general matrices.

    He then goes on to define all of the standard abelian operations of
    matrices in a nice way in the paper which I really recommend reading but
    what comes next is absolutely astonishing.

    He defines the standard matrix representation
    ________
    (T,U,V) = (a, b, c)(x,y,z)
    ________|α, β, γ|
    ________|θ, λ, μ|

    but goes on to define

    (x,y,z) = (a', b', c')(ξ,η,ζ)
    ________|α', β', γ'|
    ________|θ', λ', μ'|


    Just think about all of this so far! No magic or defining strange operations,
    matrix decomposition is natural and through it we see that matrix
    multiplication is like a kind of linear algebra chain rule, very natural.
    It is absolutely brilliant how he just morphed (x,y,z) there!

    So:

    (T,U,V) = (a, b, c)(x,y,z) = (a, b, c)_(a', b', c')(ξ,η,ζ)
    ________|α, β, γ| _____ = |α, β, γ| |α', β', γ'|
    ________|θ, λ, μ| _____ = |θ, λ, μ|_|θ', λ', μ'|

    If you read his paper you'll see that he then defines

    (a, b, c)_(a', b', c')(ξ,η,ζ) = |A _ B__C |(ξ,η,ζ)
    |α, β, γ| |α', β', γ'|_____ = |A'_ B' _C' |
    |θ, λ, μ|_|θ', λ', μ'| _____= |A''_B''_ C''|

    This is the genius of his idea and derivation, how he gets from the L.H.S.
    to the R.H.S. is what I'm going to do in the rest of my post, he skips this
    step & it took me quite a while to figure it out, if you're looking for a
    challenge read the first few pages of Caley's paper & try to figure it out
    without peeking at my solution, it's an exercise :cool:

    (NB: This could be extremely easy & I just don't recognise it)
    ----

    (T,U,V) = (a, b, c)(x,y,z) = _|a| __ |b| __ |c|
    ________|α, β, γ| _____ = x|α| + y|β| + z|γ|
    ________|θ, λ, μ| _____ = _|θ| __ |λ| __ |μ|

    (x,y,z) = (a', b', c')(ξ,η,ζ)
    ________|α', β', γ'|
    ________|θ', λ', μ'|

    x = a'ξ + b'η + c'ζ
    y = α'ξ + β'η + γ'ζ
    z = θ'ξ + λ'η + μ'ζ

    (T,U,V) = T = _|a| __ |b| __ |c| = _____________|a| + _____________|b| + _____________|c|
    ________ U = x|α| + y|β| + z|γ| = (a'ξ + b'η + c'ζ)|α| + (α'ξ + β'η + γ'ζ)|β| + (θ'ξ + λ'η + μ'ζ)|γ|
    ________ V = _|θ| __ |λ| __ |μ| = _____________|θ| + _____________|λ| + _____________|μ|

    T =
    (a'ξ + b'η + c'ζ)•a + (α'ξ + β'η + γ'ζ)•b + (θ'ξ + λ'η + μ'ζ)•c
    U = (a'ξ + b'η + c'ζ)•α + (α'ξ + β'η + γ'ζ)•β + (θ'ξ + λ'η + μ'ζ)•γ
    V = (a'ξ + b'η + c'ζ)•θ + (α'ξ + β'η + γ'ζ)•λ + (θ'ξ + λ'η + μ'ζ)•μ

    T = a'ξ•a + b'η•a + c'ζ•a + α'ξ •b+ β'η•b + γ'ζ•b + θ'ξ•c + λ'η•c + μ'ζ•c
    U = a'ξ•α + b'η•α + c'ζ•α + α'ξ•β + β'η•β + γ'ζ•β + θ'ξ•γ + λ'η•γ + μ'ζ•γ
    V = a'ξ•θ + b'η•θ + c'ζ•θ + α'ξ•λ + β'η•λ + γ'ζ•λ + θ'ξ•μ + λ'η•μ + μ'ζ•μ

    T = (a'a + α'b + θ'c)ξ + (b'a + β'b + λ'c)η + (c'a + γ'b + μ'c)ζ
    U = (a'α + α'β + θ'γ)ξ + (b'α + β'β + λ'γ)η + (c'α + γ'β + μ'γ)ζ
    V = (a'θ + α'λ + θ'μ)ξ + (b'θ + β'λ + λ'μ)η + (c'θ + γ'λ + μ'μ)ζ

    T = |(a'a + α'b + θ'c) (b'a + β'b + λ'c) (c'a + γ'b + μ'c)| (ξ,η,ζ)
    U = |(a'α + α'β + θ'γ) (b'α + β'β + λ'γ) (c'α + γ'β + μ'γ)|
    V = |(a'θ + α'λ + θ'μ) (b'θ + β'λ + λ'μ) (c'θ + γ'λ + μ'μ)|

    (Change a'a to aa' in every entry, makes more sense for what follows)

    T = |(aa' + bα' + cθ') (ab' + bβ' + cλ') (ac' + bγ' + cμ')_| (ξ,η,ζ)
    U = |(αa' + βα' + γθ') (αb' + ββ' + γλ') (αc' + βγ' + γμ') |
    V = |(θa' + λα' + μθ') (θb' + λβ' + μλ') (θc'θ + λγ' + μμ')|


    (T,U,V) = |(a,b,c)(a',α',θ') (a,b,c)(b',β',λ') (a,b,c)(c',γ',μ')_| (ξ,η,ζ)
    _____ ___|(α,β,γ)(a',α',θ') (α,β,γ)(b',β',λ') (α,β,γ)(c',γ',μ')_|
    _____ ___|(θ,λ,μ)(a',α',θ') (θ,λ,μ)(b',β',λ') (θ,λ,μ)(c',γ',μ')_|

    There we have it! :D Whadda y'all think?

    144830.jpg
    (:D^Me Right Now^:D)

    If you multiply out the R.H.S. of:

    (a, b, c)(x,y,z) = (a, b, c)_(a', b', c')(ξ,η,ζ)
    |α, β, γ| _____ = |α, β, γ| |α', β', γ'|
    |θ, λ, μ| _____ = |θ, λ, μ|_|θ', λ', μ'|

    in the normal way you do it you see it agrees with the above derivation.

    There we have it, a formal justification and it's all self-contained in a
    standard Anxn matrix exploiting the unknowns (x,y,z). Obviously this can
    be turned into a neverending rabbit hole if we want it to.


    ----

    Edit: I'll add that looking at matrix multiplication this way is an irrefutable explanation of
    the reason behind the fact that an (A)mxn matrix by an (X)nxp gives a (B)mxp matrix,
    it explains why it doesn't make sense to multiply a matrix of column "n" by a matrix
    with rows anything but "n".


  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    I really must say that I'm over the moon since I read Cayley's explanation of
    matrices, did any of you check it out? Did you's already know this & just not
    understand that this was what I was looking for? If you skipped my post you
    should read Cayley's explanation at least, it's just fascinating & I think it
    would be a great optional section in a textbook that you could really set
    as a theorem to be proven with the mini-calculation I did left as an exercise
    mid-proof :D


  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    Jesus christ here's another way to multiply matrices :eek:

    I got Hoffman/Kunze Linear Algebra 2nd hand in the TCD bookshop the
    other day & this book is apparently one of the most rigorous there is at
    this level. It's a great book so far and look what fruitful gifts it bears!

    Lets re-cap, there is

    1) The dot-product formulation,
    2) Cayley's clear explanation I've written above,
    3) What I'll explain next:

    Since two systems of linear equations are equivalent if the equations of one
    system are linear combinations of the equations in the other system, we can
    use the idea of row-equivalence (the matrix version of system equivalence) to
    construct a new matrix whose entries are simply linear combinations of the
    (matrix-representative) equations making up a certain system.

    So, if B is an mxn matrix we construct a matrix C whose rows are simply
    linear combinations of the rows of B.

    ____|β₁|
    ____|β₂|
    B = _|. |
    ____|._|
    ____|._|
    ____|β₊| (Where the β's are rows)

    and


    ____|γ₁|
    ____|γ₂|
    C =_|._|
    ____|._|
    ____|._|
    ____|γ₊|

    The γ's are linear combinations of the rows of B, so

    γ₁ = α₁β₁ + α₂β₂ + ... + α₊β₊
    γ₂ = α₁β₁ + α₂β₂ + ... + α₊β₊
    .
    .
    .
    γ₊ = α₁β₁ + α₂β₂ + ... + α₊β₊

    Now comes the conceptual leap, the α's are the column elements of a
    different matrix A :eek:

    Took me a while to figure this one out but if you take the o'th row:

    γ₀ = α₁β₁ + α₂β₂ + ... + α₊β₊

    and append the a's with the o as follows:

    γ₀ = α₀₁β₁ + α₀₂β₂ + ... + α₀₊β₊

    for every row, i.e.

    γ₁ = α₁₁β₁ + α₁₂β₂ + ... + α₁₊β₊
    γ₂ = α₂₁β₁ + α₂₂β₂ + ... + α₂₊β₊
    .
    .
    .
    γ₊ = α₂₁β₁ + α₂₂β₂ + ... + α₂₊β₊

    you've got a new way of multiplying matrices giving equivalent results &
    it's grounded in some clever theory. Remember the β's are whole rows,
    it's crazy! :D


  • Registered Users, Registered Users 2 Posts: 2,481 ✭✭✭Fremen


    Is this change of basis or is there something I've missed?

    One of the really nice results from the change of basis idea is that some matrices can be represented so that the only nonzero entries are along the main diagonal. Suppose you want to calculate B^100, for some messy matrix B. You can either let your computer grind away at it for an hour, or you can convert it to a diagonal matrix, compute the hundredth power of each entry, then convert it back. Simples.

    I promise not to use the word simples on boards ever again.


  • Registered Users, Registered Users 2 Posts: 3,038 ✭✭✭sponsoredwalk


    The khanacadey video on change of basis makes it look very similar alright
    but I got the impression from the video that they are a little bit different.
    I'll let you know when I get to CoB in this book.


Advertisement