What inspires people to define linear maps?












5












$begingroup$


I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



My understanding now is that a map is like a function that maps something to another thing.



Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



And finally, why linear maps have the additivity and homogeneity properties?










share|cite|improve this question









$endgroup$

















    5












    $begingroup$


    I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



    My understanding now is that a map is like a function that maps something to another thing.



    Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



    Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



    And finally, why linear maps have the additivity and homogeneity properties?










    share|cite|improve this question









    $endgroup$















      5












      5








      5





      $begingroup$


      I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



      My understanding now is that a map is like a function that maps something to another thing.



      Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



      Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



      And finally, why linear maps have the additivity and homogeneity properties?










      share|cite|improve this question









      $endgroup$




      I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.



      My understanding now is that a map is like a function that maps something to another thing.



      Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?



      Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.



      And finally, why linear maps have the additivity and homogeneity properties?







      linear-algebra






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Jan 12 at 2:02









      JOHN JOHN

      1538




      1538






















          3 Answers
          3






          active

          oldest

          votes


















          2












          $begingroup$

          Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



          The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



          In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



          Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



          Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



          In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



          Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



          I hope that answers your question.






          share|cite|improve this answer









          $endgroup$





















            5












            $begingroup$

            In calculus, the derivative is defined by
            $$
            f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
            $$

            Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
            $$
            tag{1} Delta f approx f'(x_0) Delta x.
            $$



            We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
            $$
            underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
            $$

            The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



            In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



            The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
            $$
            f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
            $$



            When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



            With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






            share|cite|improve this answer











            $endgroup$





















              0












              $begingroup$

              To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:




              History



              The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.



              The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.



              Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.




              At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.



              Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:




              This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.







              share|cite|improve this answer









              $endgroup$













                Your Answer





                StackExchange.ifUsing("editor", function () {
                return StackExchange.using("mathjaxEditing", function () {
                StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
                StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
                });
                });
                }, "mathjax-editing");

                StackExchange.ready(function() {
                var channelOptions = {
                tags: "".split(" "),
                id: "69"
                };
                initTagRenderer("".split(" "), "".split(" "), channelOptions);

                StackExchange.using("externalEditor", function() {
                // Have to fire editor after snippets, if snippets enabled
                if (StackExchange.settings.snippets.snippetsEnabled) {
                StackExchange.using("snippets", function() {
                createEditor();
                });
                }
                else {
                createEditor();
                }
                });

                function createEditor() {
                StackExchange.prepareEditor({
                heartbeatType: 'answer',
                autoActivateHeartbeat: false,
                convertImagesToLinks: true,
                noModals: true,
                showLowRepImageUploadWarning: true,
                reputationToPostImages: 10,
                bindNavPrevention: true,
                postfix: "",
                imageUploader: {
                brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
                contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
                allowUrls: true
                },
                noCode: true, onDemand: true,
                discardSelector: ".discard-answer"
                ,immediatelyShowMarkdownHelp:true
                });


                }
                });














                draft saved

                draft discarded


















                StackExchange.ready(
                function () {
                StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
                }
                );

                Post as a guest















                Required, but never shown

























                3 Answers
                3






                active

                oldest

                votes








                3 Answers
                3






                active

                oldest

                votes









                active

                oldest

                votes






                active

                oldest

                votes









                2












                $begingroup$

                Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                I hope that answers your question.






                share|cite|improve this answer









                $endgroup$


















                  2












                  $begingroup$

                  Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                  The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                  In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                  Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                  Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                  In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                  Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                  I hope that answers your question.






                  share|cite|improve this answer









                  $endgroup$
















                    2












                    2








                    2





                    $begingroup$

                    Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                    The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                    In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                    Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                    Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                    In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                    Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                    I hope that answers your question.






                    share|cite|improve this answer









                    $endgroup$



                    Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.



                    The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.



                    In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.



                    Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.



                    Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.



                    In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.



                    Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.



                    I hope that answers your question.







                    share|cite|improve this answer












                    share|cite|improve this answer



                    share|cite|improve this answer










                    answered Jan 12 at 2:39









                    Theo BenditTheo Bendit

                    17.2k12149




                    17.2k12149























                        5












                        $begingroup$

                        In calculus, the derivative is defined by
                        $$
                        f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                        $$

                        Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                        $$
                        tag{1} Delta f approx f'(x_0) Delta x.
                        $$



                        We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                        $$
                        underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                        $$

                        The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                        In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                        The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                        $$
                        f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                        $$



                        When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                        With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                        share|cite|improve this answer











                        $endgroup$


















                          5












                          $begingroup$

                          In calculus, the derivative is defined by
                          $$
                          f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                          $$

                          Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                          $$
                          tag{1} Delta f approx f'(x_0) Delta x.
                          $$



                          We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                          $$
                          underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                          $$

                          The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                          In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                          The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                          $$
                          f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                          $$



                          When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                          With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                          share|cite|improve this answer











                          $endgroup$
















                            5












                            5








                            5





                            $begingroup$

                            In calculus, the derivative is defined by
                            $$
                            f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                            $$

                            Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                            $$
                            tag{1} Delta f approx f'(x_0) Delta x.
                            $$



                            We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                            $$
                            underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                            $$

                            The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                            In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                            The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                            $$
                            f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                            $$



                            When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                            With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.






                            share|cite|improve this answer











                            $endgroup$



                            In calculus, the derivative is defined by
                            $$
                            f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
                            $$

                            Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
                            $$
                            tag{1} Delta f approx f'(x_0) Delta x.
                            $$



                            We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
                            $$
                            underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
                            $$

                            The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.



                            In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)



                            The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
                            $$
                            f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
                            $$



                            When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.



                            With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.







                            share|cite|improve this answer














                            share|cite|improve this answer



                            share|cite|improve this answer








                            edited Jan 12 at 2:58

























                            answered Jan 12 at 2:50









                            littleOlittleO

                            29.5k645109




                            29.5k645109























                                0












                                $begingroup$

                                To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:




                                History



                                The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.



                                The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.



                                Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.




                                At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.



                                Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:




                                This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.







                                share|cite|improve this answer









                                $endgroup$


















                                  0












                                  $begingroup$

                                  To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:




                                  History



                                  The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.



                                  The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.



                                  Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.




                                  At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.



                                  Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:




                                  This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.







                                  share|cite|improve this answer









                                  $endgroup$
















                                    0












                                    0








                                    0





                                    $begingroup$

                                    To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:




                                    History



                                    The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.



                                    The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.



                                    Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.




                                    At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.



                                    Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:




                                    This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.







                                    share|cite|improve this answer









                                    $endgroup$



                                    To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:




                                    History



                                    The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.



                                    The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.



                                    Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.




                                    At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.



                                    Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:




                                    This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.








                                    share|cite|improve this answer












                                    share|cite|improve this answer



                                    share|cite|improve this answer










                                    answered Jan 12 at 14:36









                                    CopyPasteItCopyPasteIt

                                    4,1181628




                                    4,1181628






























                                        draft saved

                                        draft discarded




















































                                        Thanks for contributing an answer to Mathematics Stack Exchange!


                                        • Please be sure to answer the question. Provide details and share your research!

                                        But avoid



                                        • Asking for help, clarification, or responding to other answers.

                                        • Making statements based on opinion; back them up with references or personal experience.


                                        Use MathJax to format equations. MathJax reference.


                                        To learn more, see our tips on writing great answers.




                                        draft saved


                                        draft discarded














                                        StackExchange.ready(
                                        function () {
                                        StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
                                        }
                                        );

                                        Post as a guest















                                        Required, but never shown





















































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown

































                                        Required, but never shown














                                        Required, but never shown












                                        Required, but never shown







                                        Required, but never shown







                                        Popular posts from this blog

                                        How to change which sound is reproduced for terminal bell?

                                        Can I use Tabulator js library in my java Spring + Thymeleaf project?

                                        Title Spacing in Bjornstrup Chapter, Removing Chapter Number From Contents