What inspires people to define linear maps?
$begingroup$
I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.
My understanding now is that a map is like a function that maps something to another thing.
Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?
Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.
And finally, why linear maps have the additivity and homogeneity properties?
linear-algebra
$endgroup$
add a comment |
$begingroup$
I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.
My understanding now is that a map is like a function that maps something to another thing.
Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?
Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.
And finally, why linear maps have the additivity and homogeneity properties?
linear-algebra
$endgroup$
add a comment |
$begingroup$
I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.
My understanding now is that a map is like a function that maps something to another thing.
Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?
Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.
And finally, why linear maps have the additivity and homogeneity properties?
linear-algebra
$endgroup$
I am currently self studying Linear Algebra Done Right. I am doing OK and currently on Chapter 3 Linear Maps.
My understanding now is that a map is like a function that maps something to another thing.
Why people like to particularly define linear maps? Just curious..maybe there are non-linear maps?
Another stupid question is that is linear maps really look like a straight line in some specific examples, like, $u, v $ is 1 dimensional vector space. And then $T(u+v) = T(u) + T(v)$ Or put it in another way why it is called linear map.
And finally, why linear maps have the additivity and homogeneity properties?
linear-algebra
linear-algebra
asked Jan 12 at 2:02
JOHN JOHN
1538
1538
add a comment |
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.
The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.
In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.
Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.
Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.
In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.
Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.
I hope that answers your question.
$endgroup$
add a comment |
$begingroup$
In calculus, the derivative is defined by
$$
f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
$$
Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
$$
tag{1} Delta f approx f'(x_0) Delta x.
$$
We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
$$
underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
$$
The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.
In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)
The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
$$
f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
$$
When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.
With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.
$endgroup$
add a comment |
$begingroup$
To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:
History
The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.
The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.
At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.
Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:
This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.
The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.
In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.
Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.
Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.
In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.
Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.
I hope that answers your question.
$endgroup$
add a comment |
$begingroup$
Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.
The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.
In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.
Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.
Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.
In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.
Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.
I hope that answers your question.
$endgroup$
add a comment |
$begingroup$
Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.
The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.
In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.
Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.
Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.
In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.
Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.
I hope that answers your question.
$endgroup$
Linearity is defined by additivity and scalar homogeneity. I can't explain why linear maps have these properties, other than to say it's inbuilt into the definition.
The separate question of as to why we care about these maps is a good one. We define linearity because it's a simple and common property for relationships to have. "Map" is indeed another word for "function": we have a relationship $T$ between two vector spaces $X$ and $Y$. We map a vector $x$ to another vector $y$. If we were to add another vector $Delta x$ to $x$, this will result in a change in $Y$, according to $T$. When $T$ is linear, we can expect the result of adding $Delta x$ to our independent variable $x$ to be very predictable: it will always add $T(Delta x)$. That is, the same change applied to our variable $x$, no matter what $x$ is beforehand, will result in the same change in $y$: specifically, adding $T(Delta x)$.
In a practical sense, this is a useful property to have, as it makes interpolations and extrapolations trivially easy. If you work $30$ hours in a week and take home $$600$ in that week, then the linear relationship will tell you very quickly that you would take home $$800$ for a $40$ hour week, or $400$ for a $20$ hour week.
Compare this to something a bit more complex. Say you're making a rocket to fly into space, and you're wondering how much fuel to put into the tank. The relationship is not linear, since putting double the fuel will not result in double the distance; each additional tonne of fuel adds weight, and more fuel will be needed to propel this extra weight into space. So, to go double the distance into space, you'll need a fair bit more than double the fuel. This is an example of a non-linear relationship, and it makes things a little more complicated.
Often, when a non-linear relationship rears its head, people will attack it with some form of linearisation. In fact, calculus is primarily about linearisation; derivatives are, in essence, about approximating a (possibly non-linear) function by a linear function.
In a theoretic sense, linear maps lend themselves to some useful theory, which hopefully you're learning about. Linear maps between finite-dimensional spaces can be represented by matrices, and almost anywhere you see a matrix, there's some intuition about linear maps behind it. Linear operators (square matrices, in the finite-dimensional setting) are of great interest to people studying dynamical systems, and form the basis for areas of study like $C^*$ algebras.
Why are they called "linear" maps? Well, in essence, they preserve lines. Actually, there is a slightly larger class of maps called "affine" maps that also preserve lines, but these are merely linear maps with a constant added on them. Linear maps are maps that preserve lines and the origin.
I hope that answers your question.
answered Jan 12 at 2:39
Theo BenditTheo Bendit
17.2k12149
17.2k12149
add a comment |
add a comment |
$begingroup$
In calculus, the derivative is defined by
$$
f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
$$
Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
$$
tag{1} Delta f approx f'(x_0) Delta x.
$$
We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
$$
underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
$$
The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.
In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)
The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
$$
f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
$$
When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.
With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.
$endgroup$
add a comment |
$begingroup$
In calculus, the derivative is defined by
$$
f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
$$
Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
$$
tag{1} Delta f approx f'(x_0) Delta x.
$$
We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
$$
underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
$$
The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.
In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)
The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
$$
f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
$$
When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.
With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.
$endgroup$
add a comment |
$begingroup$
In calculus, the derivative is defined by
$$
f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
$$
Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
$$
tag{1} Delta f approx f'(x_0) Delta x.
$$
We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
$$
underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
$$
The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.
In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)
The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
$$
f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
$$
When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.
With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.
$endgroup$
In calculus, the derivative is defined by
$$
f'(x_0) = lim_{x to x_0} frac{f(x) - f(x_0)}{x - x_0}.
$$
Intuitively, if the input changes by a small amount $Delta x$, and the corresponding change in the output is $Delta f$, then the change in the output is related to the change in the input by the equation
$$
tag{1} Delta f approx f'(x_0) Delta x.
$$
We would like to generalize the idea of the derivative to functions $f:mathbb R^n to mathbb R^m$. In this case, $Delta x in mathbb R^n$ and $Delta f in mathbb R^m$. What type of thing should $f'(x_0)$ be?
$$
underbrace{Delta f}_{m times 1} approx underbrace{f'(x_0)}_{text{?}} underbrace{Delta x}_{n times 1}
$$
The answer is that $f'(x_0)$ should be a linear transformation from $mathbb R^n$ to $mathbb R^m$. Each component of the output should be a linear combination of the components of the input. That is the simplest or most obvious way to generalize the idea of multiplying by a scalar (in equation (1)) to this new setting, where the input and output are both vectors.
In my opinion, this is the most clear way to discover the idea of linear transformations. This is why we care about them. (At least, it is a major reason why we care about them.)
The fundamental strategy of calculus is to approximate a complicated nonlinear function $f$ by a linear function:
$$
f(x) approx underbrace{f(x_0) + f'(x_0)(x - x_0)}_{L(x)}.
$$
When we replace $f$ with its local linear approximation $L$, calculations are greatly simplified, and the approximation is often good enough to be useful. Most of calculus can be derived easily using this fundamental strategy.
With this viewpoint, we see that calculus and linear algebra are connected at the most basic level.
edited Jan 12 at 2:58
answered Jan 12 at 2:50
littleOlittleO
29.5k645109
29.5k645109
add a comment |
add a comment |
$begingroup$
To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:
History
The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.
The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.
At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.
Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:
This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.
$endgroup$
add a comment |
$begingroup$
To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:
History
The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.
The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.
At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.
Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:
This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.
$endgroup$
add a comment |
$begingroup$
To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:
History
The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.
The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.
At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.
Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:
This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.
$endgroup$
To get an appreciation of what 'inspires people', the following history section has been copied from the wikipedia article on Gaussian elimination:
History
The method of Gaussian elimination appears in the Chinese mathematical text Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in eighteen problems, with two to five equations. The first reference to the book by this title is dated to 179 CE, but parts of it were written as early as approximately 150 BCE. It was commented on by Liu Hui in the 3rd century.
The method in Europe stems from the notes of Isaac Newton. In 1670, he wrote that all the algebra books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied. Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in 1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional hand computers to solve the normal equations of least-squares problems. The algorithm that is taught in high school was named for Gauss only in the 1950s as a result of confusion over the history of the subject.
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm Jordan in 1888. However, the method also appears in an article by Clasen published in the same year. Jordan and Clasen probably discovered Gauss–Jordan elimination independently.
At a very rudimentary level, linear algebra provides a compact and beautiful notation for expressing real world problems concerning a system of linear equations. And since it so beautiful, studying these mathematical structures in the abstract has lead to a deeper understanding of our universe.
Since Theo Bendit mentioned C*-algebras (the pinnacle of this mathematical abstraction), I can't resist copying the final sentence from the wikipedia article on that subject:
This C*-algebra approach is used in the Haag-Kastler axiomatization of local quantum field theory, where every open set of Minkowski spacetime is associated with a C*-algebra.
answered Jan 12 at 14:36
CopyPasteItCopyPasteIt
4,1181628
4,1181628
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3070526%2fwhat-inspires-people-to-define-linear-maps%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown