Prove that $sum_{i,j=1}^Nx_ialpha_{ij}x_j>0$ for a non-symmetric matrix
$begingroup$
Let $A=(alpha)_{i,j=1}^N$ a real square matrix. The matrix $A$ is not necessary symmetric. I want to prove that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j>0,;;;forall (x_1,ldots,x_N)inmathbb{R}^Nbackslashleftlbrace 0rightrbrace.$$
If matrix $A$ was symmetric, one way to do this, is to prove that all eigenvalues are positive. How can I deal with this case?
linear-algebra matrices
$endgroup$
add a comment |
$begingroup$
Let $A=(alpha)_{i,j=1}^N$ a real square matrix. The matrix $A$ is not necessary symmetric. I want to prove that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j>0,;;;forall (x_1,ldots,x_N)inmathbb{R}^Nbackslashleftlbrace 0rightrbrace.$$
If matrix $A$ was symmetric, one way to do this, is to prove that all eigenvalues are positive. How can I deal with this case?
linear-algebra matrices
$endgroup$
1
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22
add a comment |
$begingroup$
Let $A=(alpha)_{i,j=1}^N$ a real square matrix. The matrix $A$ is not necessary symmetric. I want to prove that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j>0,;;;forall (x_1,ldots,x_N)inmathbb{R}^Nbackslashleftlbrace 0rightrbrace.$$
If matrix $A$ was symmetric, one way to do this, is to prove that all eigenvalues are positive. How can I deal with this case?
linear-algebra matrices
$endgroup$
Let $A=(alpha)_{i,j=1}^N$ a real square matrix. The matrix $A$ is not necessary symmetric. I want to prove that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j>0,;;;forall (x_1,ldots,x_N)inmathbb{R}^Nbackslashleftlbrace 0rightrbrace.$$
If matrix $A$ was symmetric, one way to do this, is to prove that all eigenvalues are positive. How can I deal with this case?
linear-algebra matrices
linear-algebra matrices
asked Dec 28 '18 at 19:31
math_lovermath_lover
9710
9710
1
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22
add a comment |
1
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22
1
1
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22
add a comment |
2 Answers
2
active
oldest
votes
$begingroup$
Any real square matrix $A$ can be written (uniquely) as the sum of a symmetric matrix and a skew-symmetric matrix
$$A = A_{text{sym}} + A_{text{sq}}$$
where
$$A_{text{sym}} = frac{1}{2} ( A + A^t) \
A_{text{sq}} = frac{1}{2} ( A - A^t) $$
Say $A_{text{sym}}= (tilde alpha_{ij})$. It is easy to see that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j=sum_{i,j=1}^Nx_itildealpha_{ij}x_j $$ ( that is, $A$ and $A_{text{sym}}$ produce the same quadratic form). Now you are reduced to checking whether the real symmetric matrix $A_{text{sym}}$ is positive definite, which is standard.
$endgroup$
add a comment |
$begingroup$
This is false. Consider $A=-I_N$ and $x_1=x_2=cdots=x_N=1$. Then
$$sum_{i,j=1}^Nx_ia_{i,j}x_j=-N.$$
$endgroup$
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
add a comment |
Your Answer
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3055220%2fprove-that-sum-i-j-1nx-i-alpha-ijx-j0-for-a-non-symmetric-matrix%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Any real square matrix $A$ can be written (uniquely) as the sum of a symmetric matrix and a skew-symmetric matrix
$$A = A_{text{sym}} + A_{text{sq}}$$
where
$$A_{text{sym}} = frac{1}{2} ( A + A^t) \
A_{text{sq}} = frac{1}{2} ( A - A^t) $$
Say $A_{text{sym}}= (tilde alpha_{ij})$. It is easy to see that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j=sum_{i,j=1}^Nx_itildealpha_{ij}x_j $$ ( that is, $A$ and $A_{text{sym}}$ produce the same quadratic form). Now you are reduced to checking whether the real symmetric matrix $A_{text{sym}}$ is positive definite, which is standard.
$endgroup$
add a comment |
$begingroup$
Any real square matrix $A$ can be written (uniquely) as the sum of a symmetric matrix and a skew-symmetric matrix
$$A = A_{text{sym}} + A_{text{sq}}$$
where
$$A_{text{sym}} = frac{1}{2} ( A + A^t) \
A_{text{sq}} = frac{1}{2} ( A - A^t) $$
Say $A_{text{sym}}= (tilde alpha_{ij})$. It is easy to see that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j=sum_{i,j=1}^Nx_itildealpha_{ij}x_j $$ ( that is, $A$ and $A_{text{sym}}$ produce the same quadratic form). Now you are reduced to checking whether the real symmetric matrix $A_{text{sym}}$ is positive definite, which is standard.
$endgroup$
add a comment |
$begingroup$
Any real square matrix $A$ can be written (uniquely) as the sum of a symmetric matrix and a skew-symmetric matrix
$$A = A_{text{sym}} + A_{text{sq}}$$
where
$$A_{text{sym}} = frac{1}{2} ( A + A^t) \
A_{text{sq}} = frac{1}{2} ( A - A^t) $$
Say $A_{text{sym}}= (tilde alpha_{ij})$. It is easy to see that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j=sum_{i,j=1}^Nx_itildealpha_{ij}x_j $$ ( that is, $A$ and $A_{text{sym}}$ produce the same quadratic form). Now you are reduced to checking whether the real symmetric matrix $A_{text{sym}}$ is positive definite, which is standard.
$endgroup$
Any real square matrix $A$ can be written (uniquely) as the sum of a symmetric matrix and a skew-symmetric matrix
$$A = A_{text{sym}} + A_{text{sq}}$$
where
$$A_{text{sym}} = frac{1}{2} ( A + A^t) \
A_{text{sq}} = frac{1}{2} ( A - A^t) $$
Say $A_{text{sym}}= (tilde alpha_{ij})$. It is easy to see that
$$sum_{i,j=1}^Nx_ialpha_{ij}x_j=sum_{i,j=1}^Nx_itildealpha_{ij}x_j $$ ( that is, $A$ and $A_{text{sym}}$ produce the same quadratic form). Now you are reduced to checking whether the real symmetric matrix $A_{text{sym}}$ is positive definite, which is standard.
answered Dec 28 '18 at 21:14
Orest BucicovschiOrest Bucicovschi
28.6k31748
28.6k31748
add a comment |
add a comment |
$begingroup$
This is false. Consider $A=-I_N$ and $x_1=x_2=cdots=x_N=1$. Then
$$sum_{i,j=1}^Nx_ia_{i,j}x_j=-N.$$
$endgroup$
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
add a comment |
$begingroup$
This is false. Consider $A=-I_N$ and $x_1=x_2=cdots=x_N=1$. Then
$$sum_{i,j=1}^Nx_ia_{i,j}x_j=-N.$$
$endgroup$
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
add a comment |
$begingroup$
This is false. Consider $A=-I_N$ and $x_1=x_2=cdots=x_N=1$. Then
$$sum_{i,j=1}^Nx_ia_{i,j}x_j=-N.$$
$endgroup$
This is false. Consider $A=-I_N$ and $x_1=x_2=cdots=x_N=1$. Then
$$sum_{i,j=1}^Nx_ia_{i,j}x_j=-N.$$
answered Dec 28 '18 at 19:35
Ben WBen W
2,734918
2,734918
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
add a comment |
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
$begingroup$
It isn't clear, but I think the OP has a particular $alpha$ in mind or is asking about techniques for proving this sort of result in general. (Of course, symmetric does not imply positive-semidefinite.)
$endgroup$
– anomaly
Dec 28 '18 at 21:23
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3055220%2fprove-that-sum-i-j-1nx-i-alpha-ijx-j0-for-a-non-symmetric-matrix%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
Replace $alpha$ with $frac{1}{2}(alpha + alpha^t)$.
$endgroup$
– anomaly
Dec 28 '18 at 21:22