Proof of log identity for positive definite matrices












0












$begingroup$


On Wikipedia, it is claimed that $rm{tr}(log(AB)) = rm{tr}log(A) + rm{tr}log(B)$.



The result is valid only if $A$ and $B$ are positive definite.



I use this result in entropy calculations but I do not know how to prove it - in particular to show that it only works for positive definite matrices. Could someone help me with the proof?










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
    $endgroup$
    – darij grinberg
    Nov 25 '18 at 22:50
















0












$begingroup$


On Wikipedia, it is claimed that $rm{tr}(log(AB)) = rm{tr}log(A) + rm{tr}log(B)$.



The result is valid only if $A$ and $B$ are positive definite.



I use this result in entropy calculations but I do not know how to prove it - in particular to show that it only works for positive definite matrices. Could someone help me with the proof?










share|cite|improve this question









$endgroup$








  • 1




    $begingroup$
    Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
    $endgroup$
    – darij grinberg
    Nov 25 '18 at 22:50














0












0








0





$begingroup$


On Wikipedia, it is claimed that $rm{tr}(log(AB)) = rm{tr}log(A) + rm{tr}log(B)$.



The result is valid only if $A$ and $B$ are positive definite.



I use this result in entropy calculations but I do not know how to prove it - in particular to show that it only works for positive definite matrices. Could someone help me with the proof?










share|cite|improve this question









$endgroup$




On Wikipedia, it is claimed that $rm{tr}(log(AB)) = rm{tr}log(A) + rm{tr}log(B)$.



The result is valid only if $A$ and $B$ are positive definite.



I use this result in entropy calculations but I do not know how to prove it - in particular to show that it only works for positive definite matrices. Could someone help me with the proof?







linear-algebra logarithms trace






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Nov 25 '18 at 22:35









user1936752user1936752

5201513




5201513








  • 1




    $begingroup$
    Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
    $endgroup$
    – darij grinberg
    Nov 25 '18 at 22:50














  • 1




    $begingroup$
    Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
    $endgroup$
    – darij grinberg
    Nov 25 '18 at 22:50








1




1




$begingroup$
Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
$endgroup$
– darij grinberg
Nov 25 '18 at 22:50




$begingroup$
Assuming that convergence and many-valuedness are non-issues: If you apply the Baker-Campbell-Hausdorff formula to $X = log A$ and $Y = log B$, you get $logleft(ABright) = log A + log B + left(text{a sum of commutators}right)$. Now, take traces on both sides. That said, I don't know why we can assume that convergence and many-valuedness are non-issues.
$endgroup$
– darij grinberg
Nov 25 '18 at 22:50










1 Answer
1






active

oldest

votes


















2












$begingroup$

The spectra of $A,B,AB$ are included in $(0,+infty)$. Then, it suffices to prove the equality of the exponentials of RHS and LHS. Since there are no $leq 0$ eigenvalues, we use the principal logarithm.



Let $U$ be symmetric $>0$.



$exp(tr(log(U))=det(exp(log(U)))=det(U)$.



We conclude with $det(AB)=det(A)det(B)$.



Remark. i) If ,$A,B$ are symmetric but with $<0$ eigenvalues, then we cannot use the principal logarithm.



ii) if $A,B$ are invertible but not symmetric and have no $<0$ eigenvalues, then $AB$ may have $<0$ eigenvalues



example $A=B=begin{pmatrix}0&1\-1&0end{pmatrix}$.






share|cite|improve this answer









$endgroup$













    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3013512%2fproof-of-log-identity-for-positive-definite-matrices%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    2












    $begingroup$

    The spectra of $A,B,AB$ are included in $(0,+infty)$. Then, it suffices to prove the equality of the exponentials of RHS and LHS. Since there are no $leq 0$ eigenvalues, we use the principal logarithm.



    Let $U$ be symmetric $>0$.



    $exp(tr(log(U))=det(exp(log(U)))=det(U)$.



    We conclude with $det(AB)=det(A)det(B)$.



    Remark. i) If ,$A,B$ are symmetric but with $<0$ eigenvalues, then we cannot use the principal logarithm.



    ii) if $A,B$ are invertible but not symmetric and have no $<0$ eigenvalues, then $AB$ may have $<0$ eigenvalues



    example $A=B=begin{pmatrix}0&1\-1&0end{pmatrix}$.






    share|cite|improve this answer









    $endgroup$


















      2












      $begingroup$

      The spectra of $A,B,AB$ are included in $(0,+infty)$. Then, it suffices to prove the equality of the exponentials of RHS and LHS. Since there are no $leq 0$ eigenvalues, we use the principal logarithm.



      Let $U$ be symmetric $>0$.



      $exp(tr(log(U))=det(exp(log(U)))=det(U)$.



      We conclude with $det(AB)=det(A)det(B)$.



      Remark. i) If ,$A,B$ are symmetric but with $<0$ eigenvalues, then we cannot use the principal logarithm.



      ii) if $A,B$ are invertible but not symmetric and have no $<0$ eigenvalues, then $AB$ may have $<0$ eigenvalues



      example $A=B=begin{pmatrix}0&1\-1&0end{pmatrix}$.






      share|cite|improve this answer









      $endgroup$
















        2












        2








        2





        $begingroup$

        The spectra of $A,B,AB$ are included in $(0,+infty)$. Then, it suffices to prove the equality of the exponentials of RHS and LHS. Since there are no $leq 0$ eigenvalues, we use the principal logarithm.



        Let $U$ be symmetric $>0$.



        $exp(tr(log(U))=det(exp(log(U)))=det(U)$.



        We conclude with $det(AB)=det(A)det(B)$.



        Remark. i) If ,$A,B$ are symmetric but with $<0$ eigenvalues, then we cannot use the principal logarithm.



        ii) if $A,B$ are invertible but not symmetric and have no $<0$ eigenvalues, then $AB$ may have $<0$ eigenvalues



        example $A=B=begin{pmatrix}0&1\-1&0end{pmatrix}$.






        share|cite|improve this answer









        $endgroup$



        The spectra of $A,B,AB$ are included in $(0,+infty)$. Then, it suffices to prove the equality of the exponentials of RHS and LHS. Since there are no $leq 0$ eigenvalues, we use the principal logarithm.



        Let $U$ be symmetric $>0$.



        $exp(tr(log(U))=det(exp(log(U)))=det(U)$.



        We conclude with $det(AB)=det(A)det(B)$.



        Remark. i) If ,$A,B$ are symmetric but with $<0$ eigenvalues, then we cannot use the principal logarithm.



        ii) if $A,B$ are invertible but not symmetric and have no $<0$ eigenvalues, then $AB$ may have $<0$ eigenvalues



        example $A=B=begin{pmatrix}0&1\-1&0end{pmatrix}$.







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered Nov 26 '18 at 10:57









        loup blancloup blanc

        22.7k21850




        22.7k21850






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3013512%2fproof-of-log-identity-for-positive-definite-matrices%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Biblatex bibliography style without URLs when DOI exists (in Overleaf with Zotero bibliography)

            ComboBox Display Member on multiple fields

            Is it possible to collect Nectar points via Trainline?