$X,Y$ are independent iff conditional regular distribution of $X|Y$ is almost surely the same as the...












0












$begingroup$


I want to prove that $X,Y$ scalar random variables are independent iff conditional regular distribution of $X|Y$ is almost surely the same as the distribution of $X$. A simple result to prove for the... "regular" conditional distribution.



The definition employed for the regular distribution is the following.
Let $(Omega,mathcal F,P)$ be a probability space, and $(Omega',mathcal F')$ a measurable space. Now $mu:Omegatimesmathcal F'rightarrow mathbb R$ is the regular conditional distribution if:



1) For every $Ain mathcal F'$, $mu(.,A)=E(I_{Xin A}|mathcal G)$ almost surely. $mathcal Gsubseteq mathcal F$, and in this case $mathcal G=sigma (Y)$, which is the pre-image of the borel sigma algebra.



2) For almost every $omegain Omega$, $mu(omega,.)$ is a probability measure.



This definition I find very hard to use in order to prove almost anything.










share|cite|improve this question











$endgroup$

















    0












    $begingroup$


    I want to prove that $X,Y$ scalar random variables are independent iff conditional regular distribution of $X|Y$ is almost surely the same as the distribution of $X$. A simple result to prove for the... "regular" conditional distribution.



    The definition employed for the regular distribution is the following.
    Let $(Omega,mathcal F,P)$ be a probability space, and $(Omega',mathcal F')$ a measurable space. Now $mu:Omegatimesmathcal F'rightarrow mathbb R$ is the regular conditional distribution if:



    1) For every $Ain mathcal F'$, $mu(.,A)=E(I_{Xin A}|mathcal G)$ almost surely. $mathcal Gsubseteq mathcal F$, and in this case $mathcal G=sigma (Y)$, which is the pre-image of the borel sigma algebra.



    2) For almost every $omegain Omega$, $mu(omega,.)$ is a probability measure.



    This definition I find very hard to use in order to prove almost anything.










    share|cite|improve this question











    $endgroup$















      0












      0








      0





      $begingroup$


      I want to prove that $X,Y$ scalar random variables are independent iff conditional regular distribution of $X|Y$ is almost surely the same as the distribution of $X$. A simple result to prove for the... "regular" conditional distribution.



      The definition employed for the regular distribution is the following.
      Let $(Omega,mathcal F,P)$ be a probability space, and $(Omega',mathcal F')$ a measurable space. Now $mu:Omegatimesmathcal F'rightarrow mathbb R$ is the regular conditional distribution if:



      1) For every $Ain mathcal F'$, $mu(.,A)=E(I_{Xin A}|mathcal G)$ almost surely. $mathcal Gsubseteq mathcal F$, and in this case $mathcal G=sigma (Y)$, which is the pre-image of the borel sigma algebra.



      2) For almost every $omegain Omega$, $mu(omega,.)$ is a probability measure.



      This definition I find very hard to use in order to prove almost anything.










      share|cite|improve this question











      $endgroup$




      I want to prove that $X,Y$ scalar random variables are independent iff conditional regular distribution of $X|Y$ is almost surely the same as the distribution of $X$. A simple result to prove for the... "regular" conditional distribution.



      The definition employed for the regular distribution is the following.
      Let $(Omega,mathcal F,P)$ be a probability space, and $(Omega',mathcal F')$ a measurable space. Now $mu:Omegatimesmathcal F'rightarrow mathbb R$ is the regular conditional distribution if:



      1) For every $Ain mathcal F'$, $mu(.,A)=E(I_{Xin A}|mathcal G)$ almost surely. $mathcal Gsubseteq mathcal F$, and in this case $mathcal G=sigma (Y)$, which is the pre-image of the borel sigma algebra.



      2) For almost every $omegain Omega$, $mu(omega,.)$ is a probability measure.



      This definition I find very hard to use in order to prove almost anything.







      measure-theory






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Nov 25 '18 at 22:57







      Dole

















      asked Nov 25 '18 at 22:30









      DoleDole

      899514




      899514






















          1 Answer
          1






          active

          oldest

          votes


















          1












          $begingroup$

          Take $A=(-infty , x]$. If $X$ and $Y$ are independent then $mu(omega, A)=E(I_{(X in A)} |Y)=P(X^{-1}(A))=P{Xleq x}$ so the conditional distribution coincides with teh distribution of $X$. Conversely if this condition holds integrate both sides over ${Yleq y}$ to get $P{Xleq x,Yleq y} =P{Xleq x}P{Yleq y}$.






          share|cite|improve this answer









          $endgroup$













            Your Answer





            StackExchange.ifUsing("editor", function () {
            return StackExchange.using("mathjaxEditing", function () {
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            });
            });
            }, "mathjax-editing");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "69"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3013505%2fx-y-are-independent-iff-conditional-regular-distribution-of-xy-is-almost-su%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            1












            $begingroup$

            Take $A=(-infty , x]$. If $X$ and $Y$ are independent then $mu(omega, A)=E(I_{(X in A)} |Y)=P(X^{-1}(A))=P{Xleq x}$ so the conditional distribution coincides with teh distribution of $X$. Conversely if this condition holds integrate both sides over ${Yleq y}$ to get $P{Xleq x,Yleq y} =P{Xleq x}P{Yleq y}$.






            share|cite|improve this answer









            $endgroup$


















              1












              $begingroup$

              Take $A=(-infty , x]$. If $X$ and $Y$ are independent then $mu(omega, A)=E(I_{(X in A)} |Y)=P(X^{-1}(A))=P{Xleq x}$ so the conditional distribution coincides with teh distribution of $X$. Conversely if this condition holds integrate both sides over ${Yleq y}$ to get $P{Xleq x,Yleq y} =P{Xleq x}P{Yleq y}$.






              share|cite|improve this answer









              $endgroup$
















                1












                1








                1





                $begingroup$

                Take $A=(-infty , x]$. If $X$ and $Y$ are independent then $mu(omega, A)=E(I_{(X in A)} |Y)=P(X^{-1}(A))=P{Xleq x}$ so the conditional distribution coincides with teh distribution of $X$. Conversely if this condition holds integrate both sides over ${Yleq y}$ to get $P{Xleq x,Yleq y} =P{Xleq x}P{Yleq y}$.






                share|cite|improve this answer









                $endgroup$



                Take $A=(-infty , x]$. If $X$ and $Y$ are independent then $mu(omega, A)=E(I_{(X in A)} |Y)=P(X^{-1}(A))=P{Xleq x}$ so the conditional distribution coincides with teh distribution of $X$. Conversely if this condition holds integrate both sides over ${Yleq y}$ to get $P{Xleq x,Yleq y} =P{Xleq x}P{Yleq y}$.







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered Nov 25 '18 at 23:43









                Kavi Rama MurthyKavi Rama Murthy

                54.4k32055




                54.4k32055






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Mathematics Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3013505%2fx-y-are-independent-iff-conditional-regular-distribution-of-xy-is-almost-su%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Biblatex bibliography style without URLs when DOI exists (in Overleaf with Zotero bibliography)

                    ComboBox Display Member on multiple fields

                    Is it possible to collect Nectar points via Trainline?