Question about point mass prior and continuous distribution












3














Suppose we have a point mass prior,



$$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
Gamma(c,c), & prob=frac{1}{2} end{cases}$$



Then if we are asked



$lim_{c to infty} P(theta=1)$



Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.



To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$



However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$



But, by Markov, for $X sim Gamma(c,c)$



$lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$



So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$



As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.



Thanks all










share|cite|improve this question





























    3














    Suppose we have a point mass prior,



    $$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
    Gamma(c,c), & prob=frac{1}{2} end{cases}$$



    Then if we are asked



    $lim_{c to infty} P(theta=1)$



    Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.



    To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$



    However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$



    But, by Markov, for $X sim Gamma(c,c)$



    $lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$



    So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$



    As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.



    Thanks all










    share|cite|improve this question



























      3












      3








      3


      1





      Suppose we have a point mass prior,



      $$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
      Gamma(c,c), & prob=frac{1}{2} end{cases}$$



      Then if we are asked



      $lim_{c to infty} P(theta=1)$



      Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.



      To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$



      However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$



      But, by Markov, for $X sim Gamma(c,c)$



      $lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$



      So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$



      As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.



      Thanks all










      share|cite|improve this question















      Suppose we have a point mass prior,



      $$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
      Gamma(c,c), & prob=frac{1}{2} end{cases}$$



      Then if we are asked



      $lim_{c to infty} P(theta=1)$



      Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.



      To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$



      However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$



      But, by Markov, for $X sim Gamma(c,c)$



      $lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$



      So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$



      As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.



      Thanks all







      bayesian continuous-data






      share|cite|improve this question















      share|cite|improve this question













      share|cite|improve this question




      share|cite|improve this question








      edited Nov 30 at 3:15

























      asked Nov 30 at 2:55









      Learning

      729




      729






















          2 Answers
          2






          active

          oldest

          votes


















          5














          Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:



          $$begin{equation} begin{aligned}
          mathbb{P}(theta = 1|c)
          &= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
          &= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
          &= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
          &= frac{1}{2}. \[6pt]
          end{aligned} end{equation}$$



          (Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:



          $$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$



          Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)






          share|cite|improve this answer





























            2















            Disclaimer The above answer is completely fine if the question
            is about the prior probability that $θ=1$. While there is no
            indication whatsoever in the text of the question to the alternative, namely
            that the question is about the posterior probability that $θ=1$, a recent
            exchange with students about a very similar question in my book, The
            Bayesian Choice (Exercise 5.9), leads me to propose the extension
            to the posterior setting.




            In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
            begin{align*}
            mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
            &=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
            &=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
            &=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
            &=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
            end{align*}

            which converges to
            $$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
            when $c$ goes to infinity.






            share|cite|improve this answer





















              Your Answer





              StackExchange.ifUsing("editor", function () {
              return StackExchange.using("mathjaxEditing", function () {
              StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
              StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
              });
              });
              }, "mathjax-editing");

              StackExchange.ready(function() {
              var channelOptions = {
              tags: "".split(" "),
              id: "65"
              };
              initTagRenderer("".split(" "), "".split(" "), channelOptions);

              StackExchange.using("externalEditor", function() {
              // Have to fire editor after snippets, if snippets enabled
              if (StackExchange.settings.snippets.snippetsEnabled) {
              StackExchange.using("snippets", function() {
              createEditor();
              });
              }
              else {
              createEditor();
              }
              });

              function createEditor() {
              StackExchange.prepareEditor({
              heartbeatType: 'answer',
              autoActivateHeartbeat: false,
              convertImagesToLinks: false,
              noModals: true,
              showLowRepImageUploadWarning: true,
              reputationToPostImages: null,
              bindNavPrevention: true,
              postfix: "",
              imageUploader: {
              brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
              contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
              allowUrls: true
              },
              onDemand: true,
              discardSelector: ".discard-answer"
              ,immediatelyShowMarkdownHelp:true
              });


              }
              });














              draft saved

              draft discarded


















              StackExchange.ready(
              function () {
              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f379550%2fquestion-about-point-mass-prior-and-continuous-distribution%23new-answer', 'question_page');
              }
              );

              Post as a guest















              Required, but never shown

























              2 Answers
              2






              active

              oldest

              votes








              2 Answers
              2






              active

              oldest

              votes









              active

              oldest

              votes






              active

              oldest

              votes









              5














              Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:



              $$begin{equation} begin{aligned}
              mathbb{P}(theta = 1|c)
              &= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
              &= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
              &= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
              &= frac{1}{2}. \[6pt]
              end{aligned} end{equation}$$



              (Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:



              $$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$



              Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)






              share|cite|improve this answer


























                5














                Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:



                $$begin{equation} begin{aligned}
                mathbb{P}(theta = 1|c)
                &= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
                &= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
                &= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
                &= frac{1}{2}. \[6pt]
                end{aligned} end{equation}$$



                (Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:



                $$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$



                Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)






                share|cite|improve this answer
























                  5












                  5








                  5






                  Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:



                  $$begin{equation} begin{aligned}
                  mathbb{P}(theta = 1|c)
                  &= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
                  &= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
                  &= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
                  &= frac{1}{2}. \[6pt]
                  end{aligned} end{equation}$$



                  (Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:



                  $$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$



                  Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)






                  share|cite|improve this answer












                  Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:



                  $$begin{equation} begin{aligned}
                  mathbb{P}(theta = 1|c)
                  &= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
                  &= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
                  &= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
                  &= frac{1}{2}. \[6pt]
                  end{aligned} end{equation}$$



                  (Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:



                  $$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$



                  Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)







                  share|cite|improve this answer












                  share|cite|improve this answer



                  share|cite|improve this answer










                  answered Nov 30 at 3:31









                  Ben

                  21.7k224103




                  21.7k224103

























                      2















                      Disclaimer The above answer is completely fine if the question
                      is about the prior probability that $θ=1$. While there is no
                      indication whatsoever in the text of the question to the alternative, namely
                      that the question is about the posterior probability that $θ=1$, a recent
                      exchange with students about a very similar question in my book, The
                      Bayesian Choice (Exercise 5.9), leads me to propose the extension
                      to the posterior setting.




                      In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
                      begin{align*}
                      mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
                      &=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
                      &=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
                      &=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
                      &=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
                      end{align*}

                      which converges to
                      $$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
                      when $c$ goes to infinity.






                      share|cite|improve this answer


























                        2















                        Disclaimer The above answer is completely fine if the question
                        is about the prior probability that $θ=1$. While there is no
                        indication whatsoever in the text of the question to the alternative, namely
                        that the question is about the posterior probability that $θ=1$, a recent
                        exchange with students about a very similar question in my book, The
                        Bayesian Choice (Exercise 5.9), leads me to propose the extension
                        to the posterior setting.




                        In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
                        begin{align*}
                        mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
                        &=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
                        &=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
                        &=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
                        &=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
                        end{align*}

                        which converges to
                        $$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
                        when $c$ goes to infinity.






                        share|cite|improve this answer
























                          2












                          2








                          2







                          Disclaimer The above answer is completely fine if the question
                          is about the prior probability that $θ=1$. While there is no
                          indication whatsoever in the text of the question to the alternative, namely
                          that the question is about the posterior probability that $θ=1$, a recent
                          exchange with students about a very similar question in my book, The
                          Bayesian Choice (Exercise 5.9), leads me to propose the extension
                          to the posterior setting.




                          In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
                          begin{align*}
                          mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
                          &=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
                          &=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
                          &=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
                          &=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
                          end{align*}

                          which converges to
                          $$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
                          when $c$ goes to infinity.






                          share|cite|improve this answer













                          Disclaimer The above answer is completely fine if the question
                          is about the prior probability that $θ=1$. While there is no
                          indication whatsoever in the text of the question to the alternative, namely
                          that the question is about the posterior probability that $θ=1$, a recent
                          exchange with students about a very similar question in my book, The
                          Bayesian Choice (Exercise 5.9), leads me to propose the extension
                          to the posterior setting.




                          In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
                          begin{align*}
                          mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
                          &=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
                          &=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
                          &=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
                          &=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
                          end{align*}

                          which converges to
                          $$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
                          when $c$ goes to infinity.







                          share|cite|improve this answer












                          share|cite|improve this answer



                          share|cite|improve this answer










                          answered Dec 15 at 12:54









                          Xi'an

                          53.2k689345




                          53.2k689345






























                              draft saved

                              draft discarded




















































                              Thanks for contributing an answer to Cross Validated!


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              Use MathJax to format equations. MathJax reference.


                              To learn more, see our tips on writing great answers.





                              Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


                              Please pay close attention to the following guidance:


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              To learn more, see our tips on writing great answers.




                              draft saved


                              draft discarded














                              StackExchange.ready(
                              function () {
                              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f379550%2fquestion-about-point-mass-prior-and-continuous-distribution%23new-answer', 'question_page');
                              }
                              );

                              Post as a guest















                              Required, but never shown





















































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown

































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown







                              Popular posts from this blog

                              Biblatex bibliography style without URLs when DOI exists (in Overleaf with Zotero bibliography)

                              ComboBox Display Member on multiple fields

                              Is it possible to collect Nectar points via Trainline?