Question about point mass prior and continuous distribution
Suppose we have a point mass prior,
$$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
Gamma(c,c), & prob=frac{1}{2} end{cases}$$
Then if we are asked
$lim_{c to infty} P(theta=1)$
Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.
To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$
However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$
But, by Markov, for $X sim Gamma(c,c)$
$lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$
So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$
As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.
Thanks all
bayesian continuous-data
add a comment |
Suppose we have a point mass prior,
$$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
Gamma(c,c), & prob=frac{1}{2} end{cases}$$
Then if we are asked
$lim_{c to infty} P(theta=1)$
Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.
To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$
However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$
But, by Markov, for $X sim Gamma(c,c)$
$lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$
So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$
As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.
Thanks all
bayesian continuous-data
add a comment |
Suppose we have a point mass prior,
$$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
Gamma(c,c), & prob=frac{1}{2} end{cases}$$
Then if we are asked
$lim_{c to infty} P(theta=1)$
Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.
To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$
However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$
But, by Markov, for $X sim Gamma(c,c)$
$lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$
So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$
As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.
Thanks all
bayesian continuous-data
Suppose we have a point mass prior,
$$theta sim begin{cases} I(theta=1) ,& prob=frac{1}{2} \
Gamma(c,c), & prob=frac{1}{2} end{cases}$$
Then if we are asked
$lim_{c to infty} P(theta=1)$
Now here is the issue, since Gamma is a continuous distribution, it seems that in the case of gamma,we will never have $theta=1$.
To me it thus seems that regardless of the value of c, the $p(theta=1)=frac{1}{2}$
However, we also have that since expected value of a $gamma(a,b)=frac{a}{b}$ so that the expected value of the gamma is 1 when we have $a=b=c$
But, by Markov, for $X sim Gamma(c,c)$
$lim_{c to infty} Pr(|X-mu| lt epsilon) to 1$ for any $epsilon gt 0$
So is $lim_{c to infty}P(theta=1) =1$ , or is $lim_{c to infty}P(theta=1)=frac{1}{2}$
As even though the markov inequality holds, it is a continous distirbution, so we will never have it exactly equal to 1.
Thanks all
bayesian continuous-data
bayesian continuous-data
edited Nov 30 at 3:15
asked Nov 30 at 2:55
Learning
729
729
add a comment |
add a comment |
2 Answers
2
active
oldest
votes
Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:
$$begin{equation} begin{aligned}
mathbb{P}(theta = 1|c)
&= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
&= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
&= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
&= frac{1}{2}. \[6pt]
end{aligned} end{equation}$$
(Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:
$$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$
Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)
add a comment |
Disclaimer The above answer is completely fine if the question
is about the prior probability that $θ=1$. While there is no
indication whatsoever in the text of the question to the alternative, namely
that the question is about the posterior probability that $θ=1$, a recent
exchange with students about a very similar question in my book, The
Bayesian Choice (Exercise 5.9), leads me to propose the extension
to the posterior setting.
In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
begin{align*}
mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
&=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
&=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
&=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
&=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
end{align*}
which converges to
$$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
when $c$ goes to infinity.
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "65"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f379550%2fquestion-about-point-mass-prior-and-continuous-distribution%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:
$$begin{equation} begin{aligned}
mathbb{P}(theta = 1|c)
&= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
&= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
&= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
&= frac{1}{2}. \[6pt]
end{aligned} end{equation}$$
(Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:
$$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$
Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)
add a comment |
Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:
$$begin{equation} begin{aligned}
mathbb{P}(theta = 1|c)
&= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
&= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
&= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
&= frac{1}{2}. \[6pt]
end{aligned} end{equation}$$
(Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:
$$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$
Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)
add a comment |
Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:
$$begin{equation} begin{aligned}
mathbb{P}(theta = 1|c)
&= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
&= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
&= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
&= frac{1}{2}. \[6pt]
end{aligned} end{equation}$$
(Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:
$$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$
Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)
Your specified prior distribution is a mixture of a continuous and discrete part. Let $I sim text{Bern}(1/2)$ be the indicator that the parameter is taken from the gamma distribution. Then using the law of total probability you have:
$$begin{equation} begin{aligned}
mathbb{P}(theta = 1|c)
&= mathbb{P}(theta = 1|c, I=0) cdot mathbb{P}(I=0) + mathbb{P}(theta = 1|c, I=1) cdot mathbb{P}(I=1) \[6pt]
&= frac{1}{2} cdot mathbb{P}(theta = 1|c, I=0) + frac{1}{2} cdot mathbb{P}(theta = 1|c, I=1) \[6pt]
&= frac{1}{2} + frac{1}{2} cdot mathbb{P}(theta = 1| theta sim text{Ga}(c,c)) \[6pt]
&= frac{1}{2}. \[6pt]
end{aligned} end{equation}$$
(Note that the last step comes from recognising that the gamma is a continuous distribution, so the probability of a specific point is zero under this distribution.) This result holds for all values of $c$, so you are correct that it also holds in the limit:
$$lim_{c rightarrow infty} mathbb{P}(theta = 1|c) = lim_{c rightarrow infty} frac{1}{2} = frac{1}{2}.$$
Your later use of Chebychev's inequality shows that as $c rightarrow infty$ you get $theta rightarrow 1$ (convergence in probability), but this does not change the fact that $mathbb{P}(theta = 1|c) = 1/2$ for all $c > 0$. (To understand the reason for this more fully, have a read about the distinction between convergence in probability and almost-sure convergence.)
answered Nov 30 at 3:31
Ben
21.7k224103
21.7k224103
add a comment |
add a comment |
Disclaimer The above answer is completely fine if the question
is about the prior probability that $θ=1$. While there is no
indication whatsoever in the text of the question to the alternative, namely
that the question is about the posterior probability that $θ=1$, a recent
exchange with students about a very similar question in my book, The
Bayesian Choice (Exercise 5.9), leads me to propose the extension
to the posterior setting.
In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
begin{align*}
mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
&=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
&=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
&=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
&=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
end{align*}
which converges to
$$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
when $c$ goes to infinity.
add a comment |
Disclaimer The above answer is completely fine if the question
is about the prior probability that $θ=1$. While there is no
indication whatsoever in the text of the question to the alternative, namely
that the question is about the posterior probability that $θ=1$, a recent
exchange with students about a very similar question in my book, The
Bayesian Choice (Exercise 5.9), leads me to propose the extension
to the posterior setting.
In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
begin{align*}
mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
&=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
&=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
&=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
&=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
end{align*}
which converges to
$$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
when $c$ goes to infinity.
add a comment |
Disclaimer The above answer is completely fine if the question
is about the prior probability that $θ=1$. While there is no
indication whatsoever in the text of the question to the alternative, namely
that the question is about the posterior probability that $θ=1$, a recent
exchange with students about a very similar question in my book, The
Bayesian Choice (Exercise 5.9), leads me to propose the extension
to the posterior setting.
In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
begin{align*}
mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
&=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
&=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
&=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
&=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
end{align*}
which converges to
$$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
when $c$ goes to infinity.
Disclaimer The above answer is completely fine if the question
is about the prior probability that $θ=1$. While there is no
indication whatsoever in the text of the question to the alternative, namely
that the question is about the posterior probability that $θ=1$, a recent
exchange with students about a very similar question in my book, The
Bayesian Choice (Exercise 5.9), leads me to propose the extension
to the posterior setting.
In this posterior case, when $Xsim mathcal{E}(theta)$ [for instance] the probability that $θ=1$ is
begin{align*}
mathbb{P}(theta=1|x)&= frac{e^{-x}}{e^{-x}+int_0^infty theta e^{-theta x} pi_c(theta)text{d}theta}\
&=frac{e^{-x}}{e^{-x}+ int_0^infty theta e^{-theta x} c^c theta^{c-1}e^{-c theta} Gamma(c)^{-1}text{d}theta}\
&=frac{e^{-x}}{e^{-x}+c^c (c+x)^{-c-1} Gamma(c+1)Gamma(c)^{-1}}\
&=frac{e^{-x}}{e^{-x}+c^{c+1} (c+x)^{-c-1}}\
&=frac{e^{-x}}{e^{-x}+ (1+x/c)^{-c-1}}\
end{align*}
which converges to
$$frac{e^{-x}}{e^{-x}+e^{-x}}=frac{1}{2}$$
when $c$ goes to infinity.
answered Dec 15 at 12:54
Xi'an
53.2k689345
53.2k689345
add a comment |
add a comment |
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f379550%2fquestion-about-point-mass-prior-and-continuous-distribution%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown