Binomial Distribution and the Law of Iterated ExpectationBinomial distribution and expectationBounds for the ratio of probabilities of Poisson-Binomial DistributionSum of Product of Bernoulli Random VariablesCompute Variance of Poisson Binomal-esque variableShow that this expression follows a Binomial distributionBinom(n, p): estimate the probability that the number of successes is near the mode $np$Linearity Of Expectations in Bernoulli's trialsNumber of trials follows the binomial distributionExpected number of Failures within K trials for Binomial RVCalculation of expected value.

Why is this estimator biased?

What are the advantages of simplicial model categories over non-simplicial ones?

Mimic lecturing on blackboard, facing audience

Pre-mixing cryogenic fuels and using only one fuel tank

What should you do when eye contact makes your subordinate uncomfortable?

How do you make your own symbol when Detexify fails?

What if a revenant (monster) gains fire resistance?

Can a College of Swords bard use a Blade Flourish option on an opportunity attack provoked by their own Dissonant Whispers spell?

Picking the different solutions to the time independent Schrodinger eqaution

Does malloc reserve more space while allocating memory?

Multiplicative persistence

Keeping a ball lost forever

putting logo on same line but after title, latex

Unexpected behavior of the procedure `Area` on the object 'Polygon'

Why is it that I can sometimes guess the next note?

Quoting Keynes in a lecture

Why does a simple loop result in ASYNC_NETWORK_IO waits?

The IT department bottlenecks progress. How should I handle this?

Is aluminum electrical wire used on aircraft?

Why is so much work done on numerical verification of the Riemann Hypothesis?

Does IPv6 have similar concept of network mask?

Does Doodling or Improvising on the Piano Have Any Benefits?

Extract more than nine arguments that occur periodically in a sentence to use in macros in order to typset

Creepy dinosaur pc game identification



Binomial Distribution and the Law of Iterated Expectation


Binomial distribution and expectationBounds for the ratio of probabilities of Poisson-Binomial DistributionSum of Product of Bernoulli Random VariablesCompute Variance of Poisson Binomal-esque variableShow that this expression follows a Binomial distributionBinom(n, p): estimate the probability that the number of successes is near the mode $np$Linearity Of Expectations in Bernoulli's trialsNumber of trials follows the binomial distributionExpected number of Failures within K trials for Binomial RVCalculation of expected value.













1












$begingroup$


Suppose $H_i$ are random variable following Binomial($N_i, alpha$).



Conditional on $N_i$, $H_i$ is independent from past and follows Binomial($N_i, alpha$).



$N_i+1=N_i+H_i$ and $N_1=n$ (i.e., $N_1$ is not a random variable.)



I wonder whether the following holds:



$Ebig[fracH_1+H_2+...+H_tN_1+N_2+...+N_tbig]=Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_tbig]$ where $a_j$ are i.i.d. and follow Bernoulli distribution with success probability of $alpha$.



By the law of iterated expectations



$Ebig[Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_t|N_1+N_2+...+N_tbig]big]=Ebig[fracalpha(N_1+N_2+...+N_t)N_1+N_2+...+N_tbig]=alpha$










share|cite|improve this question











$endgroup$











  • $begingroup$
    You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 3:49











  • $begingroup$
    Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
    $endgroup$
    – user3509199
    Mar 15 at 3:54










  • $begingroup$
    Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 4:00











  • $begingroup$
    Great! Thank you.
    $endgroup$
    – user3509199
    Mar 15 at 4:01










  • $begingroup$
    Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
    $endgroup$
    – Michael
    Mar 15 at 4:15
















1












$begingroup$


Suppose $H_i$ are random variable following Binomial($N_i, alpha$).



Conditional on $N_i$, $H_i$ is independent from past and follows Binomial($N_i, alpha$).



$N_i+1=N_i+H_i$ and $N_1=n$ (i.e., $N_1$ is not a random variable.)



I wonder whether the following holds:



$Ebig[fracH_1+H_2+...+H_tN_1+N_2+...+N_tbig]=Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_tbig]$ where $a_j$ are i.i.d. and follow Bernoulli distribution with success probability of $alpha$.



By the law of iterated expectations



$Ebig[Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_t|N_1+N_2+...+N_tbig]big]=Ebig[fracalpha(N_1+N_2+...+N_t)N_1+N_2+...+N_tbig]=alpha$










share|cite|improve this question











$endgroup$











  • $begingroup$
    You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 3:49











  • $begingroup$
    Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
    $endgroup$
    – user3509199
    Mar 15 at 3:54










  • $begingroup$
    Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 4:00











  • $begingroup$
    Great! Thank you.
    $endgroup$
    – user3509199
    Mar 15 at 4:01










  • $begingroup$
    Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
    $endgroup$
    – Michael
    Mar 15 at 4:15














1












1








1





$begingroup$


Suppose $H_i$ are random variable following Binomial($N_i, alpha$).



Conditional on $N_i$, $H_i$ is independent from past and follows Binomial($N_i, alpha$).



$N_i+1=N_i+H_i$ and $N_1=n$ (i.e., $N_1$ is not a random variable.)



I wonder whether the following holds:



$Ebig[fracH_1+H_2+...+H_tN_1+N_2+...+N_tbig]=Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_tbig]$ where $a_j$ are i.i.d. and follow Bernoulli distribution with success probability of $alpha$.



By the law of iterated expectations



$Ebig[Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_t|N_1+N_2+...+N_tbig]big]=Ebig[fracalpha(N_1+N_2+...+N_t)N_1+N_2+...+N_tbig]=alpha$










share|cite|improve this question











$endgroup$




Suppose $H_i$ are random variable following Binomial($N_i, alpha$).



Conditional on $N_i$, $H_i$ is independent from past and follows Binomial($N_i, alpha$).



$N_i+1=N_i+H_i$ and $N_1=n$ (i.e., $N_1$ is not a random variable.)



I wonder whether the following holds:



$Ebig[fracH_1+H_2+...+H_tN_1+N_2+...+N_tbig]=Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_tbig]$ where $a_j$ are i.i.d. and follow Bernoulli distribution with success probability of $alpha$.



By the law of iterated expectations



$Ebig[Ebig[fracsum_j=1^N_1a_j+sum_j=1^N_2a_j+...+sum_j=1^N_ta_jN_1+N_2+...+N_t|N_1+N_2+...+N_tbig]big]=Ebig[fracalpha(N_1+N_2+...+N_t)N_1+N_2+...+N_tbig]=alpha$







binomial-distribution expected-value






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Mar 15 at 12:59







user3509199

















asked Mar 15 at 2:49









user3509199user3509199

416




416











  • $begingroup$
    You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 3:49











  • $begingroup$
    Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
    $endgroup$
    – user3509199
    Mar 15 at 3:54










  • $begingroup$
    Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 4:00











  • $begingroup$
    Great! Thank you.
    $endgroup$
    – user3509199
    Mar 15 at 4:01










  • $begingroup$
    Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
    $endgroup$
    – Michael
    Mar 15 at 4:15

















  • $begingroup$
    You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 3:49











  • $begingroup$
    Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
    $endgroup$
    – user3509199
    Mar 15 at 3:54










  • $begingroup$
    Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
    $endgroup$
    – Michael
    Mar 15 at 4:00











  • $begingroup$
    Great! Thank you.
    $endgroup$
    – user3509199
    Mar 15 at 4:01










  • $begingroup$
    Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
    $endgroup$
    – Michael
    Mar 15 at 4:15
















$begingroup$
You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
$endgroup$
– Michael
Mar 15 at 3:49





$begingroup$
You say the $H_i$ are i.i.d., but from their description they seem neither independent nor identically distributed.
$endgroup$
– Michael
Mar 15 at 3:49













$begingroup$
Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
$endgroup$
– user3509199
Mar 15 at 3:54




$begingroup$
Thanks for your comment Michael, I modified the statement a little bit. By i.i.d., I mean that $H_i$ is conditionally independent of the past, given $N_i$. Thanks!
$endgroup$
– user3509199
Mar 15 at 3:54












$begingroup$
Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
$endgroup$
– Michael
Mar 15 at 4:00





$begingroup$
Your calculation is good if you assume $H_i$ is conditionally independent of the past, given $N_i$. But you should not say the $H_i$ variables are "independent and identically distributed (i.i.d.)." The $H_2$ value depends on the value of $H_1$, they are not independent. Neither are they identically distributed.
$endgroup$
– Michael
Mar 15 at 4:00













$begingroup$
Great! Thank you.
$endgroup$
– user3509199
Mar 15 at 4:01




$begingroup$
Great! Thank you.
$endgroup$
– user3509199
Mar 15 at 4:01












$begingroup$
Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
$endgroup$
– Michael
Mar 15 at 4:15





$begingroup$
Sorry, I got distracted with the conditinonings. Your denominators include information about the future, not just the past. So, how do you claim $E[fracH_2N_1+N_2+N_3|N_2]=E[fracalpha N_2N_1 + N_2 + N_3|N_2]$. It looks like $N_3$ depends on $H_2$. So I should not have said "your calculation is good."
$endgroup$
– Michael
Mar 15 at 4:15











1 Answer
1






active

oldest

votes


















0












$begingroup$

No. You can consider the simple case $n=1$, $t=2$. In this case we have $N_1=1$; $H_1$ is a Bernoulli random variable with success probability $alpha$; $N_2 = 1 + H_1$. Assume $0<alpha<1$. Then:



beginalign
Eleft[fracH_1+H_2N_1+N_2right] &=Eleft[fracH_1+H_22+H_1right]\
&=Eleft[Eleft[left(fracH_1+H_22+H_1right)|H_1right]right]\
&=Eleft[fracH_1]2+H_1right]\
&=Eleft[fracH_1+alpha(1+H_1)2+H_1right]\
&=left(frac1+2alpha3right)underbraceP[H_1=1]_alpha + left(fracalpha2right)underbraceP[H_1=0]_1-alpha\
&= fracalpha(alpha+5)6 \
&neq alpha
endalign






share|cite|improve this answer









$endgroup$












  • $begingroup$
    Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
    $endgroup$
    – Michael
    Mar 15 at 16:15










  • $begingroup$
    Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
    $endgroup$
    – user3509199
    Mar 15 at 17:31











  • $begingroup$
    Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
    $endgroup$
    – Michael
    Mar 15 at 19:12











  • $begingroup$
    Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
    $endgroup$
    – user3509199
    Mar 15 at 19:15










  • $begingroup$
    Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
    $endgroup$
    – Michael
    Mar 15 at 19:17











Your Answer





StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");

StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);

else
createEditor();

);

function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);



);













draft saved

draft discarded


















StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3148833%2fbinomial-distribution-and-the-law-of-iterated-expectation%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown

























1 Answer
1






active

oldest

votes








1 Answer
1






active

oldest

votes









active

oldest

votes






active

oldest

votes









0












$begingroup$

No. You can consider the simple case $n=1$, $t=2$. In this case we have $N_1=1$; $H_1$ is a Bernoulli random variable with success probability $alpha$; $N_2 = 1 + H_1$. Assume $0<alpha<1$. Then:



beginalign
Eleft[fracH_1+H_2N_1+N_2right] &=Eleft[fracH_1+H_22+H_1right]\
&=Eleft[Eleft[left(fracH_1+H_22+H_1right)|H_1right]right]\
&=Eleft[fracH_1]2+H_1right]\
&=Eleft[fracH_1+alpha(1+H_1)2+H_1right]\
&=left(frac1+2alpha3right)underbraceP[H_1=1]_alpha + left(fracalpha2right)underbraceP[H_1=0]_1-alpha\
&= fracalpha(alpha+5)6 \
&neq alpha
endalign






share|cite|improve this answer









$endgroup$












  • $begingroup$
    Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
    $endgroup$
    – Michael
    Mar 15 at 16:15










  • $begingroup$
    Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
    $endgroup$
    – user3509199
    Mar 15 at 17:31











  • $begingroup$
    Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
    $endgroup$
    – Michael
    Mar 15 at 19:12











  • $begingroup$
    Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
    $endgroup$
    – user3509199
    Mar 15 at 19:15










  • $begingroup$
    Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
    $endgroup$
    – Michael
    Mar 15 at 19:17
















0












$begingroup$

No. You can consider the simple case $n=1$, $t=2$. In this case we have $N_1=1$; $H_1$ is a Bernoulli random variable with success probability $alpha$; $N_2 = 1 + H_1$. Assume $0<alpha<1$. Then:



beginalign
Eleft[fracH_1+H_2N_1+N_2right] &=Eleft[fracH_1+H_22+H_1right]\
&=Eleft[Eleft[left(fracH_1+H_22+H_1right)|H_1right]right]\
&=Eleft[fracH_1]2+H_1right]\
&=Eleft[fracH_1+alpha(1+H_1)2+H_1right]\
&=left(frac1+2alpha3right)underbraceP[H_1=1]_alpha + left(fracalpha2right)underbraceP[H_1=0]_1-alpha\
&= fracalpha(alpha+5)6 \
&neq alpha
endalign






share|cite|improve this answer









$endgroup$












  • $begingroup$
    Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
    $endgroup$
    – Michael
    Mar 15 at 16:15










  • $begingroup$
    Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
    $endgroup$
    – user3509199
    Mar 15 at 17:31











  • $begingroup$
    Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
    $endgroup$
    – Michael
    Mar 15 at 19:12











  • $begingroup$
    Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
    $endgroup$
    – user3509199
    Mar 15 at 19:15










  • $begingroup$
    Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
    $endgroup$
    – Michael
    Mar 15 at 19:17














0












0








0





$begingroup$

No. You can consider the simple case $n=1$, $t=2$. In this case we have $N_1=1$; $H_1$ is a Bernoulli random variable with success probability $alpha$; $N_2 = 1 + H_1$. Assume $0<alpha<1$. Then:



beginalign
Eleft[fracH_1+H_2N_1+N_2right] &=Eleft[fracH_1+H_22+H_1right]\
&=Eleft[Eleft[left(fracH_1+H_22+H_1right)|H_1right]right]\
&=Eleft[fracH_1]2+H_1right]\
&=Eleft[fracH_1+alpha(1+H_1)2+H_1right]\
&=left(frac1+2alpha3right)underbraceP[H_1=1]_alpha + left(fracalpha2right)underbraceP[H_1=0]_1-alpha\
&= fracalpha(alpha+5)6 \
&neq alpha
endalign






share|cite|improve this answer









$endgroup$



No. You can consider the simple case $n=1$, $t=2$. In this case we have $N_1=1$; $H_1$ is a Bernoulli random variable with success probability $alpha$; $N_2 = 1 + H_1$. Assume $0<alpha<1$. Then:



beginalign
Eleft[fracH_1+H_2N_1+N_2right] &=Eleft[fracH_1+H_22+H_1right]\
&=Eleft[Eleft[left(fracH_1+H_22+H_1right)|H_1right]right]\
&=Eleft[fracH_1]2+H_1right]\
&=Eleft[fracH_1+alpha(1+H_1)2+H_1right]\
&=left(frac1+2alpha3right)underbraceP[H_1=1]_alpha + left(fracalpha2right)underbraceP[H_1=0]_1-alpha\
&= fracalpha(alpha+5)6 \
&neq alpha
endalign







share|cite|improve this answer












share|cite|improve this answer



share|cite|improve this answer










answered Mar 15 at 16:12









MichaelMichael

13.1k11429




13.1k11429











  • $begingroup$
    Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
    $endgroup$
    – Michael
    Mar 15 at 16:15










  • $begingroup$
    Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
    $endgroup$
    – user3509199
    Mar 15 at 17:31











  • $begingroup$
    Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
    $endgroup$
    – Michael
    Mar 15 at 19:12











  • $begingroup$
    Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
    $endgroup$
    – user3509199
    Mar 15 at 19:15










  • $begingroup$
    Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
    $endgroup$
    – Michael
    Mar 15 at 19:17

















  • $begingroup$
    Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
    $endgroup$
    – Michael
    Mar 15 at 16:15










  • $begingroup$
    Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
    $endgroup$
    – user3509199
    Mar 15 at 17:31











  • $begingroup$
    Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
    $endgroup$
    – Michael
    Mar 15 at 19:12











  • $begingroup$
    Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
    $endgroup$
    – user3509199
    Mar 15 at 19:15










  • $begingroup$
    Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
    $endgroup$
    – Michael
    Mar 15 at 19:17
















$begingroup$
Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
$endgroup$
– Michael
Mar 15 at 16:15




$begingroup$
Note that $alpha(alpha+5)/6=alpha$ if and only if $alpha=0$ or $alpha=1$.
$endgroup$
– Michael
Mar 15 at 16:15












$begingroup$
Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
$endgroup$
– user3509199
Mar 15 at 17:31





$begingroup$
Thanks. Could you provide some intuition why the expected ratio is not equal to $alpha$? Intuitively, given the construction of $H_i$, the ratio seems to be $alpha$.
$endgroup$
– user3509199
Mar 15 at 17:31













$begingroup$
Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
$endgroup$
– Michael
Mar 15 at 19:12





$begingroup$
Your method would be correct to compute: $$Eleft[fracH_iN_1+...+N_iright] = Eleft[fracE[H_iN_1+...+N_iright]=Eleft[fracalpha N_iN_1+...+N_iright]$$ However if the denominator has information about the future (after time $i$) things are different $$ Eleft[fracH_iN_1+...+N_tright]=Eleft[fracE[H_iN_1+...+N_tright] = Eleft[fracN_i+1-N_iN_1+...+N_tright]$$
$endgroup$
– Michael
Mar 15 at 19:12













$begingroup$
Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
$endgroup$
– user3509199
Mar 15 at 19:15




$begingroup$
Thanks for the clarification. In the revised question, I used the different conditioning. For example, now, I conditioned on $N_1+...+N_t$.
$endgroup$
– user3509199
Mar 15 at 19:15












$begingroup$
Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
$endgroup$
– Michael
Mar 15 at 19:17





$begingroup$
Yes, but you also jump to the incorrect conclusion $$ E[H_i|N_1+...+N_t] = E[alpha N_i|N_1+...+N_t]$$ Knowledge of the sum $N_1+...+N_t$ gives you some information about how big the $N_i+1$ value might have been, which gives you side information about $H_i$ that skews its conditional distribution in comparison to its conditional distribution given only the value of $N_i$. It would be correct to say $E[H_i|N_i]=alpha N_i$.
$endgroup$
– Michael
Mar 15 at 19:17


















draft saved

draft discarded
















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid


  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.

Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3148833%2fbinomial-distribution-and-the-law-of-iterated-expectation%23new-answer', 'question_page');

);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

How should I support this large drywall patch? Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?How do I cover large gaps in drywall?How do I keep drywall around a patch from crumbling?Can I glue a second layer of drywall?How to patch long strip on drywall?Large drywall patch: how to avoid bulging seams?Drywall Mesh Patch vs. Bulge? To remove or not to remove?How to fix this drywall job?Prep drywall before backsplashWhat's the best way to fix this horrible drywall patch job?Drywall patching using 3M Patch Plus Primer

random experiment with two different functions on unit interval Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern)Random variable and probability space notionsRandom Walk with EdgesFinding functions where the increase over a random interval is Poisson distributedNumber of days until dayCan an observed event in fact be of zero probability?Unit random processmodels of coins and uniform distributionHow to get the number of successes given $n$ trials , probability $P$ and a random variable $X$Absorbing Markov chain in a computer. Is “almost every” turned into always convergence in computer executions?Stopped random walk is not uniformly integrable

Lowndes Grove History Architecture References Navigation menu32°48′6″N 79°57′58″W / 32.80167°N 79.96611°W / 32.80167; -79.9661132°48′6″N 79°57′58″W / 32.80167°N 79.96611°W / 32.80167; -79.9661178002500"National Register Information System"Historic houses of South Carolina"Lowndes Grove""+32° 48' 6.00", −79° 57' 58.00""Lowndes Grove, Charleston County (260 St. Margaret St., Charleston)""Lowndes Grove"The Charleston ExpositionIt Happened in South Carolina"Lowndes Grove (House), Saint Margaret Street & Sixth Avenue, Charleston, Charleston County, SC(Photographs)"Plantations of the Carolina Low Countrye