Analysis of Entropy on Two Distributions: Proving $H(X) < H(X')$Understanding the proof of the concavity of entropy.Optimization of entropy for fixed distance to uniformShannon's entropy in a set of probabilitiesEquality of sets when minimizing Shannon's EntropyHuffman codes: does less entropy imply less weighted average codeword length?Relation between Shannon Entropy and Total Variation distanceRandomly Generate Probability Mass Function With Specific EntropyExpected Entropy Based on Dirichlet DistributionDistributions with equal Renyi EntropiesDerivative of Shannon entropy for discrete distrubtionsMaximum Entropy with bounded constraints
How to secure an aircraft at a transient parking space?
List elements digit difference sort
weren't playing vs didn't play
They call me Inspector Morse
In the late 1940’s to early 1950’s what technology was available that could melt a LOT of ice?
How can I ensure my trip to the UK will not have to be cancelled because of Brexit?
Why does liquid water form when we exhale on a mirror?
In the quantum hamiltonian, why does kinetic energy turn into an operator while potential doesn't?
Is it "Vierergruppe" or "Viergruppe", or is there a distinction?
What problems would a superhuman have whose skin is constantly hot?
Why does the negative sign arise in this thermodynamic relation?
Accountant/ lawyer will not return my call
'The literal of type int is out of range' con número enteros pequeños (2 dígitos)
Was Luke Skywalker the leader of the Rebel forces on Hoth?
Declaring and defining template, and specialising them
Good for you! in Russian
Find longest word in a string: are any of these algorithms good?
What are some noteworthy "mic-drop" moments in math?
Vocabulary for giving just numbers, not a full answer
Are babies of evil humanoid species inherently evil?
Can you reject a postdoc offer after the PI has paid a large sum for flights/accommodation for your visit?
Are there historical instances of the capital of a colonising country being temporarily or permanently shifted to one of its colonies?
If I receive an SOS signal, what is the proper response?
Single word request: Harming the benefactor
Analysis of Entropy on Two Distributions: Proving $H(X)
Understanding the proof of the concavity of entropy.Optimization of entropy for fixed distance to uniformShannon's entropy in a set of probabilitiesEquality of sets when minimizing Shannon's EntropyHuffman codes: does less entropy imply less weighted average codeword length?Relation between Shannon Entropy and Total Variation distanceRandomly Generate Probability Mass Function With Specific EntropyExpected Entropy Based on Dirichlet DistributionDistributions with equal Renyi EntropiesDerivative of Shannon entropy for discrete distrubtionsMaximum Entropy with bounded constraints
$begingroup$
Let $P=p_1, p_2, p_3 ..., p_n$ and $P^'= left dfrac(p_1 + p_2)2, dfrac(p_1 + p_2)2, p_3, ..., p_nright$ be distributions on the same random variable $X$.
$1$. Show $H(X)leq H(X^')$ where $H(X)$ is Shannon's entropy formula:$$H(X) = -sum_i=1^n p_ilog_2p_i $$
This make sense since $P^'$ is closer to the uniform distribution, which maximizes entropy, but I'm not sure how to go about proving this. I believe there is some expectation that we will use the the fact that entropy can be decomposed to its binary components (with normalization along the way.)
$2$. Define $P^''=left p_1, ..., p_i-1, dfrac(p_i + p_j)2, p_i+1, ..., p_j-1, dfrac(p_i + p_j)2, p_j+1, ..., p_n right$. Use the "permutation principle" and $(a)$ to show $H(X)leq H(X^'')$
information-theory entropy
New contributor
$endgroup$
add a comment |
$begingroup$
Let $P=p_1, p_2, p_3 ..., p_n$ and $P^'= left dfrac(p_1 + p_2)2, dfrac(p_1 + p_2)2, p_3, ..., p_nright$ be distributions on the same random variable $X$.
$1$. Show $H(X)leq H(X^')$ where $H(X)$ is Shannon's entropy formula:$$H(X) = -sum_i=1^n p_ilog_2p_i $$
This make sense since $P^'$ is closer to the uniform distribution, which maximizes entropy, but I'm not sure how to go about proving this. I believe there is some expectation that we will use the the fact that entropy can be decomposed to its binary components (with normalization along the way.)
$2$. Define $P^''=left p_1, ..., p_i-1, dfrac(p_i + p_j)2, p_i+1, ..., p_j-1, dfrac(p_i + p_j)2, p_j+1, ..., p_n right$. Use the "permutation principle" and $(a)$ to show $H(X)leq H(X^'')$
information-theory entropy
New contributor
$endgroup$
1
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago
add a comment |
$begingroup$
Let $P=p_1, p_2, p_3 ..., p_n$ and $P^'= left dfrac(p_1 + p_2)2, dfrac(p_1 + p_2)2, p_3, ..., p_nright$ be distributions on the same random variable $X$.
$1$. Show $H(X)leq H(X^')$ where $H(X)$ is Shannon's entropy formula:$$H(X) = -sum_i=1^n p_ilog_2p_i $$
This make sense since $P^'$ is closer to the uniform distribution, which maximizes entropy, but I'm not sure how to go about proving this. I believe there is some expectation that we will use the the fact that entropy can be decomposed to its binary components (with normalization along the way.)
$2$. Define $P^''=left p_1, ..., p_i-1, dfrac(p_i + p_j)2, p_i+1, ..., p_j-1, dfrac(p_i + p_j)2, p_j+1, ..., p_n right$. Use the "permutation principle" and $(a)$ to show $H(X)leq H(X^'')$
information-theory entropy
New contributor
$endgroup$
Let $P=p_1, p_2, p_3 ..., p_n$ and $P^'= left dfrac(p_1 + p_2)2, dfrac(p_1 + p_2)2, p_3, ..., p_nright$ be distributions on the same random variable $X$.
$1$. Show $H(X)leq H(X^')$ where $H(X)$ is Shannon's entropy formula:$$H(X) = -sum_i=1^n p_ilog_2p_i $$
This make sense since $P^'$ is closer to the uniform distribution, which maximizes entropy, but I'm not sure how to go about proving this. I believe there is some expectation that we will use the the fact that entropy can be decomposed to its binary components (with normalization along the way.)
$2$. Define $P^''=left p_1, ..., p_i-1, dfrac(p_i + p_j)2, p_i+1, ..., p_j-1, dfrac(p_i + p_j)2, p_j+1, ..., p_n right$. Use the "permutation principle" and $(a)$ to show $H(X)leq H(X^'')$
information-theory entropy
information-theory entropy
New contributor
New contributor
edited 2 days ago
user57753
New contributor
asked 2 days ago
user57753user57753
11
11
New contributor
New contributor
1
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago
add a comment |
1
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago
1
1
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
There are several ways to attack this, some of them pointed by stochasticboy321's comment.
Another way, which looks elegant to me, but it requires you to know this (useful, and not hard to prove) property of entropy: $H(p)$ is a concave function of the distribution $p$.
Granted this, consider the two distributions $p_A=(p_1,p_2, p_3 cdots p_n)$ and $p_B=(p_2,p_1, p_3 cdots p_n)$ and let $p_C = (p_A +p_B)/2$. Clearly, $H(p_A)=H(p_B)$.
Hence, by concavity $$H(X')=H(p_C) ge fracH(p_A)+H(p_B)2= H(p_A)=H(X)$$
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function ()
return StackExchange.using("mathjaxEditing", function ()
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
);
);
, "mathjax-editing");
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "69"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
user57753 is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3141599%2fanalysis-of-entropy-on-two-distributions-proving-hx-hx%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
There are several ways to attack this, some of them pointed by stochasticboy321's comment.
Another way, which looks elegant to me, but it requires you to know this (useful, and not hard to prove) property of entropy: $H(p)$ is a concave function of the distribution $p$.
Granted this, consider the two distributions $p_A=(p_1,p_2, p_3 cdots p_n)$ and $p_B=(p_2,p_1, p_3 cdots p_n)$ and let $p_C = (p_A +p_B)/2$. Clearly, $H(p_A)=H(p_B)$.
Hence, by concavity $$H(X')=H(p_C) ge fracH(p_A)+H(p_B)2= H(p_A)=H(X)$$
$endgroup$
add a comment |
$begingroup$
There are several ways to attack this, some of them pointed by stochasticboy321's comment.
Another way, which looks elegant to me, but it requires you to know this (useful, and not hard to prove) property of entropy: $H(p)$ is a concave function of the distribution $p$.
Granted this, consider the two distributions $p_A=(p_1,p_2, p_3 cdots p_n)$ and $p_B=(p_2,p_1, p_3 cdots p_n)$ and let $p_C = (p_A +p_B)/2$. Clearly, $H(p_A)=H(p_B)$.
Hence, by concavity $$H(X')=H(p_C) ge fracH(p_A)+H(p_B)2= H(p_A)=H(X)$$
$endgroup$
add a comment |
$begingroup$
There are several ways to attack this, some of them pointed by stochasticboy321's comment.
Another way, which looks elegant to me, but it requires you to know this (useful, and not hard to prove) property of entropy: $H(p)$ is a concave function of the distribution $p$.
Granted this, consider the two distributions $p_A=(p_1,p_2, p_3 cdots p_n)$ and $p_B=(p_2,p_1, p_3 cdots p_n)$ and let $p_C = (p_A +p_B)/2$. Clearly, $H(p_A)=H(p_B)$.
Hence, by concavity $$H(X')=H(p_C) ge fracH(p_A)+H(p_B)2= H(p_A)=H(X)$$
$endgroup$
There are several ways to attack this, some of them pointed by stochasticboy321's comment.
Another way, which looks elegant to me, but it requires you to know this (useful, and not hard to prove) property of entropy: $H(p)$ is a concave function of the distribution $p$.
Granted this, consider the two distributions $p_A=(p_1,p_2, p_3 cdots p_n)$ and $p_B=(p_2,p_1, p_3 cdots p_n)$ and let $p_C = (p_A +p_B)/2$. Clearly, $H(p_A)=H(p_B)$.
Hence, by concavity $$H(X')=H(p_C) ge fracH(p_A)+H(p_B)2= H(p_A)=H(X)$$
answered 2 days ago
leonbloyleonbloy
41.5k647108
41.5k647108
add a comment |
add a comment |
user57753 is a new contributor. Be nice, and check out our Code of Conduct.
user57753 is a new contributor. Be nice, and check out our Code of Conduct.
user57753 is a new contributor. Be nice, and check out our Code of Conduct.
user57753 is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3141599%2fanalysis-of-entropy-on-two-distributions-proving-hx-hx%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
What did you try? There's a direct approach, which is just to expand $H(X') - H(X)$, cancel terms, and argue via convexity of $-log$. A more operational approach is to note that $X'$ is the output of pushing $X$ through a channel that randomises the first two symbols, and then using the data processing inequality.
$endgroup$
– stochasticboy321
2 days ago
$begingroup$
I believe we ought to use convexity in our argument, but I can not get the inequality in the proper form to argue it directly. I currently have: "It suffices to show $p_1log_2p_1 + p_2log_2p_2 geq p_1 + p_2 log_2fracp_1+p_22$, but the coefficient for the right side does not align correctly if I want to argue the convexity of f(x) = $xlog_2x$. Note that the right side occurs form having H(X) - H(X') = $-p_1log_2p_1 - p_2log_2p_2 + fracp_1+p_22 log_2fracp_1+p_22 + fracp_1+p_22 log_2fracp_1+p_22$ How can I go about this?
$endgroup$
– user57753
yesterday
$begingroup$
Right, so, by convexity of $xlog x,$ you have $$ frac(p_1 + p _2)2 log fracp_1 + p_22 le frac12 left( p_1 log p_1 + p_2 log p_2right),$$ which finishes the job. That said, I think leonbloy's suggestion is a whole lot cleaner and thus better.
$endgroup$
– stochasticboy321
8 hours ago