Relative Entropy and the Wasserstein distanceWasserstein distance from a Dirac measureThe Wasserstein distance on $mathbbR$Wasserstein distances metrize weak convergenceUnderstanding information entropyCompleteness of Wasserstein spaceAn intriguing duality gap for Wasserstein distance for Gaussian distributionsWasserstein Distance with TranslationsWasserstein distance between hyperplane and cubeWasserstein attains its infimumWasserstein distance of two flat triangles

Is exact Kanji stroke length important?

What was required to accept "troll"?

Should my PhD thesis be submitted under my legal name?

Teaching indefinite integrals that require special-casing

A known event to a history junkie

Stereotypical names

In Star Trek IV, why did the Bounty go back to a time when whales were already rare?

Meta programming: Declare a new struct on the fly

Simple recursive Sudoku solver

Partial sums of primes

Why isn't KTEX's runway designation 10/28 instead of 9/27?

The One-Electron Universe postulate is true - what simple change can I make to change the whole universe?

Is a naturally all "male" species possible?

Perfect riffle shuffles

What do you call the infoboxes with text and sometimes images on the side of a page we find in textbooks?

Are taller landing gear bad for aircraft, particulary large airliners?

How to deal with or prevent idle in the test team?

Organic chemistry Iodoform Reaction

Greatest common substring

What will be the benefits of Brexit?

How do ultrasonic sensors differentiate between transmitted and received signals?

"lassen" in meaning "sich fassen"

Are Warlocks Arcane or Divine?

Adding empty element to declared container without declaring type of element



Relative Entropy and the Wasserstein distance


Wasserstein distance from a Dirac measureThe Wasserstein distance on $mathbbR$Wasserstein distances metrize weak convergenceUnderstanding information entropyCompleteness of Wasserstein spaceAn intriguing duality gap for Wasserstein distance for Gaussian distributionsWasserstein Distance with TranslationsWasserstein distance between hyperplane and cubeWasserstein attains its infimumWasserstein distance of two flat triangles













2












$begingroup$


Can anyone give an informative example of two distributions which have a low Wasserstein distance but high relative entropy (or the other way around)? I find the Wasserstein defined (for some $p$) as



$$ W_p(mu,nu)=(inf_piinPi(mu,nu)int d^p(x,y)pi(dx,dy))^1/p $$



an intuitive, reasonable way to calculate the distance (or displacement) between two probability measures. However im struggling to see what relative entropy really tells us. I have seen from Sanov's theorem that it can be used to control the exponential rate of decay of the probability of a rare event, however I still haven't got an intuitive feel for how it works and would really appreciate a concrete example so I can compare it against the Wasserstein. I have heard that relative entropy controls the fluctuation of one distribution w.r.t another however haven't yet quite understood what this means.










share|cite|improve this question









$endgroup$
















    2












    $begingroup$


    Can anyone give an informative example of two distributions which have a low Wasserstein distance but high relative entropy (or the other way around)? I find the Wasserstein defined (for some $p$) as



    $$ W_p(mu,nu)=(inf_piinPi(mu,nu)int d^p(x,y)pi(dx,dy))^1/p $$



    an intuitive, reasonable way to calculate the distance (or displacement) between two probability measures. However im struggling to see what relative entropy really tells us. I have seen from Sanov's theorem that it can be used to control the exponential rate of decay of the probability of a rare event, however I still haven't got an intuitive feel for how it works and would really appreciate a concrete example so I can compare it against the Wasserstein. I have heard that relative entropy controls the fluctuation of one distribution w.r.t another however haven't yet quite understood what this means.










    share|cite|improve this question









    $endgroup$














      2












      2








      2





      $begingroup$


      Can anyone give an informative example of two distributions which have a low Wasserstein distance but high relative entropy (or the other way around)? I find the Wasserstein defined (for some $p$) as



      $$ W_p(mu,nu)=(inf_piinPi(mu,nu)int d^p(x,y)pi(dx,dy))^1/p $$



      an intuitive, reasonable way to calculate the distance (or displacement) between two probability measures. However im struggling to see what relative entropy really tells us. I have seen from Sanov's theorem that it can be used to control the exponential rate of decay of the probability of a rare event, however I still haven't got an intuitive feel for how it works and would really appreciate a concrete example so I can compare it against the Wasserstein. I have heard that relative entropy controls the fluctuation of one distribution w.r.t another however haven't yet quite understood what this means.










      share|cite|improve this question









      $endgroup$




      Can anyone give an informative example of two distributions which have a low Wasserstein distance but high relative entropy (or the other way around)? I find the Wasserstein defined (for some $p$) as



      $$ W_p(mu,nu)=(inf_piinPi(mu,nu)int d^p(x,y)pi(dx,dy))^1/p $$



      an intuitive, reasonable way to calculate the distance (or displacement) between two probability measures. However im struggling to see what relative entropy really tells us. I have seen from Sanov's theorem that it can be used to control the exponential rate of decay of the probability of a rare event, however I still haven't got an intuitive feel for how it works and would really appreciate a concrete example so I can compare it against the Wasserstein. I have heard that relative entropy controls the fluctuation of one distribution w.r.t another however haven't yet quite understood what this means.







      probability-theory measure-theory concentration-of-measure optimal-transport






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Mar 16 at 21:49









      MontyMonty

      36113




      36113




















          1 Answer
          1






          active

          oldest

          votes


















          2












          $begingroup$

          For an example, look at the point masses $delta_0$ and $delta_h$ supported at $0$ and $h$, respectively. The Wasserstein distance between these is $O(h)$, which is small if $h$ is small. But for $hne0$ the relative entropy is infinitely large, as the two measures are mutually singular.



          For an example the other way around, let $f$ be the density of a $U[0,N]$ rv, and let $g(x)=(1-epsilon)f(x)$ for $xin[0,N/2]$ and $g(x)=(1+epsilon)f(x)$ otherwise. The Wasserstein distance is something like $O(Nepsilon)$ (because we have to transfer like $epsilon$ of the mass over distance $N/2$, but the relative entropy is something like $O(epsilon)$ because $log f(x)/g(x) = O(epsilon)$. By proper choice of $epsilon$, we can make the Wasserstein distance big but the relative entropy small.



          The intuitive picture I have in mind, is that when one looks at the superimposed graphs of the densities of the two measures (pretending that they have densities) is that the the relative entropy measures how much they differ in a vertical sense only, but the Wasserstein metric allows for sideways nudgings of the two graphs.






          share|cite|improve this answer











          $endgroup$












            Your Answer





            StackExchange.ifUsing("editor", function ()
            return StackExchange.using("mathjaxEditing", function ()
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            );
            );
            , "mathjax-editing");

            StackExchange.ready(function()
            var channelOptions =
            tags: "".split(" "),
            id: "69"
            ;
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function()
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled)
            StackExchange.using("snippets", function()
            createEditor();
            );

            else
            createEditor();

            );

            function createEditor()
            StackExchange.prepareEditor(
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader:
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            ,
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            );



            );













            draft saved

            draft discarded


















            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3150894%2frelative-entropy-and-the-wasserstein-distance%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            2












            $begingroup$

            For an example, look at the point masses $delta_0$ and $delta_h$ supported at $0$ and $h$, respectively. The Wasserstein distance between these is $O(h)$, which is small if $h$ is small. But for $hne0$ the relative entropy is infinitely large, as the two measures are mutually singular.



            For an example the other way around, let $f$ be the density of a $U[0,N]$ rv, and let $g(x)=(1-epsilon)f(x)$ for $xin[0,N/2]$ and $g(x)=(1+epsilon)f(x)$ otherwise. The Wasserstein distance is something like $O(Nepsilon)$ (because we have to transfer like $epsilon$ of the mass over distance $N/2$, but the relative entropy is something like $O(epsilon)$ because $log f(x)/g(x) = O(epsilon)$. By proper choice of $epsilon$, we can make the Wasserstein distance big but the relative entropy small.



            The intuitive picture I have in mind, is that when one looks at the superimposed graphs of the densities of the two measures (pretending that they have densities) is that the the relative entropy measures how much they differ in a vertical sense only, but the Wasserstein metric allows for sideways nudgings of the two graphs.






            share|cite|improve this answer











            $endgroup$

















              2












              $begingroup$

              For an example, look at the point masses $delta_0$ and $delta_h$ supported at $0$ and $h$, respectively. The Wasserstein distance between these is $O(h)$, which is small if $h$ is small. But for $hne0$ the relative entropy is infinitely large, as the two measures are mutually singular.



              For an example the other way around, let $f$ be the density of a $U[0,N]$ rv, and let $g(x)=(1-epsilon)f(x)$ for $xin[0,N/2]$ and $g(x)=(1+epsilon)f(x)$ otherwise. The Wasserstein distance is something like $O(Nepsilon)$ (because we have to transfer like $epsilon$ of the mass over distance $N/2$, but the relative entropy is something like $O(epsilon)$ because $log f(x)/g(x) = O(epsilon)$. By proper choice of $epsilon$, we can make the Wasserstein distance big but the relative entropy small.



              The intuitive picture I have in mind, is that when one looks at the superimposed graphs of the densities of the two measures (pretending that they have densities) is that the the relative entropy measures how much they differ in a vertical sense only, but the Wasserstein metric allows for sideways nudgings of the two graphs.






              share|cite|improve this answer











              $endgroup$















                2












                2








                2





                $begingroup$

                For an example, look at the point masses $delta_0$ and $delta_h$ supported at $0$ and $h$, respectively. The Wasserstein distance between these is $O(h)$, which is small if $h$ is small. But for $hne0$ the relative entropy is infinitely large, as the two measures are mutually singular.



                For an example the other way around, let $f$ be the density of a $U[0,N]$ rv, and let $g(x)=(1-epsilon)f(x)$ for $xin[0,N/2]$ and $g(x)=(1+epsilon)f(x)$ otherwise. The Wasserstein distance is something like $O(Nepsilon)$ (because we have to transfer like $epsilon$ of the mass over distance $N/2$, but the relative entropy is something like $O(epsilon)$ because $log f(x)/g(x) = O(epsilon)$. By proper choice of $epsilon$, we can make the Wasserstein distance big but the relative entropy small.



                The intuitive picture I have in mind, is that when one looks at the superimposed graphs of the densities of the two measures (pretending that they have densities) is that the the relative entropy measures how much they differ in a vertical sense only, but the Wasserstein metric allows for sideways nudgings of the two graphs.






                share|cite|improve this answer











                $endgroup$



                For an example, look at the point masses $delta_0$ and $delta_h$ supported at $0$ and $h$, respectively. The Wasserstein distance between these is $O(h)$, which is small if $h$ is small. But for $hne0$ the relative entropy is infinitely large, as the two measures are mutually singular.



                For an example the other way around, let $f$ be the density of a $U[0,N]$ rv, and let $g(x)=(1-epsilon)f(x)$ for $xin[0,N/2]$ and $g(x)=(1+epsilon)f(x)$ otherwise. The Wasserstein distance is something like $O(Nepsilon)$ (because we have to transfer like $epsilon$ of the mass over distance $N/2$, but the relative entropy is something like $O(epsilon)$ because $log f(x)/g(x) = O(epsilon)$. By proper choice of $epsilon$, we can make the Wasserstein distance big but the relative entropy small.



                The intuitive picture I have in mind, is that when one looks at the superimposed graphs of the densities of the two measures (pretending that they have densities) is that the the relative entropy measures how much they differ in a vertical sense only, but the Wasserstein metric allows for sideways nudgings of the two graphs.







                share|cite|improve this answer














                share|cite|improve this answer



                share|cite|improve this answer








                edited Mar 17 at 0:37









                Ankitp

                21029




                21029










                answered Mar 16 at 22:17









                kimchi loverkimchi lover

                11.5k31229




                11.5k31229



























                    draft saved

                    draft discarded
















































                    Thanks for contributing an answer to Mathematics Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid


                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.

                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function ()
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3150894%2frelative-entropy-and-the-wasserstein-distance%23new-answer', 'question_page');

                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    Solar Wings Breeze Design and development Specifications (Breeze) References Navigation menu1368-485X"Hang glider: Breeze (Solar Wings)"e

                    Kathakali Contents Etymology and nomenclature History Repertoire Songs and musical instruments Traditional plays Styles: Sampradayam Training centers and awards Relationship to other dance forms See also Notes References External links Navigation menueThe Illustrated Encyclopedia of Hinduism: A-MSouth Asian Folklore: An EncyclopediaRoutledge International Encyclopedia of Women: Global Women's Issues and KnowledgeKathakali Dance-drama: Where Gods and Demons Come to PlayKathakali Dance-drama: Where Gods and Demons Come to PlayKathakali Dance-drama: Where Gods and Demons Come to Play10.1353/atj.2005.0004The Illustrated Encyclopedia of Hinduism: A-MEncyclopedia of HinduismKathakali Dance-drama: Where Gods and Demons Come to PlaySonic Liturgy: Ritual and Music in Hindu Tradition"The Mirror of Gesture"Kathakali Dance-drama: Where Gods and Demons Come to Play"Kathakali"Indian Theatre: Traditions of PerformanceIndian Theatre: Traditions of PerformanceIndian Theatre: Traditions of PerformanceIndian Theatre: Traditions of PerformanceMedieval Indian Literature: An AnthologyThe Oxford Companion to Indian TheatreSouth Asian Folklore: An Encyclopedia : Afghanistan, Bangladesh, India, Nepal, Pakistan, Sri LankaThe Rise of Performance Studies: Rethinking Richard Schechner's Broad SpectrumIndian Theatre: Traditions of PerformanceModern Asian Theatre and Performance 1900-2000Critical Theory and PerformanceBetween Theater and AnthropologyKathakali603847011Indian Theatre: Traditions of PerformanceIndian Theatre: Traditions of PerformanceIndian Theatre: Traditions of PerformanceBetween Theater and AnthropologyBetween Theater and AnthropologyNambeesan Smaraka AwardsArchivedThe Cambridge Guide to TheatreRoutledge International Encyclopedia of Women: Global Women's Issues and KnowledgeThe Garland Encyclopedia of World Music: South Asia : the Indian subcontinentThe Ethos of Noh: Actors and Their Art10.2307/1145740By Means of Performance: Intercultural Studies of Theatre and Ritual10.1017/s204912550000100xReconceiving the Renaissance: A Critical ReaderPerformance TheoryListening to Theatre: The Aural Dimension of Beijing Opera10.2307/1146013Kathakali: The Art of the Non-WorldlyOn KathakaliKathakali, the dance theatreThe Kathakali Complex: Performance & StructureKathakali Dance-Drama: Where Gods and Demons Come to Play10.1093/obo/9780195399318-0071Drama and Ritual of Early Hinduism"In the Shadow of Hollywood Orientalism: Authentic East Indian Dancing"10.1080/08949460490274013Sanskrit Play Production in Ancient IndiaIndian Music: History and StructureBharata, the Nāṭyaśāstra233639306Table of Contents2238067286469807Dance In Indian Painting10.2307/32047833204783Kathakali Dance-Theatre: A Visual Narrative of Sacred Indian MimeIndian Classical Dance: The Renaissance and BeyondKathakali: an indigenous art-form of Keralaeee

                    Method to test if a number is a perfect power? Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern)Detecting perfect squares faster than by extracting square rooteffective way to get the integer sequence A181392 from oeisA rarely mentioned fact about perfect powersHow many numbers such $n$ are there that $n<100,lfloorsqrtn rfloor mid n$Check perfect squareness by modulo division against multiple basesFor what pair of integers $(a,b)$ is $3^a + 7^b$ a perfect square.Do there exist any positive integers $n$ such that $lfloore^nrfloor$ is a perfect power? What is the probability that one exists?finding perfect power factors of an integerProve that the sequence contains a perfect square for any natural number $m $ in the domain of $f$ .Counting Perfect Powers