Book on Markov Decision Processes with many worked examplesWhat is the difference between all types of Markov Chains?Potential theory: discrete-time Markov processesReferences for basics of Piecewise-Deterministic Markov ProcessesReference request for stochastic process and applicationsReference request for this topicsGeneral state Markov Chains - referencesSoft Question - book recommendation - Stochastic ProcessesContinuous time Markov processes on general state spacesGood book for Integer/Non-Linear/Stochastic/Dynamic programing [Operations Research]Nonhomogenous Chains

How do spaceships determine each other's mass in space?

Factor Rings over Finite Fields

Traveling to heavily polluted city, what practical measures can I take to minimize impact?

Did Amazon pay $0 in taxes last year?

I can't die. Who am I?

Is there a math expression equivalent to the conditional ternary operator?

Can I negotiate a patent idea for a raise, under French law?

Is it possible to clone a polymorphic object without manually adding overridden clone method into each derived class in C++?

Use Mercury as quenching liquid for swords?

Is it a Cyclops number? "Nobody" knows!

Converting from "matrix" data into "coordinate" data

Smooth vector fields on a surface modulo diffeomorphisms

Computation logic of Partway in TikZ

Should we avoid writing fiction about historical events without extensive research?

Idiom for feeling after taking risk and someone else being rewarded

Leveling the sagging side of the home

Called into a meeting and told we are being made redundant (laid off) and "not to share outside". Can I tell my partner?

What will happen if my luggage gets delayed?

Create chunks from an array

How is it possible to drive VGA displays at such high pixel clock frequencies?

Too soon for a plot twist?

What is the purpose of a disclaimer like "this is not legal advice"?

Can one live in the U.S. and not use a credit card?

Having the player face themselves after the mid-game



Book on Markov Decision Processes with many worked examples


What is the difference between all types of Markov Chains?Potential theory: discrete-time Markov processesReferences for basics of Piecewise-Deterministic Markov ProcessesReference request for stochastic process and applicationsReference request for this topicsGeneral state Markov Chains - referencesSoft Question - book recommendation - Stochastic ProcessesContinuous time Markov processes on general state spacesGood book for Integer/Non-Linear/Stochastic/Dynamic programing [Operations Research]Nonhomogenous Chains













2












$begingroup$


I am looking for a book (or online article(s)) on Markov decision processes that contains lots of worked examples or problems with solutions. The purpose of the book is to grind my teeth on some problems during long commutes.



The book must...



  • have many examples using dynamic programming and the Bellman equation in discrete space and discrete time;

  • touch on policy and value iteration, and perhaps their computational complexity and implementation,

and ideally would...



  • use techniques from convex optimization, Lagrange multipliers, or combine with other computational techniques, such as sorting algorithms;

  • cover more modern examples besides the usual queueing / inventory problems, such as reinforcement learning;

  • contain lots of neat tricks and calculations.

An ideal book would not...



  • be a Theorem-Proof bible, aiming to identify the weakest conditions for optimality results

  • consider controlled Markov processes and viscosity solutions

I own Sheldon Ross's Applied probability models with optimization applications, in which there are several worked examples, a fair bit of good problems, but no solutions. I have been looking at Puterman's classic textbook Markov Decision Processes: Discrete Stochastic Dynamic Programming, but it is over 600 pages long and a bit on the "bible" side.



I'm looking for something more like Markov Chains and Mixing Times by Levin, Wilmer and Peres, but for MDPs. They have bite-sized chapters and a fair bit of explicit calculation. I like Norris's Markov Chains, which has some nice introductory exposition on potential theory, as well as the Applications chapter in David Williams's Probability with martingales. I do not mind if this "workbook" I am looking for is at an "advanced undergraduate" level, or directed at engineers or computer scientists.










share|cite|improve this question









$endgroup$
















    2












    $begingroup$


    I am looking for a book (or online article(s)) on Markov decision processes that contains lots of worked examples or problems with solutions. The purpose of the book is to grind my teeth on some problems during long commutes.



    The book must...



    • have many examples using dynamic programming and the Bellman equation in discrete space and discrete time;

    • touch on policy and value iteration, and perhaps their computational complexity and implementation,

    and ideally would...



    • use techniques from convex optimization, Lagrange multipliers, or combine with other computational techniques, such as sorting algorithms;

    • cover more modern examples besides the usual queueing / inventory problems, such as reinforcement learning;

    • contain lots of neat tricks and calculations.

    An ideal book would not...



    • be a Theorem-Proof bible, aiming to identify the weakest conditions for optimality results

    • consider controlled Markov processes and viscosity solutions

    I own Sheldon Ross's Applied probability models with optimization applications, in which there are several worked examples, a fair bit of good problems, but no solutions. I have been looking at Puterman's classic textbook Markov Decision Processes: Discrete Stochastic Dynamic Programming, but it is over 600 pages long and a bit on the "bible" side.



    I'm looking for something more like Markov Chains and Mixing Times by Levin, Wilmer and Peres, but for MDPs. They have bite-sized chapters and a fair bit of explicit calculation. I like Norris's Markov Chains, which has some nice introductory exposition on potential theory, as well as the Applications chapter in David Williams's Probability with martingales. I do not mind if this "workbook" I am looking for is at an "advanced undergraduate" level, or directed at engineers or computer scientists.










    share|cite|improve this question









    $endgroup$














      2












      2








      2





      $begingroup$


      I am looking for a book (or online article(s)) on Markov decision processes that contains lots of worked examples or problems with solutions. The purpose of the book is to grind my teeth on some problems during long commutes.



      The book must...



      • have many examples using dynamic programming and the Bellman equation in discrete space and discrete time;

      • touch on policy and value iteration, and perhaps their computational complexity and implementation,

      and ideally would...



      • use techniques from convex optimization, Lagrange multipliers, or combine with other computational techniques, such as sorting algorithms;

      • cover more modern examples besides the usual queueing / inventory problems, such as reinforcement learning;

      • contain lots of neat tricks and calculations.

      An ideal book would not...



      • be a Theorem-Proof bible, aiming to identify the weakest conditions for optimality results

      • consider controlled Markov processes and viscosity solutions

      I own Sheldon Ross's Applied probability models with optimization applications, in which there are several worked examples, a fair bit of good problems, but no solutions. I have been looking at Puterman's classic textbook Markov Decision Processes: Discrete Stochastic Dynamic Programming, but it is over 600 pages long and a bit on the "bible" side.



      I'm looking for something more like Markov Chains and Mixing Times by Levin, Wilmer and Peres, but for MDPs. They have bite-sized chapters and a fair bit of explicit calculation. I like Norris's Markov Chains, which has some nice introductory exposition on potential theory, as well as the Applications chapter in David Williams's Probability with martingales. I do not mind if this "workbook" I am looking for is at an "advanced undergraduate" level, or directed at engineers or computer scientists.










      share|cite|improve this question









      $endgroup$




      I am looking for a book (or online article(s)) on Markov decision processes that contains lots of worked examples or problems with solutions. The purpose of the book is to grind my teeth on some problems during long commutes.



      The book must...



      • have many examples using dynamic programming and the Bellman equation in discrete space and discrete time;

      • touch on policy and value iteration, and perhaps their computational complexity and implementation,

      and ideally would...



      • use techniques from convex optimization, Lagrange multipliers, or combine with other computational techniques, such as sorting algorithms;

      • cover more modern examples besides the usual queueing / inventory problems, such as reinforcement learning;

      • contain lots of neat tricks and calculations.

      An ideal book would not...



      • be a Theorem-Proof bible, aiming to identify the weakest conditions for optimality results

      • consider controlled Markov processes and viscosity solutions

      I own Sheldon Ross's Applied probability models with optimization applications, in which there are several worked examples, a fair bit of good problems, but no solutions. I have been looking at Puterman's classic textbook Markov Decision Processes: Discrete Stochastic Dynamic Programming, but it is over 600 pages long and a bit on the "bible" side.



      I'm looking for something more like Markov Chains and Mixing Times by Levin, Wilmer and Peres, but for MDPs. They have bite-sized chapters and a fair bit of explicit calculation. I like Norris's Markov Chains, which has some nice introductory exposition on potential theory, as well as the Applications chapter in David Williams's Probability with martingales. I do not mind if this "workbook" I am looking for is at an "advanced undergraduate" level, or directed at engineers or computer scientists.







      reference-request markov-chains






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked yesterday









      snarsnar

      4,2841219




      4,2841219




















          0






          active

          oldest

          votes











          Your Answer





          StackExchange.ifUsing("editor", function ()
          return StackExchange.using("mathjaxEditing", function ()
          StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
          StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
          );
          );
          , "mathjax-editing");

          StackExchange.ready(function()
          var channelOptions =
          tags: "".split(" "),
          id: "69"
          ;
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function()
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled)
          StackExchange.using("snippets", function()
          createEditor();
          );

          else
          createEditor();

          );

          function createEditor()
          StackExchange.prepareEditor(
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader:
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          ,
          noCode: true, onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          );



          );













          draft saved

          draft discarded


















          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3140533%2fbook-on-markov-decision-processes-with-many-worked-examples%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown

























          0






          active

          oldest

          votes








          0






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes















          draft saved

          draft discarded
















































          Thanks for contributing an answer to Mathematics Stack Exchange!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid


          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.

          Use MathJax to format equations. MathJax reference.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function ()
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3140533%2fbook-on-markov-decision-processes-with-many-worked-examples%23new-answer', 'question_page');

          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          How should I support this large drywall patch? Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern) Announcing the arrival of Valued Associate #679: Cesar Manara Unicorn Meta Zoo #1: Why another podcast?How do I cover large gaps in drywall?How do I keep drywall around a patch from crumbling?Can I glue a second layer of drywall?How to patch long strip on drywall?Large drywall patch: how to avoid bulging seams?Drywall Mesh Patch vs. Bulge? To remove or not to remove?How to fix this drywall job?Prep drywall before backsplashWhat's the best way to fix this horrible drywall patch job?Drywall patching using 3M Patch Plus Primer

          random experiment with two different functions on unit interval Announcing the arrival of Valued Associate #679: Cesar Manara Planned maintenance scheduled April 23, 2019 at 00:00UTC (8:00pm US/Eastern)Random variable and probability space notionsRandom Walk with EdgesFinding functions where the increase over a random interval is Poisson distributedNumber of days until dayCan an observed event in fact be of zero probability?Unit random processmodels of coins and uniform distributionHow to get the number of successes given $n$ trials , probability $P$ and a random variable $X$Absorbing Markov chain in a computer. Is “almost every” turned into always convergence in computer executions?Stopped random walk is not uniformly integrable

          Lowndes Grove History Architecture References Navigation menu32°48′6″N 79°57′58″W / 32.80167°N 79.96611°W / 32.80167; -79.9661132°48′6″N 79°57′58″W / 32.80167°N 79.96611°W / 32.80167; -79.9661178002500"National Register Information System"Historic houses of South Carolina"Lowndes Grove""+32° 48' 6.00", −79° 57' 58.00""Lowndes Grove, Charleston County (260 St. Margaret St., Charleston)""Lowndes Grove"The Charleston ExpositionIt Happened in South Carolina"Lowndes Grove (House), Saint Margaret Street & Sixth Avenue, Charleston, Charleston County, SC(Photographs)"Plantations of the Carolina Low Countrye