マルコフ決定過程ハンドブック<br>Handbook of Markov Decision Processes : Methods and Applications (International Series in Operations Research & Management Science, 40)

個数:

マルコフ決定過程ハンドブック
Handbook of Markov Decision Processes : Methods and Applications (International Series in Operations Research & Management Science, 40)

  • 提携先の海外書籍取次会社に在庫がございます。通常3週間で発送いたします。
    重要ご説明事項
    1. 納期遅延や、ご入手不能となる場合が若干ございます。
    2. 複数冊ご注文の場合、分割発送となる場合がございます。
    3. 美品のご指定は承りかねます。
  • 【入荷遅延について】
    世界情勢の影響により、海外からお取り寄せとなる洋書・洋古書の入荷が、表示している標準的な納期よりも遅延する場合がございます。
    おそれいりますが、あらかじめご了承くださいますようお願い申し上げます。
  • ◆画像の表紙や帯等は実物とは異なる場合があります。
  • ◆ウェブストアでの洋書販売価格は、弊社店舗等での販売価格とは異なります。
    また、洋書販売価格は、ご注文確定時点での日本円価格となります。
    ご注文確定後に、同じ洋書の販売価格が変動しても、それは反映されません。
  • 製本 Hardcover:ハードカバー版/ページ数 565 p.
  • 言語 ENG
  • 商品コード 9780792374596
  • DDC分類 519.233

基本説明

Contents: Part I: Finite State and Action Models. Part II: Indefinite State Models. Part III: Applications.

Full Description

Eugene A. Feinberg Adam Shwartz This volume deals with the theory of Markov Decision Processes (MDPs) and their applications. Each chapter was written by a leading expert in the re­ spective area. The papers cover major research areas and methodologies, and discuss open questions and future research directions. The papers can be read independently, with the basic notation and concepts ofSection 1.2. Most chap­ ters should be accessible by graduate or advanced undergraduate students in fields of operations research, electrical engineering, and computer science. 1.1 AN OVERVIEW OF MARKOV DECISION PROCESSES The theory of Markov Decision Processes-also known under several other names including sequential stochastic optimization, discrete-time stochastic control, and stochastic dynamic programming-studiessequential optimization ofdiscrete time stochastic systems. The basic object is a discrete-time stochas­ tic system whose transition mechanism can be controlled over time. Each control policy defines the stochastic process and values of objective functions associated with this process. The goal is to select a "good" control policy. In real life, decisions that humans and computers make on all levels usually have two types ofimpacts: (i) they cost orsavetime, money, or other resources, or they bring revenues, as well as (ii) they have an impact on the future, by influencing the dynamics. In many situations, decisions with the largest immediate profit may not be good in view offuture events. MDPs model this paradigm and provide results on the structure and existence of good policies and on methods for their calculation.

Contents

1 Introduction.- I Finite State and Action Models.- 2 Finite State and Action MDPs.- 3 Bias Optimality.- 4 Singular Perturbations of Markov Chains and Decision Processes.- II Infinite State Models.- 5 Average Reward Optimization Theory for Denumerable State Spaces.- 6 Total Reward Criteria.- 7 Mixed Criteria.- 8 Blackwell Optimality.- 9 The Poisson Equation for Countable Markov Chains: Probabilistic Methods and Interpretations.- 10 Stability, Performance Evaluation, and Optimization.- 11 Convex Analytic Methods in Markov Decision Processes.- 12 The Linear Programming Approach.- 13 Invariant Gambling Problems and Markov Decision Processes.- III Applications.- 14 Neuro-Dynamic Programming: Overview and Recent Trends.- 15 Markov Decision Processes in Finance and Dynamic Options.- 16 Applications of Markov Decision Processes in Communication Networks.- 17 Water Reservoir Applications of Markov Decision Processes.