| 000 | 01078camuu2200301 a 4500 | |
| 001 | 000045810263 | |
| 005 | 20140911111137 | |
| 008 | 140905s1996 nyu b 001 0 eng | |
| 010 | ▼a 95037683 | |
| 020 | ▼a 0387945792 (New York : hc : alk. paper) | |
| 035 | ▼a (KERIS)REF000014736745 | |
| 040 | ▼a DLC ▼c DLC ▼d DLC ▼d 211009 | |
| 050 | 0 0 | ▼a QA274.7 ▼b .H47 1996 |
| 082 | 0 0 | ▼a 003/.83/0115 ▼2 23 |
| 084 | ▼a 003.830115 ▼2 DDCK | |
| 090 | ▼a 003.830115 ▼b H557d | |
| 100 | 1 | ▼a Hernandez-Lerma, O. ▼q (Onesimo) |
| 245 | 1 0 | ▼a Discrete-time Markov control processes : ▼b basic optimality criteria / ▼c Onesimo Hernandez-Lerma, Jean Bernard Lasserre. |
| 260 | ▼a New York : ▼b Springer, ▼c c1996. | |
| 300 | ▼a xiv, 216 p. ; ▼c 25 cm. | |
| 490 | 1 | ▼a Applications of mathematics ; ▼v 30 |
| 504 | ▼a Includes bibliographical references (p. [191]-209) and index. | |
| 650 | 0 | ▼a Markov processes. |
| 650 | 0 | ▼a Discrete-time systems. |
| 700 | 1 | ▼a Lasserre, Jean-Bernard, ▼d 1953-. |
| 830 | 0 | ▼a Applications of mathematics ; ▼v 30. |
| 945 | ▼a KLPA |
소장정보
| No. | 소장처 | 청구기호 | 등록번호 | 도서상태 | 반납예정일 | 예약 | 서비스 |
|---|---|---|---|---|---|---|---|
| No. 1 | 소장처 과학도서관/Sci-Info(2층서고)/ | 청구기호 003.830115 H557d | 등록번호 121230704 (1회 대출) | 도서상태 대출가능 | 반납예정일 | 예약 | 서비스 |
컨텐츠정보
책소개
This book presents the first part of a planned two-volume series devoted to a systematic exposition of some recent developments in the theory of discrete-time Markov control processes (MCPs). Interest is mainly confined to MCPs with Borel state and control (or action) spaces, and possibly unbounded costs and noncompact control constraint sets. MCPs are a class of stochastic control problems, also known as Markov decision processes, controlled Markov processes, or stochastic dynamic pro grams; sometimes, particularly when the state space is a countable set, they are also called Markov decision (or controlled Markov) chains. Regardless of the name used, MCPs appear in many fields, for example, engineering, economics, operations research, statistics, renewable and nonrenewable re source management, (control of) epidemics, etc. However, most of the lit erature (say, at least 90%) is concentrated on MCPs for which (a) the state space is a countable set, and/or (b) the costs-per-stage are bounded, and/or (c) the control constraint sets are compact. But curiously enough, the most widely used control model in engineering and economics--namely the LQ (Linear system/Quadratic cost) model-satisfies none of these conditions. Moreover, when dealing with "partially observable" systems) a standard approach is to transform them into equivalent "completely observable" sys tems in a larger state space (in fact, a space of probability measures), which is uncountable even if the original state process is finite-valued.
This book presents the first part of a planned two-volume series devoted to a systematic exposition of some recent developments in the theory of discrete-time Markov control processes (MCPs). Interest is mainly confined to MCPs with Borel state and control (or action) spaces, and possibly unbounded costs and noncompact control constraint sets. MCPs are a class of stochastic control problems, also known as Markov decision processes, controlled Markov processes, or stochastic dynamic pro grams; sometimes, particularly when the state space is a countable set, they are also called Markov decision (or controlled Markov) chains. Regardless of the name used, MCPs appear in many fields, for example, engineering, economics, operations research, statistics, renewable and nonrenewable re source management, (control of) epidemics, etc. However, most of the lit erature (say, at least 90%) is concentrated on MCPs for which (a) the state space is a countable set, and/or (b) the costs-per-stage are bounded, and/or (c) the control constraint sets are compact. But curiously enough, the most widely used control model in engineering and economics--namely the LQ (Linear system/Quadratic cost) model-satisfies none of these conditions. Moreover, when dealing with "partially observable" systems) a standard approach is to transform them into equivalent "completely observable" sys tems in a larger state space (in fact, a space of probability measures), which is uncountable even if the original state process is finite-valued.
정보제공 :
목차
1 Introduction and Summary.- 1.1 Introduction.- 1.2 Markov control processes.- 1.3 Preliminary examples.- 1.4 Summary of the following chapters.- 2 Markov Control Processes.- 2.1 Introduction.- 2.2 Markov control processes.- 2.3 Markov policies and the Markov property.- 3 Finite-Horizon Problems.- 3.1 Introduction.- 3.2 Dynamic programming.- 3.3 The measurable selection condition.- 3.4 Variants of the DP equation.- 3.5 LQ problems.- 3.6 A consumption-investment problem.- 3.7 An inventory-production system.- 4 Infinite-Horizon Discounted-Cost Problems.- 4.1 Introduction.- 4.2 The discounted-cost optimality equation.- 4.3 Complements to the DCOE.- 4.4 Policy iteration and other approximations.- 4.5 Further optimality criteria.- 4.6 Asymptotic discount optimality.- 4.7 The discounted LQ problem.- 4.8 Concluding remarks.- 5 Long-Run Average-Cost Problems.- 5.1 Introduction.- 5.2 Canonical triplets.- 5.3 The vanishing discount approach.- 5.4 The average-cost optimality inequality.- 5.5 The average-cost optimality equation.- 5.6 Value iteration.- 5.7 Other optimality results.- 5.8 Concluding remarks.- 6 The Linear Programming Formulation.- 6.1 Introduction.- 6.2 Infinite-dimensional linear programming.- 6.3 Discounted cost.- 6.4 Average cost: preliminaries.- 6.5 Average cost: solvability.- 6.6 Further remarks.- Appendix A Miscellaneous Results.- Appendix B Conditional Expectation.- Appendix C Stochastic Kernels.- Appendix D Multifunctions and Selectors.- Appendix E Convergence of Probability Measures.- References.
정보제공 :
