Adaptive Markov Control Processes

Adaptive Markov Control Processes
Author :
Publisher : Springer Science & Business Media
Total Pages : 160
Release :
ISBN-10 : 9781441987143
ISBN-13 : 1441987142
Rating : 4/5 (43 Downloads)

Synopsis Adaptive Markov Control Processes by : Onesimo Hernandez-Lerma

This book is concerned with a class of discrete-time stochastic control processes known as controlled Markov processes (CMP's), also known as Markov decision processes or Markov dynamic programs. Starting in the mid-1950swith Richard Bellman, many contributions to CMP's have been made, and applications to engineering, statistics and operations research, among other areas, have also been developed. The purpose of this book is to present some recent developments on the theory of adaptive CMP's, i. e. , CMP's that depend on unknown parameters. Thus at each decision time, the controller or decision-maker must estimate the true parameter values, and then adapt the control actions to the estimated values. We do not intend to describe all aspects of stochastic adaptive control; rather, the selection of material reflects our own research interests. The prerequisite for this book is a knowledgeof real analysis and prob ability theory at the level of, say, Ash (1972) or Royden (1968), but no previous knowledge of control or decision processes is required. The pre sentation, on the other hand, is meant to beself-contained,in the sensethat whenever a result from analysisor probability is used, it is usually stated in full and references are supplied for further discussion, if necessary. Several appendices are provided for this purpose. The material is divided into six chapters. Chapter 1 contains the basic definitions about the stochastic control problems we are interested in; a brief description of some applications is also provided.

Adaptive Markov Control Processes

Adaptive Markov Control Processes
Author :
Publisher :
Total Pages : 190
Release :
ISBN-10 : UCAL:B4420470
ISBN-13 :
Rating : 4/5 (70 Downloads)

Synopsis Adaptive Markov Control Processes by : Onésimo Hernández-Lerma

Mathematical Theory of Adaptive Control

Mathematical Theory of Adaptive Control
Author :
Publisher : World Scientific
Total Pages : 490
Release :
ISBN-10 : 9789812701039
ISBN-13 : 9812701036
Rating : 4/5 (39 Downloads)

Synopsis Mathematical Theory of Adaptive Control by : Vladimir G. Sragovich

The theory of adaptive control is concerned with construction of strategies so that the controlled system behaves in a desirable way, without assuming the complete knowledge of the system. The models considered in this comprehensive book are of Markovian type. Both partial observation and partial information cases are analyzed. While the book focuses on discrete time models, continuous time ones are considered in the final chapter. The book provides a novel perspective by summarizing results on adaptive control obtained in the Soviet Union, which are not well known in the West. Comments on the interplay between the Russian and Western methods are also included.

Markov Processes and Controlled Markov Chains

Markov Processes and Controlled Markov Chains
Author :
Publisher : Springer Science & Business Media
Total Pages : 501
Release :
ISBN-10 : 9781461302650
ISBN-13 : 146130265X
Rating : 4/5 (50 Downloads)

Synopsis Markov Processes and Controlled Markov Chains by : Zhenting Hou

The general theory of stochastic processes and the more specialized theory of Markov processes evolved enormously in the second half of the last century. In parallel, the theory of controlled Markov chains (or Markov decision processes) was being pioneered by control engineers and operations researchers. Researchers in Markov processes and controlled Markov chains have been, for a long time, aware of the synergies between these two subject areas. However, this may be the first volume dedicated to highlighting these synergies and, almost certainly, it is the first volume that emphasizes the contributions of the vibrant and growing Chinese school of probability. The chapters that appear in this book reflect both the maturity and the vitality of modern day Markov processes and controlled Markov chains. They also will provide an opportunity to trace the connections that have emerged between the work done by members of the Chinese school of probability and the work done by the European, US, Central and South American and Asian scholars.

Stochastic Systems

Stochastic Systems
Author :
Publisher : SIAM
Total Pages : 371
Release :
ISBN-10 : 9781611974256
ISBN-13 : 1611974259
Rating : 4/5 (56 Downloads)

Synopsis Stochastic Systems by : P. R. Kumar

Since its origins in the 1940s, the subject of decision making under uncertainty has grown into a diversified area with application in several branches of engineering and in those areas of the social sciences concerned with policy analysis and prescription. These approaches required a computing capacity too expensive for the time, until the ability to collect and process huge quantities of data engendered an explosion of work in the area. This book provides succinct and rigorous treatment of the foundations of stochastic control; a unified approach to filtering, estimation, prediction, and stochastic and adaptive control; and the conceptual framework necessary to understand current trends in stochastic control, data mining, machine learning, and robotics.

Zero-Sum Discrete-Time Markov Games with Unknown Disturbance Distribution

Zero-Sum Discrete-Time Markov Games with Unknown Disturbance Distribution
Author :
Publisher : Springer Nature
Total Pages : 129
Release :
ISBN-10 : 9783030357207
ISBN-13 : 3030357201
Rating : 4/5 (07 Downloads)

Synopsis Zero-Sum Discrete-Time Markov Games with Unknown Disturbance Distribution by : J. Adolfo Minjárez-Sosa

This SpringerBrief deals with a class of discrete-time zero-sum Markov games with Borel state and action spaces, and possibly unbounded payoffs, under discounted and average criteria, whose state process evolves according to a stochastic difference equation. The corresponding disturbance process is an observable sequence of independent and identically distributed random variables with unknown distribution for both players. Unlike the standard case, the game is played over an infinite horizon evolving as follows. At each stage, once the players have observed the state of the game, and before choosing the actions, players 1 and 2 implement a statistical estimation process to obtain estimates of the unknown distribution. Then, independently, the players adapt their decisions to such estimators to select their actions and construct their strategies. This book presents a systematic analysis on recent developments in this kind of games. Specifically, the theoretical foundations on the procedures combining statistical estimation and control techniques for the construction of strategies of the players are introduced, with illustrative examples. In this sense, the book is an essential reference for theoretical and applied researchers in the fields of stochastic control and game theory, and their applications.

Selected Topics on Continuous-time Controlled Markov Chains and Markov Games

Selected Topics on Continuous-time Controlled Markov Chains and Markov Games
Author :
Publisher : World Scientific
Total Pages : 292
Release :
ISBN-10 : 9781848168497
ISBN-13 : 1848168497
Rating : 4/5 (97 Downloads)

Synopsis Selected Topics on Continuous-time Controlled Markov Chains and Markov Games by : Tomas Prieto-Rumeau

This book concerns continuous-time controlled Markov chains, also known as continuous-time Markov decision processes. They form a class of stochastic control problems in which a single decision-maker wishes to optimize a given objective function. This book is also concerned with Markov games, where two decision-makers (or players) try to optimize their own objective function. Both decision-making processes appear in a large number of applications in economics, operations research, engineering, and computer science, among other areas. An extensive, self-contained, up-to-date analysis of basic optimality criteria (such as discounted and average reward), and advanced optimality criteria (e.g., bias, overtaking, sensitive discount, and Blackwell optimality) is presented. A particular emphasis is made on the application of the results herein: algorithmic and computational issues are discussed, and applications to population models and epidemic processes are shown. This book is addressed to students and researchers in the fields of stochastic control and stochastic games. Moreover, it could be of interest also to undergraduate and beginning graduate students because the reader is not supposed to have a high mathematical background: a working knowledge of calculus, linear algebra, probability, and continuous-time Markov chains should suffice to understand the contents of the book.

Self-Learning Control of Finite Markov Chains

Self-Learning Control of Finite Markov Chains
Author :
Publisher : CRC Press
Total Pages : 315
Release :
ISBN-10 : 9781482273274
ISBN-13 : 1482273276
Rating : 4/5 (74 Downloads)

Synopsis Self-Learning Control of Finite Markov Chains by : A.S. Poznyak

Presents a number of new and potentially useful self-learning (adaptive) control algorithms and theoretical as well as practical results for both unconstrained and constrained finite Markov chains-efficiently processing new information by adjusting the control strategies directly or indirectly.

Handbook of Markov Decision Processes

Handbook of Markov Decision Processes
Author :
Publisher : Springer Science & Business Media
Total Pages : 560
Release :
ISBN-10 : 9781461508052
ISBN-13 : 1461508053
Rating : 4/5 (52 Downloads)

Synopsis Handbook of Markov Decision Processes by : Eugene A. Feinberg

Eugene A. Feinberg Adam Shwartz This volume deals with the theory of Markov Decision Processes (MDPs) and their applications. Each chapter was written by a leading expert in the re spective area. The papers cover major research areas and methodologies, and discuss open questions and future research directions. The papers can be read independently, with the basic notation and concepts ofSection 1.2. Most chap ters should be accessible by graduate or advanced undergraduate students in fields of operations research, electrical engineering, and computer science. 1.1 AN OVERVIEW OF MARKOV DECISION PROCESSES The theory of Markov Decision Processes-also known under several other names including sequential stochastic optimization, discrete-time stochastic control, and stochastic dynamic programming-studiessequential optimization ofdiscrete time stochastic systems. The basic object is a discrete-time stochas tic system whose transition mechanism can be controlled over time. Each control policy defines the stochastic process and values of objective functions associated with this process. The goal is to select a "good" control policy. In real life, decisions that humans and computers make on all levels usually have two types ofimpacts: (i) they cost orsavetime, money, or other resources, or they bring revenues, as well as (ii) they have an impact on the future, by influencing the dynamics. In many situations, decisions with the largest immediate profit may not be good in view offuture events. MDPs model this paradigm and provide results on the structure and existence of good policies and on methods for their calculation.

Finite Approximations in Discrete-Time Stochastic Control

Finite Approximations in Discrete-Time Stochastic Control
Author :
Publisher : Birkhäuser
Total Pages : 196
Release :
ISBN-10 : 9783319790336
ISBN-13 : 3319790331
Rating : 4/5 (36 Downloads)

Synopsis Finite Approximations in Discrete-Time Stochastic Control by : Naci Saldi

In a unified form, this monograph presents fundamental results on the approximation of centralized and decentralized stochastic control problems, with uncountable state, measurement, and action spaces. It demonstrates how quantization provides a system-independent and constructive method for the reduction of a system with Borel spaces to one with finite state, measurement, and action spaces. In addition to this constructive view, the book considers both the information transmission approach for discretization of actions, and the computational approach for discretization of states and actions. Part I of the text discusses Markov decision processes and their finite-state or finite-action approximations, while Part II builds from there to finite approximations in decentralized stochastic control problems. This volume is perfect for researchers and graduate students interested in stochastic controls. With the tools presented, readers will be able to establish the convergence of approximation models to original models and the methods are general enough that researchers can build corresponding approximation results, typically with no additional assumptions.