Modeling dynamic systems incurring stochastic disturbances for deriving a control policy is a ubiquitous task in engineering. However, in some instances obtaining a model of a system may be impractical or impossible. Alternative approaches have been developed using a simulation-based stochastic framework, in which the system interacts with its environment in real time and obtains information that can be processed to produce an optimal control policy. In this context, the problem of developing a policy for controlling the system’s behavior is formulated as a sequential decision-making problem under uncertainty. This paper considers the problem of deriving a control policy for a dynamic system with unknown dynamics in real time, formulated as a sequential decision-making under uncertainty. The evolution of the system is modeled as a controlled Markov chain. A new state-space representation model and a learning mechanism are proposed that can be used to improve system performance over time. The major difference between the existing methods and the proposed learning model is that the latter utilizes an evaluation function, which considers the expected cost that can be achieved by state transitions forward in time. The model allows decision-making based on gradually enhanced knowledge of system response as it transitions from one state to another, in conjunction with actions taken at each state. The proposed model is demonstrated on the single cart-pole balancing problem and a vehicle cruise-control problem.
A Real-Time Computational Learning Model for Sequential Decision-Making Problems Under Uncertainty
- Views Icon Views
- Share Icon Share
- Search Site
Malikopoulos, A. A., Papalambros, P. Y., and Assanis, D. N. (May 20, 2009). "A Real-Time Computational Learning Model for Sequential Decision-Making Problems Under Uncertainty." ASME. J. Dyn. Sys., Meas., Control. July 2009; 131(4): 041010. https://doi.org/10.1115/1.3117200
Download citation file: