Sequential Decision Making for Elevator Control
Abstract
In the last decade Reinforcement Learning (RL) has significantly changed the conventional control paradigm in many fields. RL approach is spreading with many applications such as autonomous driving and industry automation. Markov Decision Process (MDP) forms a mathematical idealized basis for RL if the explicit model is available. Dynamic programming allows to find an optimal policy for sequential decision making in a MDP. In this study we consider the elevator control as a sequential decision making problem, describe it as a MDP with finite state space and solve it using dynamic programming. At each decision making time step we aim to take the optimal action to minimize the total of hall call waiting times in the episodic task. We consider a sample 6-floor building and simulate the proposed method in comparison with the conventional Nearest Car Method (NCM).