Sequential Decision Making for Elevator Control

No Thumbnail Available

Date

2023

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

In the last decade Reinforcement Learning (RL) has significantly changed the conventional control paradigm in many fields. RL approach is spreading with many applications such as autonomous driving and industry automation. Markov Decision Process (MDP) forms a mathematical idealized basis for RL if the explicit model is available. Dynamic programming allows to find an optimal policy for sequential decision making in a MDP. In this study we consider the elevator control as a sequential decision making problem, describe it as a MDP with finite state space and solve it using dynamic programming. At each decision making time step we aim to take the optimal action to minimize the total of hall call waiting times in the episodic task. We consider a sample 6-floor building and simulate the proposed method in comparison with the conventional Nearest Car Method (NCM).

Description

Keywords

elevator control, Markov decision process, dynamic programming, optimal policy, sequential decision making

Citation

Endorsement

Review

Supplemented By

Referenced By