Partially Observable Markov Decision Processes and Robotics

Posted: 6 May 2022

See all articles by Hanna Kurniawati

Hanna Kurniawati

Australian National University (ANU)

Date Written: May 2022

Abstract

Planning under uncertainty is critical to robotics. The partially observable Markov decision process (POMDP) is a mathematical framework for such planning problems. POMDPs are powerful because of their careful quantification of the nondeterministic effects of actions and the partial observability of the states. But for the same reason, they are notorious for their high computational complexity and have been deemed impractical for robotics. However, over the past two decades, the development of sampling-based approximate solvers has led to tremendous advances in POMDP-solving capabilities. Although these solvers do not generate the optimal solution, they can compute good POMDP solutions that significantly improve the robustness of robotics systems within reasonable computational resources, thereby making POMDPs practical for many realistic robotics problems. This article presents a review of POMDPs, emphasizing computational issues that have hindered their practicality in robotics and ideas in sampling-based solvers that have alleviated such difficulties, together with lessons learned from applying POMDPs to physical robots.

Suggested Citation

Kurniawati, Hanna, Partially Observable Markov Decision Processes and Robotics (May 2022). Annual Review of Control, Robotics, & Autonomous Systems, Vol. 5, pp. 253-277, 2022, Available at SSRN: https://ssrn.com/abstract=4101059 or http://dx.doi.org/10.1146/annurev-control-042920-092451

Hanna Kurniawati (Contact Author)

Australian National University (ANU) ( email )

Canberra, Australian Capital Territory 2601
Australia

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Abstract Views
247
PlumX Metrics