Exploratory Control with Tsallis Entropy for Latent Factor Models

27 Pages Posted: 22 Nov 2022

See all articles by Ryan Francis Donnelly

Ryan Francis Donnelly

King's College London

Sebastian Jaimungal

University of Toronto - Department of Statistics

Date Written: November 14, 2022

Abstract

We study optimal control in models with latent factors where the agent controls the distribution over actions, rather than actions themselves, in both discrete and continuous time. To encourage exploration of the state space, we reward exploration with Tsallis Entropy and derive the optimal distribution over states –- which we prove is $q$-Gaussian distributed with location characterized through the solution of an FBS$\Delta$E and FBSDE in discrete and continuous time, respectively. We discuss the relation between the solutions of the optimal exploration problems and the standard dynamic optimal control solution. Finally, we develop the optimal policy in a model-agnostic setting along the lines of soft $Q$-learning. The approach may be applied in, e.g., developing more robust statistical arbitrage trading strategies.

Keywords: stochastic control, exploratory control, entropy regularization, reinforcement learning

JEL Classification: C61, C65

Suggested Citation

Donnelly, Ryan Francis and Jaimungal, Sebastian, Exploratory Control with Tsallis Entropy for Latent Factor Models (November 14, 2022). Available at SSRN: https://ssrn.com/abstract=4276921 or http://dx.doi.org/10.2139/ssrn.4276921

Ryan Francis Donnelly (Contact Author)

King's College London ( email )

Strand
London, England WC2R 2LS
United Kingdom

Sebastian Jaimungal

University of Toronto - Department of Statistics ( email )

100 St. George St.
Toronto, Ontario M5S 3G3
Canada

HOME PAGE: http://http:/sebastian.statistics.utoronto.ca

Do you have a job opening that you would like to promote on SSRN?

Paper statistics

Downloads
177
Abstract Views
430
Rank
270,500
PlumX Metrics