Stochastic Automatic Differentiation: Automatic Differentiation for Monte-Carlo Simulations
45 Pages Posted: 1 Jul 2017 Last revised: 3 Nov 2018
Date Written: June 27, 2017
In this paper we re-formulate the automatic differentiation (and in particular, the backward automatic differentiation, also known as adjoint automatic differentiation, AAD) for random variables. While this is just a formal re-interpretation it allows to investigate the algorithms in the presence of stochastic operators like expectation, conditional expectation or indicator functions.
We then specify the algorithms to efficiently incorporate (conditional) expectation operators without the need to differentiate an approximation of the (conditional) expectation. Under a comparably mild assumption it is possible to retain the simplicity of the backward automatic differentiation algorithm in the presence of (conditional) expectation operators. This simplifies important applications like - in mathematical finance - the application of backward automatic differentiation to the valuation of Bermudan options or calculation of xVA's.
In addition, the framework allows to dramatically reduce the memory requirements and improve the performance of a tapeless implementation of automatic differentiation.
Keywords: Automatic Differentiation, Adjoint Automatic Differentiation Monte Carlo Simulation, American Monte Carlo, Conditional Expectation, Indicator Function, Bermudan Option, XVA
JEL Classification: C15, G13
Suggested Citation: Suggested Citation