Optimization of Policy Evaluation and Policy Improvement Methods in Portfolio Optimization using Quasi-Monte Carlo Methods

Loading...
Thumbnail Image

Date

2024-05-24

Authors

Orok, Gavin

Advisor

Lemieux, Christiane

Journal Title

Journal ISSN

Volume Title

Publisher

University of Waterloo

Abstract

Machine learning involves many challenging integrals that can be estimated using numerical methods. One application of these methods which has been explored in recent work is the estimation of policy gradients for reinforcement learning. They found that for many standard continuous control problems, the numerical methods randomized Quasi-Monte Carlo (RQMC) and Array-RQMC that used low-discrepancy point sets improved the efficiency of both policy evaluation and policy gradient-based policy iteration compared to standard Monte Carlo (MC). We extend this work by investigating the application of these numerical methods to model-free reinforcement learning algorithms in portfolio optimization, which are of interest because they do not rely on complex model assumptions that pose difficulties to other analytical methods. We find that RQMC significantly outperforms MC under all conditions for policy evaluation and that Array-RQMC outperforms both MC and RQMC in policy iteration with a strategic choice of the reordering function.

Description

Keywords

reinforcement learning, numerical methods, quasi-Monte Carlo, portfolio optimization, continuous control

LC Keywords

Citation