Trust Region Policy Optimisation, or TRPO, is a policy gradient method in reinforcement learning that avoids parameter updates that change the policy too much with a KL-divergence constraint on the size of the policy update at each iteration.
Take the case of off-policy reinforcement learning, where the policy for collecting trajectories on rollout workers is different from the policy to optimise for. The objective function in an off-policy model measures the total advantage over the state visitation distribution and actions, while the mismatch between the training data distribution and the true policy state distribution is compensated with an importance sampling estimator:
When training on policy, theoretically the policy for collecting data is same as the policy that we want to optimise. However, when rollout workers and optimisers are running in parallel asynchronously, the behaviour policy can get stale. TRPO considers this subtle difference: It labels the behaviour policy as and thus the objective function becomes:
TRPO aims to maximise the objective function subject to a trust region constraing which enforces the distance between old and new policies measured by KL-divergence to be small enough, within a parameter :
References