online meta-critic learning for off-policy actor-critic methods · 2020-03-12 · review of...

10
Online Meta-Critic Learning for Off-Policy Actor-Critic Methods Wei Zhou *1 Yiying Li *12 Yongxin Yang 2 Huaimin Wang 1 Timothy M. Hospedales 23 Abstract Off-Policy Actor-Critic (Off-PAC) methods have proven successful in a variety of continuous con- trol tasks. Normally, the critic’s action-value func- tion is updated using temporal-difference, and the critic in turn provides a loss for the actor that trains it to take actions with higher expected re- turn. In this paper, we introduce a novel and flexi- ble meta-critic that observes the learning process and meta-learns an additional loss for the actor that accelerates and improves actor-critic learning. Compared to the vanilla critic, the meta-critic net- work is explicitly trained to accelerate the learning process; and compared to existing meta-learning algorithms, meta-critic is rapidly learned online for a single task, rather than slowly over a family of tasks. Crucially, our meta-critic framework is designed for off-policy based learners, which cur- rently provide state-of-the-art reinforcement learn- ing sample efficiency. We demonstrate that online meta-critic learning leads to improvements in a variety of continuous control environments when combined with contemporary Off-PAC methods DDPG, TD3 and the state-of-the-art SAC. 1. Introduction Off-policy Actor-Critic (Off-PAC) methods are currently central in deep reinforcement learning (RL) research due to their greater sample efficiency compared to on-policy alter- natives. On-policy requires new trajectories to be collected for each update to the policy, and is expensive as the num- ber of gradient steps and samples per step increases with task-complexity even for contemporary TRPO (Schulman et al., 2015), PPO (Schulman et al., 2017) and A3C (Mnih et al., 2016) algorithms. Off-policy methods, such as DDPG (Lillicrap et al., 2016), TD3 (Fujimoto et al., 2018) and SAC (Haarnoja et al., * Equal contribution 1 National University of Defense Technol- ogy, Changsha, China 2 The University of Edinburgh, Edinburgh, UK 3 Samsung AI Centre, Cambridge, UK. Correspondence to: <{t.hospedales}@ed.ac.uk>. 2018b) achieve greater sample efficiency due to their ability to learn from randomly sampled historical transitions with- out a time sequence requirement, thus making better use of past experience. Their critic estimates the action-value (Q-value) function using a differentiable function approx- imator, and the actor updates its policy parameters in the direction of the approximate action-value gradient. Briefly, the critic provides a loss to guide the actor, and is trained in turn to estimate the environmental action-value under the current policy via temporal-difference learning (Sutton et al., 2009). In all these cases the learning objective function is hand-crafted and fixed. Recently meta-learning, or “learning-to-learn” has become topical as a paradigm to accelerate RL by learning aspects of the learning strategy, for example, through learning fast adaptation strategies (Finn et al., 2017; Rakelly et al., 2019; Riemer et al., 2019), exploration strategies (Gupta et al., 2018), optimization strategies (Duan et al., 2016b), losses (Houthooft et al., 2018), hyperparameters (Xu et al., 2018; Veeriah et al., 2019), and intrinsic rewards (Zheng et al., 2018). However, the majority of these works perform meta- learning on a family of tasks or environments and amortize this huge cost by deploying the trained strategy for fast learning on a new task. In this paper we introduce a novel meta-critic network to enhance existing Off-PAC learning frameworks. The meta- critic is used alongside the vanilla critic to provide a loss to guide the actor’s learning. However, compared to the vanilla critic, the meta-critic is explicitly (meta)-trained to accel- erate the learning process rather than merely estimate the action-value function. Overall, the actor is trained by gradi- ents provided by both critic and meta-critic losses, the critic is trained by temporal-difference as usual, and the meta- critic is trained to generate maximum learning performance improvements in the actor. In our framework, both the critic and meta-critic use randomly sampled off-policy transitions for efficient and effective Off-PAC learning, providing su- perior sample efficiency compared to existing on-policy meta-learners. Furthermore, we demonstrate that our meta- critic can be successfully learned online within a single task. This is in contrast to the currently widely used meta-learning research paradigm – where entire task families are required to provide enough data for meta-learning, and to provide new tasks to amortize the huge cost of meta-learning. arXiv:2003.05334v1 [cs.LG] 11 Mar 2020

Upload: others

Post on 02-Aug-2020

7 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

Wei Zhou * 1 Yiying Li * 1 2 Yongxin Yang 2 Huaimin Wang 1 Timothy M. Hospedales 2 3

AbstractOff-Policy Actor-Critic (Off-PAC) methods haveproven successful in a variety of continuous con-trol tasks. Normally, the critic’s action-value func-tion is updated using temporal-difference, and thecritic in turn provides a loss for the actor thattrains it to take actions with higher expected re-turn. In this paper, we introduce a novel and flexi-ble meta-critic that observes the learning processand meta-learns an additional loss for the actorthat accelerates and improves actor-critic learning.Compared to the vanilla critic, the meta-critic net-work is explicitly trained to accelerate the learningprocess; and compared to existing meta-learningalgorithms, meta-critic is rapidly learned onlinefor a single task, rather than slowly over a familyof tasks. Crucially, our meta-critic framework isdesigned for off-policy based learners, which cur-rently provide state-of-the-art reinforcement learn-ing sample efficiency. We demonstrate that onlinemeta-critic learning leads to improvements in avariety of continuous control environments whencombined with contemporary Off-PAC methodsDDPG, TD3 and the state-of-the-art SAC.

1. IntroductionOff-policy Actor-Critic (Off-PAC) methods are currentlycentral in deep reinforcement learning (RL) research due totheir greater sample efficiency compared to on-policy alter-natives. On-policy requires new trajectories to be collectedfor each update to the policy, and is expensive as the num-ber of gradient steps and samples per step increases withtask-complexity even for contemporary TRPO (Schulmanet al., 2015), PPO (Schulman et al., 2017) and A3C (Mnihet al., 2016) algorithms.

Off-policy methods, such as DDPG (Lillicrap et al., 2016),TD3 (Fujimoto et al., 2018) and SAC (Haarnoja et al.,

*Equal contribution 1National University of Defense Technol-ogy, Changsha, China 2The University of Edinburgh, Edinburgh,UK 3Samsung AI Centre, Cambridge, UK. Correspondence to:<{t.hospedales}@ed.ac.uk>.

2018b) achieve greater sample efficiency due to their abilityto learn from randomly sampled historical transitions with-out a time sequence requirement, thus making better useof past experience. Their critic estimates the action-value(Q-value) function using a differentiable function approx-imator, and the actor updates its policy parameters in thedirection of the approximate action-value gradient. Briefly,the critic provides a loss to guide the actor, and is trained inturn to estimate the environmental action-value under thecurrent policy via temporal-difference learning (Sutton et al.,2009). In all these cases the learning objective function ishand-crafted and fixed.

Recently meta-learning, or “learning-to-learn” has becometopical as a paradigm to accelerate RL by learning aspectsof the learning strategy, for example, through learning fastadaptation strategies (Finn et al., 2017; Rakelly et al., 2019;Riemer et al., 2019), exploration strategies (Gupta et al.,2018), optimization strategies (Duan et al., 2016b), losses(Houthooft et al., 2018), hyperparameters (Xu et al., 2018;Veeriah et al., 2019), and intrinsic rewards (Zheng et al.,2018). However, the majority of these works perform meta-learning on a family of tasks or environments and amortizethis huge cost by deploying the trained strategy for fastlearning on a new task.

In this paper we introduce a novel meta-critic network toenhance existing Off-PAC learning frameworks. The meta-critic is used alongside the vanilla critic to provide a loss toguide the actor’s learning. However, compared to the vanillacritic, the meta-critic is explicitly (meta)-trained to accel-erate the learning process rather than merely estimate theaction-value function. Overall, the actor is trained by gradi-ents provided by both critic and meta-critic losses, the criticis trained by temporal-difference as usual, and the meta-critic is trained to generate maximum learning performanceimprovements in the actor. In our framework, both the criticand meta-critic use randomly sampled off-policy transitionsfor efficient and effective Off-PAC learning, providing su-perior sample efficiency compared to existing on-policymeta-learners. Furthermore, we demonstrate that our meta-critic can be successfully learned online within a single task.This is in contrast to the currently widely used meta-learningresearch paradigm – where entire task families are requiredto provide enough data for meta-learning, and to providenew tasks to amortize the huge cost of meta-learning.

arX

iv:2

003.

0533

4v1

[cs

.LG

] 1

1 M

ar 2

020

Page 2: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

Essentially our framework meta-learns an auxiliary lossfunction, which can be seen as an intrinsic motivationtowards optimum learning progress (Oudeyer & Kaplan,2009). As analogously observed in several recent meta-learning studies (Franceschi et al., 2018), our loss-learningcan be formalized as a bi-level optimization problem withthe upper level being meta-critic learning, and lower level be-ing conventional learning. We solve this joint optimizationby iteratively updating the meta-critic and base learner on-line while solving a single task. Our strategy is thus relatedto the meta-loss learning in EPG (Houthooft et al., 2018),but learned online rather than offline, and integrated withOff-PAC rather than their on-policy policy-gradient learning.The most related prior work is LIRPG (Zheng et al., 2018),which meta-learns an intrinsic reward online. However, theirintrinsic reward just provides a helpful scalar offset to theenvironmental reward for on-policy trajectory optimizationvia policy-gradient (Sutton et al., 2000). In contrast ourmeta-critic provides a loss for direct actor optimization justbased on sampled transitions, and thus achieves dramati-cally better sample efficiency than LIRPG reward learningin practice. We evaluate our framework on several con-temporary continuous control benchmarks and demonstratethat online meta-critic learning can be integrated with andimprove a selection of contemporary Off-PAC algorithmsincluding DDPG, TD3 and SAC.

2. Background and Related WorkPolicy-Gradient (PG) Methods. On-policy methods usu-ally update actor parameters in the direction of greater cumu-lative reward. However, on-policy methods need to interactwith the environment in a sequential manner to accumulaterewards and the expected reward is generally not differen-tiable due to environment dynamics. Even exploiting trickslike importance sampling and improved application of A2C(Zheng et al., 2018), the use of full trajectories is less ef-fective than off-policy transitions, as the trajectory needs aseries of continuous transitions in time. Off-policy actor-critic architectures aim to provide better sample efficiencyby reusing past experience (previously collected transitions).DDPG (Lillicrap et al., 2016) borrows two main ideas fromDeep Q Networks (Mnih et al., 2013; 2015): a big replaybuffer and a target Q network to give consistent targets dur-ing temporal-difference backups. TD3 (Twin Delayed DeepDeterministic policy gradient) (Fujimoto et al., 2018) devel-ops a variant of Double Q-learning by taking the minimumvalue between a pair of critics to limit over-estimation. SAC(Soft Actor-Critic) (Haarnoja et al., 2018a;b) proposes amaximum entropy RL framework where its stochastic actoraims to simultaneously maximize expected action-value andentropy. The latest version of SAC (Haarnoja et al., 2018b)also includes the “the minimum value between both critics”idea in its implementation.

Meta Learning for RL. Meta-learning (a.k.a. learning tolearn) (Santoro et al., 2016; Finn et al., 2017) has received aresurgence in interest recently due to its potential to improvelearning performance, and especially sample-efficiency inRL (Gupta et al., 2018). Several studies learn optimizersthat provide policy updates with respect to known loss orreward functions (Andrychowicz et al., 2016; Duan et al.,2016b; Meier et al., 2018). A few studies learn hyperparam-eters (Xu et al., 2018; Veeriah et al., 2019), loss functions(Houthooft et al., 2018; Sung et al., 2017) or rewards (Zhenget al., 2018) that steer the learning of standard optimizers.Our meta-critic framework is in the category of loss-functionmeta-learning, but unlike most of these we are able to meta-learn the loss function online in parallel to learning a singleextrinsic task rather. No costly offline learning on a taskfamily is required as in Houthooft et al. (2018); Sung et al.(2017). Most current Meta-RL methods are based on on-policy policy-gradient, limiting the sample efficiency. Forexample, while LIRPG (Zheng et al., 2018) is one of the fewprior works to attempt online meta-learning, it is ineffectivein practice due to only providing a scalar reward incrementrather than a loss for direct optimization. A few meta-RLstudies have begun to address off-policy RL, for conven-tional multi-task meta-learning (Rakelly et al., 2019) andfor optimising transfer vs forgetting in continual learning ofmultiple tasks (Riemer et al., 2019). The contribution of ourMeta-Critic is to enhance state-of-the-art Off-PAC RL withsingle-task online meta-learning.

Loss Learning. Loss learning has been exploited in ‘learn-ing to teach’ (Wu et al., 2018) and surrogate loss learning(Huang et al., 2019; Grabocka et al., 2019) where a teachernetwork predicts the parameters of a manually designed lossin supervised learning. In contrast our meta-critic is itselfa differentiable loss, and is designed for use in reinforce-ment learning. Other applications learn losses that improvemodel robustness to out of distribution samples (Li et al.,2019; Balaji et al., 2018). Our loss learning architectureis related to (Li et al., 2019), but designed for acceleratingsingle-task Off-PAC RL rather than improving robustnessin multi-domain supervised learning.

3. MethodologyWe aim to learn a meta-critic that provides an auxiliary lossLauxω to assist the actor’s learning of a task. The auxiliary

loss parameters ω are optimized in a meta-learning process.The main policy loss Lmain and auxiliary loss Laux

ω train theactor πφ off-policy via stochastic gradient descent.

3.1. Review of Off-Policy Actor-Critic RL

Reinforcement learning involves an agent interacting withthe environment E. At each time t, the agent receivesan observation st, takes a (possibly stochastic) action at

Page 3: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

based on its policy π : S → A, and receives a scalarreward rt and new state of the environment st+1. We call(st, at, rt, st+1) as a single point transition. The objectiveof RL is to find the optimal policy πφ, which maximizes theexpected cumulative return J .

In on-policy RL, J is defined as the discountedepisodic return based on a sequential trajectory overthe horizon H: (s0, a0, r0, · · · , sH , aH , rH). J =

Ert,st∼E,at∼π[∑H

t=0 γtrt

]. In the usual implementation

of A2C, r is represented by a surrogate state-value V (st)from its critic. Since J is only a scalar value, the gradient ofJ with respect to policy parameters φ has to be optimizedunder the policy gradient theorem (Sutton et al., 2000):∇φJ(φ) = E [J ∇φ log πφ(at|st)].

In off-policy RL (e.g., DDPG, TD3, SAC) which is ourfocus in this paper, parameterized policies πφ can be directlyupdated by defining the actor loss in terms of the expectedreturn J(φ) and taking its gradient ∇φJ(φ), where J(φ)depends on the action-value Qθ(s, a). The main loss Lmain

provided by the vanilla critic is thus

Lmain = −J(φ) = −Es∼pπQθ(s, a)|a=πφ(s), (1)

where we follow the notation in TD3 and SAC that φ and θdenote actors and critics respectively.

The main loss is calculated by a mini-batch of transitionsrandomly sampled from the replay buffer. The actor’s pol-icy network is updated as ∆φ = α∇φLmain, following thecritic’s gradient to increase the likelihood of actions thatachieve a higher Q-value. Meanwhile, the critic uses Q-learning updates to estimate the action-value function:

θ ← arg minθ

(Qθ(st, at)− rt − γQθ(st+1, π(st+1))2.

(2)

3.2. Algorithm Overview

Our meta-learning goal is to train an auxiliary meta-criticnetwork Laux

ω that in turn enhances actor learning. Specif-ically, it should lead to the actor φ having improved per-formance on the main task Lmain when following gradientsprovided by the meta-critic as well as those provided bythe main task. This can be seen as a bi-level optimizationproblem1 (Franceschi et al., 2018; Rajeswaran et al., 2019)of the form:

ω = arg minω

Lmeta(dval;φ∗)

s.t. φ∗ = arg minφ

(Lmain(dtrn;φ) + Lauxω (dtrn;φ)),

(3)

where we can assume Lmeta(·) = Lmain(·) for now. Here thelower-level optimization trains the actor φ to minimize both

1See Franceschi et al. (2018) for a discussion on convergenceof bi-level algorithms.

Figure 1. Meta-critic for Off-PAC. The agent uses data sampledfrom the replay buffer during meta-train and meta-test. Actor pa-rameters are first updated using only vanilla critic, or both vanilla-and meta-critic. Meta-critic parameters are updated by meta-loss.

the main task and meta-critic-provided losses on some train-ing samples. The upper-level optimization further requiresthe meta-critic ω to have produced a learned actor φ∗ thatminimizes a meta-loss that measures the actor’s main taskperformance on a second set of validation samples, afterbeing trained by the meta-critic. Note that in principle thelower-level optimization could purely rely on Laux

ω analo-gously to the procedure in EPG (Houthooft et al., 2018), butwe find that optimizing their sum greatly increases learningstability and speed. Eq. (3) is satisfied when the meta-criticsuccessfully trains the actor for good performance on themain task as measured by validation meta loss. Note thatthe vanilla critic update is also in the lower loop, but as itupdates as usual, so we focus on the actor and meta-criticoptimization for simplicity of exposition.

In this setup the meta-critic is a neural network hω(dtrn;φ)that takes as input some featurisation of the actor φ and thestates and actions in dtrn. This auxiliary neural networkmust produce a scalar output, which we can then treat asa loss Laux

ω := hω, and must be differentiable with respectto φ. We next discuss the overall optimization flow, anddiscuss the specific meta-critic architecture later.

Meta-Optimization Flow. To optimize Eq. (3), we itera-tively update the meta-critic parameters ω (upper-level) andactor and vanilla-critic parameters φ and θ (lower-level). Ateach iteration, we perform: (i) Meta-train: Sample a mini-batch of transitions and putatively update policy φ accordingto the main Lmain and meta-critic Laux

ω losses. (ii) Meta-test:Sample another mini-batch of transitions to evaluate theperformance of the updated policy according to Lmeta. (iii)Meta-optimization: Update the meta-critic parameters ωto maximize the performance on the validation batch, andperform the real actor update according to both losses. Inthis way the meta-critic is trained online and in parallel to

Page 4: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

Algorithm 1 Online Meta-Critic Learning for Off-PAC RLInput: Initialized parameters φ, θ, ω, replay buffer D ← ∅Output: optimized φ

1 begin2 for each iteration do3 for each environment step do4 at ∼ πφ(at|st)5 st+1 ∼ p(st+1|st, at), rt6 D ← D ∪ {(st, at, rt, st+1)}7 for each gradient step do8 Sample mini-batch dtrn from D9 θ ← θ − λ∇θJQ(θ) // Update critic

10 Meta-train:11 Lmain ← Eqs. (1), (8) or (9) // Main actor12 loss13 Laux

ω ← Eqs. (6) or (7) // Auxiliary actor loss14 φold = φ− α∇φLmain

15 φnew = φold − α∇φLauxω

16 meta-test:17 Sample mini-batch dval from D18 Lmeta(dval;φold, φnew)← Eq. (4)19 meta-optimization:20 φ← φ− η(∇φLmain +∇φLaux

ω ) // Update21 actor22 ω ← ω − η∇ωLmeta // Update meta-critic

the actor so that they co-evolve. Figure 1 and Alg. ?? sum-marize the process and the details of each step are explainednext.

Updating Actor Parameters (φ). During meta-train,we randomly sample a mini-batch of transitions dtrn ={(si, ai, ri, si+1)} with batch size N from the replay bufferD. We then update the policy using both losses as: φnew =

φ − η ∂ Lmain(dtrn)∂φ − η ∂ L

auxω (dtrn)∂φ . We also compute a sep-

arate update φold = φ− η ∂Lmain(dtrn)∂φ that only makes use

of the vanilla loss. If the meta-critic provided a beneficialsource of loss, φnew should be a better parameter than φ,and in particular it should be a better parameter than φold.We will use this comparison in the next meta-test step.

Updating Meta-Critic Parameters (ω). To train the meta-critic network, we sample another mini-batch of transitions:dval = {(sval

i , avali , r

vali , s

vali+1)} with batch size M . The

use of a validation batch for bi-level meta-optimization(Franceschi et al., 2018; Rajeswaran et al., 2019) ensuresthe meta-learned component does not overfit. Since ourframework is off-policy, this does not incur any sample-efficiency cost. The meta-critic is then updated by a metaloss ω ← arg minω L

meta(dval;φnew), which could in prin-ciple be the same as the main loss Lmeta = Lmain. However,

we find it helpful for optimization efficiency to optimizethe difference between the updates with- and without meta-critic’s input. Specifically, we use

Lmeta = tanh(Lmain(dval;φnew)−Lmain(dval;φold)), (4)

which is simply a monotonic re-centering and re-scaling ofLmain2. This leads to

ω ← arg minω

tanh(Lmain(dval;φnew)−Lmain(dval;φold)).

(5)Note that here the updated actor φnew has dependence onthe feedback given by meta-critic ω and φold does not. Thusonly the first term is optimized for ω. In his setup theLmain(dval;φnew) term should obtain high reward/low losson the validation batch and the latter provides a baseline,analogous to the baseline widely used to accelerate and sta-bilize policy-gradient RL. The use of tanh reflects the ideaof diminishing marginal utility, and ensures that the meta-loss range is always nicely distributed in (−1, 1). In essence,the meta-loss is for the agent to ask itself the question: “Didmeta-critic improve validation performance?”, and adjuststhe meta-critic (auxiliary task) parameters accordingly.

Designing Meta-Critic (hω). The meta-critic network hωimplements the auxiliary loss for the actor. The design-space for hω has several requirements: (i) Its input mustdepend on the policy parameters φ, because this auxiliaryloss is also used to update policy network. (ii) It should bepermutation invariant to transitions in dtrn, i.e., it shouldnot make a difference if we feed the randomly sampledtransitions indexed [1,2,3] or [3,2,1]. The naivest way toachieve (i) is given in MetaReg (Balaji et al., 2018) whichmeta-learns a parameter regularizer: hω(φ) =

∑i ωi|φi|.

Although this form of hω acts directly on φ, it does notexploit state information, and introduces a large number ofparameters in hω, as φ may be a high-dimensional neuralnetwork. Therefore, we design a more efficient and effec-tive form of hω that also meets both of these requirements.Similar to the feature extractor in supervised learning, theactor needs to analyse and extract information from statesfor decision-making. We assume the policy network can berepresented as πφ(s) = π(π(s)) and decomposed into thefeature extraction πφ and decision-making πφ (i.e., the lastlayer of the full policy network) modules. Thus the outputof the penultimate layer of full policy network is just the out-put of feature extraction πφ(s), and such output of featurejointly encodes φ and s. Given this encoding, we implementhw(dtrn;φ) as a three-layer multi-layer perceptron (MLP)whose input is the extracted feature from πφ(s). Here weconsider two designs for meta-critic (hω): using our jointfeature alone (Eq. (6)) or augmenting the joint feature with

2Note that the parameter ω that minimises Lmeta as Eq. 4 is alsothe minimum of Lmain and vice-versa.

Page 5: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

states and actions (Eq. (7)):

(i) hw(dtrn;φ) =1

N

N∑i=1

MLPω(πφ(si)), (6)

(ii) hw(dtrn;φ) =1

N

N∑i=1

MLPω(πφ(si), si, ai).

(7)

hω is to work out the auxiliary loss based on such batch-wiseset-embdedding (Zaheer et al., 2017) of our joint actor-statefeature. That is to say, dtrn is a randomly sampled mini-batch transitions from the replay buffer, and then the s (anda) of the transitions are inputted to the hω network in apermutation invariant way, and finally we can obtain theauxiliary loss for this batch dtrn. Here, our design of Eq. (7)also includes the cues features in LIRPG and EPG where siand ai are used as the input of their learned reward and lossrespectively. We set a softplus activation to the final layerof hω , following the idea in TD3 that the vanilla critic mayover-estimate and so the introduction of a non-negative actorauxiliary loss can mitigate such over-estimation. Moreover,we point out that only si (and ai) from dtrn are used whencalculating Lmain and Laux

ω for the actor, while si, ai, ri andsi+1 are all used for optimizing the vanilla critic.

Implementation on DDPG, TD3 and SAC. Our meta-critic module can be incorporated in the main Off-PACmethods DDPG, TD3 and SAC. In our framework, thesealgorithms differ only in their definitions of Lmain, and themeta-critic implementation is otherwise exactly the samefor each. Further implementation details can be found in thesupplementary material.

TD3 (Fujimoto et al., 2018) borrows the Double Q-learningidea and use the minimum value between both critics tomake unbiased value estimations. At the same time, compu-tational cost is obtained by using a single actor optimizedwith respect to Qθ1 . Thus the corresponding Lmain for actorbecomes:

Lmain = −Es∼pπQθ1(s, a)|a=πφ(s). (8)

In SAC, two key ingredients are considered for the actor:maximizing the policy entropy and automatic temperaturehyper-parameter regulation. At the same time, the latestversion of SAC (Haarnoja et al., 2018b) also draws lessonsfrom “taking the minimum value between both critics”. TheLmain for SAC actor is:Lmain = Es∼pπ [α log (πφ(a|s))−Qθ(s, a)|a=πφ(s)]. (9)

4. Experiments and EvaluationThe goal of our experimental evaluation is to demonstratethe versatility of our meta-critic module in integration withseveral prior Off-PAC algorithms, and its efficacy in improv-ing their respective performance. We use the open-source

implementations of DDPG, TD3 and SAC algorithms asour baselines, and denote their enhancements by meta-criticas DDPG-MC, TD3-MC, SAC-MC respectively. All -MCagents have both their built-in vanilla critic, and the meta-critic that we propose. We take Eq. (6) as the default meta-critic architecture hω , and we compare the alternative in thelater ablation study. For our implementation of meta-critic,we use a three-layer neural network with an input dimen-sion of π (300 in DDPG and TD3, 256 in SAC), two hiddenfeed-forward layers of 100 hidden nodes each, and ReLUnon-linearity between layers.

We evaluate the methods on a suite of seven MuJoCo con-tinuous control tasks (Todorov et al., 2012) in OpenAI Gym(Brockman et al., 2016), two MuJoCo tasks in rllab (Duanet al., 2016a), and the simulated racing car environmentTORCS (Loiacono et al., 2013). For MuJoCo-Gym, we usethe latest V2 tasks instead of V1 used in TD3 and the old-SAC (Haarnoja et al., 2018a) work without any modificationto their original environment or reward.

Implementation Details. For DDPG, we use the open-source implementation “OurDDPG” 3 which is the re-tunedversion of DDPG implemented in Fujimoto et al. (2018)with the same hyper-parameters of the actor and critic forMuJoCo tasks. For TD3 and SAC, we use the open-sourceimplementations of TD3 4 and SAC 5. In MuJoCo cases weintegrate our meta-critic with learning rate 0.001. The hyper-parameters for TORCS can be found in the supplementarymaterial.

4.1. Evaluation of Meta-Critic Off-PAC learning

DDPG Figure 2 shows the learning curves of DDPG andDDPG-MC. The experimental results corresponding to eachtask are averaged over 5 random seeds (trials) and networkinitialisations, and the standard deviation confidence inter-vals are represented as shaded regions over the time steps.Following (Fujimoto et al., 2018), curves are uniformlysmoothed for clarity (window size=10 for TORCS, 30 forothers). We run the gym-MuJoCo experiments for 1-10million depending on to environment, rllab experiments for3 million steps and TORCS experiment for 100 thousandsteps. Every 1000 steps we evaluate our policy over 10episodes with no exploration noise.

From the learning curves in Figure 2, we can see that DDPG-MC generally outperforms the corresponding DDPG base-line in terms of the learning speed and asymptotic perfor-mance. Furthermore, it usually has smaller variance. Thesummary results for all tasks in terms of max average re-turn are given in Table 1. -MC usually provides consis-

3https://github.com/sfujim/TD3/blob/master/OurDDPG.py4https://github.com/sfujim/TD3/blob/master/TD3.py5https://github.com/pranz24/pytorch-soft-actor-critic

Page 6: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0

2000

4000

6000

8000

10000

Aver

age

Rewa

rdHalfCheetah-v2

DDPGDDPG-MC

0.0 0.2 0.4 0.6 0.8 1.0Time Steps (1e6)

0250500750

10001250150017502000

Aver

age

Rewa

rd

Hopper-v2DDPGDDPG-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0500

100015002000250030003500

Aver

age

Rewa

rd

Walker2d-v2DDPGDDPG-MC

0 2 4 6 8 10Time Steps (1e6)

1000

0

1000

2000

3000

4000

Aver

age

Rewa

rd

Ant-v2DDPGDDPG-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0100020003000400050006000

Aver

age

Rewa

rd

HalfCheetah (rllab)

DDPGDDPG-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0

500

1000

1500

2000

2500

Aver

age

Rewa

rd

Ant (rllab)DDPGDDPG-MC

0 2 4 6 8 10Time Steps(1e4)

0

2000

4000

6000

8000

Aver

age

Rewa

rd

TORCSDDPGDDPG-MC

Figure 2. Learning curve Mean and Std-Deviation of vanilla DDPG and meta-critic enhanced DDPG-MC for continuous control tasks.

Table 1. Max Average Return over 5 trials over all time steps. Max value for each task is bolded.

Environment DDPG DDPG-MC TD3 TD3-MC SAC SAC-MC PPO PPO-LIRPGHalfCheetah 8440.2 10187.5 12735.7 15064.0 16651.8 16815.9 2061.5 1882.6Hopper 2097.5 3253.6 3807.0 3854.3 3610.6 3738.4 3762.0 2750.0Walker2d 2920.1 3753.7 5942.7 5955.5 6398.8 7164.9 4432.6 3652.9Ant 2375.4 3661.1 5914.8 6280.0 6954.4 7204.3 684.2 23.6Reacher -3.6 -3.7 -3.0 -2.9 -2.8 -2.7 -6.08 -7.53InvPend 1000.0 1000.0 1000.0 1000.0 1000.0 1000.0 988.2 971.6InvDouPend 9307.5 9326.5 9357.4 9358.8 9359.6 9359.6 7266.0 6974.9HalfCheetah(rllab) 5860.8 6254.6 8029.6 8552.1 10011.0 10597.0 - -Ant(rllab) 2300.8 2721.1 3672.6 4776.8 8014.8 8353.8 - -TORCS 6188.1 9353.3 14841.7 33684.2 25020.6 32869.0 - -

tently higher max return. We select the seven tasks shownin Figure 2 for plotting, because the other MuJoCo tasks“Reacher”, “InvertedPendulum” and “InvertedDoublePen-dulum” have environmental reward upper bounds which allmethods reach quickly without obvious differences.

TD3 and SAC Figure 3 reports the learning curves for TD3.For some tasks vanilla TD3 performance declines in the longrun, while our TD3-MC shows improved stability with muchhigher asymptotic performance. Generally speaking, thelearning curves show that TD3-MC providing comparableor better learning performance in each case, while Table 1shows the clear improvement in the max average return.

Figure 4 report the learning curves of SAC. Note that weuse the most recent update of SAC (Haarnoja et al., 2018b),which can be regarded as the combination SAC+TD3. Al-though this SAC+TD3 is arguably the strongest existingmethod, SAC-MC still gives a clear boost on the asymptoticperformance for several of the tasks.

Comparison vs PPO-LIRPG Intrinsic Reward Learningfor PPO (Zheng et al., 2018) is the most related method toour work in performing online single-task meta-learning ofan auxiliary reward/loss via a neural network. The originalPPO-LIRPG study evaluated on a modified environment

with hidden rewards. Here we apply it to the standard un-modified learning tasks that we aim to improve. The resultsin Table 1 demonstrate that: (i) In this conventional setting,PPO-LIRPG worsens rather than improves basic PPO per-formance. (ii) Overall Off-PAC methods generally performbetter than on-policy PPO for most environments. Thisshows the importance of our meta-learning contribution tothe off-policy setting. In general Meta-Critic is preferredcompared to PPO-LIRPG because the latter only provides ascalar reward bonus only influences the policy indirectly viahigh-variance policy-gradient updates, while Meta-Criticprovides a direct loss.

Summary Table 1 and Figure 5 summarize all the resultsin terms of max average return. We can see that SAC-MCgenerally performs best; the Meta-Critic-enhanced methodsare generally comparable or better than their correspond-ing vanilla alternatives; and Meta-Critic usually providesimproved variance in return compared to the baselines.

4.2. Further Analysis

Loss Analysis. To analyse the learning dynamics of ouralgorithm, we take a simple learning problem, tabular MDP(Duan et al., 2016b) (|S| = 2, |A| = 2) as an example, and

Page 7: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

0 2 4 6 8 10Time Steps (1e6)

02000400060008000

10000120001400016000

Aver

age

Rewa

rdHalfCheetah-v2

TD3TD3-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0500

100015002000250030003500

Aver

age

Rewa

rd

Hopper-v2

TD3TD3-MC

0.0 0.2 0.4 0.6 0.8 1.0Time Steps (1e6)

0

1000

2000

3000

4000

5000

Aver

age

Rewa

rd

Walker2d-v2

TD3TD3-MC

0 2 4 6 8 10Time Steps (1e6)

0100020003000400050006000

Aver

age

Rewa

rd

Ant-v2

TD3TD3-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0

2000

4000

6000

8000

Aver

age

Rewa

rd

HalfCheetah (rllab)

TD3TD3-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

0

1000

2000

3000

4000

Aver

age

Rewa

rd

Ant (rllab)TD3TD3-MC

0 2 4 6 8 10Time Steps(1e4)

05000

1000015000200002500030000

Aver

age

Rewa

rd

TORCSTD3TD3-MC

Figure 3. Learning curve Mean and Std-Deviation of vanilla TD3 and meta-critic enhanced TD3-MC for continuous control tasks.

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

02000400060008000

10000120001400016000

Aver

age

Rewa

rd

HalfCheetah-v2

SACSAC-MC

0.0 0.2 0.4 0.6 0.8 1.0Time Steps (1e6)

50010001500200025003000

Aver

age

Rewa

rd

Hopper-v2

SACSAC-MC

0 2 4 6 8 10Time Steps (1e6)

01000200030004000500060007000

Aver

age

Rewa

rd

Walker2d-v2

SACSAC-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0Time Steps (1e6)

01000200030004000500060007000

Aver

age

Rewa

rd

Ant-v2

SACSAC-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

02000400060008000

10000

Aver

age

Rewa

rd

HalfCheetah (rllab)

SACSAC-MC

0.0 0.5 1.0 1.5 2.0 2.5 3.0Time Steps (1e6)

010002000300040005000600070008000

Aver

age

Rewa

rd

Ant (rllab)

SACSAC-MC

0 2 4 6 8 10Time Steps(1e4)

05000

1000015000200002500030000

Aver

age

Rewa

rd

TORCSSACSAC-MC

Figure 4. Learning curve Mean and Std-Deviation of vanilla SAC and meta-critic enhanced SAC-MC for continuous control tasks.

compare DDPG vs DDPG-MC. Figure 6 reports the mainloss Lmain curves of actor and the loss curve of hω (i.e.,Lauxω ) and Lmeta over 5 trials for DDPG-MC. In addition,

we plot the model optimization trajectories (pink dots) viaa 2D weight-space slice in Figure 7. These are plottedover the average reward surface for this slice. Followingthe the neural network visualization method of Li et al.(2018), we calculate the subspace to plot as: Let φi denotemodel parameters at episode i and the final estimate asφn (we set n = 100). We apply PCA to the matrix M =[φ0−φn, . . . , φn−1−φn], and take the two most explanatorydirections of this optimization path. Model parameters arethen projected onto the plane defined by these directions forplotting; and models at each point on this plane are denselyevaluated to calculate average reward.

We see some interesting behavior in these results. Figure 6shows: (i) DDPG-MC shows faster convergence to a lowervalue of Lmain, demonstrating the auxiliary loss’s ability toaccelerate learning. (ii) The meta-loss (which correspondsto the success of the meta-critic in improving actor learning)shows a pattern: ‘positive’→ ‘negative’→ ‘converging to

zero’. This pattern is expected because: At the start, hωis randomly initialised and knows little about how to helpthe actor, thus φold-based model outperforms φnew-basedmodel. Then as ω is trained by the meta-loss, it begins tomake φnew better than φold. In the late stage, meta-loss goestowards zero, which indicates that all of hω’s knowledgehas been distilled to help the actor. (iii) The auxiliary lossconverges smoothly under the supervision of the meta-loss.In Figure 7 (iv) DDPG-MC has a very direct optimizationtrajectory to the high reward zone, while the vanilla DDPGmodel moves slowly through the low reward space andbefore finally finding the direction to the high-reward zone.

Ablation on hω design. To analyse the designs of hω, werun Walker2d experiments under SAC-MC with the alterna-tive hω architecture from Eq. (7) or MetaReg (Balaji et al.,2018) format (input actor parameters directly). As shown inTable 2, we record the max average return and sum averagereturn (regarded as the area under the average reward curve)of all evaluations during all time steps. Eq. (7) achievesthe highest max average return and our default hω (Eq. (6))attains the highest mean average return. We can also see

Page 8: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG

2000400060008000

10000120001400016000

HalfCheetah-v2

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG0

1000

2000

3000

4000Hopper-v2

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG0

1000200030004000500060007000

Walker2d-v2

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG0

1000200030004000500060007000

Ant-v2

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG5000

6000

7000

8000

9000

10000

11000HalfCheetah(rllab)

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG

20003000400050006000700080009000 Ant(rllab)

DDPGDDPG-MC TD3

TD3-MC SACSAC-MC PPO

PPO-LIRPG5000

10000150002000025000300003500040000

TORCS

Figure 5. Box plots of the Max Average Return over 5 trials of all time steps.

Table 2. Ablation study on different designs of meta-critic (auxiliary-loss) and meta-loss applied to SAC on Walker2d. Max and SumAverage Return over 5 trials of all time steps. Max value in each row is bolded.

SACLmeta : φnew − φold Lmeta : φnew

hω(πφ) hω(πφ, s, a) hω(φ) hω(πφ)

Max Average Return 6398.8 ± 289.2 7164.9 ± 151.3 7423.8 ± 780.2 6644.3 ± 1815.6 6456.1 ± 424.8Sum Average Return 53,695,678 61,672,039 57,364,405 58,875,184 52,446,717

0 2000 4000 6000 8000 10000Time Steps

10080604020

0

Mai

n Lo

ss o

f Act

or

DDPGDDPG-MC

(a) Main loss

0 2000 4000 6000 8000 10000Time Steps

1.041.051.061.071.081.091.10

Aux

Loss

(b) Auxiliary loss

0 2000 4000 6000 8000 10000Time Steps

0.1500.1250.1000.0750.0500.0250.0000.0250.050

Met

a Lo

ss

(c) Meta-loss

Figure 6. Loss analysis of our algorithm.

some improvement for hω(φ) using MetaReg format, butthe huge number (73484) of parameters is expensive. Over-all, all meta-critic module designs provides at least a smallimprovement on vanilla SAC.

Ablation on baseline in meta-loss. In Eq. (4), weuse Lmain(dval;φold) as a baseline to improve numeri-cal stability of the gradient update. To evaluate this de-sign, we remove the φold baseline and optimize ω ←arg minω tanh(Lmain(dval;φnew)). The last column in Ta-ble 2 shows that this barely improves on vanilla SAC, vali-dating our design choice to use a baseline.

5. ConclusionWe present Meta-Critic, an auxiliary critic module for Off-PAC methods that can be meta-learned online during singletask learning. The meta-critic is trained to generate gra-

10 8 6 4 2 0 2 4

4

2

0

2

4

6

8

1011

0.000

110.000

110.000

110.000

110.000 125.0

00

140.0

00

155.0

00

170.0

00

185.0

00

200.0

00

80

95

110

125

140

155

170

185

200

215

4 2 0 2 4 6 8 10

4

2

0

2

4

6

8

10

95.00

0

110.000

110.

000

110.000110.

000110.000

110.000

110.000

110.000

125.

000

125.

000

140.0

00

140.

000

155.0

00

155.

000

170.000

170.

000

185.0

00

185.

000

200.000

200.

000

80

95

110

125

140

155

170

185

200

215

Figure 7. Visualization of optimisation dynamics. 2D projectionof model trajectories overlaid on top of actual reward contours.Left: vanilla DDPG; Right: DDPG-MC. Meta-Critic enables moredirect movement to high-reward zone of parameter space.

dients that improve the actor’s learning performance overtime, and leads to long run performance gains in continuouscontrol. The meta-critic module can be flexibly incorpo-rated into various contemporary Off-PAC methods to boostperformance. In future work, we plan to apply the meta-critic to conventional meta-learning with multi-task andmulti-domain RL.

ReferencesAndrychowicz, M., Denil, M., Gomez, S., Hoffman, M. W.,

Pfau, D., Schaul, T., Shillingford, B., and Freitas, N. D.Learning to learn by gradient descent by gradient descent.

Page 9: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

In NIPS, 2016.

Balaji, Y., Sankaranarayanan, S., and Chellappa, R. Metareg:Towards domain generalization using meta-regularization.In NeurIPS, 2018.

Brockman, G., Cheung, V., Pettersson, L., Schneider, J.,Schulman, J., Tang, J., and Zaremba, W. Openai gym. InarXiv, 2016.

Duan, Y., Chen, X., Houthooft, R., Schulman, J., andAbbeel, P. Benchmarking deep reinforcement learningfor continuous control. In ICML, 2016a.

Duan, Y., Schulman, J., Chen, X., Bartlett, P. L., Sutskever,I., and Abbeel, P. Rl 2: Fast reinforcement learning viaslow reinforcement learning. In arXiv, 2016b.

Finn, C., Abbeel, P., and Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In ICML,2017.

Franceschi, L., Frasconi, P., Salzo, S., Grazzi, R., and Pontil,M. Bilevel programming for hyperparameter optimizationand meta-learning. In ICML, 2018.

Fujimoto, S., van Hoof, H., and Meger, D. Addressingfunction approximation error in actor-critic methods. InICML, 2018.

Grabocka, J., Scholz, R., and Schmidt-Thieme, L. Learningsurrogate losses. In arXiv, 2019.

Gupta, A., Mendonca, R., Liu, Y., Abbeel, P., and Levine,S. Meta-reinforcement learning of structured explorationstrategies. In NeurIPS, 2018.

Haarnoja, T., Zhou, A., Abbeel, P., and Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcementlearning with a stochastic actor. In ICML, 2018a.

Haarnoja, T., Zhou, A., Hartikainen, K., Tucker, G., Ha, S.,Tan, J., Kumar, V., Zhu, H., Gupta, A., Abbeel, P., andLevine, S. Soft actor-critic algorithms and applications.In arXiv, 2018b.

Houthooft, R., Chen, R. Y., Isola, P., Stadie, B. C., Wolski,F., Ho, J., and Abbeel, P. Evolved policy gradients. InNeurIPS, 2018.

Huang, C., Zhai, S., Talbott, W., Bautista, M. A., Sun, S.,Guestrin, C., and Susskind, J. Addressing the loss-metricmismatch with adaptive loss alignment. In ICML, 2019.

Li, H., Xu, Z., Taylor, G., Studer, C., and Goldstein, T.Visualizing the loss landscape of neural nets. In NeurIPS,2018.

Li, Y., Yang, Y., Zhou, W., and Hospedales, T. M. Feature-critic networks for heterogeneous domain generalization.In ICML, 2019.

Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T.,Tassa, Y., Silver, D., and Wierstra, D. Continuous controlwith deep reinforcement learning. In ICLR, 2016.

Loiacono, D., Cardamone, L., and Lanzi, P. L. Simulatedcar racing championship: Competition software manual.In arXiv, 2013.

Meier, F., Kappler, D., and Schaal, S. Online learning of amemory for learning rates. In ICRA, 2018.

Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A.,Antonoglou, I., Wierstra, D., and Riedmiller, M. Playingatari with deep reinforcement learning. In arXiv, 2013.

Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Ve-ness, J., Bellemare, M. G., Graves, A., Riedmiller, M.,Fidjeland, A. K., Ostrovski, G., Petersen, S., Beattie, C.,Sadik, A., Antonoglou, I., King, H., Kumaran, D., Wier-stra, D., Legg, S., and Hassabis, D. Human-level controlthrough deep reinforcement learning. Nature, 518(7540):529–533, 2015.

Mnih, V., Badia, A. P., Mirza, M., Graves, A., Lillicrap,T., Harley, T., Silver, D., and Kavukcuoglu, K. Asyn-chronous methods for deep reinforcement learning. InICML, 2016.

Oudeyer, P.-Y. and Kaplan, F. What is intrinsic motivation?a typology of computational approaches. Frontiers inneurorobotics, 1:6, 2009.

Rajeswaran, A., Finn, C., Kakade, S., and Levine, S. Meta-learning with implicit gradients. In NeurIPS, 2019.

Rakelly, K., Zhou, A., Quillen, D., Finn, C., and Levine,S. Efficient off-policy meta-reinforcement learning viaprobabilistic context variables. In ICML, 2019.

Riemer, M., Cases, I., Ajemian, R., Liu, M., Rish, I., Tu,Y., and Tesauro, G. Learning to learn without forgettingby maximizing transfer and minimizing interference. InICLR, 2019.

Santoro, A., Bartunov, S., Botvinick, M., Wierstra, D., andLillicrap, T. Meta-learning with memory-augmented neu-ral networks. In ICML, 2016.

Schulman, J., Levine, S., Abbeel, P., Jordan, M., and Moritz,P. Trust region policy optimization. In ICML, 2015.

Schulman, J., Wolski, F., Dhariwal, P., Radford, A., andKlimov, O. Proximal policy optimization algorithms. InarXiv, 2017.

Page 10: Online Meta-Critic Learning for Off-Policy Actor-Critic Methods · 2020-03-12 · Review of Off-Policy Actor-Critic RL Reinforcement learning involves an agent interacting with the

Online Meta-Critic Learning for Off-Policy Actor-Critic Methods

Sung, F., Zhang, L., Xiang, T., Hospedales, T., and Yang,Y. Learning to learn: meta-critic networks for sampleefficient learning. In arXiv, 2017.

Sutton, R. S., McAllester, D. A., Singh, S. P., and Mansour,Y. Policy gradient methods for reinforcement learningwith function approximation. In NIPS, 2000.

Sutton, R. S., Maei, H. R., Precup, D., Bhatnagar, S., Silver,D., Szepesvri, C., and Wiewiora, E. Fast gradient-descentmethods for temporal-difference learning with linear func-tion approximation. In ICML, 2009.

Todorov, E., Erez, T., and Tassa, Y. Mujoco: A physicsengine for model-based control. In IROS, 2012.

Veeriah, V., Hessel, M., Xu, Z., Lewis, R., Rajendran, J., Oh,J., van Hasselt, H., Silver, D., and Singh, S. Discovery ofuseful questions as auxiliary tasks. In NeurIPS, 2019.

Wu, L., Tian, F., Xia, Y., Fan, Y., Qin, T., Jian-Huang,L., and Liu, T.-Y. Learning to teach with dynamic lossfunctions. In NeurIPS, 2018.

Xu, Z., van Hasselt, H., and Silver, D. Meta-gradient rein-forcement learning. In NeurIPS, 2018.

Zaheer, M., Kottur, S., Ravanbakhsh, S., Poczos, B.,Salakhutdinov, R. R., and Smola, A. J. Deep sets. InNIPS. 2017.

Zheng, Z., Oh, J., and Singh, S. On learning intrinsicrewards for policy gradient methods. In NeurIPS, 2018.