The use of deep neural networks in reinforcement learning (RL) often suffers from performance
degradation as model size increases. While soft mixtures of experts (SoftMoEs) have recently shown
promise in mitigating this issue for online RL, the reasons behind their effectiveness remain
largely unknown. In this work we provide an in-depth analysis identifying the key factors driving
this performance gain. We discover the surprising result that tokenizing the encoder output, rather
than the use of multiple experts, is what is behind the efficacy of SoftMoEs. Indeed, we demonstrate
that even with an appropriately scaled single expert, we are able to maintain the performance gains,
largely thanks to tokenization.
The loss of plasticity in learning agents, analogous to the solidification of neural pathways in biological brains,
significantly impedes learning and adaptation in reinforcement learning due to its non-stationary nature. To address
this fundamental challenge, we propose a novel approach, Neuroplastic Expansion (NE), inspired by cortical expansion
in cognitive science. NE maintains learnability and adaptability throughout the entire training process by dynamically
growing the network from a smaller initial size to its full dimension. Our method is designed with three key
components: (1) elastic neuron generation based on potential gradients, (2) dormant neuron pruning to optimize network
expressivity, and (3) neuron consolidation via experience review to strike a balance in the plasticity-stability
dilemma. Extensive experiments demonstrate that NE effectively mitigates plasticity loss and outperforms
state-of-the-art methods across various tasks in MuJoCo and DeepMind Control Suite environments. NE enables more
adaptive learning in complex, dynamic environments, which represents a crucial step towards transitioning deep
reinforcement learning from static, one-time training paradigms to more flexible, continually adapting models.
Mixtures of Experts (MoEs) have gained prominence in (self-)supervised learning due to their enhanced
inference efficiency, adaptability to distributed training, and modularity. Previous research has illustrated
that MoEs can significantly boost Deep Reinforcement Learning (DRL) performance by expanding the network's
parameter count while reducing dormant neurons, thereby enhancing the model's learning capacity and ability
to deal with non-stationarity. In this work, we shed more light on MoEs' ability to deal with non-stationarity
and investigate MoEs in DRL settings with "amplified" non-stationarity via multi-task training, providing further
evidence that MoEs improve learning capacity. In contrast to previous work, our multi-task results allow us to
better understand the underlying causes for the beneficial effect of MoE in DRL training, the impact of the various
MoE components, and insights into how best to incorporate them in actor-critic-based DRL networks. Finally,
we also confirm results from previous work.
Deep reinforcement learning (deep RL) has achieved tremendous success on various domains through a combination
of algorithmic design and careful selection of hyper-parameters. Algorithmic improvements are often the result
of iterative enhancements built upon prior approaches, while hyper-parameter choices are typically inherited from
previous methods or fine-tuned specifically for the proposed technique. Despite their crucial impact on performance,
hyper-parameter choices are frequently overshadowed by algorithmic advancements. This paper conducts an extensive
empirical study focusing on the reliability of hyper-parameter selection for value-based deep reinforcement learning
agents, including the introduction of a new score to quantify the consistency and reliability of various hyper-parameters.
Our findings not only help establish which hyper-parameters are most critical to tune, but also help clarify which tunings
remain consistent across different training regimes.
Recent work has shown that deep reinforcement learning agents have difficulty in effectively using their network parameters.
We leverage prior insights into the advantages of sparse training techniques and demonstrate that gradual magnitude pruning
enables value-based agents to maximize parameter effectiveness. This results in networks that yield dramatic performance
improvements over traditional networks, using only a small fraction of the full network parameters.
Mixtures of Experts Unlock Parameter Scaling for Deep RL
Johan Obando-Ceron*, Ghada Sokar*, Timon Willi*,
Clare Lyle, Jesse Farebrother, Jakob Nicolaus Foerster, Gintare Karolina Dziugaite,
Doina Precup and Pablo Samuel Castro
In Internation Conference on Machine Learning, 2024
The recent rapid progress in (self) supervised learning models is in large part predicted by empirical
scaling laws: a model's performance scales proportionally to its size. Analogous scaling laws remain elusive
for reinforcement learning domains, however, where increasing the parameter count of a model often hurts its
final performance. In this paper, we demonstrate that incorporating Mixture-of-Expert (MoE) modules, and in
particular Soft MoEs (Puigcerver et al., 2023), into value-based networks results in more parameter-scalable models,
evidenced by substantial performance increases across a variety of training regimes and model sizes.
This work thus provides strong empirical evidence towards developing scaling laws for reinforcement learning.
This paper introduces JaxPruner, an open-source JAX-based pruning and sparse training library for machine learning research.
JaxPruner aims to accelerate research on sparse neural networks by providing concise implementations of popular pruning and sparse
training algorithms with minimal memory and latency overhead. Algorithms implemented in JaxPruner use a common API and work seamlessly
with the popular optimization library Optax, which, in turn, enables easy integration with existing JAX based libraries. We demonstrate
this ease of integration by providing examples in four different codebases: Scenic, t5x, Dopamine and FedJAX and provide baseline
experiments on popular benchmarks. Jaxpruner is hosted at github. com/google-research/jaxpruner
In value-based deep reinforcement learning with replay memories, the batch size parameter specifies how
many transitions to sample for each gradient update. Although critical to the learning process, this value
is typically not adjusted when proposing new algorithms. In this work we present a broad empirical study that
suggests reducing the batch size can result in a number of significant performance gains; this is surprising,
as the general tendency when training neural networks is towards larger batch sizes for improved performance.
We complement our experimental findings with a set of empirical analyses towards better understanding this phenomenon..
We introduce a value-based RL agent, which we call BBF, that achieves super-human performance in the
Atari 100K benchmark. BBF relies on scaling the neural networks used for value estimation, as well as a
number of other design choices that enable this scaling in a sample-efficient manner. We conduct extensive
analyses of these design choices and provide insights for future work. We end with a discussion about
updating the goalposts for sample-efficient RL research on the ALE. We make our code and data publicly
available at https://github. com/google-research/google-research/tree/master/bigger_better_faster.
Since the introduction of DQN, a vast majority of reinforcement learning research has focused on reinforcement learning with deep neural
networks as function approximators. New methods are typically evaluated on a set of environments that have now become standard, such as
Atari 2600 games. While these benchmarks help standardize evaluation, their computational cost has the unfortunate side effect of widening
the gap between those with ample access to computational resources, and those without. In this work we argue that, despite the community’s
emphasis on large-scale environments, the traditional small-scale environments can still yield valuable scientific insights and can help reduce
the barriers to entry for underprivileged communities. To substantiate our claims, we empirically revisit the paper which introduced the
Rainbow algorithm [Hessel et al., 2018] and present some new insights into the algorithms used by Rainbow.