Papers
Topics
Authors
Recent
Search
2000 character limit reached

XQC: Well-conditioned Optimization Accelerates Deep Reinforcement Learning

Published 29 Sep 2025 in cs.LG and cs.AI | (2509.25174v1)

Abstract: Sample efficiency is a central property of effective deep reinforcement learning algorithms. Recent work has improved this through added complexity, such as larger models, exotic network architectures, and more complex algorithms, which are typically motivated purely by empirical performance. We take a more principled approach by focusing on the optimization landscape of the critic network. Using the eigenspectrum and condition number of the critic's Hessian, we systematically investigate the impact of common architectural design decisions on training dynamics. Our analysis reveals that a novel combination of batch normalization (BN), weight normalization (WN), and a distributional cross-entropy (CE) loss produces condition numbers orders of magnitude smaller than baselines. This combination also naturally bounds gradient norms, a property critical for maintaining a stable effective learning rate under non-stationary targets and bootstrapping. Based on these insights, we introduce XQC: a well-motivated, sample-efficient deep actor-critic algorithm built upon soft actor-critic that embodies these optimization-aware principles. We achieve state-of-the-art sample efficiency across 55 proprioception and 15 vision-based continuous control tasks, all while using significantly fewer parameters than competing methods.

Summary

  • The paper introduces XQC, a framework that combines batch normalization, weight normalization, and categorical cross-entropy loss to create a well-conditioned optimization landscape for DRL.
  • The paper demonstrates that stabilizing the critic's Hessian eigenspectrum improves convergence, resulting in state-of-the-art performance on diverse continuous control tasks.
  • The paper shows that XQC achieves comparable or better performance with about 4.5× fewer parameters and 5× less compute, ensuring scalability and robustness.

XQC: Well-conditioned Optimization Accelerates Deep Reinforcement Learning

Introduction

The paper "XQC: Well-conditioned Optimization Accelerates Deep Reinforcement Learning" introduces an innovative framework designed to enhance sample efficiency in deep reinforcement learning (DRL) by improving the optimization landscape of the critic network. Leveraging insights from the eigenspectrum and condition number analyses of the critic's Hessian, the authors propose a new approach that combines batch normalization (BN), weight normalization (WN), and categorical cross-entropy (CE) loss. This configuration results in significantly smaller condition numbers, yielding better sample efficiency with fewer parameters and reduced computational requirements.

Methodology

The foundation of XQC lies in its well-conditioned optimization landscape, achieved through meticulous architectural choices. The proposed framework incorporates key elements that collectively contribute to improved training dynamics:

  • Normalization Techniques: By utilizing BN and WN, XQC ensures a stable and well-bounded optimization process, which plays a critical role in maintaining effective learning rates under non-stationary targets.
  • Categorical Cross-Entropy Loss: The incorporation of a distributional CE loss over the traditional mean squared error (MSE) enhances the landscape's condition, fostering better convergence properties.

These design principles underpin the XQC algorithm, which is implemented as an extension to the soft actor-critic (SAC) framework. The resulting architecture outperforms existing baselines across a multitude of tasks without resorting to larger models or increased computational resources.

Empirical Results

Figure 1

Figure 1: Well-conditioned network architectures yield state-of-the-art RL performance. Our algorithm, XQC with a BN and WN-based architecture and a CE loss, achieves competitive performance against state-of-the-art baselines across 55 proprioceptive continuous control tasks.

The empirical evaluation showcases XQC's capability to deliver state-of-the-art performance on 55 proprioceptive and 15 vision-based continuous control tasks. The algorithm matches or surpasses more complex approaches in terms of sample efficiency, leveraging fewer parameters (approximately 4.5× fewer) and computational resources (5× less compute) compared to the closest competitors like simba-v.

Optimization Landscape Analysis

Figure 2

Figure 2: When performing gradient-based optimization, the condition number (κ\kappa) of the objective's Hessian significantly impacts convergence.

Through comprehensive analysis, the authors demonstrate that the optimization landscape's conditioning strongly correlates with DRL performance. Architectures incorporating BN exhibit more compact and stable eigenspectra, avoiding outlier modes that can destabilize training. Figure 3

Figure 3: Eigenvalues and condition numbers on dog-trot over 5 seeds for different critic architectures during training.

The empirical eigenvalue analysis revealed that the combination of BN, WN, and CE loss substantially improves landscape conditioning over traditional MSE losses. The compactness and stability in the eigenspectra are crucial for efficient convergence and enhanced sample efficiency.

Parameter and Compute Efficiency

XQC achieves competitive sample efficiency with significant parameter and compute savings. This efficiency can be attributed to XQC's well-conditioned architectural design, which demonstrates robustness and scalability across varying update-to-date (UTD) ratios, layer widths, and depths. Figure 4

Figure 4: XQC is significantly more parameter and compute efficient... Robust to scaling.

The detailed breakdown underscores the architecture's capability to scale gracefully, maintaining or improving performance even as compute and model capacity are increased—a critical aspect for real-world applications where computational resources are constrained.

Conclusion

The XQC framework emphasizes the importance of well-conditioned optimization landscapes in DRL, where stability and sample efficiency are paramount. By focusing on improving the critic's Hessian properties through BN, WN, and CE loss, XQC achieves significant performance gains with reduced complexity and resource requirements. This paradigm shift from added complexity to optimizing the landscape presents a promising direction for future exploration in reinforcement learning research, potentially unlocking further improvements in efficiency and capability across diverse applications.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 7 tweets with 17 likes about this paper.