Papers
Topics
Authors
Recent
Search
2000 character limit reached

Decoupled Entropy Minimization

Published 5 Nov 2025 in cs.LG, cs.CV, cs.IT, math.IT, math.ST, stat.ML, and stat.TH | (2511.03256v1)

Abstract: Entropy Minimization (EM) is beneficial to reducing class overlap, bridging domain gap, and restricting uncertainty for various tasks in machine learning, yet its potential is limited. To study the internal mechanism of EM, we reformulate and decouple the classical EM into two parts with opposite effects: cluster aggregation driving factor (CADF) rewards dominant classes and prompts a peaked output distribution, while gradient mitigation calibrator (GMC) penalizes high-confidence classes based on predicted probabilities. Furthermore, we reveal the limitations of classical EM caused by its coupled formulation: 1) reward collapse impedes the contribution of high-certainty samples in the learning process, and 2) easy-class bias induces misalignment between output distribution and label distribution. To address these issues, we propose Adaptive Decoupled Entropy Minimization (AdaDEM), which normalizes the reward brought from CADF and employs a marginal entropy calibrator (MEC) to replace GMC. AdaDEM outperforms DEM*, an upper-bound variant of classical EM, and achieves superior performance across various imperfectly supervised learning tasks in noisy and dynamic environments.

Summary

  • The paper decouples classical entropy minimization into CADF and GMC, addressing reward collapse and easy-class bias in deep neural networks.
  • It introduces AdaDEM, which normalizes CADF using the L1-norm and employs a dynamic marginal entropy calibrator to eliminate hyperparameter tuning.
  • Empirical results demonstrate AdaDEM’s superior performance in test-time adaptation, semi-supervised learning, unsupervised domain adaptation, and reinforcement learning.

Decoupled Entropy Minimization: Theory, Algorithms, and Empirical Analysis

Introduction and Motivation

Entropy Minimization (EM) is a foundational self-supervised learning principle, widely adopted for reducing class overlap, bridging domain gaps, and calibrating uncertainty in deep neural networks (DNNs). Despite its broad utility in clustering, semi-supervised learning (SSL), unsupervised domain adaptation (UDA), and test-time adaptation (TTA), classical EM exhibits inherent limitations that restrict its effectiveness. This work rigorously reformulates EM, decoupling it into two distinct components: the Cluster Aggregation Driving Factor (CADF) and the Gradient Mitigation Calibrator (GMC). The analysis reveals two critical failure modes—reward collapse and easy-class bias—arising from the coupled formulation, and introduces Adaptive Decoupled Entropy Minimization (AdaDEM) to address these issues without hyperparameter tuning.

Theoretical Reformulation of Entropy Minimization

Classical EM minimizes the conditional entropy of model predictions:

H(z)=i=1Cpi(z)logpi(z)H(\mathbf{z}) = -\sum_{i=1}^C p_i(\mathbf{z}) \log p_i(\mathbf{z})

where pi(z)p_i(\mathbf{z}) is the softmax probability for class ii given logits z\mathbf{z}. The authors decouple this into:

H(z)=i=1Cpi(z)ziCADF+logi=1CeziGMCH(\mathbf{z}) = \underbrace{-\sum_{i=1}^C p_i(\mathbf{z}) z_i}_{\text{CADF}} + \underbrace{\log \sum_{i=1}^C e^{z_i}}_{\text{GMC}}

CADF rewards dominant classes, promoting peaked output distributions, while GMC penalizes high-confidence predictions, regularizing the optimization. The partial derivatives with respect to logits reveal that CADF amplifies rewards for high-probability classes, whereas GMC imposes penalties proportional to confidence.

Limitations of Classical EM

Two major limitations are identified:

  • Reward Collapse: As predicted probabilities approach 1.0, the gradient magnitude for high-certainty samples collapses to zero, impeding their contribution to learning.
  • Easy-Class Bias: The model systematically favors dominant/easy classes, misaligning output and true label distributions, especially under noise or imbalance. Figure 1

Figure 1

Figure 1: (Left) Reward collapse in classical EM; (Right) Easy-class bias in output distributions after adaptation.

Decoupled Entropy Minimization (DEM) and Hyperparameterization

To mitigate these issues, DEM introduces temperature τ\tau to soften CADF and weight α\alpha to scale GMC:

H(z)=i=1Cpτi(z)zi+αlogi=1CeziH(\mathbf{z}) = -\sum_{i=1}^C p_{\tau i}(\mathbf{z}) z_i + \alpha \log \sum_{i=1}^C e^{z_i}

where pτi(z)p_{\tau i}(\mathbf{z}) is the softmax with temperature τ\tau. DEM* searches for optimal (τ,α)(\tau^*, \alpha^*) via TPE, establishing an upper bound for classical EM performance. Figure 2

Figure 2

Figure 2

Figure 2: (Left) Reward curves for varying τ\tau; (Center) Correlation between optimal τ\tau and average predicted probability; (Right) TTA results across domains.

Figure 3

Figure 3

Figure 3

Figure 3: (Left) Reward curves for varying α\alpha; (Center) Effects of α\alpha on static/dynamic shifts; (Right) TTA results for optimal α\alpha.

Adaptive Decoupled Entropy Minimization (AdaDEM)

AdaDEM eliminates hyperparameter tuning by:

  • Normalizing CADF Rewards: Uses the L1-norm of CADF gradients (δ\delta) to normalize sample contributions, directly addressing reward collapse.
  • Marginal Entropy Calibrator (MEC): Replaces GMC with a dynamically estimated marginal entropy, penalizing dominant/easy classes without requiring label priors.

H(z)=1δi=1C(pi(z)pkit)ziH(\mathbf{z}) = -\frac{1}{\delta} \sum_{i=1}^C (p_i(\mathbf{z}) - \overline{\mathfrak{p}_{k i}^t}) z_i

where pkit\overline{\mathfrak{p}_{k i}^t} is the moving average of marginal class probabilities.

Empirical Evaluation

AdaDEM and DEM* are evaluated on TTA, SSL, UDA, and RL tasks, consistently outperforming classical EM and prior SOTA methods. Figure 4

Figure 4

Figure 4: (Left) Ablation studies for single-domain and continual TTA; (Right) AdaDEM exhibits 10x expanded learning rate tolerance compared to classical EM.

Test-Time Adaptation

AdaDEM yields substantial gains in both single-domain and continual TTA, improving robustness to distribution shifts and reducing catastrophic forgetting.

Semi-Supervised Learning

AdaDEM enhances performance across SSL benchmarks, especially under severe class imbalance, outperforming classical EM and improving existing SSL algorithms when integrated. Figure 5

Figure 5

Figure 5: AdaDEM improves per-class and average F1 scores on class-imbalanced benchmarks.

Unsupervised Domain Adaptation

AdaDEM bridges domain gaps more effectively, improving mIoU in semantic segmentation and yielding sharper object boundaries. Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6

Figure 6: AdaDEM reduces pixel entropy and sharpens segmentation boundaries compared to baselines.

Reinforcement Learning

AdaDEM is compatible with entropy maximization strategies in RL, maintaining or improving average returns in Minigrid environments. Figure 7

Figure 7

Figure 7

Figure 7

Figure 7

Figure 7: AdaDEM improves RL performance across Minigrid tasks.

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8

Figure 8: Average return curves for AdaDEM and baselines in Minigrid environments.

Ablation and Sensitivity Analyses

AdaDEM demonstrates reduced sensitivity to optimizer choice and learning rate, robustly handling noisy, imbalanced, and dynamic environments. Figure 9

Figure 9

Figure 9: AdaDEM and its variants show reduced learning rate sensitivity compared to classical EM.

Ablation studies confirm the necessity of L1-norm normalization and dynamic marginal entropy estimation. Additional scaling of MEC and alternative norm types are explored, with L1-norm providing optimal stability. Figure 10

Figure 10

Figure 10: Ablation of α\alpha scaling for MEC in AdaDEM.

Figure 11

Figure 11

Figure 11: Ablation of momentum parameter π\pi for prior label distribution estimation in MEC.

Figure 12

Figure 12

Figure 12: Ablation of different norm types for CADF reward normalization.

Implementation Details

AdaDEM and DEM* are implemented as drop-in replacements for classical EM in existing frameworks. Pseudocode is provided for PyTorch-style integration, and hyperparameter search for DEM* is automated via TPE. AdaDEM requires no labeled data or hyperparameter tuning, strictly adhering to unsupervised learning paradigms.

Implications and Future Directions

The decoupling of EM into CADF and GMC provides a principled framework for analyzing and improving self-supervised objectives. AdaDEM's plug-and-play design enables robust adaptation in real-world, imperfectly supervised scenarios, including online learning, domain adaptation, and RL. The elimination of hyperparameter tuning and improved stability are particularly relevant for deployment in dynamic and resource-constrained environments.

Theoretically, the work motivates further exploration of disentangled self-supervised objectives and dynamic regularization strategies. Practically, AdaDEM's robustness to distribution shift and class imbalance suggests utility in federated, continual, and open-world learning settings.

Conclusion

Decoupled Entropy Minimization offers a rigorous theoretical and empirical advancement over classical EM, addressing reward collapse and easy-class bias through principled reformulation and adaptive normalization. AdaDEM achieves superior performance across diverse machine learning tasks, with strong robustness and practical deployability. While the approach does not resolve all challenges in self-supervised learning, it establishes a solid foundation for future research in adaptive, imperfectly supervised systems.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 1 like about this paper.