Analytic Dynamic Imbalance Rectifier (ADIR)
- ADIR is a closed-form algorithm that rebalances class contributions using weighted ridge regression to address intra- and inter-task imbalances.
- It employs a dynamic regularization strategy based on the Gini coefficient to adaptively mitigate catastrophic forgetting.
- ADIR eliminates raw exemplar storage, enabling memory-efficient and privacy-compliant incremental learning in applications like sound source localization.
The Analytic Dynamic Imbalance Rectifier (ADIR) is a closed-form, exemplar-free algorithmic module designed to address dynamic class-imbalance and catastrophic forgetting in generalized class-incremental learning (GCIL) settings, particularly for sound source localization (SSL) and related continual learning domains. ADIR operates by analytically re-weighting class contributions and adaptively regularizing classifier updates based on evolving cross-task class distributions, eliminating the need for raw exemplar storage and enabling privacy-compliant, memory-efficient incremental adaptation (Fan et al., 26 Jan 2026, Fang et al., 2024).
1. Motivation and Context
Many real-world class-incremental and generalized class-incremental learning (GCIL) problems exhibit two pervasive forms of imbalance: intra-task (long-tailed distribution of labels within a task) and inter-task (distribution skews and overlaps across tasks). In GCIL, classes (e.g., discrete directions-of-arrival for SSL) arrive in a non-monotonic sequence, may reappear or disappear, and are often distributed unevenly across time. Uncorrected, this leads to two major issues:
- Majority bias: Classifiers become biased toward classes with abundant recent samples, neglecting minority and previously encountered classes.
- Catastrophic forgetting: The classifier's accuracy on earlier classes deteriorates as it is updated for new classes or tasks.
ADIR is developed to explicitly correct these phenomena in an analytic, computationally efficient, memory-light, and privacy-preserving manner—critical in scenarios where storing raw data for replay is infeasible or prohibited (Fan et al., 26 Jan 2026, Fang et al., 2024).
2. Theoretical Formulation
ADIR frames incremental classifier learning as a weighted ridge regression problem, in which class contributions are rebalanced inversely to their total sample counts, and the regularization coefficient is dynamically adapted to class-imbalance severity. Let be the total number of tasks, the current task, and the number of classes observed so far.
Given accumulated per-class statistics:
- : total samples of class
- (auto-correlation)
- (cross-correlation)
where is the feature representation (from a frozen extractor) and is a smoothed one-hot target, the ADIR loss for weight matrix is: Here, (per-class inverse frequency); is the adaptive regularization.
The analytic optimal solution is: with , (Fan et al., 26 Jan 2026, Fang et al., 2024).
3. Dynamic Regularization and Class-Imbalance Quantification
ADIR incorporates a novel dynamic regularization strategy responsive to the instantaneous skew of the class distribution:
- Gini coefficient: For class prior ,
- Adaptive regularization: The regularization parameter is set as
where and are hyperparameters (e.g., , ) (Fan et al., 26 Jan 2026). Higher skew (Gini ) results in exponentially increased regularization, which suppresses overfitting to head classes.
4. Incremental Update Algorithm
ADIR operates solely via incremental maintenance of per-class auto- and cross-correlation matrices and per-class sample counts. The update algorithm, executed at each new phase or task, comprises:
- For each new sample :
- Extract feature
- Update
- Update
- Increment
- Compute for all classes with
- Compute class proportions , then Gini coefficient and set adaptively
- Form weighted global sums ,
- Solve for
- Use for classification in subsequent inference (Fan et al., 26 Jan 2026, Fang et al., 2024).
5. Memory and Computational Complexity
ADIR does not retain raw exemplars, relying exclusively on first- and second-order class statistics. The memory footprint is thus for -dimensional features and classes—tractable for common dimensionalities (e.g., , in SSL (Fan et al., 26 Jan 2026)). Each phase update requires for a single matrix inversion and for accumulations, which is comparable or superior to gradient-based approaches that require repeated backpropagation across epochs (Fang et al., 2024).
6. Empirical Validation and Comparative Analysis
Ablation experiments on the SSLR benchmark (Fan et al., 26 Jan 2026) and standard CIL datasets (Fang et al., 2024) demonstrate the effectiveness of ADIR for mitigating catastrophic forgetting and long-tailed imbalance. In SSL, core metrics include accuracy (ACC), mean absolute error (MAE), and backward transfer (BWT):
| Setting | ACC (%) | MAE (°) | BWT |
|---|---|---|---|
| Baseline (no GDA, no ADIR) | 72.0 | — | –17.7 |
| + ADIR only | 82.4 | — | +1.4 |
| + GDA only | 75.0 | — | –15.8 |
| Full model (+ADIR & +GDA) | 89.0 | 5.3 | +1.6 |
ADIR alone yields a 10% increase in accuracy and reverses catastrophic forgetting (BWT becomes positive). When combined with intra-task augmentation (GDA), state-of-the-art results are achieved under both clean and noisy conditions. Similarly, AIR achieves last-phase accuracy on CIFAR-100 above 79.7% with invariance to class order and matches or exceeds the performance of replay-based approaches with large memory.
Both studies confirm that ADIR eliminates head-class bias, enhances tail-class representation, and is robust to dynamic, nonstationary class distributions (Fan et al., 26 Jan 2026, Fang et al., 2024).
7. Connections and Extensions
ADIR/AIR generalizes to various continual learning modalities beyond SSL, including computer vision and other domains where non-stationary, imbalanced incremental learning is essential. The subclass of methods characterized by:
- Closed-form, weighted ridge regression reflection of per-class sample distributions
- Memoryless, privacy-compliant operation (via sufficient statistics)
- Task-adaptive analytic control of regularization
forms a scalable foundation for exemplar-free continual learning. A plausible implication is that further extensions could integrate per-sample weighting or adaptive feature extraction, although this is not covered in the referenced works.
ADIR stands as a rigorously validated, theoretically principled, and empirically effective approach for dynamically rectifying class imbalance and preventing catastrophic forgetting in generalized class-incremental learning (Fan et al., 26 Jan 2026, Fang et al., 2024).