Proxy-Aligned Co-Evolution Mechanism
- Proxy-Aligned Co-Evolution is an optimization framework where multiple proxy models are simultaneously updated under explicit alignment constraints to closely mimic non-differentiable ground-truth systems.
- The mechanism employs differentiable neural proxies and embedding caches, using loss functions and coupled or alternating optimization to prevent proxy drift and improve system robustness.
- Applications include hardware design reinforcement (RL-MORPH) and multimodal out-of-distribution detection (CoEvo), achieving improved convergence and calibration metrics in empirical evaluations.
The proxy-aligned co-evolution mechanism is a class of rigorous optimization and adaptive realignment procedures in machine learning and artificial intelligence that simultaneously evolve two or more proxy models or signal caches under explicit alignment constraints. Recent implementations span hardware design-reinforcement co-optimization and zero-shot out-of-distribution (OOD) detection in multimodal representation spaces, exemplified by RL-MORPH for design-policy joint search (He et al., 2023) and CoEvo for vision-language OOD uncertainty calibration (Tang et al., 13 Jan 2026). These instantiations integrate differentiable proxies (typically neural networks or embedding caches) with non-differentiable or black-box ground-truth models by penalizing proxy drift, synchronizing proxy updates, and using alternating or coupled optimization routines to yield robust, high-performance solutions.
1. Architectural Foundations of Proxy Models
Proxy-aligned co-evolution deploys at least two model representations:
- For hardware co-design (RL-MORPH), the true environment is defined by a possibly non-differentiable physics model , indexed by design parameters (e.g., link lengths, joint angles). A differentiable proxy neural network is concurrently evolved: implemented as a multilayer MLP architecture (e.g., 3 hidden layers, ReLU activations, output dimension ), it approximates the cumulative effect of on task-space effect , enabling gradient-based optimization.
- For cross-modal representation (CoEvo), dual proxy caches of textual and visual embeddings are initialized. The textual proxy is derived from encoded prompts for in-distribution (ID) classes ( dimensionality), while the negative textual proxy () grows at test time by mining semantic negatives. The visual proxies and are and queues, respectively, continually refined with empirical exemplars aligned to respective textual proxies.
This dual-proxy setup is foundational, enabling proxy alignment and mutual evolution to support robust optimization even in the presence of non-differentiable or drifting features and distributions.
2. Alignment Objectives and Loss Formulations
A core element is the explicit alignment loss that penalizes the discrepancy between proxy outputs and ground-truth or reference signals:
- In RL-MORPH, the alignment objective is , using a buffer of input pairs collected during training to ensure the neural proxy remains close to the physics model.
- In CoEvo, alignment is implicit through shared embedding spaces and update gating. All proxies and embeddings are normalized in , and cosine similarity is used for scoring. Updates are gated through adaptive thresholds and confidence margins, preventing noisy alignment near uncertain regions.
Alignment loss terms are balanced with task or detection objectives via trade-off hyperparameters (e.g., in RL-MORPH), providing a regularization effect that tethers the proxy to a reliable reference while allowing for adaptive evolution, minimizing risk of drift or misalignment.
3. Co-Evolution Mechanism and Optimization Algorithms
Proxy-aligned co-evolution proceeds via coupled or alternating updates driven by shared loss landscapes and proxy-task interdependence:
- For RL-MORPH, the total loss is jointly minimized: PPO/A2C is used for policy gradients ; for proxy gradients, destructive interference between and is mitigated via PCGrad projection. Specifically, if (computed as inner product normalized by norm), the RL gradient is projected to remove the component aligned with proxy loss, then summed with weighted proxy gradient for SGD/Adam update.
- In CoEvo, a test-time loop conditions proxy evolution on sample-specific scores: after computing preliminary multimodal scores , new negatives are mined (near/far semantics) based on confidence-gated thresholds. Visual caches are synchronously expanded and refined by entropy-based selection, keeping high-confidence, low-entropy exemplars. To prevent premature misalignment, updates only occur away from the adaptive threshold, as regulated by margin .
4. Dynamic Re-Weighting and Decision Fusion
Both RL-MORPH and CoEvo employ dynamic fusion of proxy signals for robust decision-making:
- In RL-MORPH, proxy accuracy and actionable policy are balanced via . Ablating yields a U-shaped performance curve: low produces proxy drift and poor returns, high impedes task reward learning.
- CoEvo utilizes a fusion parameter () to combine textual and visual proxy scores. Pre-update fusion () privileges stable textual cues; post-update () leverages evolved, contextually aligned visual features. Empirically, yield optimal calibration (Tang et al., 13 Jan 2026).
This fusion strategy calibrates OOD detection and balances task performance with proxy reliability, enabling adaptability to distributional shift and data imbalance.
5. Experimental Setups, Results, and Ablations
Empirical evaluation on RL-MORPH encompassed 2D Reacher (5-link planar arm with forward kinematics physics) and 3D multi-finger manipulation with full Mujoco dynamics, using episodic return as the metric. Baselines included RL-noHWOpt, outer-loop CMA-ES, and Transform2Act. RL-MORPH converged to high-return designs under all tested tasks, with learning collapse observed in the absence of proxy alignment or proper gradient projection.
For CoEvo, experiments on ImageNet-1K (ID) and four OOD sets (iNat, SUN, Places, Textures) demonstrated AUROC improvement to (vs. for AdaNeg) and FPR95 reduction to (vs. ), with ablation showing dual-modal evolution is superior to textual-only or visual-only variants. Performance was robust under data imbalance and varying test set sizes (Tang et al., 13 Jan 2026).
6. Practical Recommendations and Limitations
Reported best practices are:
- For RL-MORPH, always include a small alignment term () and employ PCGrad or similar conflict mitigation between RL and alignment gradients. A 3-4 layer MLP proxy with widths sufficed for various robotic configurations. Sample proxy training data every updates to maintain buffer freshness. For design search, use CMA-ES for low-dimensional or SGD for differentiable ; scalability to topological changes (e.g., variable numbers of robot links) remains an open direction (He et al., 2023).
- For CoEvo, optimal adaptation was achieved with fusion weight and mining batch size . High-confidence gating () and entropy-based exemplar selection were critical for stable evolution. Both textual and visual caches should be iteratively updated and realigned for best OOD calibration (Tang et al., 13 Jan 2026).
Limitations include parametric design constraints (fixed topology) in RL-MORPH and absence of gradient-based test-time supervision in CoEvo. For sim-to-real transfer, real-world rollout data can be incorporated into alignment objectives.
7. Impact and Research Directions
Proxy-aligned co-evolution mechanisms decompose otherwise intractable joint optimization into modular, tightly coupled subproblems, facilitating efficient gradient-based search and adaptation under complex, non-differentiable, or distributionally unstable regimes. Extending these mechanisms to topological and structure search (e.g., graph-NN proxies), as well as incorporating real-world feedback for sim-to-real transfer, represent salient directions for future research. The proxy-aligned co-evolution paradigm is empirically validated to enable robust joint optimization and adaptive detection in open-world scenarios (He et al., 2023, Tang et al., 13 Jan 2026).