Papers
Topics
Authors
Recent
Search
2000 character limit reached

Directional Conductance Divergence (DCD)

Updated 8 February 2026
  • DCD is a metric quantifying asymmetric functional coverage in vision–language models and anisotropic conductance in graphene nanostructures.
  • In VLMs, DCD computes layerwise conductance using entropy-regularized softmax, offering model-specific transferability insights through block importance deviations.
  • In strained graphene, DCD captures the divergence of ballistic conductance along specific crystallographic directions, signaling emergent directional superconductivity near critical strain.

Directional Conductance Divergence (DCD) denotes two distinct concepts in contemporary research literature: (i) an asymmetric, task- and model-specific metric for assessing functional similarity and transferability between visual tasks in vision–LLMs (VLMs), and (ii) the divergence of ballistic conductance along specific crystallographic directions in strained graphene at a critical deformation. The former is central to few-shot model selection in large VLM zoos; the latter describes the emergence of “directional superconductivity” in graphene nanostructures. Both share the feature that conductance (or functional coverage) becomes highly anisotropic and, in a rigorous sense, diverges or saturates along preferred axes as determined by system-specific criteria (Yang et al., 1 Feb 2026, Soodchomshom, 2011).

1. Asymmetric Task Similarity in Vision–LLMs

DCD in the context of VLMs encapsulates the need for an asymmetric, entropy-regularized divergence to measure how a pretrained source representation covers blocks critical to a target task. For a given model MM, the visual encoder is partitioned into dd coarse-grained blocks. Each image xx from task TT yields a layerwise conductance vector gM(x)=(C1(x;M),,Cd(x;M))g_M(x) = (C_1(x;M), \dots, C_d(x;M))^\top, where

C(x;M)=i=1pFM(x)h,iC_\ell(x;M) = \sum_{i=1}^p \left|\frac{\partial F_M(x)}{\partial h_{\ell,i}}\right|

with FM(x)=fM(x)2F_M(x) = \lVert f_M(x)\rVert_2 for output embedding fM(x)Rpf_M(x)\in\mathbb{R}^p. The mean profile U^M,T\hat U_{M,T} over NN images allows model-level summaries on both source (SS) and target (TT).

The normalized activation uM,Tu_{M,T} is obtained as

uM,T=U^M,Tmax(U^M,T2,ϵ)u_{M,T} = \frac{\hat U_{M,T}}{\max(\lVert\hat U_{M,T}\rVert_2,\epsilon)}

and the importance distribution aM,Ta_{M,T} arises via the entropy-regularized softmax: aM,T(i)=exp(nuM,T,i)j=1dexp(nuM,T,j),n>0.a_{M,T}(i) = \frac{\exp(n\,u_{M,T,i})}{\sum_{j=1}^d\exp(n\,u_{M,T,j})}, \qquad n>0. This quantifies target-specific saliency of encoder blocks.

The directional deviation between source and target is

δi(ST)=uM,T,iuM,S,iuM,T,i+ϵ\delta_i(S \to T) = \frac{|u_{M,T,i} - u_{M,S,i}|}{u_{M,T,i}+\epsilon}

weighted by target saliency, inducing the metric

DCDM(ST)=i=1daM,T(i)δi(ST)\mathrm{DCD}_M(S\to T) = \sum_{i=1}^d a_{M,T}(i)\,\delta_i(S\to T)

which is generally non-symmetric due to the directional weighting aM,Ta_{M,T}. The degree to which SS covers the blocks salient for TT determines the inferred model transferability (Yang et al., 1 Feb 2026).

2. Entropy-Regularized Alignment and Importance Weighting

Entropy-regularized alignment ensures that the block importance distribution aM,Ta_{M,T} is both sharp where TT displays significant conductance and widespread to preserve statistical stability. Formally, this amounts to maximizing

p,uM,T+1nH(p)\langle p, u_{M,T} \rangle + \frac{1}{n}H(p)

for pp in the dd-simplex, with Shannon entropy H(p)H(p). The resulting softmax has tunable “attention intensity” nn, interpolating between uniform weighting for n0n\to 0 and sharp focus on the maximal block(s) as nn\to\infty. This framework underlines the model- and target-specific directionality intrinsic to DCD, precluding symmetric elementary divergences such as cosine or Jensen–Shannon, which do not sufficiently account for functional asymmetry between tasks. Ablations confirm a 10%\geq 10\% deficit in NDCG@5 for symmetric proxies (Yang et al., 1 Feb 2026).

3. DCD in Ballistic Transport of Strained Graphene

In the context of uniaxially zigzag-strained graphene, DCD refers to the physical divergence of conductance along the armchair direction as critical strain is approached. The system is governed by the modified tight-binding Hamiltonian

H(k;ε)=s=13ts(ε)eikδs(ε)+h.c.H(\vec{k};\varepsilon) = -\sum_{s=1}^3 t_s(\varepsilon)e^{i\vec{k}\cdot\vec{\delta}_s(\varepsilon)} + \mathrm{h.c.}

with anisotropic hopping parameters. Below a critical strain εc\varepsilon_c, the band-structure remains gapless; at εc\varepsilon_c (n(εc)=2n(\varepsilon_c) = 2), Dirac points merge. The energy spectrum is an anisotropic Weyl form: E(k;ε)=±vx(ε)2kx2+vy(ε)2ky2E(\vec{k}; \varepsilon) = \pm\hbar\sqrt{v_x(\varepsilon)^2 k_x^2 + v_y(\varepsilon)^2 k_y^2} with vx0v_x \to 0 as εεc\varepsilon \to \varepsilon_c. The Landauer conductance for carrier propagation at angle θ\theta is determined by the number of transverse modes N(ε,EF;θ)N(\varepsilon,E_F;\theta), producing (Soodchomshom, 2011): Gy(ε)=2e2hNy(EF)G_y(\varepsilon) = \frac{2e^2}{h}N_y(E_F) where

limεεcGy(ε)\lim_{\varepsilon\to\varepsilon_c}G_y(\varepsilon)\to\infty

and GxG_x remains finite. This physical “divergence” exemplifies directional (anisotropic) electronic transport and provides an analogy to superconductivity along selected directions.

4. Computational Workflow and Algorithm

For model selection in VLMs, the complete DCD computation is:

  1. Layerwise Conductance Extraction: For each block \ell and sample xx in source and target, compute C(x;M)C_\ell(x;M).
  2. Profile Averaging: Average conductance over all samples to obtain U^M,S\hat U_{M,S} and U^M,T\hat U_{M,T}.
  3. Normalization: Obtain uM,Su_{M,S} and uM,Tu_{M,T} via 2\ell_2 normalization with ϵ\epsilon regularization.
  4. Block Importance: Compute aM,T(i)a_{M,T}(i) via softmax over uM,Tu_{M,T}.
  5. Relative Deviations: Evaluate per-block deviations δi\delta_i.
  6. Metric Aggregation: Sum aM,T(i)δia_{M,T}(i)\,\delta_i to yield DCDM(ST)\mathrm{DCD}_M(S\to T).

In large-scale evaluation, rankings for held-out target tasks are predicted by aggregating known source task ranks, weighted by exponentiated negative DCD differences. The principal metrics are NDCG@5 and Kendall's $\tau@5$ (Yang et al., 1 Feb 2026).

For strained graphene, the analytical calculation derives from evaluating Landauer-mode integrals as vx0v_x\to 0, yielding formally divergent results for armchair conductance, while zigzag remains regular (Soodchomshom, 2011).

5. Experimental Evidence and Comparative Benchmarks

On 48 open-source VLMs and 21 image benchmarks (classification, OCR, satellite, medical, etc.), DCD-based selection achieves a 14.7% NDCG@5 improvement over symmetric and data-expensive baselines, with performance saturating at \sim25 source images per task. The approach demonstrates consistent gains in few-shot settings (one target image, 25 source images): DCD yields NDCG@5 = 0.707 versus SWAB’s 0.616, and $\tau@5 = 0.365$ vs. 0.318. Ablations confirm the necessity of both directionality and entropy-regularized block alignment (Yang et al., 1 Feb 2026).

For graphene physics, the analytical divergence is predicted under idealized conditions (zero temperature, ballistic regime, tight-binding model), with distinct physical signatures such as diverging density of states, vanishing resistance Ry0R_y\to 0, and a synthetic superconducting analogy as the bandgap opens at εc\varepsilon_c (Soodchomshom, 2011).

6. Interpretation, Importance, and Scope

DCD, as formalized in VLM selection, is distinctively asymmetric, model-aware, and target-driven. It rectifies the limitations of previous proxies by quantifying transferability in a manner rooted in internal model dynamics rather than in textual or distributional similarity, enabling data- and compute-efficient model selection absent direct inference. The link between coverage of salient functional blocks and predicted transferability is a plausible mechanism underpinning effective transfer in modern multimodal architectures.

In condensed matter, DCD manifests physically as a divergence in conductance, arising from strong anisotropy and band-structure engineering. The analogy to superconductivity is justified by the vanishing resistance along the armchair axis and the opening of an excitation gap, albeit in a ballistic non-interacting regime.

Both utilizations of DCD reflect a broader recognition of directionality, asymmetry, and anisotropy as fundamental features—whether in information transfer across neural architectures, or in charge transport under symmetry-breaking structural perturbations. Future work may extend DCD metrics to broader classes of models or complex materials, contingent on analogous notions of saliency or mode-count divergence.


Key References:

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Directional Conductance Divergence (DCD).