Papers
Topics
Authors
Recent
Search
2000 character limit reached

Agentic AI-driven SemSteCom

Updated 30 January 2026
  • Agentic AI-driven SemSteCom is a framework that integrates AI-based semantic extraction, token-controlled reference image generation, and coverless steganography to secure wireless communications.
  • It leverages a modular architecture comprising semantic extraction, EDICT-based steganography, a semantic JSCC codec, and optional enhancement to optimize transmission and recovery.
  • Benchmark results demonstrate significant improvements in PSNR, SSIM, and MSE, confirming robust semantic security and resilience against eavesdropping.

Agentic AI-driven SemSteCom (AgentSemSteCom) denotes a paradigm of secure and efficient semantic steganographic communication for future intellicise wireless networks, in which agentic artificial intelligence orchestrates semantic extraction, generative steganography, semantic channel coding, and task-based enhancement modules. AgentSemSteCom systematically eliminates the requirements for both cover images and private semantic keys, thereby maximizing steganographic capacity and transmission security. The technical foundation integrates agent-controlled semantic feature extraction, digital token–regulated reference image synthesis, mathematically exact invertible diffusion sampling, and advanced semantics-aware joint source-channel coding under agentic feedback and optimization (Meng et al., 23 Jan 2026, Yu et al., 1 Oct 2025, Gao et al., 29 Dec 2025).

1. System Architecture and Workflow

AgentSemSteCom architecture comprises five principal modules: semantic extraction, digital-token-controlled reference image generation, coverless steganography, semantic codec, and optional task-oriented enhancement. Decision-making and orchestration across modules are governed by agentic AI entities such as LLMs, large vision models (LVMs), and RL-based control agents. The workflow proceeds as follows (Meng et al., 23 Jan 2026):

  1. Semantic Extraction: Agentic AI extracts a public semantic key KpubK_{pub} (via LLM) and implicit semantic features xfeatx_{feat} (via domain-specific deep networks) from the secret image xsx_s and requirements RR.
  2. Digital Token Reference Generation: A user-provided token kk is hashed to seed a latent noise zDz_D. A ControlNet-augmented diffusion model, conditioned on KpubK_{pub} and xfeatx_{feat}, generates a reference image xrefx_{ref} via DDIM sampling and VAE decoding.
  3. Coverless Steganography: The secret xsx_s is VAE-encoded to latent zsz_s. EDICT-based invertible forward-noising is applied, token-guided perturbation via binary mask M(s,η)M(s, \eta) is performed, then conditional reverse EDICT denoising (with decoupled cross-attention on KpubK_{pub} and xrefx_{ref}) reconstructs the stego-image xstegox_{stego}.
  4. Semantic Codec Transmission: xstegox_{stego} is encoded by a semantic JSCC codec, transmitted over AWGN channels, and decoded at the receiver by DsemD_{sem}.
  5. Optional Enhancement: Agentic AI selects enhancement operator(s) (e.g., face denoiser, super-resolution) to fulfill RR on recovered secret xs∗x_s^*.

The diagram below outlines the principal order:

Step Module Output
1 Semantic Extraction KpubK_{pub}, xfeatx_{feat}
2 Token Reference Gen. xrefx_{ref}
3 Steganography xstegox_{stego}
4 Semantic Codec Transmitted xstegox_{stego}
5 Enhancement (optional) xs∗x_s^*

2. Mathematical Models

Semantic Feature Extraction

  • Public Semantic Key (KpubK_{pub}): Autoregressive LLM yields qâ„“q_{\ell} at each layer via:

qℓ′=CSA(LN(qℓ−1))+qℓ−1 qℓ=MLP(LN(qℓ′))+qℓ′q'_{\ell} = CSA(LN(q_{\ell-1})) + q_{\ell-1} \ q_{\ell} = MLP(LN(q'_{\ell})) + q'_{\ell}

  • Token Sampling: For i=1,…,Ti=1,\ldots,T,

K^i=Softmax(Wpub⋅LN(qLi)),Ki∼K^i\hat{K}_i = \text{Softmax}(W_{pub} \cdot LN(q_L^i)), \quad K_i \sim \hat{K}_i

Form Kpub=(K1,…,KT)K_{pub} = (K_1, \ldots, K_T).

  • Implicit Features (xfeatx_{feat}): xfeat=G(xs,R)x_{feat}=\mathcal{G}(x_s, R), e.g., OneFormer segmentation for general, OpenPose for faces, xfeat=∅x_{feat}=\varnothing for style-only.

Reference Image Generation via Diffusion

  • Token-controlled latent zD=Randn(s)z_D = \text{Randn}(s), s=H(k)s = \mathcal{H}(k) (hashed user token).
  • Diffusion sampling: ControlNet-augmented U-Net backbone guided by KpubK_{pub} and xfeatx_{feat}:
    • DDIM iterative update:

    zt−1=αt−1(zt−1−αtϵ^θαt)+1−αt−1ϵ^θz_{t-1} = \sqrt{\alpha_{t-1}} \left( \frac{z_t - \sqrt{1-\alpha_t} \hat{\epsilon}_\theta}{\sqrt{\alpha_t}} \right) + \sqrt{1-\alpha_{t-1}} \hat{\epsilon}_\theta - Final xref=DVAE(z0)x_{ref} = D_{VAE}(z_0).

  • Training Loss:

L=Ez0,t,xfeat,Kpub,ϵ[∥ϵ−ϵθ(zt,t,xfeat,Kpub)∥22]\mathcal{L} = \mathbb{E}_{z_0, t, x_{feat}, K_{pub}, \epsilon}[ \| \epsilon - \epsilon_\theta(z_t, t, x_{feat}, K_{pub}) \|_2^2 ]

EDICT-based Steganography

  • VAE encoding: zs=EVAE(xs)z_s = E_{VAE}(x_s)

  • Forward EDICT invertible noising: (z0,u0)=(zs,zs)(z_0,u_0) = (z_s,z_s), affine coupling and mixing as per eqs. 14a–14d (Meng et al., 23 Jan 2026)

  • Token-guided perturbation: zT′=zT⊙(1−M)−zT⊙Mz'_T = z_T \odot (1-M) - z_T \odot M

  • Decoupled cross-attention (reverse):

    • ZtextZ_{text} and ZimgZ_{img} calculated with KpubK_{pub} and xrefx_{ref} embeddings at each U-Net layer, summed to give ZconZ_{con}.
  • EDICT-DDIM update:

zt−1=αt(pzt+(1−p)ut)+βtϵ~t,cz_{t-1} = \alpha_t (p z_t + (1-p) u_t) + \beta_t \tilde{\epsilon}_{t,c}

with ϵ~t,c\tilde{\epsilon}_{t,c} compounded by KpubK_{pub}, xrefx_{ref}.

JSCC Semantic Codec

  • Encoder: S=Esem(xstego)S = E_{sem}(x_{stego})
  • Channel Model: S^=hâ‹…S+n\hat{S} = h \cdot S + n, n∼CN(0,σw2)n \sim \mathcal{CN}(0,\sigma_w^2)
  • Decoder: x^stego=Dsem(S^)\hat{x}_{stego} = D_{sem}(\hat{S})
  • Objective: Analog variant minimizes Lsem=E[d(xstego,x^stego)+λ∥S∥2]\mathcal{L}_{sem} = \mathbb{E}[d(x_{stego}, \hat{x}_{stego}) + \lambda \|S\|^2]

3. Integrated Agentic AI Control and Semantic Communication

Agentic AI in SemSteCom fulfills dual objectives: strategic module selection (semantic extraction, enhancement) and secure protocol orchestration (token-key management, attack resistance). Control agents include:

  • LLM-Driven Key Extraction: Ensures KpubK_{pub} derives only public, non-reversible semantics.
  • Digital Token Management: User token kk deterministically seeds reference generation and stego encoding; wrong token collapses EDICT invertibility.
  • Feature Selection Strategy: Contextual networks such as segmentation or pose models augment extraction based on task RR.
  • JSCC Codec Orchestration: Embedded agents optimize encoding parameters for semantics fidelity and channel conditions (Yu et al., 1 Oct 2025, Gao et al., 29 Dec 2025).

A plausible implication is that future extensions may permit active negotiation of semantic protection levels and adaptive stego capacities in agentic networks.

4. Security Analysis and Steganographic Capacity

AgentSemSteCom addresses semantic eavesdropping threats by:

  • Elimination of Cover Images: Stego image xstegox_{stego} bears no overt payload markers, and external cover is unnecessary.
  • Token-Guided Invertibility: Recovery of xsx_s requires precise token kk and semantic features; an incorrect token results in non-recoverable outputs (ΔPSNR up to 6.5 dB between legitimate receiver and agentic-AI eavesdropper at SNR=10 dB for facial class) (Meng et al., 23 Jan 2026).
  • EDICT Mathematical Exactness: At each step, affine updates in EDICT chains are bijective, ensuring forward noising is invertible only with correct agentic keys.
  • Steganographic Capacity: Full-resolution semantics are embeddable; capacity is quantifiable as Cup≈Hâ‹…Wâ‹…channelsC_{up} \approx H \cdot W \cdot \text{channels}.

Privacy is further enhanced by adding differential privacy noise to semantic embeddings s(x)s(x), subject to distortion Δsem≤ϵpriv\Delta_{sem} \leq \epsilon_{priv} constraints (Yu et al., 1 Oct 2025).

5. Quantitative Performance and Benchmarks

On the UniStega dataset (100 images, three content classes):

Method PSNR (dB) SSIM MSE (×10−2\times 10^{-2}) LPIPS
AgentSemSteCom 21.59 0.5643 0.799 0.4030
SemSteDiff 18.86 0.5183 1.21 0.4046

AgentSemSteCom demonstrates +14.29% PSNR, +8.88% SSIM, −43.75% MSE, and −5.1% LPIPS improvements over SemSteDiff (Meng et al., 23 Jan 2026). For eavesdropper scenarios (wrong token), PSNR drops by ~6.5 dB, confirming significant resilience to semantic interception. Transmission over AWGN channels at SNR ∈ {5,10,15,20} dB further validates robust reconstruction.

The AKB-JSCC framework (source/channel KB) is shown to stably outperform NTSCC and traditional digital codecs in MS-COCO image task, yielding higher PSNR under both bandwidth and channel SNR constraints (Gao et al., 29 Dec 2025).

6. Applications, Deployment, and Future Directions

AgentSemSteCom is applicable to autonomous drone swarms, multi-vehicle perception, and multi-robot rescue in 6G and intellicise networks:

  • Autonomous Drones: Fine-tuning of semantic encoders via GAN synthesis, lightweight pruned/quantized models, and hierarchical RL for spectrum allocation maximize coverage/QoE (Yu et al., 1 Oct 2025).
  • Secure Multi-Agent Coordination: Coverless steganography allows secret intent or commands to be embedded and exchanged securely, with optional post-recovery enhancement based on downstream tasks.
  • Challenges: Heterogeneity of agents and rapid topology changes are mitigated by backward-compatible model distillation and warm-started model banks.
  • Standardization: Interoperability requires unified semantics-API, signaling protocols, and agent KPIs.

Extension to broader semantic communication scenarios includes continual learning for evolving knowledge bases, modular testbeds for reproducibility, and integrating federated privacy protocols (Gao et al., 29 Dec 2025).

7. Summary and Outlook

AgentSemSteCom manifests the convergence of agentic AI with semantic communication, generating a knowledge-augmented, security-hardened, and resource-optimal framework for task-oriented wireless intelligence. The principal advancements—coverless steganography via token-controlled invertible diffusion, agentic orchestration of feature extraction and codec adaptation, and exact semantic recovery—set a new benchmark for confidential, high-capacity, low-latency information exchange. Emerging research focuses on robust semantic intent signaling, modular agent deployment, and dynamic evolution of semantic-code and knowledge-base structures (Meng et al., 23 Jan 2026, Yu et al., 1 Oct 2025, Gao et al., 29 Dec 2025).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Agentic AI-driven SemSteCom (AgentSemSteCom).