Papers
Topics
Authors
Recent
Search
2000 character limit reached

3P-ADMM-PC2: Parallel Privacy-Preserving ADMM

Updated 28 January 2026
  • The paper introduces 3P-ADMM-PC2, a protocol that integrates ADMM with homomorphic encryption and quantization for secure, distributed LASSO optimization.
  • It employs a three-phase structure—initialization, secure data sharing, and parallel privacy-computing—with adaptive GPU acceleration to reduce computation and communication overhead.
  • Experimental outcomes demonstrate near-lossless accuracy and significant runtime improvements, showcasing scalability in large-scale edge networks compared to traditional CPU implementations.

Three-Phase Parallel Collaborative ADMM Privacy Computing (3P-ADMM-PC2) is a cryptographically enhanced distributed optimization protocol designed for edge networks, addressing the need to simultaneously reduce computational burden, minimize information leakage, and enable privacy-preserving model training over split data. It integrates the alternating direction method of multipliers (ADMM) with Paillier homomorphic encryption (HE), a real-to-integer quantization strategy, and adaptive GPU acceleration for efficient and private solution of high-dimensional, distributed LASSO problems (Xia et al., 21 Jan 2026).

1. Protocol Structure: The Three Phases

3P-ADMM-PC2 operates in three sequential phases that collectively enable secure, distributed ADMM optimization on partitioned data:

  1. Initialization Phase: The master node partitions the large global LASSO problem,

minxRN12yAx22+λx1,\min_{x\in\mathbb R^N}\frac12\|y - A x\|_2^2 + \lambda\|x\|_1,

column-wise into KK smaller subproblems. For each edge kk, the master transmits {AkTAk,ρ}\{A_k^T A_k,\,\rho\}, and edge kk precomputes Bk=(AkTAk+ρI)1B_k=(A_k^T A_k + \rho I)^{-1} and quantized Bˉk=Γ2(Bkρ)\bar B_k=\Gamma_2(B_k \rho).

  1. Data Security Sharing Phase: The master quantizes and Paillier-encrypts the sensitive vector BkAkTyB_k A_k^T y for each edge as α^k=fen(Γ1(BkAkTy))\hat\alpha_k = f_{en}(\Gamma_1(B_k A_k^T y)). Edge nodes download and locally store these encrypted values for reuse.
  2. Parallel Privacy-Computing Phase: For t=1,2,t=1,2,\dots, master and edges collaboratively perform an ADMM iteration on encrypted data:
    • Master computes z(t)=Sλ/ρ(v(t1)+x(t1))z^{(t)} = S_{\lambda/\rho}(v^{(t-1)} + x^{(t-1)}) and v(t)=v(t1)+x(t1)z(t)v^{(t)} = v^{(t-1)} + x^{(t-1)} - z^{(t)}. It quantizes and Paillier-encrypts zk(t)z_k^{(t)}, vk(t)-v_k^{(t)} as z^k(t)=E(Γ2(zk(t)))\hat z_k^{(t)} = E(\Gamma_2(z_k^{(t)})) and v^k(t)=E(Γ2(vk(t)))\hat v_k^{(t)} = E(\Gamma_2(-v_k^{(t)})).
    • Edge kk performs, under Paillier homomorphism,

    x^k(t)=α^k[Γ2(Bˉk)(z^k(t1)(v^k(t1)))],\hat x_k^{(t)} = \hat\alpha_k \oplus \left[\Gamma_2(\bar B_k) \otimes (\hat z_k^{(t-1)} \oplus (-\hat v_k^{(t-1)}))\right],

    then returns x^k(t)\hat x_k^{(t)} to the master, which decrypts and inverse-quantizes to recover xk(t)x_k^{(t)}.

Each node operates exclusively on low-dimensional data, ensuring privacy and reducing communication overhead (Xia et al., 21 Jan 2026).

2. ADMM Update Mechanism

The protocol is anchored in ADMM for the LASSO regression objective:

minx,z  12yAx22+λz1subject to xz=0,\min_{x,z}\; \tfrac12\|y-Ax\|_2^2 + \lambda\|z\|_1\quad \text{subject to } x-z=0,

with augmented Lagrangian,

Lρ(x,z,u)=12yAx22+λz1+uT(xz)+ρ2xz22.\mathcal L_\rho(x,z,u) = \tfrac12\|y-Ax\|_2^2 + \lambda\|z\|_1 + u^T(x-z) + \tfrac\rho2\|x-z\|_2^2.

Centralized ADMM has the update rules:

  • xx-update: (ATA+ρI)1(ATy+ρ(z(t1)u(t1)/ρ))(A^T A + \rho I)^{-1}(A^T y + \rho(z^{(t-1)} - u^{(t-1)}/\rho))

  • zz-update: Sλ/ρ(x(t)+u(t1)/ρ)S_{\lambda/\rho}(x^{(t)} + u^{(t-1)}/\rho)

  • uu-update: u(t1)+ρ(x(t)z(t))u^{(t-1)} + \rho(x^{(t)} - z^{(t)})

The 3P-ADMM-PC2 distributed form upper-bounds yAkxk2\|y-\sum A_k x_k\|^2, leading to per-node subproblems,

xk(t)=(AkTAk+ρI)1(AkTy/K+ρ(zk(t1)vk(t1))),x_k^{(t)} = (A_k^T A_k + \rho I)^{-1}(A_k^T y/K + \rho(z_k^{(t-1)}-v_k^{(t-1)})),

with synchronized global zz and vv updates.

This matrix partitioning and update design enables independent encrypted computations at each edge, with secure aggregation by the master (Xia et al., 21 Jan 2026).

3. Quantization for Real-Valued Encryption

Because Paillier HE only supports integer arithmetic, real-valued vectors are mapped into finite integer intervals for encryption:

  • For vector u[umin,umax]u\in[u_{\min}, u_{\max}], use

Γ2(u)=Δuuminumaxumin[0,Δ]\Gamma_2(u) = \left\lfloor \Delta \frac{u-u_{\min}}{u_{\max}-u_{\min}} \right\rceil \in [0, \Delta]

  • For matrix-vector or two-term operations, use squared scaling:

Γ1(w)=Δ2wwmin(wmaxwmin)2\Gamma_1(w) = \left\lfloor \Delta^2 \frac{w-w_{\min}}{(w_{\max}-w_{\min})^2} \right\rceil

Rounding error per entry is at most $1/2$, so worst-case reconstruction error scales as O(1/Δ)O(1/\Delta). The decrypted output after inverse quantization differs from the true real value by O((zmaxzmin)2/Δ2)O((z_{\max}-z_{\min})^2/\Delta^2). With practical Δ\Delta, this quantization error becomes negligible, ensuring near-lossless privacy-preserving updates (Xia et al., 21 Jan 2026).

4. Paillier Homomorphic Encryption Scheme

3P-ADMM-PC2 applies the Paillier cryptosystem:

  • Key generation involves large primes p,qp,q, modulus n=pqn=pq, special gZn2g \in \mathbb{Z}_{n^2}^*, and computation of λ\lambda and μ\mu.

  • Encryption of a message mm:

E(m)=gmrnmodn2,rZnE(m) = g^m r^n \bmod n^2,\quad r\in\mathbb{Z}_n^*

  • Decryption of ciphertext cc:

D(c)=[L(cλ ⁣modn2)μ]modn,L(x)=(x1)/nD(c) = [L(c^\lambda\!\bmod n^2)\mu] \bmod n,\quad L(x) = (x-1)/n

  • Homomorphic properties:
    • E(m1)×E(m2)=E(m1+m2modn)E(m_1)\times E(m_2) = E(m_1+m_2\bmod n)
    • E(m)k=E(kmmodn)E(m)^k = E(k m\bmod n)

These properties realize the secure sum and scalar-multiply required in the edge-side ADMM subproblem,

x^k(t)=α^k[Γ2(Bˉk)(z^k(t1)(v^k(t1)))].\hat x_k^{(t)} = \hat\alpha_k \oplus \left[\Gamma_2(\bar B_k) \otimes (\hat z_k^{(t-1)} \oplus ( -\hat v_k^{(t-1)}))\right].

This procedure never exposes the raw data vectors, ensuring full Paillier-level confidentiality during collaborative computations (Xia et al., 21 Jan 2026).

5. Adaptive GPU Acceleration

Due to the computational intensity of large-integer modular exponentiation, 3P-ADMM-PC2 adopts several GPU-specific optimizations:

  1. CRT Decomposition: Modular exponentiation modn2\bmod\, n^2 is decomposed into computations modulo p2p^2 (edges) and q2q^2 (master), combined via the Chinese Remainder Theorem.
  2. GPU FFT-accelerated Multiplication: Large integers are represented as digit-vectors; multiplications are performed with FFT and IFFT in parallel on the GPU, with modular reduction (e.g., Barrett reduction) applied via low-bitwidth arithmetic.
  3. Parallel ModExp in GPU Kernels: Each GPU streaming multiprocessor loads low-bitwidth digit chunks; a bitwise loop performs modular multiplications in parallel, using FFT routines and Barrett reduction.
  4. Three-Round CRT Computation: CRT computations are distributed: edges handle computations modp2\bmod\,p^2, master handles modq2\bmod\,q^2, and the result is combined such that neither party ever operates directly on the full n2n^2 modulus.

This strategy achieves substantial speedup: with a 4096-bit Paillier key, GPU throughput for modular exponentiation is approximately 20×\times that of a 64-core CPU (Xia et al., 21 Jan 2026).

6. Computational Complexity and Solution Quality

The chief computational bottleneck is large-integer modular exponentiation (ModExp), which under FFT-based multiplication has complexity O(LlogL)O(L\log L) for LL-digit integers, with each exponentiation requiring O(logn)O(\log n) such multiplications—yielding overall costs per ciphertext of O(logn×LlogL)O(\log n \times L\log L).

Per ADMM iteration, computation is dominated by the O(KNk)O(KN_k) ciphertext operations plus fixed O(KNk3)O(KN_k^3) for local solves. GPU acceleration yields a per-ModExp throughput more than 20 times higher than CPU implementation on long keys.

Convergence analysis (as per approximate ADMM theory) holds so long as quantization and data splitting errors remain uniformly bounded (O(1/Δ)O(1/\Delta) and O(1/M)O(1/\sqrt{M}) respectively). Mean squared error (MSE) deviation from non-private distributed ADMM is on the order of O(1/Δ)1015O(1/\Delta)\approx 10^{-15} in typical settings.

Empirically, on LASSO problems with AR3000×27000A\in\mathbb{R}^{3000\times27000} and K=3K=3, wall-clock times are:

  • CPU-based distributed HE-ADMM: 29,800 s (1024-bit), 41,000 s (2048-bit), 79,300 s (4096-bit)
  • GPU-accelerated 3P-ADMM-PC2: 11,700 s (1024-bit), 20,500 s (2048-bit), 34,900 s (4096-bit) This yields $2$–3×3\times speedup with matching accuracy (Xia et al., 21 Jan 2026).

7. Experimental Outcomes and Topology Variation

Experiments evaluate accuracy, computational efficiency, and scalability:

  • Accuracy: 3P-ADMM-PC2 closely tracks non-private distributed ADMM (Dis-ADMM) with MSE error within 101410^{-14}; by contrast, DP-ADMM (differential privacy) incurs \sim0.2 units higher MSE.
  • Edge node count: For N=65536N=65536, M=10000M=10000, increasing edge count (K=3K=3 to K=10K=10) reduces per-iteration wall-clock but marginally increases partition-induced MSE.
  • GPU latency reduction: Per iteration, node waiting times drop with GPU acceleration—master waiting 12\sim12s (vs. \sim30s on CPU), edges $1$–$2$s (vs. $10$s).
  • Application: Power-network reconstruction: On large-scale MATPOWER (13,569-bus) benchmarks, 3P-ADMM-PC2 achieves AUROC/AUPRC parity with Dis-ADMM, confirming no quality loss.

Combined, these results document that 3P-ADMM-PC2 attains secure, nearly lossless privacy-preserving distributed optimization with significant runtime gains in heterogeneous, large-scale edge network settings (Xia et al., 21 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Three-Phase Parallel Collaborative ADMM Privacy Computing (3P-ADMM-PC2).