Learning-Augmented PSMM Protocol
- The paper presents the Learning-Augmented PSMM protocol that integrates information-theoretic privacy via polynomial masking with computational efficiency gained from low-rank tensor decompositions.
- It guarantees perfect privacy under a semi-honest model with up to t-1 colluding agents and achieves optimal recovery thresholds through careful algebraic encoding and polynomial interpolation.
- By leveraging learned tensor decompositions, the protocol significantly reduces local computation costs—demonstrating up to an 80% speedup for large matrices—while maintaining rigorous security guarantees.
Learning-Augmented PSMM (Perfectly Secure Collaborative Matrix Multiplication) is a protocol designed for secure multiparty computation (MPC) of matrix products, specifically over finite fields $\F$, under strict storage and privacy constraints. The core innovation integrates information-theoretic secrecy based on polynomial-masking techniques with computational speedup through learning-based, low-rank tensor decompositions, yielding substantial improvements in local computation while maintaining perfect security guarantees (He et al., 14 Jan 2026).
1. Problem Framework and Security Model
Given semi-honest agents, connected to a controller and able to store at most a $1/k$ fraction of each input matrix (i.e., one block of and ), the protocol considers the scenario where at most agents may collude. The objective is for a controller to compute exactly, ensuring:
- Information-theoretic privacy: Any coalition of up to agents obtains no information about the inputs.
- Local storage constraint: Each agent holds exactly one block of each matrix plus masking randomness.
- Optimal recovery threshold: The number of agents achieves matching lower bounds for polynomial-sharing-based secure matrix multiplication.
This setting adheres to the standard security definitions in MPC and coded computing, with explicit attention to storage and collusion bounds, and assumes a trusted source and private authenticated channels (He et al., 14 Jan 2026).
2. Algebraic Structure: Polynomial Masking and Coefficient Alignment
Input matrices $A, B \in \F^{m \times m}$ are partitioned into column blocks:
$A = [A_1~\cdots~A_k],~~B = [B_1~\cdots~B_k],~~A_i, B_j \in \F^{m \times (m/k)}.$
Each block is encoded as a sparse masking polynomial; for example: \begin{align*} g_A(x) &= \sum_{i=1}k A_i x{i-1} + \sum_{\ell=1}{t-1} R{(A)}_\ell x{k2+\ell-1}, \ g_B(x) &= \sum_{j=1}k B_j x{k(j-1)} + \sum_{\ell=1}{t-1} R{(B)}_\ell x{k2+\ell-1}. \end{align*} The “signal support” term contains input blocks; the “masking tail” terms (, drawn uniformly and independently over $\F^{m \times (m/k)}$) ensure information-theoretic security.
For each publicly chosen $\alpha_n \in \F$, agent receives . Local computation yields
which, as a polynomial in , decomposes into coefficients such that the indices directly recover the products , while all other coefficients are linear combinations involving random masks.
The scheme is reminiscent of Beaver-style MPC but is realized via algebraic encoding: the masking tail effectively plays the role of classical Beaver triples, ensuring that any set of evaluations (i.e., views of colluding agents) reveals zero information about the signals, as formalized by the polynomial masking lemma (He et al., 14 Jan 2026).
3. Recovery Thresholds and Information-Theoretic Privacy
The number of nonzero coefficients in determines the minimal number of agents required for recovery:
By assigning agents and choosing randomly, the protocol constructs a block-Vandermonde interpolation system of full rank with high probability. This approach guarantees both perfect privacy against up to colluders and optimal recovery, matching known information-theoretic lower bounds for polynomial-sharing protocols, under the given storage and privacy parameters (He et al., 14 Jan 2026).
4. Learning-Augmented Protocol: Tensorization and Low-Rank Methods
The learning-augmented extension—LA-PSMM—replaces each agent’s dense multiplication by a lower-rank, learned tensor decomposition. Conventional local operations require finite-field operations, which becomes prohibitive for large . Instead, matrices are multiplied in bilinear tensorized form:
where is the rank of the decomposition; for Strassen’s method, for matrix multiplication, while learned decompositions (e.g., via AlphaTensor) achieve larger , enabling scalable reductions in local computation.
The local step thus becomes:
$\hat{M}(\alpha_n) = \sum_{r=1}^{T_l} \langle u_r, \operatorname{vec}(g_A(\alpha_n)) \rangle \langle v_r, \operatorname{vec}(g_B(\alpha_n)) \rangle \mat(w_r),$
with learned from tensor decomposition, and the learned rank (He et al., 14 Jan 2026).
Operator-invariance: If the local bilinear mapping is exactly equivalent to for all , it preserves the distribution of signal and masking coefficients in , thus does not compromise privacy or the recovery threshold.
5. Protocol Workflow and Computational Complexity
Protocol Workflow
The LA-PSMM protocol consists of the following steps:
| Step | Actor | Operation |
|---|---|---|
| Partition | Source plant | Split into blocks |
| Masking | Source plant | Sample |
| Polynomial encoding | Source plant | Form |
| Point selection | Source plant | Choose |
| Share distribution | Source plant | Send |
| Local multiplication | Agent | Compute via learned expansion |
| Upload results | Agent | Send to controller |
| Interpolation | Controller | Reconstruct from block coefficients |
Complexity
- PSMM: Encoding involves operations; local multiply scales as per agent; decoding uses (naive) or near-linear (fast) methods.
- LA-PSMM: Encoding/decoding overheads are unchanged. Local multiplication reduces to , provided .
Speedup is observed when , with empirical results reaching up to local speedup for , and speedup scaling approximately linearly with (e.g., a reduction for ). This suggests LA-PSMM achieves substantial gains for large matrix dimensions and moderate partition factors (He et al., 14 Jan 2026).
6. Security Analysis and Theoretical Guarantees
The privacy of LA-PSMM is founded on:
- The masking lemma, ensuring that any polynomial evaluations are statistically independent of , thus any subset of colluding agents observes fully random, independent shares.
- Operator-invariance, as learned bilinear expansions in LA-PSMM are constructed to be exactly equivalent to the standard multiplication for all field inputs, ensuring that the polynomial masking structure and critical “signal” exponents (those from which are recovered) remain unchanged.
- The recovery threshold remains , with the Vandermonde interpolation problem remaining full-rank due to the polynomial structure and random block selection.
The LA-PSMM protocol thus inherits the perfect privacy, optimal recovery, and correctness guarantees of the original PSMM, while introducing no new vulnerabilities (He et al., 14 Jan 2026).
7. Empirical Evaluation and Implications
Experimental benchmarks, conducted over $\F_p$ for large prime and square matrices of size , demonstrate:
- For , , and learned rank (using AlphaTensor-style reinforcement learning), LA-PSMM local computation times are reduced to of conventional PSMM for .
- Speedup, defined as PSMM time divided by LA-PSMM time, increases almost linearly with , reaching approximately for .
- Wall-clock times (excluding network latency) confirm the scalability of LA-PSMM’s local computation cost advantage for increasing matrix dimensions.
A plausible implication is that the practical cost of perfect secrecy can now be substantially reduced in large-scale collaborative or distributed settings, provided suitable low-rank bilinear decompositions are available and exact (He et al., 14 Jan 2026).
Learning-augmented PSMM synthesizes block-masked, information-theoretically secure MPC protocols with advances in learning-based tensor decompositions. This union enables scalable, perfectly secure collaborative matrix multiplication under strong adversarial models, with computational efficiency improvements that scale with problem size, offering a highly practical primitive for coded computing with robust privacy-preserving guarantees (He et al., 14 Jan 2026).