An Erdős problem on random subset sums in finite abelian groups
Abstract: Let $f(N)$ denote the least integer $k$ such that, if $G$ is an abelian group of order $N$ and $A \subseteq G$ is a uniformly random $k$-element subset, then with probability at least $\tfrac12$ the subset-sum set ${ \sum_{x \in S} x : S \subseteq A }$ equals $G$. In 1965, Erdős and Rényi proved that for all $N$, $$ f(N) \le \log_2 N + \left(\frac{1}{\log 2}+o(1)\right)\log\log N. $$ Erdős later conjectured that this bound cannot be improved to $f(N)\le \log_2 N+o(\log\log N)$. In this paper we confirm this conjecture by showing that, for primes $p$, $$ f(p)\ge \log_2 p+\left(\frac{1}{2\log 2}+o(1)\right)\log\log p. $$ This work is an outcome of human--AI collaboration: the original qualitative proof was generated autonomously by ChatGPT-5.2 Pro, while the quantitative refinement was developed by the authors.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
Overview
Think of a big clock with hours, where is a prime number. You pick different “times” on the clock. Now look at all possible sums you can make by adding some of your picked times (you can use each picked time at most once), and wrap around the clock whenever you pass (this is called mod addition). The main question of the paper is: how large does have to be so that, with good probability, these sums hit every hour on the clock?
The authors prove that you need a little more than just “enough” combinations to cover all hours. In fact, even though there are different subset-sums, that isn’t quite enough: you also need an extra “log log” factor. This confirms a prediction (a conjecture) made by the famous mathematician Paul Erdős.
What questions were they asking?
- If you pick random elements from a group of size (you can picture hours on a clock), and then take all possible sums of some of those elements, when do you get every element of the group?
- Let be the smallest so that, with probability at least 1/2, you cover all elements by these subset sums. Erdős and Rényi had shown long ago that
Erdős later wondered if you could remove most of that “” and still be okay.
- This paper shows the answer is no: you cannot shrink that “” too much. In fact, for prime , you still need roughly half of it.
How did they study the problem?
Here is the setup in everyday language, with simple analogies:
- The group: Imagine a clock with hours. Adding times is just moving around the clock and wrapping around at . Mathematicians call this the group under addition mod .
- Picking numbers: Choose random hours on the clock.
- Subset sums: Make every sum you can by adding some of your chosen hours (each used at most once). There are such sums (including the empty sum, which gives 0).
- Goal: Do these sums hit every hour on the clock?
Two key ideas make the proof work:
- Random throws into buckets (Poisson analogy)
- Picture the hours as buckets.
- There are nonempty subset sums. If the chosen hours are random, each subset sum acts “roughly like” a random throw into one of the buckets.
- The average number of throws per bucket is .
- If is only barely bigger than , then is modest (not huge). In that case, it’s quite likely some buckets stay empty (some hours are never hit).
- In probability, the number of hits per bucket behaves like a Poisson random variable with mean . The chance a fixed bucket is empty is about .
- Second moment method (controlling fluctuations)
- Let be the total number of empty buckets (hours that none of the sums hit).
- The authors compute the average of and also control how spread out is (its variance).
- If the average number of empty buckets is large and the spread is not too big, then with high probability there really are empty buckets. That means you do not cover the whole clock.
Behind the scenes, they:
- Switch to a convenient “independent” model (where repeated picks are allowed) and show it behaves the same when is small compared to .
- Prove carefully that the “Poisson-like” behavior holds for one bucket and for two buckets at once (this matters for controlling the variance).
- Use inclusion–exclusion (Bonferroni) and “factorial moments” to justify the Poisson approximation.
- Use linear algebra over and over to count and control rare “structured” dependencies that could break the randomness.
You can think of all this as: “We have raindrops (subset sums) falling on tiles (group elements). When is not big enough, many tiles stay dry, and we can quantify exactly how likely that is.”
What did they find, and why does it matter?
Main finding (for primes ):
- If you choose
with any constant , then the probability that your subset sums hit all of goes to 0 as grows. In short: with that small , you almost surely miss some hours on the clock.
- Therefore,
The “” means a term that goes to 0 as grows.
Why important:
- This confirms Erdős’s conjecture from 1973: you cannot improve the old upper bound all the way down to . You really do need a “”-sized extra term.
- It sharpens our understanding of how many random elements you need so that their subset sums cover everything. The basic “” rule-of-thumb isn’t enough; collisions and randomness force you to add a specific extra cushion of size about (or more).
The role of AI in this research
The project was a human–AI collaboration:
- The authors asked a LLM to find a counterexample to the optimistic bound. The LLM produced a proof idea with a gap.
- After prompting, the LLM patched its argument to give a valid qualitative disproof (showing the bound can’t be that small).
- The human authors then strengthened and refined the argument to get the sharp quantitative result above for prime .
- The paper documents this workflow and provides references to public records of the process.
This is a nice example of AI suggesting a promising path and humans turning it into a rigorous, precise result.
Implications and what’s next
- The exact “best” second term is still open. Define
The paper shows
Heuristics suggest the upper value might be the truth, but new ideas are needed to prove it.
- Structure matters: in some special groups (like ), the story can be different, and you might be able to cover everything with smaller . Understanding which group structures help or hurt coverage is an ongoing direction.
- Methodological impact: The paper shows how probabilistic thinking (Poisson approximations, second moments) and linear algebra can be combined, and it highlights how AI can assist in modern mathematical discovery.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
Below is a consolidated list of what remains unresolved or underexplored, focusing on concrete directions for future work:
- Closing the constant gap in the second-order term: determine the exact constant c* = limsup_{N→∞} (f(N) − log_2 N)/log log N, currently bounded by 1/(2 log 2) ≤ c* ≤ 1/log 2. Heuristics suggest c* = 1/log 2, but the present method cannot exceed 1/(2 log 2).
- Extension beyond prime orders: the quantitative lower bound is proved only for G ≅ F_p. Develop techniques for composite N (especially prime powers and general finite abelian groups), where linear-algebra over fields is unavailable and the current rank-stability arguments over F_p break.
- Group-structure dependence: characterize for which abelian groups G of order N a random k-set with k = log_2 N + o(log log N) typically covers G, and identify the worst-case groups for which the larger second-order term is necessary.
- Sharp threshold and window: identify the sharp threshold k_c(N) for the property Σ(A) = G and the width of the transition window around k ≈ log_2 N + Θ(log log N). Determine whether the window is O(1), O(√log log N), or something else.
- High-probability regime: the paper proves P(Σ(A) = G) → 0 when k = log_2 p + c log log p with c < 1/(2 log 2), and prior work gives a 1/2-probability upper bound near c = 1/log 2. Establish thresholds for P → 1 (rather than ≥ 1/2), and quantify the dependence on the target probability level.
- Distributional limit for the number of misses U: beyond the first two moments used to show P(U = 0) → 0, determine the limiting law of U (e.g., Poisson, normal, or compound Poisson after appropriate centering/scaling) and establish concentration around its mean p e{−λ}.
- Stronger Poisson approximation and joint independence: Proposition 3.1 gives Poisson-like estimates only for m ∈ {1, 2}. Extend to fixed m ≥ 3 (or m growing slowly with p), control mixed factorial moments uniformly, and develop a Chen–Stein framework (or dependency-graph approach) for the joint vector (1_{X_x=0})_{x∈F_p\times}.
- Improving the “3/4” barrier in Lemma 2.4 (and Lemma 3.3): the bound |W ∩ {0,1}r| ≤ (3/4)·2d under the “no 1- or 2-sparse orthogonal” condition is a key bottleneck. Find stronger anticoncentration/entropy bounds (possibly via Littlewood–Offord-type or additive combinatorics) to reduce the constant below 3/4 and thereby push the lower-bound constant c beyond 1/(2 log 2).
- Rank-stability range (Lemma 2.1): the equality rank_Q(V) = rank_{F_p}(V) is proved only for r ≤ (log p)β with β < 1/2 using Hadamard’s bound. Optimize this to the natural barrier r up to ≈ (2 + o(1))·log p / log log p (the threshold where s{s/2} < p) to enable higher-moment control and potentially stronger constants.
- Low-rank incidence counting (Lemma 3.3): the upper bound T_{r,d} ≤ 2{r2}·((3/4)·2d)k is crude. Develop sharper enumeration of low-rank 0–1 incidence patterns (e.g., via refined column-space structure, coding-theoretic bounds, or spectral methods) to reduce the low-rank mass and strengthen the main term.
- Beyond coupling with replacement (Section 3.1): the reduction from the subset model to the i.i.d. model uses a union bound and only controls o(1) differences when k = O(log p). For finer distributional results (or larger k), quantify and reduce the model discrepancy more precisely.
- Larger truncation R in Bonferroni: analysis truncates inclusion–exclusion at R = ⌊(log p)β⌋. Pushing R higher (with matching rank-stability and low-rank counting) would enable control of more factorial moments and improve constants.
- Binomial selection model: analyze the analogous threshold when elements of G are included independently with probability q (rather than choosing a fixed-size k-set), and relate the two models quantitatively at the subleading order.
- Sensitivity to multiple-target coverage: study the joint event that a prescribed finite set B ⊂ G is missed (or covered) and extend the uniformity in Proposition 3.1 to |B| growing slowly with p; this is a step toward Poisson approximation for U.
- Refined upper bounds: go beyond the Erdős–Rényi upper bound f(N) ≤ log_2 N + (1/log 2)·log log N + O(1) by identifying the exact O(1) term (or logarithmic corrections) and matching it with a corresponding lower bound.
- Non-abelian analogues: while the paper focuses on abelian groups and subset sums, explore analogous questions for subset products in non-abelian groups (with appropriate definitions), and determine whether group structure similarly dictates second-order thresholds.
- Robustness to modeling choices: investigate variants such as excluding the empty sum, restricting subset sizes, or bounding the number of summands allowed, and determine how these constraints shift the threshold.
Practical Applications
Overview
The paper settles a long-standing Erdős conjecture on random subset sums in finite abelian groups by proving a quantitative lower bound: for prime p, a uniformly random k-element subset A ⊂ F_p fails to cover F_p by subset sums with probability tending to 1 unless k ≥ log₂ p + (1/(2 log 2) + o(1)) log log p. Methodologically, it develops a Poisson-like estimate via factorial moments and Bonferroni inequalities, combines it with a second-moment argument, and uses structural lemmas about ranks of 0–1 matrices over Q vs F_p. The paper also documents a human–AI collaboration workflow where an LLM produced a qualitative argument later refined by the authors.
Below are actionable, sector-linked applications derived from the findings and techniques, grouped by deployment horizon. Each item notes assumptions and dependencies affecting feasibility.
Immediate Applications
The following applications can be deployed now, primarily in research, cryptography practice, and software tooling.
- Additive combinatorics and probability method toolset (Academia)
- Application: Use the paper’s Poisson-like “miss” approximation (via factorial moments and inclusion–exclusion) and second-moment analysis as a template for threshold problems where many weakly dependent events collectively govern a coverage phenomenon (e.g., sumset coverage, random linear combinations spanning a target set, occupancy-like models with dependencies).
- Tools/workflows:
- PoissonMomentApprox: a small library that computes truncated factorial moments and Bonferroni bounds for dependent indicator sums, with sanity checks against Monte Carlo.
- CoverageThresholdExplorer: scripts to simulate “coverage vs. k” curves for group-sum models and related combinatorial coverage processes.
- Assumptions/dependencies: Asymptotic regime (large p), ability to estimate low-order factorial moments up to r ≤ (log p)β, and mild structural conditions to control low-rank configurations.
- Rank-stability diagnostics for small 0–1 systems (Academia, Software)
- Application: Deploy Lemma-level insights that for 0–1 matrices with O((log p)β) rows, rank over Q matches rank over F_p for large p. Useful when transferring linear-algebraic arguments between fields in probabilistic combinatorics and coding constructions with small constraint sets.
- Tools/workflows:
- RankStabilityChecker: a utility that, given bounds on row size and p, certifies that rank computations over Q suffice for F_p analysis.
- Assumptions/dependencies: Row count subpolynomial in p (here ≤ (log p)β with β < 1/2); entries in {0,1}; sufficiently large prime p.
- Guidance for cryptographic parameter intuition (Cryptography, Software)
- Application: Informally calibrate expectations about the number of random elements needed so that subset sums over F_p “look” surjective. This cautions against assuming full coverage from too-small random bases in constructions that rely on subset-sum mixing (e.g., toy schemes, randomized hashing over cyclic groups).
- Tools/workflows:
- SubsetSumCoverageEstimator: a simulator that estimates coverage probability P(Σ(A) = F_p) for given p, k, reporting whether one is below the proven barrier region.
- Assumptions/dependencies: Results are asymptotic and proven for primes; real systems should not rely solely on these bounds for security claims.
- Modeling without-replacement vs with-replacement sampling (Software, Data Science)
- Application: Use the paper’s coupling between the uniform-k subset model and the i.i.d. model (when k = O(log p)) as a principled approximation to simplify analysis and simulation of sampling without replacement.
- Tools/workflows:
- IIDCouplingApprox: a reusable component offering error bounds when replacing uniform-k sampling by i.i.d. sampling for small k relative to population size.
- Assumptions/dependencies: k ≪ √p; here k = O(log p) ensures collision probability o(1).
- Teaching and research training modules (Education, Academia)
- Application: Incorporate worked examples on factorial-moment/Bonferroni techniques, second-moment arguments, and the Q vs F_p rank transfer into advanced probability/combinatorics curricula; use the paper’s human–AI workflow as a case study in AI-assisted theorem discovery and verification.
- Tools/workflows:
- Classroom notebooks illustrating: (i) constructing X_B and its factorial moments, (ii) Bonferroni truncation error control, (iii) low-rank pattern counting, (iv) simulation vs theory overlays.
- Assumptions/dependencies: None beyond curricular fit.
- AI-in-research transparency protocol (Policy, Academia, Research Management)
- Application: Adopt the paper’s disclosure pattern (segregating AI-generated content, independent verification, public artifacts) as a template for responsible AI use in mathematical research.
- Tools/workflows:
- AI-Contrib-Statement template; structured audit trails for proof provenance; repository checklists for AI-generated drafts and human verification notes.
- Assumptions/dependencies: Institutional readiness to accept and standardize AI contribution statements; version-controlled artifacts.
Long-Term Applications
The following require further research, scaling, or integration to mature into robust practice.
- Sharpening threshold constants for coverage phenomena (Academia, Coding Theory)
- Application: Extend methods to determine the exact constant c* in f(N) = log₂ N + c* log log N + o(log log N), and adapt to other groups (e.g., vector spaces over small fields) and to coding-theoretic questions about when random linear combinations or parity checks guarantee coverage/spanning or syndrome surjectivity.
- Potential products:
- ThresholdFinder: a research platform combining symbolic bounds, rank-pattern counting, and Monte Carlo to conjecture and test sharp constants in random coverage/spanning problems.
- Assumptions/dependencies: Improved counting of low-rank configurations beyond the paper’s Lemma-level bottleneck; possibly new combinatorial geometry/entropy methods.
- Security analyses for systems using subset-sum-like mixing (Cryptography)
- Application: Guide the design and analysis of cryptographic or hashing primitives that implicitly rely on coverage/surjectivity of random linear combinations modulo p (e.g., certain accumulator constructions, probabilistic encodings, randomized commitment schemes).
- Potential products:
- MixCoverageAuditor: static-analysis tooling that flags parameter regimes where subset-sum coverage is statistically unlikely, prompting parameter increases or alternative designs.
- Assumptions/dependencies: Formal reductions linking coverage properties to concrete security notions; bridging asymptotic results to finite-size guarantees with tight finite-sample bounds.
- Data structures and hashing with algebraic mixing (Software, Systems)
- Application: Inform the design of algebraic filters or sketches (e.g., XOR-like filters over F_p) where successful initialization or query accuracy hinges on solving linear systems or achieving coverage in group-sum constructions; predict failure thresholds and rehash rates.
- Potential products:
- AlgebraicFilterDesigner: planner that chooses p and k to balance space/time against initialization success probabilities derived from coverage thresholds.
- Assumptions/dependencies: Mapping between specific data-structure invariants and the subset-sum coverage model; adaptation from F_2 to general F_p when beneficial.
- Automated theorem discovery pipelines (AI for Science, Policy)
- Application: Systematize the paper’s human–AI collaboration pattern into a reproducible pipeline: LLMs for conjecture generation and sketch proofs; programmatic gap-finding; human quantitative refinement; public audit artifacts.
- Potential products:
- ProofPilot: an orchestration framework integrating LLMs, computer algebra, SAT/SMT solvers, and proof assistants to iterate between heuristic proofs and verified refinements.
- ProvenanceLedger: standardized metadata and documentation for AI-assisted results suitable for journals and grant reporting.
- Assumptions/dependencies: Continued advances in LLM mathematical reliability, integration with formal proof systems, community standards for credit and reproducibility.
- General Poisson-approximation frameworks for dependent coverage (Academia, Quantitative Finance, Network Reliability)
- Application: Extend factorial-moment/Bonferroni control to other dependent-coverage settings (e.g., network link activation leading to connectivity, rare-event portfolio coverage, sensor coverage problems) where classical Poisson approximation is delicate.
- Potential products:
- DependentPoissonKit: a library for bounding deviation between dependent indicator sums and Poisson laws using moment truncations plus combinatorial structure controls.
- Assumptions/dependencies: Problem-specific structure to bound low-rank or high-dependence configurations; careful translation of asymptotic proofs into non-asymptotic, domain-appropriate error bars.
- Standards and governance for AI disclosures in mathematics (Policy)
- Application: Develop community norms and publishing standards for AI contribution statements, artifact availability, and verification practices inspired by this paper’s transparency.
- Potential products:
- Journal policy templates; funder guidelines; conference artifact-evaluation tracks for AI-in-the-loop mathematics.
- Assumptions/dependencies: Community consensus and coordination among societies, journals, and funders.
Cross-cutting assumptions and dependencies to monitor
- Asymptotic-to-finite translation: Many guarantees depend on p → ∞; deriving sharp finite-size bounds is often problem-specific and requires additional effort.
- Group structure: The main quantitative lower bound is proved for primes (cyclic groups of order p); extensions to arbitrary finite abelian groups may require new ideas.
- Independence approximations: The i.i.d. reduction relies on k = O(log p) to keep collisions negligible; outside this regime, coupling errors may dominate.
- Low-rank counting bottlenecks: Pushing constants beyond 1/(2 log 2) likely hinges on stronger bounds for the number of low-rank incidence patterns; progress here unlocks multiple long-term applications.
- AI reliability: The viability of AI-assisted discovery pipelines depends on improved model validity, verification tooling, and agreed-upon reporting standards.
Glossary
- Abelian group: A group in which the group operation is commutative. "if is an abelian group of order "
- Additive group: A group structure using addition as the operation. "and here we only view it as an additive group."
- Bernoulli distribution: A distribution on {0,1} with a given success probability; here, with parameter 1/2. "where are independent and ."
- Bonferroni inequalities: Alternating-sum bounds derived from inclusion–exclusion, used to estimate probabilities. "proceeds by applying Bonferroni inequalities as well as estimating factorial moments"
- Chebyshev's inequality: A bound relating variance to tail probabilities. "and Chebyshev's inequality gives"
- Column space: The subspace spanned by the columns of a matrix. "Let denote the column space of over ."
- Cyclic group: A group generated by a single element. "so has to be isomorphic to the cyclic group of order "
- Falling factorial: The product for integer . "let denote the falling factorial."
- Factorial moments: Expectations of falling factorials of a random variable, used in Poisson approximations. "estimating factorial moments of an associated Poisson-like random variable"
- Finite field: A field with finitely many elements; here with prime. "The notion usually stands for the field of elements"
- Hadamard's inequality: An upper bound on the absolute value of a determinant in terms of column norms. "By Hadamard's inequality, "
- Homogeneous linear form: A linear functional with no constant term. "there exists a homogeneous linear form "
- i.i.d. model: A model where variables are independent and identically distributed. "In the independent (i.i.d.)\ model, let be i.i.d.\ uniform on "
- Inclusion–exclusion principle: A combinatorial identity for counting unions via alternating sums. "which arise from the inclusion-exclusion principle on factorial moments"
- Incidence matrix: A 0–1 matrix encoding membership of elements in sets. "we write for the incidence matrix"
- Indicator vector: A 0–1 vector indicating membership of elements in a subset. "let be its indicator vector."
- Isomorphism: A bijective structure-preserving map between algebraic structures. "such that the projection is an isomorphism."
- Kernel (of a linear map): The set of inputs mapped to zero by a linear transformation. "every fiber has size ."
- Law of total probability: Decomposes probabilities conditioned on a partition of events. "By the law of total probability, for any event we have"
- Orthogonal complement: The set of vectors orthogonal to a given subspace. "Then its orthogonal complement contains no nonzero vector supported on at most two coordinates."
- Poisson random variable: A count variable with distribution determined by its mean; arises as a limit of rare events. "comparing them with those of a Poisson random variable of mean "
- Projection (linear): A linear map that extracts selected coordinates/components. "the projection "
- Rank (matrix): The dimension of the column space of a matrix. "$\rank_{Q}(V)=\rank_{F_p}(V).$"
- Second moment method: A probabilistic technique using first and second moments to bound probabilities. "\keywords{random subset sums, Erd\H{o}s problems, abelian groups, second moment method}"
- Stirling's formula: An asymptotic approximation for factorials. "Stirling's formula yields"
- Subset-sum set: The set of all sums of distinct elements from a subset. "the subset-sum set equals ."
- Surjective (map): A function whose image equals its codomain. "the map (defined in the proof of Lemma~\ref{lem:rank-prob}) is surjective"
- Support (of a vector): The set of indices where the vector has nonzero entries. "supported on at most two coordinates"
- Union bound: An inequality upper-bounding the probability of a union by the sum of probabilities. "a union bound gives"
- Vinogradov's asymptotic notation: Notation such as , , , , for comparing growth rates. "We use Vinogradov's asymptotic notation."
Collections
Sign up for free to add this paper to one or more collections.