Papers
Topics
Authors
Recent
Search
2000 character limit reached

Resolution of Erdős Problem #728: a writeup of Aristotle's Lean proof

Published 12 Jan 2026 in math.NT | (2601.07421v1)

Abstract: We provide a writeup of a resolution of Erdős Problem #728; this is the first Erdos problem (a problem proposed by Paul Erdős which has been collected in the Erdos Problems website) regarded as fully resolved autonomously by an AI system. The system in question is a combination of GPT-5.2 Pro by OpenAI and Aristotle by Harmonic, operated by Kevin Barreto. The final result of the system is a formal proof written in Lean, which we translate to informal mathematics in the present writeup for wider accessibility. The proved result is as follows. We show a logarithmic-gap phenomenon regarding factorial divisibility: For any constants $0<C_1<C_2$ there exist infinitely many triples $(a,b,n)\in\mathbb N3$ such that [ a!\,b!\mid n!\,(a+b-n)!\qquad\text{and}\qquad C_1\log n < a+b-n < C_2\log n. ] The argument reduces this to a binomial divisibility $\binom{m+k}{k}\mid\binom{2m}{m}$ and studies it prime-by-prime. By Kummer's theorem, $ν_p\binom{2m}{m}$ translates into a carry count for doubling $m$ in base $p$. We then employ a counting argument to find, in each scale $[M,2M]$, an integer $m$ whose base-$p$ expansions simultaneously force many carries when doubling $m$, for every prime $p\le 2k$, while avoiding the rare event that one of $m+1,\dots,m+k$ is divisible by an unusually high power of $p$. These "carry-rich but spike-free" choices of $m$ force the needed $p$-adic inequalities and the divisibility. The overall strategy is similar to results regarding divisors of $\binom{2n}{n}$ studied earlier by Erdős and by Pomerance.

Summary

  • The paper establishes an infinite family of solutions where the factorial divisibility holds with the gap k constrained within a logarithmic window.
  • It reduces the problem to a binomial coefficient condition, employing Kummer’s theorem and p-adic valuation analysis to track carry events in digital expansions.
  • The proof integrates AI-driven theorem proving and Lean formal verification, setting a new standard for rigor in combinatorial number theory.

Resolution of Erdős Problem #728: A Formalization of Factorial Divisibility with Logarithmic Gap

Introduction and Problem Context

The paper "Resolution of Erdős Problem #728: a writeup of Aristotle's Lean proof" (2601.07421) presents a rigorous, formal proof of a factorial divisibility problem originally posed by Paul Erdős. The central object of study is the magnitude of the gap k=a+bnk = a + b - n for which the divisibility a!b!n!k!a!\,b! \mid n!\,k! holds. Classic approaches have dealt with configurational constraints on aa, bb, and nn, often eliminating trivial cases through imposed bounds. The present work demonstrates an infinite family of solutions for which kk consistently sits inside a strictly logarithmic window in nn, i.e., for any 0<C1<C20 < C_1 < C_2, there are infinitely many (a,b,n)N3(a, b, n) \in \mathbb{N}^3 such that

a!b!n!(a+bn)!andC1logn<a+bn<C2logn.a!\,b! \mid n!\,(a+b-n)! \quad \text{and} \quad C_1\log n < a+b-n < C_2\log n.

The proof proceeds via a reduction to binomial coefficient divisibility:

(m+kk)(2mm),\binom{m+k}{k}\mid\binom{2m}{m},

and studies the associated pp-adic valuations through the lens of Kummer's theorem, leveraging carry counts in digital expansions. The resolution is notable for its reliance on automated theorem proving by AI systems—specifically, OpenAI GPT-5.2 Pro and Harmonic Aristotle, with the formal Lean development acting as a verifiable artifact.

Prior Results and Literature

This result expands upon a body of work dealing with central binomial coefficients and their divisors. Erdős previously established upper bounds on a+ba+b in relation to nn when a!b!n!a!\,b!\mid n!, with asymptotically sharp results and tight logarithmic terms [Erdos557]. Pomerance [Pomerance2015] showed that for each fixed kk, the set of nn for which n+k(2nn)n+k \mid \binom{2n}{n} has density 1. Ford and Konyagin [FordKonyagin2021] developed divisibility results for powers of nn in (2nn)\binom{2n}{n}, establishing positive asymptotic densities and explicit formulas.

Recent advances by Croot, Mousavi, Schmidt [CrootMousaviSchmidt2024] and Bloom, Croot [BloomCroot2025] have focused on the distribution of carries and the prime power divisibility of binomial coefficients, often constructing integers with "carry-poor" properties. In sharp contrast, this work enforces "carry-rich" expansions, exploiting the rare structure of digital carry events to satisfy stringent divisibility criteria within a logarithmic gap.

Technical Framework and Reduction

After parametrization—n=2mn=2m, b=mb=m, a=m+ka=m+k—the factorial divisibility condition reduces elegantly to the binomial divisibility (m+kk)(2mm)\binom{m+k}{k}\mid\binom{2m}{m}, with kclogmk \doteq c\log m. An mm is sought in [M,2M][M,2M] so that kk lands precisely inside the desired window.

The pp-adic analysis is central. The pp-adic valuation νp\nu_p of (2mm)\binom{2m}{m}, denoted κp(m)\kappa_p(m), is equivalent (per Kummer's theorem) to the number of carries in the base-pp addition m+mm+m. The core inequality requiring proof for all primes pp is:

Vp(m,k):=max1ikνp(m+i)κp(m).V_p(m,k) := \max_{1 \leq i \leq k} \nu_p(m + i) \leq \kappa_p(m).

This separation allows the problem to be split into two prime regimes: "large primes" (p>2kp > 2k) and "small primes" (p2kp \leq 2k).

  • For large primes, the divisibility is straightforward: only one of m+im+i can be deeply divisible by pjp^j, and carry counts easily outmatch rare divisibility spikes.
  • For small primes, the construction leverages base-pp digit patterns, demanding a high count of large digits to force many carries, and ensures that exceptionally high divisibility (dubbed "spikes") of m+im+i is excluded via careful enumeration arguments.

Standard probabilistic tools, notably Chernoff bounds, permit bounding undesirable residue classes corresponding to carry failures, whereas residue-class counting manages spike avoidance. The beam search for mm in [M,2M][M,2M] then proceeds by a union bound over bad events, guaranteeing the existence of an mm that clears all aforementioned obstacles.

Main Theorem and Quantitative Bounds

The main theorem asserts the existence of infinitely many (a,b,n)(a, b, n) such that a!b!n!(a+bn)!a!\,b! \mid n! (a+b-n)! while k=a+bnk = a + b - n falls strictly within a logarithmic window:

C1logn<k<C2logn,C_1\log n < k < C_2\log n,

for any fixed 0<C1<C20 < C_1 < C_2. The proof is fully formalized in Lean and cross-mapped to corresponding lemma structures. Notably, Terence Tao has observed this bound is suboptimal, suggesting logarithmic gaps up to k=exp(clogn)k = \exp(c\sqrt{\log n}) may be obtainable with further advances.

Implications and Future Directions

This research represents the confluence of combinatorial number theory and AI-driven theorem proving. On the mathematical side, it sharpens understanding of the distribution and structure of binomial divisibility, particularly in the rare, high-carry regime. The construction bypasses straightforward counterexamples typical in extremal parameterizations by imposing the carry-rich criterion.

Practically, the proof demonstrates that autonomous AI systems, equipped with probabilistic and combinatorial insight, can formalize intricate results in arithmetic divisibility and produce Lean-verifiable outputs—a clear step in AI-involved foundational mathematics. The dual process—initial informal reasoning followed by strict formal verification—highlights a path toward robust, reproducible mathematical research involving AI collaboration.

Future studies will likely investigate tighter upper bounds on kk in the divisibility window, potentially with sharper digit and carry pattern constructions. Additionally, the methods in residue-class analysis, probabilistic quantification in digital expansions, and spike avoidance could find further application in related combinatorial and analytic number-theoretic problems.

Conclusion

The paper provides a rigorous, AI-assisted proof of Erdős Problem #728, demonstrating infinite occurrence of factorial divisibility within precise logarithmic gaps. The approach utilizes advanced pp-adic and combinatorial techniques, explicitly tracks carry structures in digital expansions, and employs probabilistic bounds to ensure the required inequalities. Formalization in Lean enhances confidence in correctness and opens avenues for future work on both mathematical and AI fronts. The methodology and formal rigor exemplified in this work signify an important milestone in autonomous mathematical problem solving and divisibility theory.

Whiteboard

Explain it Like I'm 14

What is this paper about?

This paper solves a number puzzle first asked by the famous mathematician Paul Erdős (Problem #728). It studies when one big product of numbers (a factorial) divides another and shows there are infinitely many times this works with a gap that’s about as big as a constant times the logarithm of a number. It’s also notable because the solution was first found and fully checked by AI systems, then rewritten here in human-friendly math.

Before we begin, here are two quick ideas:

  • The factorial of a number, written n!n!, means 1×2×3××n1 \times 2 \times 3 \times \cdots \times n.
  • XX divides YY” means YY can be evenly split into XX without leftovers.

The main question (in simple terms)

We look at three whole numbers aa, bb, and nn. Consider the divisibility:

  • Does a!b!a!\,b! divide n!(a+bn)!n!\,(a+b-n)!?

The “gap” is k=a+bnk = a + b - n. The question asks: How large can kk be while the divisibility still holds?

The paper proves that there are infinitely many choices of a,b,na, b, n where this works and the gap kk is between two constant multiples of logn\log n:

  • For any fixed constants 0<C1<C20 < C_1 < C_2, there are infinitely many triples (a,b,n)(a,b,n) with

a!b!n!(a+bn)!andC1logn<a+bn<C2logn.a!\,b!\mid n!\,(a+b-n)!\quad\text{and}\quad C_1\log n < a+b-n < C_2\log n.

Here, logn\log n is a slowly growing function (roughly: like the number of digits of nn in a fixed base).

The key objectives (in everyday language)

The paper aims to:

  • Show that the gap k=a+bnk = a + b - n can be as large as a constant times logn\log n, infinitely often.
  • Do this by converting the factorial divisibility into a cleaner statement about binomial coefficients (the “NN choose kk” numbers from combinatorics).
  • Prove the statement by checking, prime number by prime number, that each prime divides the left side no more than it divides the right side.

How the proof works (the approach and ideas)

The authors use number bases, carries in addition, and some clever counting to find the needed examples. Here are the main steps, with analogies and simple explanations.

Step 1: Turn factorials into binomial coefficients

A binomial coefficient is written (Nk)\binom{N}{k} and means “the number of ways to choose kk things from NN things.” It’s related to factorials by

(Nk)=N!k!(Nk)!.\binom{N}{k} = \frac{N!}{k!(N-k)!}.

The authors show the original divisibility is equivalent to:

(m+kk)  (2mm)\binom{m+k}{k}\ \mid\ \binom{2m}{m}

for certain choices of mm and kk (they set n=2mn=2m, b=mb=m, a=m+ka=m+k, so that a+bn=ka+b-n=k). They plan to take kk about clogmc\log m for some constant cc.

Why this helps: Binomial coefficients are easier to analyze prime-by-prime.

Step 2: Check prime factors one at a time

To prove a division like XYX \mid Y, it’s enough to check that for every prime pp, pp doesn’t appear “more times” in XX than in YY. In math-speak, for each prime pp, compare how many copies of pp divide XX versus YY.

So the paper reduces the goal to: For every prime pp,

  • the exponent of pp in (m+kk)\binom{m+k}{k} is at most the exponent of pp in (2mm)\binom{2m}{m}.

Step 3: Carries in base-pp addition (Kummer’s Theorem)

Here’s a very cool fact: If you write the number mm in base pp (like base 10, but with digit size pp), then the number of carries you get when adding m+mm + m (i.e., doubling mm) in base pp is exactly the exponent of pp in (2mm)\binom{2m}{m}.

Analogy:

  • Think about normal base-10 addition. If you add 57 + 57, at the ones place you get 14, so you write 4 and “carry” 1 to the tens place. Carries happen where a digit sum overflows the base.
  • In base pp, the same idea holds, but the threshold is pp instead of 10.
  • Kummer’s theorem says: “How many carries happen when doubling mm in base pp?” equals “How many times pp divides (2mm)\binom{2m}{m}.”

So to make (2mm)\binom{2m}{m} have lots of pp-power, the authors try to force many carries when doubling mm in base pp.

Step 4: Forcing carries by choosing the digits of mm

If a base-pp digit of mm is at least half of pp (like 5,6,7,8,9 in base 10), then when you double that digit, it forces a carry at that digit position (because the sum will be at least pp).

So the plan is to choose mm so that, for each small prime pp (those with p2kp \le 2k), many of the first few base-pp digits of mm are “big” (at least half of p).Thisguaranteesmanycarries,whichmeansalargep”). This guarantees many carries, which means a largeppowerin-power in\binom{2m}{m}.</p><p>Tofindsuchan.</p> <p>To find such an m,theauthorslookinalonginterval, the authors look in a long interval [M, 2M]anduseacounting/probabilityargument:amongallnumbers and use a counting/probability argument: among all numbers minthatinterval,agoodchunkwillhaveabouttheexpectednumberofbigdigitsineachbase in that interval, a good chunk will have “about the expected” number of big digits in each base p.Thatmakesthemcarryrich.</p><h3class=paperheadingid=step5avoidingrarespikes>Step5:Avoidingrarespikes</h3><p>Theresapotentialproblem:thenumber. That makes them “carry-rich.”</p> <h3 class='paper-heading' id='step-5-avoiding-rare-spikes'>Step 5: Avoiding rare “spikes”</h3> <p>There’s a potential problem: the number \binom{m+k}{k}containstheproduct contains the product (m+1)\cdots(m+k)dividedby divided by k!.If,forsomeprime. If, for some prime p,oneofthenumbers, one of the numbers m+1,\ldots,m+kisdivisiblebyanunusuallylargepowerof is divisible by an unusually large power of p(say (say p^Jwithbig with big J),itcouldruintheinequalityforthatprime.Theauthorscallsuchrareeventsspikes.</p><p>Butspikesarerarebecausetheyforce), it could ruin the inequality for that prime. The authors call such rare events “spikes.”</p> <p>But spikes are rare because they force mintoaveryspecificpositionmodulo into a very specific “position” modulo p^J(likelandingonasinglenarrowlaneonabighighway).Usingsimplecounting,theyshowthatamong (like landing on a single narrow lane on a big highway). Using simple counting, they show that among m \in [M, 2M],veryfew, very few mproducespikesforanygiven produce spikes for any given p.</p><h3class=paperheadingid=step6puttingittogethersmallprimesversuslargeprimes>Step6:Puttingittogether(smallprimesversuslargeprimes)</h3><ul><li>Forlargeprimes.</p> <h3 class='paper-heading' id='step-6-putting-it-together-small-primes-versus-large-primes'>Step 6: Putting it together (small primes versus large primes)</h3> <ul> <li>For large primes p > 2k:itseasytoshowtheinequalityholds(largeprimerangeissimplebecauseatmostonenumberin: it’s easy to show the inequality holds (“large-prime range” is simple because at most one number in m+1,\dots,m+kisdivisibleby is divisible by por or p^j).</li><li>Forsmallprimes).</li> <li>For small primes p \le 2k:choose: choose mwithmanybigbase with many big base-pdigits(forcingmanycarries)andwithnospikes.Aunionbound(astandardcountingtrick)showsthatatleastone digits (forcing many carries) and with no spikes. A union bound (a standard counting trick) shows that at least one min in [M,2M]worksforalltheseprimesatonce.</li></ul><p>Finally,theauthorstake works for all these primes at once.</li> </ul> <p>Finally, the authors take k \approx c \log Mandlet and let Mgrow.Thisgivesthepromisedlogarithmicgap grow. This gives the promised “logarithmic gap” kbetween between C_1 \log nand and C_2 \log n,infinitelyoften.</p><h2class=paperheadingid=themainfindingsandwhytheymatter>Themainfindingsandwhytheymatter</h2><ul><li>Mainresult:Foranyfixed, infinitely often.</p> <h2 class='paper-heading' id='the-main-findings-and-why-they-matter'>The main findings and why they matter</h2> <ul> <li>Main result: For any fixed 0 < C_1 < C_2,thereareinfinitelymanytriples, there are infinitely many triples (a,b,n)suchthat</li></ul><p> such that</li> </ul> <p>a!\,b!\mid n!\,(a+b-n)!\quad\text{and}\quad C_1\log n < a+b-n < C_2\log n.</p><p>Inwords:Thegap</p> <p>In words: The gap k = a+b-ncanbeasbigasaconstanttimes can be as big as a constant times \log nandthiskeepshappeninginfinitelyoften.</p><ul><li>Whythisisinteresting:<ul><li>Itpreciselylocateshowlargethegap—and this keeps happening infinitely often.</p> <ul> <li>Why this is interesting: <ul> <li>It precisely locates how large the gap kcanbewhilekeepingthedivisibilitytrue,matchingalogarithmicscale(aslowgrowingbutunboundedsize).</li><li>Themethodoffersaneatblendofcombinatorics(binomialcoefficients),numbertheory(primefactors),andadigitbasedviewpoint(carriesindifferentbases).</li><li>Theproofshowsacarryrichphenomenon:bycontrollingthebase can be while keeping the divisibility true, matching a “logarithmic” scale (a slow-growing but unbounded size).</li> <li>The method offers a neat blend of combinatorics (binomial coefficients), number theory (prime factors), and a digit-based viewpoint (carries in different bases).</li> <li>The proof shows a “carry-rich” phenomenon: by controlling the base-pdigitsof digits of m,youcanforcemanycarriesindoubling,whichunlockstheprimefactorbehaviorofthecentralbinomialcoefficient, you can force many carries in doubling, which unlocks the prime-factor behavior of the central binomial coefficient \binom{2m}{m}$.

What’s the impact?

  • Mathematical impact:
    • It advances our understanding of how factorials and binomial coefficients divide each other and how prime factors distribute in these numbers.
    • It builds on classic insights (like Kummer’s theorem) and recent ideas about “digits behaving randomly,” turning them into a concrete tool to produce infinitely many examples with controlled gap size.
    • The strategy is flexible and relates to other results about divisibility of central binomial coefficients, connecting to work by Erdős and more recent research.
  • Broader significance:
    • The proof was first found and formalized by AI systems (OpenAI’s GPT-5.2 Pro and Harmonic’s Aristotle, producing a Lean formal proof), then translated into a human-style writeup. This marks a significant milestone: a recognized Erdős problem resolved autonomously by AI and certified in a proof assistant.
    • It hints at a future where AI can not only discover proofs but also help guide human-understandable explanations and connect with existing mathematical literature.

Overall, the paper combines clever number-base thinking (carries), prime-by-prime reasoning, and probabilistic counting to settle Erdős Problem #728 with a clean “logarithmic gap” answer, and it showcases how AI can contribute to modern mathematics.

Knowledge Gaps

Knowledge gaps, limitations, and open questions

Below is a concise list of concrete gaps and questions the paper leaves unresolved, focusing on what remains missing, uncertain, or unexplored, and framed so that future work can act on them.

  • True order of magnitude of the maximal gap: The paper proves infinitely many triples with k = a + b − n in a logarithmic window, and remarks that the method’s bottleneck might allow k up to exp(c sqrt(log n)). It remains open to determine the correct growth rate of the maximal achievable k (as a function of n), and to establish matching upper bounds.
  • Formal upgrade beyond logarithmic k: The remark states the union-bound step k * e{−μ̃(M)/8} = o(1) can be made to work up to k = exp(c sqrt(log n)), but the paper does not provide a complete proof or a formal theorem at that scale. Making this precise (with rigorous parameter choices, explicit constants, and a Lean formalization) is an actionable target.
  • Lack of upper bounds: The paper gives lower bounds (existence) for k but no nontrivial upper bounds on k under natural side conditions (e.g., a, b ≤ (1 − ε)n). Establishing upper bounds that match or nearly match the achievable lower bounds is open.
  • Density of good m (quantitative counting): The proof only claims existence of at least one m ∈ [M, 2M] for each large M. The bounds in Lemma 4.13 imply the bad set is o(M) (hence a 1 − o(1) density of good m is plausible), but this stronger density statement is not stated nor quantified. Providing explicit asymptotics or a positive-density result would strengthen the conclusion.
  • Per-n guarantees: The construction fixes n = 2m and shows infinitely many even n. It remains open whether for each sufficiently large n (or for a positive density of n), one can find a, b with k ≥ c log n (or larger), and to quantify how often such triples exist as n varies.
  • Explicit proximity of a and b to n/2: The introduction claims “a and b are extremely close to n/2,” but the paper does not quantify this proximity. Providing explicit bounds on |a − n/2| and |b − n/2| (in terms of k or n) would clarify the strength of the construction.
  • Constructive/algorithmic aspects: The argument is non-constructive (via counting). Is there an efficient randomized or deterministic algorithm that, given M (or n), finds suitable m (hence a, b) with k in the desired range, with provable runtime and success probability guarantees?
  • Parameter optimization: The proof uses fixed choices (e.g., η = 1/10 and t(M) = ⌈10 log log M⌉), coarse bounds θ(p) ≥ 1/3, and a union bound across primes. Optimizing these parameters (and using sharper tail or sieve bounds, or dependency tooling such as the Lovász Local Lemma) could increase the admissible k; the best obtainable k via this method is not analyzed.
  • Threshold for the “large prime” regime: Lemma 3.1 imposes p > 2k to force carries. It is unclear whether the threshold 2k is optimal for this argument; tightening this (e.g., to p > (1 + o(1))k) could further relax constraints and enlarge k.
  • Correlations across primes: The analysis treats primes independently and unions the bad events. Understanding and exploiting correlations between carry events in different bases (or demonstrating a form of pseudo-independence) could enable larger k beyond what union bounding permits.
  • Classification problem for divisibility: The paper proves existence of many (m, k) with C(m + k, k) | C(2m, m). A structural characterization (or threshold behavior) of all such pairs (m, k), or an asymptotic for their count up to X, remains open.
  • Generalizations beyond n = 2m: Can analogous results be obtained for more general linear families (e.g., C(m + k, k) | C((1 + δ)m, m) for fixed δ ≠ 1) or for broader triples (a, b, n) subject to a, b ≤ (1 − ε)n? Extending the carry analysis beyond doubling is not explored.
  • Beyond binomial coefficients: Are there analogous “carry-rich implies divisibility” phenomena for multinomial coefficients or for q-analogues (Gaussian binomial coefficients)? The extent to which Kummer-type interpretations can be adapted remains unexplored.
  • Sharpening carry lower bounds: The proof enforces many large base-p digits among the first L_p digits. Refined carry-count lower bounds (e.g., tracking incoming carries or using precise digit-sum/carry-process martingale inequalities) might yield stronger results for k.
  • Spike control improvements: The “no-spike” condition V_p(m, k) < J_p + t(M) uses a crude union bound over i ≤ k and primes p ≤ 2k. Can sharper divisor-distribution results (or dense-set avoidance techniques) reduce t(M) and thereby enlarge k?
  • Explicit constants and ranges: While the theorem allows any 0 < C1 < C2, the proof does not provide explicit upper limits on admissible C2 as a function of the chosen parameters (η, t(M), etc.). Making the allowable range of constants explicit would aid applications.
  • Multi-k robustness for a fixed m: For a given good m, for how many k in 1, K does the divisibility hold simultaneously? Establishing such “interval-robustness” would parallel the interval-product divisibility results in the literature.
  • Stronger “almost all n” analogues: Erdős’s Aufgabe 557 has an “almost all n” refinement. An analogous statement here (e.g., for almost all m in [M, 2M] the divisibility holds with k ≍ log m) is not stated and would be valuable to formalize and prove.
  • Methodological barriers: The paper does not delineate inherent limits of the “carry-rich but spike-free” strategy. Pinpointing where this approach must fail (e.g., via entropy or second-moment barriers) would guide the search for new ideas to push k further.

Practical Applications

Immediate Applications

Below are concrete use cases that can be deployed now, drawing on the paper’s mathematical results and the demonstrated AI+Lean workflow.

  • Carry-count API and binomial divisibility checker for large integers
    • Sector: software (computational mathematics, CAS), education
    • What: Implement functions that compute ν_p(2m choose m) via Kummer’s carry interpretation and check divisibility relations like (m+k)! m! | (2m)! k! using prime-by-prime valuation comparisons rather than constructing huge integers.
    • Tools/products/workflows:
    • A Python/Rust library module “CarryCount” that exposes κ_p(m) via carry counts.
    • A “BinomialDivisibility” module that relies on Legendre’s formula and the paper’s valuation reductions (W_p ≤ κ_p + v_p(k!)).
    • CAS plugins (e.g., for SageMath, Mathematica) to simplify factorial ratios and binomial divisibility questions exactly.
    • Assumptions/dependencies: Efficient base-p digit extraction; prime enumeration up to O(log n); correctness relies on Kummer’s theorem and Legendre’s formula (well-established).
  • Stress-test generator for integer addition (base-2 carry chains)
    • Sector: semiconductors, systems software (BigInt libraries), HPC
    • What: Generate integers that force dense carry propagation on doubling/addition in base 2, to benchmark and validate adder implementations, CPU micro-ops, and arbitrary-precision integer libraries.
    • Tools/products/workflows:
    • “CarryStress” generator using the paper’s digit-threshold idea specialized to p=2 (require many low-order bits equal to 1) to produce worst-case carry chains.
    • Assumptions/dependencies: For hardware, base-2 is the only relevant base; while simple patterns (e.g., 2L−1) already exist, the paper’s probabilistic construction provides families inside arbitrary ranges [M, 2M], which is useful for randomized stress suites.
  • Residue-class filtering utilities for modular constraints
    • Sector: software testing, cryptography tooling (auxiliary), optimization
    • What: Use the paper’s residue-class counting and union bounds to generate integers within [M, 2M] that simultaneously avoid “rare spikes” (high power divisibility by many small primes).
    • Tools/products/workflows:
    • A utility that, given primes p ≤ P and thresholds J_p, filters m in a range to avoid m+i ≡ 0 (mod p{J_p}) for small i, enabling robust test-case generation for arithmetic routines and modular algorithms.
    • Assumptions/dependencies: Prime list and moduli sizes must satisfy Q ≤ M{1−η} for uniformity bounds; efficiency depends on fast residue checks.
  • Lean-to-readable proof translation pipeline
    • Sector: academia, publishing, education
    • What: A semi-automated workflow to convert formal Lean proofs into human-readable papers, preserving proof provenance and correctness.
    • Tools/products/workflows:
    • “Lean-to-Text Translator” combining LLM summarization with Lean proof artifacts (lemmas, dependencies) to produce accessible expositions.
    • Assumptions/dependencies: Quality assurance and human editorial passes; availability of robust mathlib references.
  • Provenance and reproducibility practices for AI-generated proofs
    • Sector: academic policy, journals, research integrity
    • What: Adopt a workflow to archive AI prompts, Lean scripts, and proof artifacts to establish transparent provenance for AI-driven results.
    • Tools/products/workflows:
    • A “ProofProvenance Ledger” storing artifacts (LLM inputs/outputs, Lean files, version hashes) and links to public repos.
    • Assumptions/dependencies: Journal and community acceptance; repository sustainability; minimal privacy issues over model interactions.
  • Curriculum modules and interactive demos on carries and valuations
    • Sector: education (secondary, undergraduate)
    • What: Hands-on visualizations of Kummer’s theorem (carry counts) and Legendre’s formula; exercises on binomial divisibility via digit patterns.
    • Tools/products/workflows:
    • Interactive web apps to visualize base-p addition and carries; classroom-ready notebooks showing why ν_p(2m choose m) equals the carry count of m+m in base p.
    • Assumptions/dependencies: None beyond standard web tooling; content alignment with curricula.
  • Fast valuation-aware checks for combinatorial simplification in CAS
    • Sector: software (symbolic algebra)
    • What: Use valuation inequalities (W_p ≤ κ_p + v_p(k!)) to decide exact integrality/divisibility in simplification routines, avoiding expensive expansion of factorials/binomials.
    • Tools/products/workflows:
    • CAS heuristics that short-circuit simplifications by valuation checks; e.g., deciding whether interval products divide central binomial coefficients.
    • Assumptions/dependencies: Correct mapping of combinatorial expressions to valuation bounds; prime enumeration up to needed thresholds.

Long-Term Applications

These use cases require further research, scaling, engineering, or consensus-building before broad deployment.

  • End-to-end autonomous mathematical research assistants
    • Sector: academia, research software
    • What: Generalize the demonstrated GPT+Aristotle pipeline into systems that propose conjectures, refine strategies, formalize, and publish with formal verification, for broad classes of problems.
    • Tools/products/workflows:
    • Integrated environments combining exploratory LLMs, formal proof systems (Lean/Isabelle/Coq), literature search, and provenance tracking.
    • Assumptions/dependencies: Stronger automated tactic libraries, better interaction models, and community buy-in for AI-authored work.
  • Standards for AI-generated proof acceptance and credit attribution
    • Sector: policy, scholarly publishing
    • What: Policies that define acceptable evidence for autonomy, set provenance requirements, and establish credit frameworks for human–AI collaboration.
    • Tools/products/workflows:
    • Journal policies, audit protocols for AI proofs, and citation standards referencing formal artifacts.
    • Assumptions/dependencies: Community consensus, legal/IP clarity, infrastructure for audits.
  • CAS and compiler-level optimization using valuation- and carry-aware reasoning
    • Sector: software (CAS, compilers for numerical kernels)
    • What: Embed valuation/carry logic into simplification and code generation to avoid expensive numeric paths when divisibility can be decided symbolically.
    • Tools/products/workflows:
    • Compiler passes or CAS kernels that translate combinatorial identities into valuation constraints, reducing computational overhead.
    • Assumptions/dependencies: Robust symbolic-to-valuation mapping and guaranteed correctness; benchmarking on real workloads.
  • Hardware design analytics informed by carry distributions
    • Sector: semiconductors
    • What: Use probabilistic digit models (and thresholds ensuring “carry-rich” events) to bound worst-case carry propagation, guide adder and pipeline depth design, and predict performance tails.
    • Tools/products/workflows:
    • Statistical models for carry behavior leveraged in microarchitectural simulations; stress-suite generation guided by multi-digit constraints.
    • Assumptions/dependencies: Translating multi-base theory to base-2 specifics; validation against silicon or cycle-accurate simulators.
  • Side-channel and timing analysis using controlled carry behavior
    • Sector: cryptography/security
    • What: Evaluate whether carry propagation patterns (e.g., in multi-precision addition/doubling) introduce measurable timing/leakage and design tests with carry-rich inputs to probe defenses.
    • Tools/products/workflows:
    • Security test harnesses generating controlled inputs; analytic models linking carry density to leakage/timing signatures.
    • Assumptions/dependencies: Demonstrable linkage between carry patterns and side-channel signals; careful experimental design.
  • Multi-modulus constraint solvers for residue number systems (RNS) and number-theoretic algorithms
    • Sector: cryptography, HPC
    • What: Adapt the “rare spike elimination” plus union-bound methodology to choose integers meeting many simultaneous modular constraints (across several coprime bases) for testing, scheduling, or algorithmic tuning.
    • Tools/products/workflows:
    • RNS-aware integer selection tools controlling ν_p patterns across a prime set; performance tuners that exploit predictable modular behavior.
    • Assumptions/dependencies: Mapping carry-based criteria to RNS (which generally avoids carries); demonstration of practical gains.
  • Integrated education platforms blending number theory and probability
    • Sector: education
    • What: Coherent curricula that unify base expansions, carries, p-adic valuations, and Chernoff bounds to teach cross-disciplinary reasoning.
    • Tools/products/workflows:
    • Modular courseware, auto-graded problem sets, and visual labs aligned with standards.
    • Assumptions/dependencies: Curriculum development and teacher training; assessment integration.
  • Data-driven discovery of combinatorial divisibility phenomena
    • Sector: academia (number theory, combinatorics)
    • What: Generalize the carry-rich/spike-free construction to uncover new divisibility windows (e.g., extend beyond logarithmic to exp(c√log n) scales) and related phenomena.
    • Tools/products/workflows:
    • Search pipelines combining randomized digit constraints, modular filters, and formal verification to propose and test new results.
    • Assumptions/dependencies: Further theoretical breakthroughs; computational experiments; integration with formal systems.

Notes on feasibility and dependencies

  • The divisibility result itself is existential and asymptotic; while constructive search within [M, 2M] is feasible using residue-class filtering, producing “closed-form” m is not guaranteed by the proof.
  • Prime handling is modest since p ≤ 2k and k ≈ c log n; practical implementations rely on enumerating small primes and base-p digit extraction up to L_p ≈ Θ(log n / log p).
  • The AI+Lean pipeline is already shown viable for this problem; scaling to broader classes requires stronger automation, tactic libraries, and editorial workflows to ensure readability and trust.

Glossary

  • Asymptotic density: The limiting proportion of integers up to x that satisfy a property, as x grows. "has a positive asymptotic density cc_\ell"
  • Asymptotic formula: An expression describing the leading behavior of a quantity as a parameter tends to infinity. "and they obtain an asymptotic formula for cc_\ell as \ell\to\infty."
  • Base-pp digits: The digits of an integer when written in base p. "Let Xp(m)X_p(m) be the number of the first LpL_p base-pp digits of mm which are p/2\ge \lceil p/2\rceil."
  • Big-O notation: A notation that bounds the growth rate of a function up to constant factors. "Erd\H{o}s proved an upper bound a+b<n+O(logn)a+b<n+O(\log n)"
  • Binomial divisibility: The property that one binomial coefficient divides another. "The argument reduces this to a binomial divisibility (m+kk)(2mm)\binom{m+k}{k}\mid\binom{2m}{m}"
  • Carry count: The number of carry operations produced when adding numbers in a given base. "By Kummer's theorem, νp(2mm)\nu_p\binom{2m}{m} translates into a carry count for doubling mm in base pp."
  • Carry-poor: Having few carries in base-pp addition (e.g., when doubling a number). "constructs integers nn that are carry-poor for doubling in each of the bases p1,,prp_1,\dots,p_r simultaneously."
  • Carry-rich: Having many carries in base-pp addition (e.g., when doubling a number). "In contrast, the present proof is ``carry-rich'', forcing many carries simultaneously."
  • Central binomial coefficient: The binomial coefficient (2nn)\binom{2n}{n}, central in the nth row of Pascal’s triangle. "divisors of the central binomial coefficient (2nn)\binom{2n}{n}."
  • Chernoff inequality: A tail bound for sums of independent Bernoulli random variables. "This is the standard Chernoff inequality P(X(1δ)μ)eδ2μ/2\mathbb{P}(X\le (1-\delta)\mu)\le e^{-\delta^2\mu/2} with δ=1/2\delta=1/2."
  • Erdős Problem: A problem posed by Paul Erdős, often part of a curated collection of number-theoretic questions. "We provide a writeup of a resolution of Erd\H{o}s Problem #728;"
  • Interval product divisibility: Divisibility of a product of consecutive integers by another structured integer. "one has for almost all nn the interval product divisibility [(n+1)\cdots(n+a)\ \mid\ \binom{2n}{n}]."
  • Kummer's theorem: A characterization of the pp-adic valuation of binomial coefficients via carries in base-pp addition. "Kummer's theorem identifies κp(m)\kappa_p(m) with the number of carries when adding m+mm+m in base~pp."
  • Lean (theorem prover): An interactive proof assistant used to construct formal, machine-checked proofs. "The final result of the system is a formal proof written in Lean, which we translate to informal mathematics in the present writeup for wider accessibility."
  • Legendre's formula: A formula for the pp-adic valuation of factorials: vp(n!)=j1n/pjv_p(n!)=\sum_{j\ge1}\lfloor n/p^j\rfloor. "By Legendre's formula, vp(k!)=j=1Jk/pjv_p(k!) = \sum_{j=1}^{J}\lfloor k/p^j\rfloor."
  • Little-oh notation: A notation indicating a function grows strictly slower than another (e.g., f(x)=o(g(x))f(x)=o(g(x))). "the main constraint is keμ~(M)/8=o(1)k \cdot e^{-\tilde{\mu}(M)/8} = o(1)"
  • p-adic valuation: The exponent of a prime p in the prime factorization of an integer. "usually governed by Kummer's carry interpretation of pp-adic valuations."
  • Prime counting function: The function π(x)\pi(x) that counts the number of primes less than or equal to xx. "where π()\pi(\cdot) is the prime counting function."
  • Residue class: An equivalence class of integers modulo a given modulus. "Each residue class modulo QQ occurs at most (M+1)/Q+1(M+1)/Q+2\lceil (M+1)/Q\rceil+1\le (M+1)/Q+2 times in the interval."

Open Problems

We found no open problems mentioned in this paper.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 23 tweets with 6421 likes about this paper.