High‑probability conditional likelihood bound for large‑support Poisson priors
Establish that, in the Poisson empirical Bayes setting with true prior G0 supported on [0, A] where A = A_n may grow up to O(n^2), and with a training prior‑on‑prior that draws k = \widetilde{\Theta}(\sqrt{A}) atom locations uniformly on [0, A] with Dirichlet(1,…,1) weights, the training conditional distribution Π_{X_n|X^{n−1}} satisfies −log Π_{X_n|X^{n−1}} = O(polylog(n)) with probability at least 1 − (n\sqrt{A})^{−1} over X^n ∼ f_{G0}^{\otimes n}. Proving this bound would imply, together with known metric entropy and Hellinger–regret inequalities, that the pretrained estimator achieves the minimax regret rate \widetilde{\Theta}(A^{1.5}/n) for A = O(n^2).
References
This regret bound can be recovered by a pretrained transformer with $k=\widetilde{\Theta}(\sqrt{A})$ atoms drawn uniformly at random on $[0,A]$, provided that one can establish $-\log \Pi_{X_n|X{n-1} = O(\mathsf{polylog}(n))$ with probability at least $1-(n\sqrt{A}){-1}$ over the randomness of the test data $Xn\sim f_{G_0}{\otimes n}$. … The desired high-probability upper bound on $-\log \Pi_{X_n|X{n-1}$ precisely corresponds to choosing $\rho = \exp(-O(\mathsf{polylog}(n)))$. However, we find it difficult to rigorously establish such a bound, and therefore leave it as a conjecture.