Papers
Topics
Authors
Recent
Search
2000 character limit reached

Discretizing Distributions with Exact Moments: Error Estimate and Convergence Analysis

Published 17 Aug 2013 in math.NA | (1308.3753v3)

Abstract: The maximum entropy principle is a powerful tool for solving underdetermined inverse problems. This paper considers the problem of discretizing a continuous distribution, which arises in various applied fields. We obtain the approximating distribution by minimizing the Kullback-Leibler information (relative entropy) of the unknown discrete distribution relative to an initial discretization based on a quadrature formula subject to some moment constraints. We study the theoretical error bound and the convergence of this approximation method as the number of discrete points increases. We prove that (i) the theoretical error bound of the approximate expectation of any bounded continuous function has at most the same order as the quadrature formula we start with, and (ii) the approximate discrete distribution weakly converges to the given continuous distribution. Moreover, we present some numerical examples that show the advantage of the method and apply to numerically solving an optimal portfolio problem.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.