Papers
Topics
Authors
Recent
Search
2000 character limit reached

Neural Monte Carlo Renormalization Group

Published 12 Oct 2020 in cond-mat.dis-nn, cond-mat.stat-mech, and hep-th | (2010.05703v1)

Abstract: The key idea behind the renormalization group (RG) transformation is that properties of physical systems with very different microscopic makeups can be characterized by a few universal parameters. However, finding the optimal RG transformation remains difficult due to the many possible choices of the weight factors in the RG procedure. Here we show, by identifying the conditional distribution in the restricted Boltzmann machine (RBM) and the weight factor distribution in the RG procedure, an optimal real-space RG transformation can be learned without prior knowledge of the physical system. This neural Monte Carlo RG algorithm allows for direct computation of the RG flow and critical exponents. This scheme naturally generates a transformation that maximizes the real-space mutual information between the coarse-grained region and the environment. Our results establish a solid connection between the RG transformation in physics and the deep architecture in machine learning, paving the way to further interdisciplinary research.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.