LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

Efficient modeling of trivializing maps for lattice ϕ4 theory using normalizing flows: A first look at scalability

Photo from wikipedia

General-purpose Markov Chain Monte Carlo sampling algorithms suffer from a dramatic reduction in efficiency as the system being studied is driven towards a critical point through, for example, taking the… Click to show full abstract

General-purpose Markov Chain Monte Carlo sampling algorithms suffer from a dramatic reduction in efficiency as the system being studied is driven towards a critical point through, for example, taking the continuum limit. Recently, a series of seminal studies suggested that normalizing flows — a class of deep generative models — can form the basis of a sampling strategy that does not suffer from this ‘critical slowing down’. The central idea is to use machine learning techniques to build (approximate) trivializing maps, i.e. field transformations that map the theory of interest into a ‘simpler’ theory in which the degrees of freedom decouple. These trivializing maps provide a representation of the theory in which all its non-trivial aspects are encoded within an invertible transformation to a set of field variables whose statistical weight in the path integral is given by a distribution from which sampling is easy. No separate process is required to generate training data for such models, and convergence to the desired distribution is guaranteed through a reweighting procedure such as a Metropolis test. From a theoretical perspective, this approach has the potential to become more efficient than traditional sampling since the statistical efficiency of the sampling algorithm is decoupled from the correlation length of the system. The caveat to all of this is that, somehow, the costs associated with the highly non-trivial task of sampling from the path integral of an interacting field theory are transferred to the training of a model to perform this transformation. In a proof-of-principle demonstration on two-dimensional φ theory, Albergo et al. [1] modelled the trivializing map as a sequence of pointwise affine transformations. We pick up this thread, with the aim of quantifying how well we can expect this approach to scale as we increase the number of degrees of freedom in the system. We make several modifications to the original design that allow our models learn more efficient representations of trivializing maps using much smaller neural networks, which leads to a large reduction in the computational cost required to train models of equivalent quality. After making these changes, we find that sampling efficiency is almost entirely dictated by how extensively a model has been trained, while being unresponsive to further alterations that increase model flexibility. However, as we move towards the continuum limit the training costs scale extremely quickly, which urgently requires further work to fully understand and mitigate.

Keywords: efficient modeling; maps lattice; theory; trivializing maps; modeling trivializing; normalizing flows

Journal Title: Physical Review D
Year Published: 2021

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.