Phase transitions and complexity in computer science: an overview of the statistical physics approach to the random satisfiability problem

2002 ◽  
Vol 306 ◽  
pp. 381-394 ◽  
Author(s):  
Giulio Biroli ◽  
Simona Cocco ◽  
Rémi Monasson
Author(s):  
Fabrizio Altarelli ◽  
Rémi Monasson ◽  
Guilhem Semerjian ◽  
Francesco Zamponi

This chapter surveys a part of the intense research activity that has been devoted by theoretical physicists to the study of randomly generated k-SAT instances. It can be at first sight surprising that there is a connection between physics and computer science. However low-temperature statistical mechanics concerns precisely the behaviour of the low-lying configurations of an energy landscape, in other words the optimization of a cost function. Moreover the ensemble of random k-SAT instances exhibit phase transitions, a phenomenon mostly studied in physics (think for instance at the transition between liquid and gaseous water). Besides the introduction of general concepts of statistical mechanics and their translations in computer science language, the chapter presents results on the location of the satisfiability transition, the detailed picture of the satisfiable regime and the various phase transitions it undergoes, and algorithmic issues for random k-SAT instances.


Computer science and physics have been closely linked since the birth of modern computing. In recent years, an interdisciplinary area has blossomed at the junction of these fields, connecting insights from statistical physics with basic computational challenges. Researchers have successfully applied techniques from the study of phase transitions to analyze NP-complete problems such as satisfiability and graph coloring. This is leading to a new understanding of the structure of these problems, and of how algorithms perform on them. Computational Complexity and Statistical Physics will serve as a standard reference and pedagogical aid to statistical physics methods in computer science, with a particular focus on phase transitions in combinatorial problems. Addressed to a broad range of readers, the book includes substantial background material along with current research by leading computer scientists, mathematicians, and physicists. It will prepare students and researchers from all of these fields to contribute to this exciting area.


Author(s):  
Lorenza Saitta ◽  
Attilio Giordana ◽  
Antoine Cornuejols

1991 ◽  
Vol 02 (01) ◽  
pp. 201-208
Author(s):  
ROBERT H. SWENDSEN

Monte Carlo simulations of thermodynamic phase transitions are usually hampered by long relaxation times due to the phenomenon of “critical slowing down.” Using a mapping due to Fortuin and Kasteleyn, a cluster approach to Monte Carlo simulations has been developed, which greatly reduces relaxation times, improving efficiency by up to two or three orders of magnitude. New developments and extensions of this approach are also discussed.


2019 ◽  
Vol 116 (12) ◽  
pp. 5451-5460 ◽  
Author(s):  
Jean Barbier ◽  
Florent Krzakala ◽  
Nicolas Macris ◽  
Léo Miolane ◽  
Lenka Zdeborová

Generalized linear models (GLMs) are used in high-dimensional machine learning, statistics, communications, and signal processing. In this paper we analyze GLMs when the data matrix is random, as relevant in problems such as compressed sensing, error-correcting codes, or benchmark models in neural networks. We evaluate the mutual information (or “free entropy”) from which we deduce the Bayes-optimal estimation and generalization errors. Our analysis applies to the high-dimensional limit where both the number of samples and the dimension are large and their ratio is fixed. Nonrigorous predictions for the optimal errors existed for special cases of GLMs, e.g., for the perceptron, in the field of statistical physics based on the so-called replica method. Our present paper rigorously establishes those decades-old conjectures and brings forward their algorithmic interpretation in terms of performance of the generalized approximate message-passing algorithm. Furthermore, we tightly characterize, for many learning problems, regions of parameters for which this algorithm achieves the optimal performance and locate the associated sharp phase transitions separating learnable and nonlearnable regions. We believe that this random version of GLMs can serve as a challenging benchmark for multipurpose algorithms.


1992 ◽  
Vol 06 (13) ◽  
pp. 773-784 ◽  
Author(s):  
R. HILFER

The recent classification theory for phase transitions (R. Hilfer, Physica Scripta 44, 321 (1991)) and its relation with the foundations of statistical physics is reviewed. First it is outlined how Ehrenfests classification scheme can be generalized into a general thermodynamic classification theory for phase transitions. The classification theory implies scaling and multiscaling thereby eliminating the need to postulate the scaling hypothesis as a fourth law of thermodynamics. The new classification has also led to the discovery and distinction of nonequilibrium transitions within equilibrium statistical physics. Nonequilibrium phase transitions are distinguished from equilibrium transitions by orders less than unity and by the fact that equilibrium thermodynamics and statistical mechanics become inapplicable at the critical point. The latter fact requires a change in the Gibbs assumption underlying the canonical and grandcanonical ensembles in order to recover the thermodynamic description in the critical limit.


2021 ◽  
Vol 18 (176) ◽  
Author(s):  
Mirta Galesic ◽  
Henrik Olsson ◽  
Jonas Dalege ◽  
Tamara van der Does ◽  
Daniel L. Stein

Belief change and spread have been studied in many disciplines—from psychology, sociology, economics and philosophy, to biology, computer science and statistical physics—but we still do not have a firm grasp on why some beliefs change more easily and spread faster than others. To fully capture the complex social-cognitive system that gives rise to belief dynamics, we first review insights about structural components and processes of belief dynamics studied within different disciplines. We then outline a unifying quantitative framework that enables theoretical and empirical comparisons of different belief dynamic models. This framework uses a statistical physics formalism, grounded in cognitive and social theory, as well as empirical observations. We show how this framework can be used to integrate extant knowledge and develop a more comprehensive understanding of belief dynamics.


2015 ◽  
Vol 4 (1) ◽  
pp. 1-27
Author(s):  
L´eon Brenig

This essay corresponds to the content of three lectures about statistical physics delivered to the audience of the 2014 section of the R. A. Salmeron School of Physics, at the UnB. Our starting point was very simple statistical models (lattice gas, spin-1/2 ferromagnet), used as illustrations of the competencies and methods in statistical physics. Thus we introduce the Gibbs ensembles, defining a connection with thermodynamics and discussing the role played by fluctuations and large numbers. We present phenomenological aspects of phase transitions and critical phenomena in simple fluids and in uniaxial ferromagnets, emphasizing the universal character of the critical exponents. We describe the phenomenological van der Waals and Curie-Weiss theories and the Landau expansion, which are present-day relevant methods, despite the fact that such theories give rise to critical exponents in disagreement with experiments. We present then the paradigmatic Ising model, which points us to a way to overcome the phenomenological results. A brief presentation of the scale phenomenological methods and the contemporaneous renormalization group are considered at the end of these lectures.


Sign in / Sign up

Export Citation Format

Share Document