Group Seminars
Year 2024
Seminar of Gianluca Teza: 21/11/2024
Title: Nonequilibrium shortcuts and anomalous thermal relaxations: the Mpemba effect
Abstract:
Most of our intuition about the behavior of physical systems is shaped by observations at or near thermal equilibrium. However, even a phenomenon as basic as a thermal quench leads to states far from any thermal equilibrium, where counterintuitive, anomalous, effects can occur. A prime example of anomalous thermal relaxation is the Mpemba effect – a phenomenon in which a system prepared at a hot temperature cools down to the temperature of the cold environment faster than an identical system prepared at a warm temperature. Although originally witnessed in water by Aristotle more than 2000 years ago, perspectives towards the design of optimal heating/cooling protocols and several observations in a variety of systems pushed the development of a high-level characterization in the framework of nonequilibrium statistical mechanics. In this talk, I will review the phenomenology of this and related anomalous relaxation effects, in which nonmonotonic relaxation times act as the common denominator. With a focus on Ising systems, I will provide some insight on the physical mechanisms that enable the emergence of these effects. I will show how they can survive boundary and even arbitrarily weak couplings to the thermal bath, highlighting the role played by equilibrium and dynamical features. If time allows, I will showcase some preliminary results in the attempt to observe these out-of-equilibrium effects in quantum simulators.
Seminar of Fernando Martínez-García: 29/10/2024
Title: Learning Generalized Statistical Mechanics with Matrix Product State
Abstract:
Generative models can be used to learn complex probability distributions. While methods based on neural networks and restricted Boltzmann machines have been the subject of considerable work, recent efforts have focused on the study and development of tensor network generative models. In this talk, I will introduce basic concepts about tensor networks that will allow us to define a tensor network based variational algorithm. This model is trained by minimizing a generalized free energy defined using Tsallis entropy instead of the standard Gibbs entropy. The resulting trained model can generate probability distributions associated with generalized statistical mechanics. This training is efficient, since the resulting free energy and its gradient can be calculated exactly through tensor network contractions, as opposed to other methods which require estimating the Gibbs entropy by sampling. Moreover, we devise a variational annealing scheme by ramping up the inverse temperature, which allows us to train the model while avoiding getting trapped in local minima. We show the validity of our approach in Ising spin-glass problems by comparing it to exact numerical results and quasi-exact analytical approximations. Our work opens up new possibilities for studying generalized statistical physics and solving combinatorial optimization problems with tensor networks.
Link to article: arXiv:2409.08352
Seminar of Carlos Gandarilla: 12/06/2024
Title: Inverse Statistical Physics Algorithms for Inferring Protein-Protein Interactions
Abstract:
Predicting protein-protein interactions from sequences is an important task in computational biology, crucial for a systems-level understanding of the cell. Phylogeny and residue coevolution present in the multiple sequence alignments of both protein families are utilized for this purpose. Based on inverse statistical physics, the Direct Coupling Analysis (DCA) algorithm exploits coevolution to accurately address two conceptually related inference tasks: finding interacting protein pairs and identifying pairs of residues forming contacts between interacting proteins. In both tasks, we observe a quadratic scaling relationship between the quality and size of the dataset, consistent with quadratic noise and linearly added signal as the dataset size increases. We developed the GA-IPA method that combines phylogeny and residue coevolution to enhance the performance of inferring interaction pairs between paralogs. This method employs the alignment of sequence similarity graphs (GA) from the two families to produce a robust partial matching, which is used to seed an iterative matching algorithm (IPA) based on coevolution. The improvement achieved by this method is impressive in challenging cases where the average number of paralogs per species is large or where the total number of sequences is modest.
Link to article: Plos Comp. Bio 2023
Seminar of Gianluca Manzan: 09/04/2024
Title: Efficiency limits of Restricted Boltzmann Machines in a Teacher-Student Framework
Abstract:
Unsupervised Machine learning with Boltzmann machines is the inverse problem of finding a suitable Gibbs measure to approximate an unknown probability distribution from a training set consisting of a large amount of samples. The minimum size of the training set necessary for a good estimation depends on both the properties of the data and of the machine. We investigate this problem in a controlled environment where a Teacher Restricted Boltzmann machine (T-RBM) is used to generate the dataset and another Student machine (S-RBM) is trained with it. We consider different classes of unit priors and weight regularizers and we analyze both the informed and mismatched cases, viewed as the amount of information the Student receives about the Teacher model. We describe the results in terms of phase transitions in the Student posterior distribution, interpreted as a statistical mechanics system. In the analysis we give special attention to the Hopfield model scenario, where the problem is expressed in terms of phase diagrams, describing the zoology of the possible working regimes of the entire environment. In this present case it is possible to observe the differences between memorization and learning approach. When data become large and confused the learning methodology overcomes memorization.
Link to article: Applied mathematics and Computation
Seminar of Misaki Ozawa: 23/01/2024
Title: Renormalization Group Approach for Machine Learning Hamiltonian
Abstract:
Reconstructing, or generating, Hamiltonian associated with high dimensional probability distributions starting from data is a central problem in machine learning and data sciences. We will present a method —The Wavelet Conditional Renormalization Group —that combines ideas from physics (renormalization group theory) and computer science (wavelets, Monte-Carlo sampling, etc.). The Wavelet Conditional Renormalization Group allows reconstructing in a very efficient way classes of Hamiltonians and associated high dimensional distributions hierarchically from large to small length scales. We will present the method and then show its applications to data from statistical physics and cosmology.
Link to article: PRX 2023
Year 2023
Seminar of Claudio Chilin: 20/11/2023
Title: The Hopfield model towards modern generalisations
Abstract:
The Hopfield model is one of the few examples of neural computation systems that allows for an analytical solution via the tools of statistical physics. The original formulation makes use of the inefficient Hebb rule, that can store a number P of uncorrelated examples that is P=0.138N, where N is the number of neurons composing the network. In the light of the results of modern machine learning, some improvements to the usual treatment can be suggested: the use of correlated data and of improved learning protocols - the classic Hebbian unlearning algorithm and a modern non-destructive version, the Daydreaming algorithm -. The combination of these elements suggests the presence of previously unobserved behaviours of this model.
Link to article: openreview
Seminar of Carlo Lucibello: 16/10/2023
Title: The Exponential Capacity of Dense Associative Memories
Seminar of Matteo Negri: 05/07/2023
Title: Storage, Learning and Daydreaming in the Random-Features Hopfield Model
Seminar of Nicolas Béreux: 24/05/2023
Title: Learning a restricted Boltzmann machine using biased Monte Carlo sampling
In our work, we investigate a biased sampling method named Tethered Monte Carlo (TMC) to overcome this limitation. This method allows to properly sample such low dimensional datasets in a significantly shorter time, leading to a more accurate likelihood gradient during training, allowing the RBM to accurately learn such datasets. This method can also be used to retrieve the distribution learned by the RBM after training, allowing to assess the quality of the training.
However, this method breaks the intra-layer independence of the RBMs which forbids the parallelisation of the MCMC updates, limiting the size of the model we can use.
Seminar of Miguel Ruiz García: 24/02/2023
Title: Loss function landscapes, how their structure can determine the fate of constraint satisfaction problems and machine learning
Seminar of Federico Ricci-Tersenghi: 31/01/2023
Title: The spin glass physics behind hard inference problems
Year 2022
Seminar of Cyril Furtlehner: 14/12/2022
Title: Free Dynamics of Feature Learning Processes
Seminar of Ilaria Paga: 22/11/2022
Title: Memory and rejuvenation in spin glasses: numerical simulations meet experiments.
Seminar of Lorenzo Rosset: 20/10/2022
Title: Exploiting the learning dynamics of the Restricted Boltzmann Machine to construct relational trees of data
In this talk, I will present you with a new and general method for building relational trees of data by leveraging the learning dynamics of the Restricted Boltzmann Machine (RBM). In the spirit of explainable Machine Learning, this method has its roots in the Mean Field approach developed in the context of disordered systems. The proposed algorithm can also be used to categorize the data by relying on minimal knowledge of the dataset. This approach yielded encouraging results when applied to three different real-world datasets and yet offers several possible directions of improvement.
Seminar of Elisabeth Agoritsas: 10/10/2022
Title: Towards a unifying picture of driven disordered systems
Disorder is ubiquitous in physical systems, and can radically alter their physical properties compared to their ‘pure’ counterparts. For instance, amorphous materials such as emulsions, foams, metallic glasses or biological tissues are all structurally disordered, and this has key implications for their rheological, mechanical or transport properties. Nevertheless, theoretical descriptions of such ‘driven' amorphous materials remain challenging, despite decades of extensive analytical and computational studies. The difficulties pertain to the interplay of competing sources of stochasticity, and to the resulting out-of-equilibrium nature of these systems. A standard model for amorphous materials, which allows one to focus on the key role of their structural (positional) disorder, is provided by dense many-body systems of pairwise interacting particles. Here I will introduce an exact Dynamical Mean-Field Theory (DMFT) for these many-body systems, derived in the limit of infinite spatial dimension. In this framework, the many- body Langevin dynamics of the whole problem can be exactly reduced to a single scalar effective stochastic process, and dynamical observables such as pressure or shear stress can be computed for arbitrary driving protocols. Using this DMFT, we were in particular able to establish a direct equivalence between a global forcing (external shear) and a random local forcing (reminiscent of active matter), upon a simple rescaling of the control parameter (the accumulated strain). In this framework, global shear is thus simply a special case of a much broader family of local forcing that can be explored by tuning its spatial correlations. Our predictions were moreover found to be in remarkably good agreement with two-dimensional numerical simulations. These results hint at a unifying framework for establishing rigorous analogies, at the mean-field level, between different families of driven disordered systems, such as sheared granular materials and active matter, or machine-learning algorithms.
Link to articles: Granular system, Out-of-Equilibrium dynamical system
Seminar of Antonio Lasanta: 21/09/2022
Title: Relajación anómala en sistemas lejos del equilibrio
En esta charla presentaré algunos resultados recientes tanto teóricos como experimentales sobre la relajación de sistemas sujetos a uno o dos procesos de enfriamiento repentino. En particular, mostraré que durante la evolución transitoria de esos sistemas y antes de alcanzar el estado de equilibrio o estacionario y bajo algunas condiciones concretas, aparecen algunos fenómenos sorprendentes y contraintuitivos
Seminar of Edoardo Sarti: 27/05/2022
Title: Understanding the function of paralogous protein sequences
One of the main ways organisms evolve new functional proteins is via a duplication event in their genome. When two copies of a gene are present, either the organism benefits from a larger concentration of the produced protein or the sequence of one of the two copies will accumulate mutations and diverge in evolution, often developing new functions. Annotating the function of paralogous sequences has always been very challenging both in small-scale, expert-guided assays and in large-scale bioinformatics studies, where paralogs are the most important source of functional annotation errors.
ProfileView is a novel computational method designed to functionally classify sets of homologous sequences. It constructs a library of probabilistic models accurately representing the functional variability of protein families, and extracts biologically interpretable information from the classification process. We have tested it on the 11 proteins composing the Calvin-Benson cycle, and obtained fully consistent results on 8 of them, and partially consistent results on other 2. The knowledge about paralog function annotation in the CBC is being now employed for matching same-function paralog sequences for producing joint MSAs for protein-protein interaction studies
Seminar of David Yllanes: 06/05/2022
Title: Geometric control of thermalised elastic sheets: crumpling and buckling
Seminar of Giovanni Catania: 22/04/2022
Title: Approximate inference on discrete graphical models: message-passing, loop-corrected methods and applications
Year 2021
Seminar of Tony Bonnaire: 27/10/2021
Title: The principal graph of the cosmic web: learning patterns in point cloud datasets