Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
RBMs are generative models capable of fitting complex dataset's probability distributions. Thanks to their simple structure, they are particularly well suited for interpretability and pattern extraction, a feature particularly appealing for scientific use. In this talk, we show that RBMs operate in two distinct regimes, depending on the procedure followed to estimate the log-likelihood gradient during the training. Short sampling times fit machines that are trained to reproduce exactly the dynamics followed to train them, long samplings (as compared to the MCMC mixing time) are need to learn a good model for the data. The non-equilibrium regime should be used to generate high quality samples in short learning and sampling times, but cannot be used to extract the unnormalized data probability of the data necessary for interpretability. In practice, it is hard to extract good equilibrium models for structured datasets (which is the typical case in biological applications) due to a divergence of the Monte Carlo mixing times. In this work, we show this barrier can be surmounted using biased Monte Carlo methods.
[-]
RBMs are generative models capable of fitting complex dataset's probability distributions. Thanks to their simple structure, they are particularly well suited for interpretability and pattern extraction, a feature particularly appealing for scientific use. In this talk, we show that RBMs operate in two distinct regimes, depending on the procedure followed to estimate the log-likelihood gradient during the training. Short sampling times fit ...
[+]
68T07 ; 82C44 ; 65C05