The modeling of large-scale data is essentially probabilistic. Model learning, inference and the generation of new data require sampling these probability distributions. Impressive results have been achieved using neural networks to generate images, sounds, texts or physical fields. We will follow a path from the mathematical foundations to the algorithmic frontiers of random generation by score diffusion.
The lecture introduces the mathematical framework of Monte Carlo learning and statistical inference, as well as random generation algorithms. Consideration is given to Markov field models, which explain the conditional independence of variables, and are characterized by Gibbs energies. The lecture presents rejection sampling algorithms, and those based on Markov chains, notably the Metropolis-Hastings algorithm. It concludes with a presentation of score-diffusion generation algorithms, which enable complex probability distributions to be sampled, by estimating the score with neural networks.