In this lecture, we pivot from our motivation from the Simulated Annealing optimization metaheuristic to thinking about how to sample from microstates within the physically inspired search process. This requires us to introduce the concept of entropy, a quantity which measures the number of microstates in a coarse-grained "macrostate" description of a system. Within the constraints of a system, we seek a distribution of microstates that represents only those constraints and not any additional information. This is the maximal entropy distribution for those constraints. We provide a few formalities on how to make this a little more rigorous and then introduce Maximum Entropy (MaxEnt) methods once popular in NLP that remain to be popular in Species Distribution Modeling and archaeology. We will use MaxEnt to help us define the Boltzmann–Gibbs distribution (and Monte Carlo methods to sample from it) next time.
Whiteboard notes for this lecture can be found at:
https://www.dropbox.com/scl/fi/01pfdkj3d3ilk7wiyu79a/IEE598-Lecture5B-2026-03-24-From_Entropy_to_Maximum_Entropy_MaxEnt_Methods-Notes.pdf?rlkey=xfe1pie4sxu0qklg871czuc05&dl=0