Sampling by Divergence Minimization Abstract: We introduce a family of Markov Chain Monte Carlo (MCMC) methods designed to sample from target distributions with irregular geometry using an adaptive scheme. In cases where targets exhibit non-Gaussian behaviour, we propose that adaption should be regional in nature as opposed to global. Our algorithms minimize the information projection side of the Kullback-Leibler (KL) divergence between the proposal distribution class and the target to encourage proposals distributed similarly to the regional geometry of the target. Unlike traditional adaptive MCMC, this procedure rapidly adapts to the geometry of the current position as it explores the space without the need for a large batch of samples. We extend this approach to multimodal targets by introducing a heavily tempered chain to enable faster mixing between regions of interest. The divergence minimization algorithms are tested on target distributions with multiple irregularly shaped modes and we provide results demonstrating the effectiveness of our methods. The paper in its entirety can be accessed [here](https://arxiv.org/abs/2105.00520). Supporting code and model documentation can be found in the [ScoutMCMC Github repository](https://github.com/AmeerD/Scout-MCMC).