The Kullback-Leibler divergence (or relative entropy) is a measure of how a probability distribution differs from another reference probability distribution. I want to know what connection it has to the maximum entropy principle, which says that the uniform ($1/N$) distribution has max entropy.
If the reference distribution is the uniform distribution, and I minimize the KL-divergence of some empirical data's probabilities to that reference, can this be viewed somehow as an attempt to attain maximum entropy in that a KL-divergence of 0 means the empirical is identical (does not diverge from) the target uniform distribution?