In An Introduction to Statistical Learning, the following statement is made comparing leave-one-out cross validation to using a single validation set:
LOOCV has a couple of major advantages over the validation set approach. First, it has far less bias. In LOOCV, we repeatedly fit the statistical learning method using training sets that contain $n − 1$ observations, almost as many as are in the entire data set. This is in contrast to the validation set approach, in which the training set is typically around half the size of the original data set. Consequently, the LOOCV approach tends not to overestimate the test error rate as much as the validation set approach does.
I would think that LOOCV actually just provides a better estimate than using a single validation set, since it is able to fit using more data, and therefore would have a lower variance rather than a lower bias. Why would the bias be lower?