I'm glad I asked – ordinal data can be treated rather differently than nominal data. ordinal is a subtype of categorical-data, and differs from nominal in that levels are ordered (hover over these tags for excerpts of the tag wikis, or click on the "info" links for the full spiel). Preserving this information often improves predictions because it better reflects the nature of the predictor variable.
The binary predictors don't need to be treated differently than nominal data, but the polytomous ordinal predictors would be handled well by a penalized regression model. This is essentially a smoothing method that prevents adjacent levels of the ordinal variable from having drastically different dummy coefficients. This is a non-issue with binary variables, because they only require one dummy coefficient apiece. See "Continuous dependent variable with ordinal independent variable" for more. I think regression splines and LASSO / elastic net have some application in linear models based on ordinal predictors as well, but I know less about those myself.
For a little appeal to intuition, compare ordinal variables to continuous variables in your modeling mind. With continuous variables in a linear model, straight lines or smooth curves often make the best regression models; even local regression applies smoothing to prevent bumpy models, as these would likely be overfitted. Many ordinal variables really represent grouped, latent, continuous variables, and as such shouldn't produce really spiky models either. It generally makes more sense for nominal data to be heterogeneously related to metric response variables. To prevent overestimating the difference in relationships between the response variable and various levels of an ordinal variable, some penalization of starkly different coefficients for adjacent levels of ordinal predictors often helps.