You want to make sure that your analysis properly weights cases and that you capture the error variance appropriately.
The advantage of the Poisson generalized linear model is that it has a chance of doing all that at once. The standard logarithmic link between the linear predictor and the counts means that you don't take the log of the counts yourself. You adjust for the log of the population size in an offset
term* that forces its regression coefficient to be exactly one. The results are thus equivalent to an analysis of rate. As the variance of a Poisson distribution must equal its mean, the analysis properly weights samples and takes into account the error variance, if the Poisson distribution is an adequate model.
The first option you present would equally weight a population of 10 members having one event and a population of 10,000 members with 1000 events. That wouldn't make much sense, as the rate estimate from the larger population is much more precise. The second, as written, would model the log of deaths as a linear function of the population size, which I don't think that you want if you are essentially modeling rates. Using log of the population would help, but unless you treat that as an offset you would not be modeling rates (and I'm not immediately sure how well that would work if you used the inverse hyperbolic sine).
It probably makes the most sense to try the Poisson model, see if the assumptions underlying the model hold well enough, and if not use one of related models like quasipoisson or negative binomial.
*The example in the linked page is for events over time, but the principle holds for events among a population.