Simulation shows very similar efficiency for these estimators, even when $a=0.99$.
> a<-0.99
> b<-10
> r<-replicate(1000,{
+ x<-numeric(1000+2000)
+ x[1]<-rnorm(1)
+ for(i in 2:3000) x[i]<-x[i-1]*a+b+rnorm(1)
+ x<-x[-(1:1000)]
+ m<-lm(x[-1]~x[-2000])
+ c(coef(m)[1]/(1-coef(m)[2]),mean(x))
+ })
> apply(r,1,mean)
(Intercept)
1000.091 1000.068
> apply(r,1,sd)
(Intercept)
2.289759 2.209923
Setting $\sigma^2$ be small doesn't help, either: with $\sigma^2=0.01$
> apply(r,1,mean)
(Intercept)
999.9974 999.9953
> apply(r,1,sd)
(Intercept)
0.2182076 0.2109398
The other thing that shows up in simulation is that the correlation between the two estimators increases with the length of the simulation. It is 0.972 for the simulation above, with 2000 points, but 0.998 for 20000 points.
So, it looks possible that the estimators are the same (for infinite time).
Now consider the OLS regression. Define $z_d =x_{d-1}-\bar x_{[1:(n-1)}$ and $y_d=x_d-\bar x_{[2:n]}$, ie, center the two variables in the regression about their respective sample means. Write $\hat\beta$ for the intercept and $\hat\alpha$ for the fitted intercept and slope in this new regression.
We have $\hat\beta=0$ and $\hat\alpha=\hat a$, just as a fact about OLS. So the estimator $\hat\beta/(1-\hat\alpha)$ for the mean of $y$ is identical to the estimator $\bar y$ as both are identically zero.
Reparametrising back to the original scale, we shift the intercept to the left by $\bar x_{[2:n]}$ and up by $\bar x_{[1:(n-1)}$ to
$$\hat b= \hat\beta+\bar x_{[1:(n-1)} -\hat \alpha\times\bar x_{[2:n]}$$
For large $n$, the two means are approximately the same, so
$$\hat b= (1-\hat a)\times \bar x_n +O_p(n^{-1})$$
and the estimated mean is
$$\frac{\hat b}{1-\hat a}=\bar x_n +O_p(n^{-1})$$
The two estimators are asymptotically equivalent (to first order) for large $n$.
I will just conclude by noting this is the sort of result that is quite hard to prove if you can't imagine that it might be true, and simulation is a good way to come to imagine that it might be true.
UPDATE
In the special case where the errors are Normal, we can also look at maximum likelihood estimation. The standard results for independent data don't imply efficiency, but we can expect at least pretty good efficiency.
Consider the model $X\sim N(\mu, \Xi)$, where $\mu$ is the mean we're interested in and $\Xi$ is AR-1 covariance model implied by the generating equation. The deviance is
$$d= -2\ell(\mu,\Xi)= \log |Xi| +(x-\mu)^T\Xi^{-1}(x-\mu)$$
Differentiating wrt $\mu$ gives
$$0 = -1^T\Xi^{-1}(y-\mu) - (y-\mu)^T\Xi^{-1}1$$
Write $\xi^{ij}$ for the $(i,j)$ element of $\Xi^{-1}$, and we have
$$ \hat\mu = \frac{\sum_{i,j}\xi^{ij}x_i}{\sum_{i,j}\xi^{ij}}$$
Now, except for edge effects, $\xi^{ij}$ depends only on $|i-j|$. In fact, $\Xi^{-1}$ is tridiagonal, so $\xi^{ij}$ is non-zero only when $|i-j|\leq 1$. So (again, up to edge effects)
$$\sum_{i,j=1}^n\xi^{ij}x_i\approx\sum_{k=1}^n \sum_{l=-1}^1\xi^{k,k+l}x_k\approx \sum_{k=1}^n \left(\sum_{l=-1}^1\xi^{k,k+l}\right)x_k$$
Now, $\left(\sum_{l=-1}^1\xi^{k,k+l}\right)$ is constant in $k$ (except for edge effects). Call it $A$.
$$\hat\mu\approx\frac{A\sum_i x_i}{An}=\bar x$$
So the MLE is also asymptotically equivalent to the sample average, to first order.