Consider the model estimated in levels (also assume this is the true population model):
$$y_t = x_t\beta + e_t$$
As usual we have the dependent variable $y$, independent $x$, the error term $e$, and the to be estimated parameter $\beta$.
Take away $y_{t-1}$ from both sides of the equation and substitute the second term of the first equation to the righ hand side to obtain regression in differences:
$$y_t - y_t = x_t \beta - x_{t-1} \beta +e_t - e_{t-1}$$ $$ \Delta y_t = \Delta x_t \beta + \Delta e_t$$
Since the only difference in the models is the autocorrelation of the error term, you would have to assume that the transformation doesn't generate bias. However, one of the estimators is more efficient, because of the different autocorrelation structure of the error term. From this we can conclude that the betas are exactly the same, only reason why one should be different from the other asymptotically is bias.
Yet, the betas are supposed to differ from one another and it is said that by estimating the difference regression you can't imply levels differences. (This is one motivation for (V)ECMs; however, omitted variable bias could be another motivation.)
My question is, what is the actual relationship between the betas of the levels and differences regression? Is the divergence of the parameter estimates in fact only a result of breaking other Gauss-Markov assumptions (than the autocorrelation)? For example, if the variables have a relationship, but also behave like I(1) processes, then levels regression obviously has bias.
Of course, it's problematic for panel first-differences models, too, if only short-run dynamics can be implied from differenced variables. What is the truth of the matter?
For example, in this video the argument is made that the estimates of beta won't be the same in the two equations. The results are even simulated and I can't see any faults in the simulation, although the betas are not presented.