Let $\{X_k\}$ be a sequence of dependent random variables with mean 0. Define $\bar{Y}_k = \frac{1}{\sqrt k}\sum_{i=1}^k X_i$.
Let $\{W_k\}$ be a sequence of i.i.d. random variables with mean 1 and variance 1. Define $\bar{Z}_k = \frac{1}{\sqrt k}\sum_{i=1}^k W_i X_i$.
It is known that $\bar{Y}_k \implies \mathcal{N}(0,V)$, where $\implies$ denotes convergence in distribution, and $V$ is some covariance matrix.
Can we also say that $\bar{Z}_k$ converges to the same distribution? How would one prove this rigorously? And does this imply that $\bar{Y}_k$ and $\bar{Z}_k$ are asymptotically equivalent, in some sense?'
The main problem I see here is the dependence in the sequence $\{X_k\}$. If they were i.i.d., then the variance of the sum becomes the sum of variances, and the proof becomes straightforward.