thanks to kjetil, I now understand two things that were erroneous in my question.
First off, $np(1-p) \neq np - p^2$, it's equal to $np - np^2$
$np(1-p)=\lambda(1-p) = \lambda$. Why does $\lambda(1-p) = \lambda$? Because it's part of our assumptions that $p \to 0$ for the poisson distribution.
This provides some pretty great discussion on how to conceptualize the poisson.
The discussion in my own words.
The most enlightening to me was thinking of the equation $p = \lambda/n$. The probability $p$ of an event occurring is the number_of_occurrences_over_some_interval_or_space/ number_of_occurrences_for_smaller_partitions_of_that_interval_or_space.
Really what we're trying to ask is if our event in our probability space can show up anywhere on the space with the same probability. We call how much we expect to see it $\lambda$ and then we can scale that does to some arbitrarily small measure.
At this point, it doesn't make any sense to even think about $n$. It's just a variable we used to conceptualize what the poisson distribution really is. What's the probability that your "arbitrarily small sample" contains the event? 0. There's no way it could happen. How often do you expect it to happen if you were to scale up that epsilon large sample ($\epsilon=p=\lambda/n$ where $n\to\infty$) by n? $\lambda$ because that's what we defined it to be.
Now we have the poisson distribution.
You'll notice at this point, $n$ amd $p$ don't seem to have much to do with the poisson distribution. Well yeah... That's why they're not variables when we're describing it's density function.