In a course it was stated that the boundaries determined by linear discriminant analysis are the same as the boundaries determined by linear regression models if we seperate two classes that are of the same size.
Intuitively, the statement is clear. However, I don't manage to proof or understand it formally/ on the mathematical point of view.
In discriminant analysis, the boundaries are determined by taking the log on $P(x|Y=Class_c, \theta)$ and equalise it with $P(x|Y=Class_{c_1}, \theta)$, i.e. linear discriminant analysis determines all points for the two classes $Class_c$ and $Class_{c_1}$ for which $$ \{x \in \mathrm{IR}^p: P(x|Y=Class_{c_1}, \theta) = P(x|Y=Class_c, \theta)\}$$ where $p$ is the number of features in the data set and $P(x|Y=Class_{c_1}, \theta)$ ~ $N(x|\mu_{Class_{c_1}}, \Sigma_{Class_{c_1}})$
In linear regression, the two models are described by $$P(Y_{Class_{c_1}}|x,\theta) \sim N(y|x^T\beta_{Class_{c_1}},\sigma^2)$$
But how is it now evident that these boundaries fall together? And how are the "equal sized training sets" considered in the proof? I don't really know how to put this together...
Thanks four your help and thought-provoking impulses! :)