Questions tagged [proof]

For mathematical proofs or derivations of results.

191 questions
79
votes
5 answers

How exactly did statisticians agree to using (n-1) as the unbiased estimator for population variance without simulation?

The formula for computing variance has $(n-1)$ in the denominator: $s^2 = \frac{\sum_{i=1}^N (x_i - \bar{x})^2}{n-1}$ I've always wondered why. However, reading and watching a few good videos about "why" it is, it seems, $(n-1)$ is a good unbiased…
PhD
  • 13,429
  • 19
  • 45
  • 47
38
votes
8 answers

Is it possible to prove a null hypothesis?

As the question states - Is it possible to prove the null hypothesis? From my (limited) understanding of hypothesis, the answer is no but I can't come up with a rigorous explanation for it. Does the question have a definitive answer?
Pulkit Sinha
  • 491
  • 1
  • 4
  • 6
28
votes
3 answers

Proof that moment generating functions uniquely determine probability distributions

Wackerly et al's text states this theorem "Let $m_x(t)$ and $m_y(t)$ denote the moment-generating functions of random variables X and Y, respectively. If both moment-generating functions exist and $m_x(t) = m_y(t)$ for all values of t, then X and Y…
23
votes
2 answers

Uniform random variable as sum of two random variables

Taken from Grimmet and Stirzaker: Show that it cannot be the case that $U=X+Y$ where $U$ is uniformly distributed on [0,1] and $X$ and $Y$ are independent and identically distributed. You should not assume that X and Y are continuous variables. A…
20
votes
2 answers

Deriving the bivariate Poisson distribution

I've recently encountered the bivariate Poisson distribution, but I'm a little confused as to how it can be derived. The distribution is given by: $P(X = x, Y = y) = e^{-(\theta_{1}+\theta_{2}+\theta_{0})}…
19
votes
4 answers

Prove the equivalence of the following two formulas for Spearman correlation

From wikipedia, Spearman's rank correlation is calculated by converting variables $X_i$ and $Y_i$ into ranked variables $x_i$ and $y_i$, and then calculating Pearson's correlation between the ranked variables: However, the article goes on to state…
Alex
  • 3,728
  • 3
  • 25
  • 46
16
votes
3 answers

Explanation of formula for median closest point to origin of N samples from unit ball

In Elements of Statistical Learning, a problem is introduced to highlight issues with k-nn in high dimensional spaces. There are $N$ data points that are uniformly distributed in a $p$-dimensional unit ball. The median distance from the origin to…
user64773
  • 163
  • 1
  • 5
14
votes
4 answers

Question about a normal equation proof

How can you prove that the normal equations: $(X^TX)\beta = X^TY$ have one or more solutions without the assumption that X is invertible? My only guess is that it has something to do with generalized inverse, but I am totally lost.
ryati
  • 243
  • 3
  • 8
13
votes
1 answer

Quantile regression estimator formula

I have seen two different representations of the quantile regression estimator which are $$Q(\beta_{q}) = \sum^{n}_{i:y_{i}\geq x'_{i}\beta} q\mid y_i - x'_i \beta_q \mid + \sum^{n}_{i:y_{i}< x'_{i}\beta} (1-q)\mid y_i - x'_i \beta_q…
AlexH
  • 946
  • 1
  • 9
  • 18
13
votes
2 answers

Sum of two normal products is Laplace?

It is apparently the case that if $X_i \sim N(0,1)$, then $X_1 X_2 + X_3 X_4 \sim \mathrm{Laplace(0,1)}$ I've seen papers on arbitrary quadratic forms, which always results in horrible non-central chi-squared expressions. The above simple…
Corvus
  • 4,573
  • 1
  • 27
  • 58
11
votes
3 answers

The concept of 'proven statistically'

When the news talk about things been 'proven statistically' are they using a well-defined concept of statistics correctly, using it wrong, or just using an oxymoron? I imagine that a 'statistical proof' is not actually something performed to prove a…
Quora Feans
  • 384
  • 2
  • 3
  • 10
11
votes
1 answer

Regression Proof that the point of averages (x,y) lies on the estimated regression line

How do you show that the point of averages (x,y) lies on the estimated regression line?
Justin Meltzer
  • 357
  • 1
  • 3
  • 7
10
votes
1 answer

Expectation, Variance and Correlation of a bivariate Lognormal distribution

If $Y \sim N(\mu,\sigma^2)$ is normally distributed, then $X=\mathrm{e}^Y$ is lognormally distributed. To get the log-$\mu$ and log-$\sigma$ of this lognormal distribution you calculate $$\sigma^2 = \ln\left( \frac{\mathit{Var}}{E^2} + 1 \right)$$…
spore234
  • 1,323
  • 1
  • 15
  • 31
10
votes
1 answer

Show that if $X \sim Bin(n, p)$, then $E|X - np| \le \sqrt{npq}.$

Currently stuck on this, I know I should probably use the mean deviation of the binomial distribution but I can't figure it out.
10
votes
2 answers

Is there an elegant/insightful way to understand this linear regression identity for multiple $R^2$?

In linear regression I have come across a delightful result that if we fit the model $$E[Y] = \beta_1 X_1 + \beta_2 X_2 + c,$$ then, if we standardize and centre the $Y$, $X_1$ and $X_2$ data, $$R^2 = \mathrm{Cor}(Y,X_1) \beta_1 + \mathrm{Cor}(Y,…
Corvus
  • 4,573
  • 1
  • 27
  • 58
1
2 3
12 13