Questions tagged [bias]

The difference between the expected value of a parameter estimator & the true value of the parameter. Do NOT use this tag to refer to the [bias-term] / [bias-node] (ie the [intercept]).

Bias, in a statistical framework, means that an estimate of a parameter has an expected value that is not equal to the actual parameter value. The bias of an estimator can be evaluated with the mean squared error: $$MSE(\widehat{\theta}) = E[(\widehat{\theta} - \theta)^2]$$ which can be decomposed into the sum of the squared bias and the variance of an estimator.

One common example of using a biased estimator is ridge regression; ridge regression can be useful when there is collinearity. The estimators are biased (unlike OLS estimates) but have much lower variance.

962 questions
122
votes
8 answers

Bias and variance in leave-one-out vs K-fold cross validation

How do different cross-validation methods compare in terms of model variance and bias? My question is partly motivated by this thread: Optimal number of folds in $K$-fold cross-validation: is leave-one-out CV always the best choice?. The answer…
61
votes
10 answers

What does "Scientists rise up against statistical significance" mean? (Comment in Nature)

The title of the Comment in Nature Scientists rise up against statistical significance begins with: Valentin Amrhein, Sander Greenland, Blake McShane and more than 800 signatories call for an end to hyped claims and the dismissal of possibly…
uhoh
  • 685
  • 1
  • 6
  • 10
59
votes
7 answers

Intuitive explanation of the bias-variance tradeoff?

I am looking for an intuitive explanation of the bias-variance tradeoff, both in general and specifically in the context of linear regression.
NPE
  • 5,351
  • 5
  • 33
  • 44
50
votes
4 answers

When is a biased estimator preferable to unbiased one?

It's obvious many times why one prefers an unbiased estimator. But, are there any circumstances under which we might actually prefer a biased estimator over an unbiased one?
39
votes
7 answers

What are the most common biases humans make when collecting or interpreting data?

I am an econ/stat major. I am aware that economists have tried to modify their assumptions about human behavior and rationality by identifying situations in which people don't behave rationally. For example, suppose I offer you a 100% chance of a…
Stan Shunpike
  • 3,623
  • 2
  • 27
  • 36
38
votes
4 answers

(Why) do overfitted models tend to have large coefficients?

I imagine that the larger a coefficient on a variable is, the more ability the model has to "swing" in that dimension, providing an increased opportunity to fit noise. Although I think I've got a reasonable sense of the relationship between the…
David Marx
  • 6,647
  • 1
  • 25
  • 43
38
votes
4 answers

When is the bootstrap estimate of bias valid?

It is often claimed that bootstrapping can provide an estimate of the bias in an estimator. If $\hat t$ is the estimate for some statistic, and $\tilde t_i$ are the bootstrap replicas (with $i\in\{1,\cdots,N\}$), then the bootstrap estimate of bias…
Bootstrapped
  • 381
  • 1
  • 3
  • 5
29
votes
4 answers

Intuitive reasoning behind biased maximum likelihood estimators

I have a confusion on biased maximum likelihood (ML) estimators. The mathematics of the whole concept is pretty clear to me but I cannot figure out the intuitive reasoning behind it. Given a certain dataset which has samples from a distribution,…
ssah
  • 391
  • 3
  • 5
27
votes
5 answers

No regularisation term for bias unit in neural network

According to this tutorial on deep learning, weight decay (regularization) is not usually applied to the bias terms b why? What is significance (intuition) behind it?
Harshit
  • 371
  • 1
  • 3
  • 4
26
votes
6 answers

Deep learning : How do I know which variables are important?

In terms of neural network lingo (y = Weight * x + bias) how would I know which variables are more important than others? I have a neural network with 10 inputs, 1 hidden layer with 20 nodes, and 1 output layer which has 1 node. I'm not sure how to…
user1367204
  • 831
  • 3
  • 10
  • 18
26
votes
2 answers

Bias of moment estimator of lognormal distribution

I am doing some numerical experiment that consists in sampling a lognormal distribution $X\sim\mathcal{LN}(\mu, \sigma)$, and trying to estimate the moments $\mathbb{E}[X^n]$ by two methods: Looking at the sample mean of the $X^n$ Estimating $\mu$…
user29918
  • 363
  • 2
  • 6
25
votes
5 answers

Why does a decision tree have low bias & high variance?

Questions Does it depend on whether the tree is shallow or deep? Or can we say this irrespective of the depth/levels of the tree? Why is bias low & variance high? Please explain intuitively and mathematically
GeorgeOfTheRF
  • 5,063
  • 14
  • 42
  • 51
24
votes
5 answers

Why is bias affected when a clinical trial is terminated at an early stage?

An interim analysis is an analysis of the data at one or more time points prior the official close of the study with the intention of, e.g., possibly terminating the study early. According to Piantadosi, S. (Clinical trials - a methodologic…
ocram
  • 19,898
  • 5
  • 76
  • 77
24
votes
1 answer

Incidental parameter problem

I always struggle to get the true essence of the incidental parameter problem. I read in several occasions that the fixed effects estimators of nonlinear panel data models can be severely biased because of the "well-known" incidental parameter…
emeryville
  • 628
  • 1
  • 4
  • 17
24
votes
3 answers

What intuitively is "bias"?

I'm struggling to grasp the concept of bias in the context of linear regression analysis. What is the mathematical definition of bias? What exactly is biased and why/how? Illustrative example?
Fabian
  • 1,341
  • 4
  • 12
  • 13
1
2 3
64 65