Questions tagged [bayesian]

Bayesian inference is a method of statistical inference that relies on treating the model parameters as random variables and applying Bayes' theorem to deduce subjective probability statements about the parameters or hypotheses, conditional on the observed dataset.

Overview

Bayesian inference is a method of statistical inference that treats model parameters as if they were random variables in order to rely on probability calculus and produces complete and unified probabilistic statements about these parameters.

This approach starts with choosing a reference or prior probability distribution on the parameters and then applies Bayes' Theorem to deduce probability statements about parameters or hypotheses, conditional on the data, treating the likelihood function as a conditional density of the data given the (random) parameter. Bayes' Theorem asserts that the conditional density of the parameter $\theta$ given the data, $P(\theta|d)$, can be expressed in terms of the density of the data given $\theta$ as

$$P(\theta|d) = \dfrac{P(d|\theta)P(\theta)}{P(d)}.$$

$P(\theta|d)$ is called the posterior probability. $P(d|\theta)$ is often called the likelihood function and denoted $L(\theta|d)$. The distribution of $\theta$ itself, given by $P(\theta)$, is called the prior or the reference measure. It encodes previous or prior beliefs about $\theta$ within a model appropriate for the data. There is necessarily a part of arbitrariness or subjectivity in the choice of that prior, which means that the resulting inference is impacted by this choice (or conditional to it). This also means that two different choices of priors lead to two different posterior distributions, which are not directly comparable.

The marginal distribution of the data, $P(d)$ (which appears as a normalization factor), is also called the evidence, as it is directly used for Bayesian model comparison through the notions of Bayes factors and model posterior probabilities.

The comparison of two models (including two opposed hypotheses about the parameters) in the Bayesian framework indeed proceeds by taking the ratio of the evidences for these two models under comparisons, $$ B_{12} = P_1(d)\big/P_2(d)\,. $$ This is called the Bayes factor and it is usually compared to $1$.

Bayes' formula can be used as an updating procedure: as more data become available, the posterior can be updated successively, becoming the prior for the next step.

References

The following threads contain lists of references:

The following journal is dedicated to research in Bayesian statistics:

7099 questions
414
votes
14 answers

Bayesian and frequentist reasoning in plain English

How would you describe in plain English the characteristics that distinguish Bayesian from Frequentist reasoning?
Daniel Vassallo
  • 4,249
  • 3
  • 15
  • 7
292
votes
10 answers

What's the difference between a confidence interval and a credible interval?

Joris and Srikant's exchange here got me wondering (again) if my internal explanations for the difference between confidence intervals and credible intervals were the correct ones. How you would explain the difference?
261
votes
11 answers

How would you explain Markov Chain Monte Carlo (MCMC) to a layperson?

Maybe the concept, why it's used, and an example.
Neil McGuigan
  • 9,292
  • 13
  • 54
  • 62
231
votes
38 answers

What is the best introductory Bayesian statistics textbook?

Which is the best introductory textbook for Bayesian statistics? One book per answer, please.
Shane
  • 11,961
  • 17
  • 71
  • 89
147
votes
15 answers

Amazon interview question—probability of 2nd interview

I got this question during an interview with Amazon: 50% of all people who receive a first interview receive a second interview 95% of your friends that got a second interview felt they had a good first interview 75% of your friends that DID NOT…
Rick
  • 1,431
  • 2
  • 11
  • 9
145
votes
3 answers

Help me understand Bayesian prior and posterior distributions

In a group of students, there are 2 out of 18 that are left-handed. Find the posterior distribution of left-handed students in the population assuming uninformative prior. Summarize the results. According to the literature 5-20% of people are…
Bob
  • 1,451
  • 3
  • 10
  • 3
129
votes
14 answers

What's wrong with XKCD's Frequentists vs. Bayesians comic?

This xkcd comic (Frequentists vs. Bayesians) makes fun of a frequentist statistician who derives an obviously wrong result. However it seems to me that his reasoning is actually correct in the sense that it follows the standard frequentist…
repied2
  • 1,577
  • 2
  • 10
  • 10
116
votes
10 answers

ASA discusses limitations of $p$-values - what are the alternatives?

We already have multiple threads tagged as p-values that reveal lots of misunderstandings about them. Ten months ago we had a thread about psychological journal that "banned" $p$-values, now American Statistical Association (2016) says that with our…
Tim
  • 108,699
  • 20
  • 212
  • 390
94
votes
12 answers

Who Are The Bayesians?

As one becomes interested in statistics, the dichotomy "Frequentist" vs. "Bayesian" soon becomes commonplace (and who hasn't read Nate Silver's The Signal and the Noise, anyway?). In talks and introductory courses, the point of view is…
Antoni Parellada
  • 23,430
  • 15
  • 100
  • 197
93
votes
9 answers

Are there any examples where Bayesian credible intervals are obviously inferior to frequentist confidence intervals

A recent question on the difference between confidence and credible intervals led me to start re-reading Edwin Jaynes' article on that topic: Jaynes, E. T., 1976. `Confidence Intervals vs Bayesian Intervals,' in Foundations of Probability Theory,…
Dikran Marsupial
  • 46,962
  • 5
  • 121
  • 178
87
votes
11 answers

Why should I be Bayesian when my model is wrong?

Edits: I have added a simple example: inference of the mean of the $X_i$. I have also slightly clarified why the credible intervals not matching confidence intervals is bad. I, a fairly devout Bayesian, am in the middle of a crisis of faith of…
Guillaume Dehaene
  • 2,137
  • 1
  • 10
  • 18
87
votes
4 answers

What is an "uninformative prior"? Can we ever have one with truly no information?

Inspired by a comment from this question: What do we consider "uninformative" in a prior - and what information is still contained in a supposedly uninformative prior? I generally see the prior in an analysis where it's either a frequentist-type…
Fomite
  • 21,264
  • 10
  • 78
  • 137
83
votes
14 answers

When (if ever) is a frequentist approach substantively better than a Bayesian?

Background: I do not have an formal training in Bayesian statistics (though I am very interested in learning more), but I know enough--I think--to get the gist of why many feel as though they are preferable to Frequentist statistics. Even the…
jsakaluk
  • 5,006
  • 1
  • 20
  • 45
81
votes
2 answers

XKCD's modified Bayes theorem: actually kinda reasonable?

I know this is from a comic famous for taking advantage of certain analytical tendencies, but it actually looks kind of reasonable after a few minutes of staring. Can anyone outline for me what this "modified Bayes theorem" is doing?
eric_kernfeld
  • 4,828
  • 1
  • 16
  • 41
79
votes
2 answers

Bayes regression: how is it done in comparison to standard regression?

I got some questions about the Bayesian regression: Given a standard regression as $y = \beta_0 + \beta_1 x + \varepsilon$. If I want to change this into a Bayesian regression, do I need prior distributions both for $\beta_0$ and $\beta_1$ (or…
TinglTanglBob
  • 878
  • 1
  • 8
  • 13
1
2 3
99 100