Questions tagged [statistical-learning]

SL refers to the statistical perspective on machine learning.

Statistical learning refers to the statistical perspective on machine learning. It does not cover all ML topics; it focuses primarily on predictive modeling (regression and classification).

95 questions
128
votes
5 answers

How does a Support Vector Machine (SVM) work?

How does a Support Vector Machine (SVM) work, and what differentiates it from other linear classifiers, such as the Linear Perceptron, Linear Discriminant Analysis, or Logistic Regression? * (* I'm thinking in terms of the underlying motivations for…
tdc
  • 7,289
  • 5
  • 32
  • 62
28
votes
4 answers

Supervised clustering or classification?

The second question is that I found in a discussion somewhere on the web talking about "supervised clustering", as far as I know, clustering is unsupervised, so what is exactly the meaning behind "supervised clustering" ? What is the difference with…
16
votes
4 answers

What is data blending?

This term appears frequently in the method-related threads. Is blending a specific method in data-mining and statistical learning? I cannot get a relevant result from google. It seems blending is mixing up outcomes from many models and resulting in…
14
votes
2 answers

How can a vector of variables represent a hyperplane?

I am reading Elements of Statistical Learning and on page 12 (section 2.3) a linear model is notated as: $$\widehat{Y} = X^{T} \widehat{\beta}$$ ...where $X^{T}$ is the transpose of a column vector of the predictors / independent variables / inputs.…
Scott
  • 599
  • 1
  • 5
  • 12
14
votes
1 answer

Why should we discuss convergence behaviors of different estimators in different topologies?

In the first chapter of the book Algebraic Geometry and Statistical Learning Theory which talks about the convergence of estimations in different functional space, it mentions that the Bayesian estimation corresponds to the Schwartz distribution…
ziyuang
  • 1,536
  • 8
  • 32
13
votes
2 answers

Is Random Forest a good option for unbalanced data Classification?

Despite the resembling and other increasing data variability approaches, can the random forest "as an algorithm" be considered a good option for the unbalanced data classification?
12
votes
5 answers

Difference between Bias and Error?

In statistics, what is the difference between Bias and Error? You can say, Bias is a type of error? or Bias is an error with some tendency?
Darwin PC
  • 459
  • 2
  • 4
  • 10
11
votes
1 answer

Clarification in information geometry

This question is concerned with the paper Differential Geometry of Curved Exponential Families-Curvatures and Information Loss by Amari. The text goes as follows. Let $S^n=\{p_{\theta}\}$ be an $n$-dimensional manifold of probability distributions…
11
votes
1 answer

Confused by Derivation of Regression Function

I just got a copy of The Elements of Statistical Learning by Hastie, Tibshirani, and Friedman. In chapter 2 (Overview of Supervised Learning) section 4 (Statistical Decision Theory), he gives a derivation of the regression function. Let $X \in…
Orangutango
  • 213
  • 1
  • 5
11
votes
1 answer

Logistic regression for data from Poisson distributions

From some machine learning notes talking about some discriminative classification methods, in particular logistic regression, where y is the class label (0 or 1) and x is the data, it is said that: if $x|y = 0 \sim \mathrm{Poisson}(λ_0)$, and $x|y…
11
votes
1 answer

How are piecewise cubic spline bases constructed?

There are words from the The Elements of Statistical Learning on page 119: It is not hard to show that the following basis represents a cubic spline with knots at $\xi_1$ and $\xi_2$: $h_1(X)=1$…
jerry_sjtu
  • 435
  • 4
  • 10
11
votes
1 answer

Extending 2-class models to multi-class problems

This paper on Adaboost gives some suggestions and code (page 17) for extending 2-class models to K-class problems. I would like to generalize this code, such that I can easily plug in different 2-class models and compare the results. Because most…
Zach
  • 22,308
  • 18
  • 114
  • 158
10
votes
1 answer

Statistical learning theory VS computational learning theory?

What relations and differences are between statistical learning theory and computational learning theory? Are they about the same topic? Solve the same problems, and use the same methods? For example, the former says it is the theory of prediction…
Tim
  • 1
  • 29
  • 102
  • 189
8
votes
1 answer

Fat-shattering dimension

A set of points $X = \{x\}$ is $\gamma$-shattered by a set of functions $\mathcal{F}$ if there are real numbers $r_x$ indexed by $x$ such that for any binary vector $b$ defining labeling of points from $X$ we can find a function $f \in \mathcal{F}$…
Leo
  • 2,484
  • 3
  • 22
  • 29
8
votes
1 answer

Understanding the different formulations for SVM

I've been working with kernlab for more than a year now, but I always stickied to the vanilla cost (C-svc) formulation for classification. Of course, kernlab includes some other formulations. In the manual, some classification formulations are…
Firebug
  • 15,262
  • 5
  • 60
  • 127
1
2 3 4 5 6 7