Questions tagged [pre-training]

Unsupervised pre-training initializes a discriminative neural network from one which was trained using an unsupervised criterion, such as a deep belief network or a deep autoencoder.

47 questions
38
votes
3 answers

What is pre training a neural network?

Well the question says it all. What is meant by "pre training a neural network"? Can someone explain in pure simple English? I can't seem to find any resources related to it. It would be great if someone can point me to them.
Machina333
  • 863
  • 2
  • 9
  • 10
11
votes
3 answers

What is pretraining and how do you pretrain a neural network?

I understand that pretraining is used to avoid some of the issues with conventional training. If I use backpropagation with, say an autoencoder, I know I'm going to run into time issues because backpropagation is slow, and also that I can get stuck…
9
votes
1 answer

Is Greedy Layer-Wise Training of Deep Networks necessary for successfully training or is stochastic gradient descent enough?

Is it possible to achieve state of the art results by using back-propagation only (without pre-training) ? Or is it so that all record breaking approaches use some form of pre-training ? Is back-propagation alone good enough ?
user70990
8
votes
1 answer

Is initializing the weights of autoencoders still a difficult problem?

I was wondering if initializing the weights of autoencoders is still difficult and what the most recent strategies are for it. I have been reading different articles. In one of Hinton's papers (2006), it says: With large initial weights,…
7
votes
3 answers

Feature Selection in unbalanced data

I was always taught 3 things: Training algorithms (rf, trees, etc) don't perform well with unbalanced data. I should balance data only after performing feature selection (mainly to keep variables independent) Feature selection algorithms usually…
7
votes
0 answers

Change image input size of a pre-trained convnet

maybe this question will sound a bit as a newbie one but I'd like to have some clarification. I'm using a VGG16-like convnet, pre-trained with VGG16 weights and edited top layers to work with my classification problem; specifically I removed the…
7
votes
1 answer

How to pretrain Convolution filter

I was implementing convolutional neural network, For classification of natural images like face, car, flower etc of about 10 categories. I read(from Andrew NG notes) that pre trained convolutional filter are much efficient and less resource…
5
votes
2 answers

Why does pre-training help avoid the vanishing gradient problem?

I read that a problem with the Classic approach to deep NN is the vanishing gradient, which is caused by the derivative of the logistic activation function - broadly speaking, the update flowing down through the network becomes ever more small. In…
3
votes
1 answer

Rationale for different activation function neural network pretraining vs. supervised training?

I was reading a paper that used neural networks to predict protein conformation, Improving prediction of secondary structure, local backbone angles, and solvent accessible surface area of proteins by iterative deep learning: Linear activation…
3
votes
0 answers

State of the Art Status of Deep Boltzmann Machine and Pretraining

I have been reading some old papers by Hinton on deep Boltzmann machine and deep belief networks, but I wonder what the current status is regarding these models: Are DBM and DBN totally outdated? I can understand they do not work as well as CNN and…
DiveIntoML
  • 1,583
  • 1
  • 11
  • 21
3
votes
1 answer

Visualizing model trajectories for Neural Networks using function approximator

Erhan et al. in their 2010 paper discusses how pre-training improves deep networks: http://www.jmlr.org/papers/volume11/erhan10a/erhan10a.pdf#page=15 In there, they compare different neural network models by visualizing the function representation…
3
votes
1 answer

Using pretrained segmantation network for unseen motives

For a research project, I need to do a segmentation on images. Since the motivation is nothing any of the big networks was ever trained on, I would ask if it still makes sense to use pretrained segmentation networks like SegNet to do the…
2
votes
1 answer

Transfer learning for regression problems

I have trained a regression model with 7 features for a given problem. Now, I have another regression problem (quite similar to the previous one) where I have only 6 samples in hand, but with 3 more features than the first model (7+3). The…
2
votes
1 answer

How to resume training in neural networks properly?

I'm working on training a network to identify different kinds of cells. For each experimental batch, I would take my previous model weight, and then train a few new pictures on it. Since the model is for the same kind of data, thus this is not a…
2
votes
0 answers

Pre-training without seeing data

Is there a solid reference on pre-training methods in deep neural networks which never see the actual inputs? Any such known thing in literature? I guess a more correct term is "initialization using gradient methods" instead of "pre-training". I see…
1
2 3 4