Skip to main content

Questions tagged [regularization]

For questions about application of regularization techniques.

0 votes
1 answer
102 views

The typical objective function in regression problems like Lasso or Ridge includes a Residual Sum of Squares (RSS) term added to a penalty term based on a norm of the coefficients. What are the ...
BigMistake's user avatar
2 votes
0 answers
48 views

I am faced with a problem which I bet was already solved before, but that I had never seen. Perhaps by discussing it abstractly, someone can point me to relevant literature. It goes like this: I have ...
Alek Fröhlich's user avatar
0 votes
1 answer
115 views

I am creating a neural network using batchnorm as a regularization method to enable deep models and prevent overfitting. I understand that batchnorming supresses the internal covariance shift ...
Quantum's user avatar
  • 121
4 votes
1 answer
2k views

I am optimizing a neural network with Adam using 3 different losses. Their scale is very different, and the current method is to either sum the losses and clip the gradient or to manually weight them ...
Simon's user avatar
  • 273
1 vote
1 answer
157 views

I, i have a doubt about making validation using early stopping given two NN models. Suppose I have two models M1 and M2 and a Training set TS and Test set TS. Take the TS and consider TS_80% and TS_20%...
PwNzDust's user avatar
  • 113
0 votes
1 answer
133 views

I was watching a video by Andrew Ng about regularization in logistic regression and neural network models. He uses the following term for regularization to (the sum is over the weights in the network)....
martinkunev's user avatar
1 vote
1 answer
113 views

In the original paper on dropout, in section 7.3.2, we see that while keeping $pn$ constant, we get a (test) error increase by decreasing retainment below 0.6. Why would that happen? If $pn$ is ...
Apples14's user avatar
3 votes
1 answer
4k views

I've searched for an answer to this, and read several scientific articles on the subject, but I can't find a practical explanation of how Dropout actually drops nodes in an algorithm. I've read that ...
Connor's user avatar
  • 133
1 vote
0 answers
378 views

I’m working on a classification problem (500 classes). My NN has 3 fully connected layers, followed by an LSTM layer. I use nn.CrossEntropyLoss() as my loss ...
helloworld's user avatar
0 votes
1 answer
371 views

Let's say I have data records looking like that: (x1, x2, x3, x4, ..., x100), where each x can be either ...
leleogere's user avatar
  • 101
1 vote
1 answer
100 views

Dropout is a regularization technique used in neural networks. It is useful in preventing overfitting by making a neural network as good as an ensemble system. In dropout, we switch off $p$ percent of ...
hanugm's user avatar
  • 4,302
0 votes
1 answer
524 views

I have the following question on "Regularization vs. Constrained Optimization" : In the context of statistical modelling, we are often taught about "Regularization" as a method of ...
stats_noob's user avatar
4 votes
2 answers
480 views

As I understand, this is the general summary of the Regularization-Overfitting Problem: The classical "Bias-Variance Tradeoff" suggests that complicated models (i.e. models with more ...
stats_noob's user avatar
2 votes
0 answers
78 views

I just had a very strange experience where I was training an 8 layer deep & pretty wide (max: 512 neurons) neural network for a regression task. I had assumed since it was big enough that it would ...
profPlum's user avatar
  • 586
1 vote
1 answer
160 views

Accuracy of my regularized model is higher for training set than for validation set. The situation improves when regularization coeefficient is reduced: What does this really imply? From my ...
Aadith Ramia's user avatar

15 30 50 per page