Ex2 - Early Stopping Midway Through Epoch #107
-
This is probably more of a general conceptual discussion about early stopping. From what I've read and learned so far, the theory is to check against the validation set every x epochs (being a full run across all training data) to make sure overfitting doesn't occur. However, my reading of the implementation provided is actually to check partially through the epoch and to stop early if performance is reducing. Could we clarify what the effects are of stopping partially through and starting a new epoch again? Wouldn't that mean that some of the training data isn't even seen by the model - or does the batch process give a random batch when If the former, then what are the implications of a model not looking at some of the training data at all? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
Hello, you are right, with early stopping one evaluates against the validation set. However one can do that not only after every x epochs, but after a certain number of training steps. Best, |
Beta Was this translation helpful? Give feedback.
Hello,
you are right, with early stopping one evaluates against the validation set. However one can do that not only after every x epochs, but after a certain number of training steps.
One does not stop partially through the epoch and then start a new epoch, the whole training is finished when early stopping terminates. The reasoning behind that is that one not trains until the whole end of the specified epoch number, but that the training stops, if the performance does not increase after n additional training steps.
Maybe this article also helps a bit to understand how it works.
Best,
Sophia