Background
I am watching a popular YouTube crash course on machine learning.
At 3:35:50, he mentions that the model is likely overfit, so fits it again with less epochs.
Since he didn't reinstantiate the model, isn't this equivalent to fitting the model with that same data, thereby continuing to overtrain it?
My Question
Assume you have a model created and data ready to go.
You run:
model.fit(train_images, train_labels, epochs=10)
model.fit(train_images, train_labels, epochs=8)
Is this equivalent to running:
model.fit(train_images, train_labels, epochs=18)
Or:
model.fit(train_images, train_labels, epochs=8)
If previously fitted data is overwritten, why does running model.fit
a second time begin with the accuracy of the previous model?
In multiple other questions regarding saving and training models, the accepted solutions are to load the previously trained model, and run model.fit
again.
If this will overwrite the pre-existing weights, doesn't that defeat the purpose of saving the model in the first place? Wouldn't training the model for the first time on the new data be equivalent?
What is the appropriate way to train a model across multiple, similar datasets while retaining accuracy across all of the data?