Accuracy on Validation and Test set, Overfit?

Just a quick question, I am building a ML model right now however I am receiving very similar (72.2 and 72.4 for example)% for both Accuracy and F1-Score on my Validation Dataset and my unseen Test Set respectively. This is occuring on most of the baseline models I have produced for my problem right now.

Is this showing that my model is completely overfitting or just acting completely random and getting lucky.

Thanks

Topic f1score overfitting dataset machine-learning

Category Data Science


If the training set, validation set and the unseen test set (as you put it) have the same score, but lower than you expected then the model has not overfitted.

An overfitted model would have higher scores for the training data at least, and depending on how you optimised the hyper parameters on the validation data but lower for the unseen test set.

The more likely outcome is that your model has underfitted i.e. low but consistent scores across all 3 sets of data.

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.