Overfitting or Underfitting the Training Data

Overfitting means that the model performs well on the training data, but it does not generalize well, but underfitting is the opposite of overfitting. It occurs when your model is too simple to learn the underlying structure of the data.

Overfitting

overfitting occurs when a model learns "too well" from the training data, including noise and outliers. This makes it perform excellently on the training dataset but poorly on new, unseen data.

Indication

High accuracy on training data but low accuracy on validation data or un-seen data.

Cause of overfitting

How to mitigate overfitting:

Underfitting

when a model is too simple to capture the underlying patterns in the data. It fails to represent the training data well.

Indication

Low accuracy on both training and validation data.

How to mitigate underfitting: