Skip to content

Commit 7317d00

Browse files
Update machine-learning/03-Training and AutoML.ipynb
Co-authored-by: Luis Quintanilla <46974588+luisquintanilla@users.noreply.github.com>
1 parent b08b1a2 commit 7317d00

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

machine-learning/03-Training and AutoML.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@
4646
"\n",
4747
"The process of finding the best configuration for your trainer is known as hyper-parameter optimization (HPO). Like the process of choosing your trainer it involves a lot of trial and error. The built-in Automated ML (AutoML) capabilities in ML.NET simplify the HPO process.\n",
4848
"\n",
49-
"### OverFitting and UnderFitting\n",
49+
"### Overfitting and Underfitting\n",
5050
"Overfitting and underfitting are the two most common problems you encounter when training a model. Underfitting means the selected trainer is not capable enough to fit training dataset and usually result in a high loss during training and low score/metric on test dataset. To resolve this you need to either select a more powerful model or perform more feature engineering. Overfitting is the opposite, which happens when model learns the training data too well. This usually results in low loss metric during training but high loss on test dataset.\n".
5151

5252
A good analogy for these concepts is studying for an exam. Let's say you knew the questions and answers ahead of time. After studying, you take the test and get a perfect score. Great news! However, when you're given the exam again with the questions rearranged and with slightly different wording you get a lower score. That suggests you memorized the answers and didn't actually learn the concepts you were being tested on. This is an example of overfitting. Underfitting is the opposite where the study materials you were given don't accurately represent what you're evaluated on for the exam. As a result, you resort to guessing the answers since you don't have enough knowledge to answer correctly.

0 commit comments

Comments
 (0)