Skip to content

Commit d09396c

Browse files
update
1 parent bc54df9 commit d09396c

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

machine-learning/03-Training and AutoML.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@
3939
"metadata": {},
4040
"source": [
4141
"## Trainers in ML.Net\n",
42-
"ML.NET provides a variety of trainers. You can find most of them under the [StandardTrainersCatalog](https://docs.microsoft.com/dotnet/api/microsoft.ml.standardtrainerscatalog?view=ml-dotnet). Examples of trainers include linear trainers like `SDCA`, `Lbfgs`, `LinearSvm` and tree-based non-linear trainerd like `FastTree`, `RandomForest` and `LightGbm`. Generally, each trainer's capability is different. Non-linear models sometimes have better training performance (lower loss) than linear ones, but it doesn't always mean they are always the better choice. Picking the right trainer to build the best model for your data requires many attempts of trial and error.\n",
42+
"ML.NET provides a variety of trainers. You can find most of them under the [StandardTrainersCatalog](https://docs.microsoft.com/dotnet/api/microsoft.ml.standardtrainerscatalog?view=ml-dotnet). Examples of trainers include linear trainers like `SDCA`, `Lbfgs`, `LinearSvm` and tree-based non-linear trainers like `FastTree`, `RandomForest` and `LightGbm`. Generally, each trainer's capability is different. Non-linear models sometimes have better training performance (lower loss) than linear ones, but it doesn't always mean they are always the better choice. Picking the right trainer to build the best model for your data requires many attempts of trial and error.\n",
4343
"\n",
4444
"### Hyper-parameter optimization\n",
4545
"Other than difference in trainers, different hyper-parameter in one trainer also have a huge impact over the final training performance, especially for tree-base trainers. This is because the capability of these trainers to fit a specific dataset is largly depends on their hyper parameters. For example, larger `numberOfLeaves` in `LightGbm` results to a larger model and usually enable it to fit on a more complex dataset, but it might have countereffect on small dataset and cause overfitting. On the contrary, if the dataset is complex but you set a small `numberOfLeaves`, it might impair `LightGbm`'s ability on fitting that dataset and cause underfit.\n",

0 commit comments

Comments
 (0)