Skip to content

Commit 9d59e25

Browse files
committed
Pushing the docs to dev/ for branch: main, commit 0f8a7775ad248b9aa4be63291ae71d9212a46e6c
1 parent 9647edd commit 9d59e25

File tree

1,317 files changed

+7211
-7207
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,317 files changed

+7211
-7207
lines changed

dev/.buildinfo

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
# Sphinx build info version 1
22
# This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done.
3-
config: 10004e4ad06b5587b4adde4c3a2e0879
3+
config: a65a6fe570c78743acc0f5705db216d7
44
tags: 645f666f9bcd5a90fca523b33c5a78b7
Binary file not shown.

dev/_downloads/2f3ef774a6d7e52e1e6b7ccbb75d25f0/plot_gradient_boosting_quantile.py

+1
Original file line numberDiff line numberDiff line change
@@ -191,6 +191,7 @@ def highlight_min(x):
191191
# outliers and overfits less.
192192
#
193193
# .. _calibration-section:
194+
#
194195
# Calibration of the confidence interval
195196
# --------------------------------------
196197
#
Binary file not shown.

dev/_downloads/8452fc8dfe9850cfdaa1b758e5a2748b/plot_gradient_boosting_early_stopping.ipynb

+2-2
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@
6565
"cell_type": "markdown",
6666
"metadata": {},
6767
"source": [
68-
"## Visualize Comparision\nIt includes three subplots:\n1. Plotting training errors of both models over boosting iterations.\n2. Plotting validation errors of both models over boosting iterations.\n3. Creating a bar chart to compare the training times and the estimator used\nof the models with and without early stopping.\n\n"
68+
"## Visualize Comparison\nIt includes three subplots:\n\n1. Plotting training errors of both models over boosting iterations.\n2. Plotting validation errors of both models over boosting iterations.\n3. Creating a bar chart to compare the training times and the estimator used\n of the models with and without early stopping.\n\n\n"
6969
]
7070
},
7171
{
@@ -90,7 +90,7 @@
9090
"cell_type": "markdown",
9191
"metadata": {},
9292
"source": [
93-
"## Summary\nIn our example with the :class:`~sklearn.ensemble.GradientBoostingRegressor`\nmodel on the California Housing Prices dataset, we have demonstrated the\npractical benefits of early stopping:\n\n- **Preventing Overfitting:** We showed how the validation error stabilizes\nor starts to increase after a certain point, indicating that the model\ngeneralizes better to unseen data. This is achieved by stopping the training\nprocess before overfitting occurs.\n\n- **Improving Training Efficiency:** We compared training times between\nmodels with and without early stopping. The model with early stopping\nachieved comparable accuracy while requiring significantly fewer\nestimators, resulting in faster training.\n\n"
93+
"## Summary\nIn our example with the :class:`~sklearn.ensemble.GradientBoostingRegressor`\nmodel on the California Housing Prices dataset, we have demonstrated the\npractical benefits of early stopping:\n\n- **Preventing Overfitting:** We showed how the validation error stabilizes\n or starts to increase after a certain point, indicating that the model\n generalizes better to unseen data. This is achieved by stopping the training\n process before overfitting occurs.\n- **Improving Training Efficiency:** We compared training times between\n models with and without early stopping. The model with early stopping\n achieved comparable accuracy while requiring significantly fewer\n estimators, resulting in faster training.\n\n"
9494
]
9595
}
9696
],

dev/_downloads/b5ac5dfd67b0aab146fcb9faaac8480c/plot_gradient_boosting_quantile.ipynb

+1-1
Original file line numberDiff line numberDiff line change
@@ -173,7 +173,7 @@
173173
"cell_type": "markdown",
174174
"metadata": {},
175175
"source": [
176-
"Errors are higher meaning the models slightly overfitted the data. It still\nshows that the best test metric is obtained when the model is trained by\nminimizing this same metric.\n\nNote that the conditional median estimator is competitive with the squared\nerror estimator in terms of MSE on the test set: this can be explained by\nthe fact the squared error estimator is very sensitive to large outliers\nwhich can cause significant overfitting. This can be seen on the right hand\nside of the previous plot. The conditional median estimator is biased\n(underestimation for this asymmetric noise) but is also naturally robust to\noutliers and overfits less.\n\nCalibration of the confidence interval\n--------------------------------------\n\nWe can also evaluate the ability of the two extreme quantile estimators at\nproducing a well-calibrated conditional 90%-confidence interval.\n\nTo do this we can compute the fraction of observations that fall between the\npredictions:\n\n"
176+
"Errors are higher meaning the models slightly overfitted the data. It still\nshows that the best test metric is obtained when the model is trained by\nminimizing this same metric.\n\nNote that the conditional median estimator is competitive with the squared\nerror estimator in terms of MSE on the test set: this can be explained by\nthe fact the squared error estimator is very sensitive to large outliers\nwhich can cause significant overfitting. This can be seen on the right hand\nside of the previous plot. The conditional median estimator is biased\n(underestimation for this asymmetric noise) but is also naturally robust to\noutliers and overfits less.\n\n\n## Calibration of the confidence interval\n\nWe can also evaluate the ability of the two extreme quantile estimators at\nproducing a well-calibrated conditional 90%-confidence interval.\n\nTo do this we can compute the fraction of observations that fall between the\npredictions:\n\n"
177177
]
178178
},
179179
{

dev/_downloads/be911e971b87fe80b6899069dbcfb737/plot_gradient_boosting_early_stopping.py

+11-10
Original file line numberDiff line numberDiff line change
@@ -112,13 +112,15 @@
112112
val_errors_with.append(mean_squared_error(y_val, val_pred))
113113

114114
# %%
115-
# Visualize Comparision
116-
# ---------------------
115+
# Visualize Comparison
116+
# --------------------
117117
# It includes three subplots:
118+
#
118119
# 1. Plotting training errors of both models over boosting iterations.
119120
# 2. Plotting validation errors of both models over boosting iterations.
120121
# 3. Creating a bar chart to compare the training times and the estimator used
121-
# of the models with and without early stopping.
122+
# of the models with and without early stopping.
123+
#
122124

123125
fig, axes = plt.subplots(ncols=3, figsize=(12, 4))
124126

@@ -170,11 +172,10 @@
170172
# practical benefits of early stopping:
171173
#
172174
# - **Preventing Overfitting:** We showed how the validation error stabilizes
173-
# or starts to increase after a certain point, indicating that the model
174-
# generalizes better to unseen data. This is achieved by stopping the training
175-
# process before overfitting occurs.
176-
#
175+
# or starts to increase after a certain point, indicating that the model
176+
# generalizes better to unseen data. This is achieved by stopping the training
177+
# process before overfitting occurs.
177178
# - **Improving Training Efficiency:** We compared training times between
178-
# models with and without early stopping. The model with early stopping
179-
# achieved comparable accuracy while requiring significantly fewer
180-
# estimators, resulting in faster training.
179+
# models with and without early stopping. The model with early stopping
180+
# achieved comparable accuracy while requiring significantly fewer
181+
# estimators, resulting in faster training.

0 commit comments

Comments
 (0)