Skip to content

Commit b98e08e

Browse files
committed
Pushing the docs to dev/ for branch: main, commit 74e23aab06d2456401470c020034142e766b10b4
1 parent d92be47 commit b98e08e

File tree

1,287 files changed

+5209
-5209
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,287 files changed

+5209
-5209
lines changed

dev/.buildinfo

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
# Sphinx build info version 1
22
# This file hashes the configuration used when building these files. When it is not found, a full rebuild will be done.
3-
config: 14fe89ee948c902f97edbd045f4b1a63
3+
config: a28be30aafd91c9eaba1819985c6c865
44
tags: 645f666f9bcd5a90fca523b33c5a78b7
Binary file not shown.

dev/_downloads/10754505339af88c10b8a48127535a4a/plot_quantile_regression.ipynb

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -76,7 +76,7 @@
7676
"cell_type": "markdown",
7777
"metadata": {},
7878
"source": [
79-
"With the heteroscedastic Normal distributed target, we observe that the\nvariance of the noise is increasing when the value of the feature `x` is\nincreasing.\n\nWith the asymmetric Pareto distributed target, we observe that the positive\nresiduals are bounded.\n\nThese types of noisy targets make the estimation via\n:class:`~sklearn.linear_model.LinearRegression` less efficient, i.e. we need\nmore data to get stable results and, in addition, large outliers can have a\nhuge impact on the fitted coefficients. (Stated otherwise: in a setting with\nconstant variance, ordinary least squares estimators converge much faster to\nthe *true* coefficients with increasing sample size.)\n\nIn this asymmetric setting, the median or different quantiles give additional\ninsights. On top of that, median estimation is much more robust to outliers\nand heavy tailed distributions. But note that extreme quantiles are estimated\nby very view data points. 95% quantile are more or less estimated by the 5%\nlargest values and thus also a bit sensitive outliers.\n\nIn the remainder of this tutorial, we will show how\n:class:`~sklearn.linear_model.QuantileRegressor` can be used in practice and\ngive the intuition into the properties of the fitted models. Finally,\nwe will compare the both :class:`~sklearn.linear_model.QuantileRegressor`\nand :class:`~sklearn.linear_model.LinearRegression`.\n\n## Fitting a `QuantileRegressor`\n\nIn this section, we want to estimate the conditional median as well as\na low and high quantile fixed at 5% and 95%, respectively. Thus, we will get\nthree linear models, one for each quantile.\n\nWe will use the quantiles at 5% and 95% to find the outliers in the training\nsample beyond the central 90% interval.\n\n"
79+
"With the heteroscedastic Normal distributed target, we observe that the\nvariance of the noise is increasing when the value of the feature `x` is\nincreasing.\n\nWith the asymmetric Pareto distributed target, we observe that the positive\nresiduals are bounded.\n\nThese types of noisy targets make the estimation via\n:class:`~sklearn.linear_model.LinearRegression` less efficient, i.e. we need\nmore data to get stable results and, in addition, large outliers can have a\nhuge impact on the fitted coefficients. (Stated otherwise: in a setting with\nconstant variance, ordinary least squares estimators converge much faster to\nthe *true* coefficients with increasing sample size.)\n\nIn this asymmetric setting, the median or different quantiles give additional\ninsights. On top of that, median estimation is much more robust to outliers\nand heavy tailed distributions. But note that extreme quantiles are estimated\nby very few data points. 95% quantile are more or less estimated by the 5%\nlargest values and thus also a bit sensitive outliers.\n\nIn the remainder of this tutorial, we will show how\n:class:`~sklearn.linear_model.QuantileRegressor` can be used in practice and\ngive the intuition into the properties of the fitted models. Finally,\nwe will compare the both :class:`~sklearn.linear_model.QuantileRegressor`\nand :class:`~sklearn.linear_model.LinearRegression`.\n\n## Fitting a `QuantileRegressor`\n\nIn this section, we want to estimate the conditional median as well as\na low and high quantile fixed at 5% and 95%, respectively. Thus, we will get\nthree linear models, one for each quantile.\n\nWe will use the quantiles at 5% and 95% to find the outliers in the training\nsample beyond the central 90% interval.\n\n"
8080
]
8181
},
8282
{
@@ -170,7 +170,7 @@
170170
"cell_type": "markdown",
171171
"metadata": {},
172172
"source": [
173-
"On the training set, we see that MAE is lower for\n:class:`~sklearn.linear_model.QuantileRegressor` than\n:class:`~sklearn.linear_model.LinearRegression`. In contrast to that, MSE is\nlower for :class:`~sklearn.linear_model.LinearRegression` than\n:class:`~sklearn.linear_model.QuantileRegressor`. These results confirms that\nMAE is the loss minimized by :class:`~sklearn.linear_model.QuantileRegressor`\nwhile MSE is the loss minimized\n:class:`~sklearn.linear_model.LinearRegression`.\n\nWe can make a similar evaluation but looking at the test error obtained by\ncross-validation.\n\n"
173+
"On the training set, we see that MAE is lower for\n:class:`~sklearn.linear_model.QuantileRegressor` than\n:class:`~sklearn.linear_model.LinearRegression`. In contrast to that, MSE is\nlower for :class:`~sklearn.linear_model.LinearRegression` than\n:class:`~sklearn.linear_model.QuantileRegressor`. These results confirms that\nMAE is the loss minimized by :class:`~sklearn.linear_model.QuantileRegressor`\nwhile MSE is the loss minimized\n:class:`~sklearn.linear_model.LinearRegression`.\n\nWe can make a similar evaluation by looking at the test error obtained by\ncross-validation.\n\n"
174174
]
175175
},
176176
{
Binary file not shown.

dev/_downloads/8a712694e4d011e8f35bfcb1b1b5fc82/plot_quantile_regression.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@
9393
# In this asymmetric setting, the median or different quantiles give additional
9494
# insights. On top of that, median estimation is much more robust to outliers
9595
# and heavy tailed distributions. But note that extreme quantiles are estimated
96-
# by very view data points. 95% quantile are more or less estimated by the 5%
96+
# by very few data points. 95% quantile are more or less estimated by the 5%
9797
# largest values and thus also a bit sensitive outliers.
9898
#
9999
# In the remainder of this tutorial, we will show how
@@ -281,7 +281,7 @@
281281
# while MSE is the loss minimized
282282
# :class:`~sklearn.linear_model.LinearRegression`.
283283
#
284-
# We can make a similar evaluation but looking at the test error obtained by
284+
# We can make a similar evaluation by looking at the test error obtained by
285285
# cross-validation.
286286
from sklearn.model_selection import cross_validate
287287

dev/_downloads/scikit-learn-docs.zip

11.2 KB
Binary file not shown.
-7 Bytes
-168 Bytes
55 Bytes
-248 Bytes

0 commit comments

Comments
 (0)