Skip to content

Commit b43216e

Browse files
committed
Pushing the docs for revision for branch: master, commit 317dea8a05b0087f83318f318d112976e90566ff
1 parent 882d301 commit b43216e

File tree

800 files changed

+2736
-2742
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

800 files changed

+2736
-2742
lines changed

dev/_downloads/document_clustering.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
Two algorithms are demoed: ordinary k-means and its more scalable cousin
2828
minibatch k-means.
2929
30-
Additionally, latent sematic analysis can also be used to reduce dimensionality
30+
Additionally, latent semantic analysis can also be used to reduce dimensionality
3131
and discover latent patterns in the data.
3232
3333
It can be noted that k-means (and minibatch k-means) are very sensitive to

dev/_downloads/plot_ard.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@
3232
n_samples, n_features = 100, 100
3333
# Create Gaussian data
3434
X = np.random.randn(n_samples, n_features)
35-
# Create weigts with a precision lambda_ of 4.
35+
# Create weights with a precision lambda_ of 4.
3636
lambda_ = 4.
3737
w = np.zeros(n_features)
3838
# Only keep 10 weights of interest

dev/_downloads/plot_bayesian_ridge.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,11 +25,11 @@
2525
from sklearn.linear_model import BayesianRidge, LinearRegression
2626

2727
###############################################################################
28-
# Generating simulated data with Gaussian weigthts
28+
# Generating simulated data with Gaussian weights
2929
np.random.seed(0)
3030
n_samples, n_features = 100, 100
3131
X = np.random.randn(n_samples, n_features) # Create Gaussian data
32-
# Create weigts with a precision lambda_ of 4.
32+
# Create weights with a precision lambda_ of 4.
3333
lambda_ = 4.
3434
w = np.zeros(n_features)
3535
# Only keep 10 weights of interest

dev/_downloads/plot_calibration_curve.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -125,10 +125,10 @@ def plot_calibration_curve(est, name, fig_index):
125125

126126
plt.tight_layout()
127127

128-
# Plot calibration cuve for Gaussian Naive Bayes
128+
# Plot calibration curve for Gaussian Naive Bayes
129129
plot_calibration_curve(GaussianNB(), "Naive Bayes", 1)
130130

131-
# Plot calibration cuve for Linear SVC
131+
# Plot calibration curve for Linear SVC
132132
plot_calibration_curve(LinearSVC(), "SVC", 2)
133133

134134
plt.show()

dev/_downloads/plot_compare_calibration.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,10 +10,10 @@
1010
0.8, approx. 80% actually belong to the positive class.
1111
1212
LogisticRegression returns well calibrated predictions as it directly
13-
optimizes log-loss. In contrast, the other methods return biased probilities,
13+
optimizes log-loss. In contrast, the other methods return biased probabilities,
1414
with different biases per method:
1515
16-
* GaussianNaiveBayes tends to push probabilties to 0 or 1 (note the counts in
16+
* GaussianNaiveBayes tends to push probabilities to 0 or 1 (note the counts in
1717
the histograms). This is mainly because it makes the assumption that features
1818
are conditionally independent given the class, which is not the case in this
1919
dataset which contains 2 redundant features.
@@ -35,7 +35,7 @@
3535
trained with random forests have relatively high variance due to feature
3636
subseting." As a result, the calibration curve shows a characteristic sigmoid
3737
shape, indicating that the classifier could trust its "intuition" more and
38-
return probabilties closer to 0 or 1 typically.
38+
return probabilities closer to 0 or 1 typically.
3939
4040
* Support Vector Classification (SVC) shows an even more sigmoid curve as
4141
the RandomForestClassifier, which is typical for maximum-margin methods

dev/_downloads/plot_compare_cross_decomposition.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -135,7 +135,7 @@
135135
y = X[:, 0] + 2 * X[:, 1] + np.random.normal(size=n * 1) + 5
136136
pls1 = PLSRegression(n_components=3)
137137
pls1.fit(X, y)
138-
# note that the number of compements exceeds 1 (the dimension of y)
138+
# note that the number of components exceeds 1 (the dimension of y)
139139
print("Estimated betas")
140140
print(np.round(pls1.coef_, 1))
141141

dev/_downloads/plot_kernel_approximation.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -169,7 +169,7 @@
169169

170170
X = pca.transform(data_train)
171171

172-
# Gemerate grid along first two principal components
172+
# Generate grid along first two principal components
173173
multiples = np.arange(-2, 2, 0.1)
174174
# steps along first component
175175
first = multiples[:, np.newaxis] * pca.components_[0, :]

dev/_downloads/plot_kmeans_silhouette_analysis.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
a way to assess parameters like number of clusters visually. This measure has a
1010
range of [-1, 1].
1111
12-
Silhoette coefficients (as these values are referred to as) near +1 indicate
12+
Silhouette coefficients (as these values are referred to as) near +1 indicate
1313
that the sample is far away from the neighboring clusters. A value of 0
1414
indicates that the sample is on or very close to the decision boundary between
1515
two neighboring clusters and negative values indicate that those samples might
@@ -43,7 +43,7 @@
4343
print(__doc__)
4444

4545
# Generating the sample data from make_blobs
46-
# This particular setting has one distict cluster and 3 clusters placed close
46+
# This particular setting has one distinct cluster and 3 clusters placed close
4747
# together.
4848
X, y = make_blobs(n_samples=500,
4949
n_features=2,
@@ -110,7 +110,7 @@
110110
ax1.set_xlabel("The silhouette coefficient values")
111111
ax1.set_ylabel("Cluster label")
112112

113-
# The vertical line for average silhoutte score of all the values
113+
# The vertical line for average silhouette score of all the values
114114
ax1.axvline(x=silhouette_avg, color="red", linestyle="--")
115115

116116
ax1.set_yticks([]) # Clear the yaxis labels / ticks

dev/_downloads/plot_learning_curve.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
def plot_learning_curve(estimator, title, X, y, ylim=None, cv=None,
2828
n_jobs=1, train_sizes=np.linspace(.1, 1.0, 5)):
2929
"""
30-
Generate a simple plot of the test and traning learning curve.
30+
Generate a simple plot of the test and training learning curve.
3131
3232
Parameters
3333
----------

dev/_downloads/plot_mlp_alpha.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
in a decision boundary plot that appears with lesser curvatures.
1414
Similarly, decreasing alpha may fix high bias (a sign of underfitting) by
1515
encouraging larger weights, potentially resulting in a more complicated
16-
decision boundery.
16+
decision boundary.
1717
"""
1818
print(__doc__)
1919

0 commit comments

Comments
 (0)