Skip to content

Commit 4ac818d

Browse files
committed
Pushing the docs to dev/ for branch: master, commit 502261bd9deca6ba57b4a051da2b37f170f28e88
1 parent 85a0495 commit 4ac818d

File tree

1,447 files changed

+2939
-2941
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,447 files changed

+2939
-2941
lines changed
-7 Bytes
Binary file not shown.
-7 Bytes
Binary file not shown.

dev/_downloads/plot_mlp_training_curves.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626
},
2727
"outputs": [],
2828
"source": [
29-
"print(__doc__)\nimport matplotlib.pyplot as plt\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn import datasets\n\n# different learning rate schedules and momentum parameters\nparams = [{'solver': 'sgd', 'learning_rate': 'constant', 'momentum': 0,\n 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'constant', 'momentum': .9,\n 'nesterovs_momentum': False, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'constant', 'momentum': .9,\n 'nesterovs_momentum': True, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': 0,\n 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': .9,\n 'nesterovs_momentum': True, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': .9,\n 'nesterovs_momentum': False, 'learning_rate_init': 0.2},\n {'solver': 'adam', 'learning_rate_init': 0.01}]\n\nlabels = [\"constant learning-rate\", \"constant with momentum\",\n \"constant with Nesterov's momentum\",\n \"inv-scaling learning-rate\", \"inv-scaling with momentum\",\n \"inv-scaling with Nesterov's momentum\", \"adam\"]\n\nplot_args = [{'c': 'red', 'linestyle': '-'},\n {'c': 'green', 'linestyle': '-'},\n {'c': 'blue', 'linestyle': '-'},\n {'c': 'red', 'linestyle': '--'},\n {'c': 'green', 'linestyle': '--'},\n {'c': 'blue', 'linestyle': '--'},\n {'c': 'black', 'linestyle': '-'}]\n\n\ndef plot_on_dataset(X, y, ax, name):\n # for each dataset, plot learning for each learning strategy\n print(\"\\nlearning on dataset %s\" % name)\n ax.set_title(name)\n X = MinMaxScaler().fit_transform(X)\n mlps = []\n if name == \"digits\":\n # digits is larger but converges fairly quickly\n max_iter = 15\n else:\n max_iter = 400\n\n for label, param in zip(labels, params):\n print(\"training: %s\" % label)\n mlp = MLPClassifier(verbose=0, random_state=0,\n max_iter=max_iter, **param)\n mlp.fit(X, y)\n mlps.append(mlp)\n print(\"Training set score: %f\" % mlp.score(X, y))\n print(\"Training set loss: %f\" % mlp.loss_)\n for mlp, label, args in zip(mlps, labels, plot_args):\n ax.plot(mlp.loss_curve_, label=label, **args)\n\n\nfig, axes = plt.subplots(2, 2, figsize=(15, 10))\n# load / generate some toy datasets\niris = datasets.load_iris()\ndigits = datasets.load_digits()\ndata_sets = [(iris.data, iris.target),\n (digits.data, digits.target),\n datasets.make_circles(noise=0.2, factor=0.5, random_state=1),\n datasets.make_moons(noise=0.3, random_state=0)]\n\nfor ax, data, name in zip(axes.ravel(), data_sets, ['iris', 'digits',\n 'circles', 'moons']):\n plot_on_dataset(*data, ax=ax, name=name)\n\nfig.legend(ax.get_lines(), labels=labels, ncol=3, loc=\"upper center\")\nplt.show()"
29+
"print(__doc__)\nimport matplotlib.pyplot as plt\nfrom sklearn.neural_network import MLPClassifier\nfrom sklearn.preprocessing import MinMaxScaler\nfrom sklearn import datasets\n\n# different learning rate schedules and momentum parameters\nparams = [{'solver': 'sgd', 'learning_rate': 'constant', 'momentum': 0,\n 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'constant', 'momentum': .9,\n 'nesterovs_momentum': False, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'constant', 'momentum': .9,\n 'nesterovs_momentum': True, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': 0,\n 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': .9,\n 'nesterovs_momentum': True, 'learning_rate_init': 0.2},\n {'solver': 'sgd', 'learning_rate': 'invscaling', 'momentum': .9,\n 'nesterovs_momentum': False, 'learning_rate_init': 0.2},\n {'solver': 'adam', 'learning_rate_init': 0.01}]\n\nlabels = [\"constant learning-rate\", \"constant with momentum\",\n \"constant with Nesterov's momentum\",\n \"inv-scaling learning-rate\", \"inv-scaling with momentum\",\n \"inv-scaling with Nesterov's momentum\", \"adam\"]\n\nplot_args = [{'c': 'red', 'linestyle': '-'},\n {'c': 'green', 'linestyle': '-'},\n {'c': 'blue', 'linestyle': '-'},\n {'c': 'red', 'linestyle': '--'},\n {'c': 'green', 'linestyle': '--'},\n {'c': 'blue', 'linestyle': '--'},\n {'c': 'black', 'linestyle': '-'}]\n\n\ndef plot_on_dataset(X, y, ax, name):\n # for each dataset, plot learning for each learning strategy\n print(\"\\nlearning on dataset %s\" % name)\n ax.set_title(name)\n X = MinMaxScaler().fit_transform(X)\n mlps = []\n if name == \"digits\":\n # digits is larger but converges fairly quickly\n max_iter = 15\n else:\n max_iter = 400\n\n for label, param in zip(labels, params):\n print(\"training: %s\" % label)\n mlp = MLPClassifier(verbose=0, random_state=0,\n max_iter=max_iter, **param)\n mlp.fit(X, y)\n mlps.append(mlp)\n print(\"Training set score: %f\" % mlp.score(X, y))\n print(\"Training set loss: %f\" % mlp.loss_)\n for mlp, label, args in zip(mlps, labels, plot_args):\n ax.plot(mlp.loss_curve_, label=label, **args)\n\n\nfig, axes = plt.subplots(2, 2, figsize=(15, 10))\n# load / generate some toy datasets\niris = datasets.load_iris()\ndigits = datasets.load_digits()\ndata_sets = [(iris.data, iris.target),\n (digits.data, digits.target),\n datasets.make_circles(noise=0.2, factor=0.5, random_state=1),\n datasets.make_moons(noise=0.3, random_state=0)]\n\nfor ax, data, name in zip(axes.ravel(), data_sets, ['iris', 'digits',\n 'circles', 'moons']):\n plot_on_dataset(*data, ax=ax, name=name)\n\nfig.legend(ax.get_lines(), labels, ncol=3, loc=\"upper center\")\nplt.show()"
3030
]
3131
}
3232
],

dev/_downloads/plot_mlp_training_curves.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,5 +85,5 @@ def plot_on_dataset(X, y, ax, name):
8585
'circles', 'moons']):
8686
plot_on_dataset(*data, ax=ax, name=name)
8787

88-
fig.legend(ax.get_lines(), labels=labels, ncol=3, loc="upper center")
88+
fig.legend(ax.get_lines(), labels, ncol=3, loc="upper center")
8989
plt.show()

dev/_downloads/scikit-learn-docs.pdf

-28.8 KB
Binary file not shown.

0 commit comments

Comments
 (0)