Skip to content

Commit 5268aad

Browse files
committed
copy-edit: change 'GMM model'
1 parent 46cfb1c commit 5268aad

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

notebooks/05.12-Gaussian-Mixtures.ipynb

+6-6
Original file line numberDiff line numberDiff line change
@@ -374,7 +374,7 @@
374374
"cell_type": "markdown",
375375
"metadata": {},
376376
"source": [
377-
"With this in place, we can take a look at what the GMM model gives us for our initial data:"
377+
"With this in place, we can take a look at what the four-component GMM gives us for our initial data:"
378378
]
379379
},
380380
{
@@ -561,8 +561,8 @@
561561
"metadata": {},
562562
"source": [
563563
"Here the mixture of 16 Gaussians serves not to find separated clusters of data, but rather to model the overall *distribution* of the input data.\n",
564-
"This is a generative model of the distribution, meaning that the GMM model gives us the recipe to generate new random data distributed similarly to our input.\n",
565-
"For example, here are 400 new points drawn from this 16-component GMM model to our original data:"
564+
"This is a generative model of the distribution, meaning that the GMM gives us the recipe to generate new random data distributed similarly to our input.\n",
565+
"For example, here are 400 new points drawn from this 16-component GMM fit to our original data:"
566566
]
567567
},
568568
{
@@ -604,7 +604,7 @@
604604
"The fact that GMM is a generative model gives us a natural means of determining the optimal number of components for a given dataset.\n",
605605
"A generative model is inherently a probability distribution for the dataset, and so we can simply evaluate the *likelihood* of the data under the model, using cross-validation to avoid over-fitting.\n",
606606
"Another means of correcting for over-fitting is to adjust the model likelihoods using some analytic criterion such as the [Akaike information criterion (AIC)](https://en.wikipedia.org/wiki/Akaike_information_criterion) or the [Bayesian information criterion (BIC)](https://en.wikipedia.org/wiki/Bayesian_information_criterion).\n",
607-
"Scikit-Learn's ``GMM`` model actually includes built-in methods that compute both of these, and so it is very easy to operate on this approach.\n",
607+
"Scikit-Learn's ``GMM`` estimator actually includes built-in methods that compute both of these, and so it is very easy to operate on this approach.\n",
608608
"\n",
609609
"Let's look at the AIC and BIC as a function as the number of GMM components for our moon dataset:"
610610
]
@@ -725,8 +725,8 @@
725725
"cell_type": "markdown",
726726
"metadata": {},
727727
"source": [
728-
"We have nearly 1,800 digits in 64 dimensions, and we can build a GMM model on top of these to generate more.\n",
729-
"GMM can have difficulty converging in such a high dimensional space, so we will start with an invertible dimensionality reduction algorithm on the data.\n",
728+
"We have nearly 1,800 digits in 64 dimensions, and we can build a GMM on top of these to generate more.\n",
729+
"GMMs can have difficulty converging in such a high dimensional space, so we will start with an invertible dimensionality reduction algorithm on the data.\n",
730730
"Here we will use a straightforward PCA, asking it to preserve 99% of the variance in the projected data:"
731731
]
732732
},

0 commit comments

Comments
 (0)