You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: notebooks/05.04-Feature-Engineering.ipynb
+1-1
Original file line number
Diff line number
Diff line change
@@ -86,7 +86,7 @@
86
86
"It turns out that this is not generally a useful approach in Scikit-Learn: the package's models make the fundamental assumption that numerical features reflect algebraic quanitites.\n",
87
87
"Thus such a mapping would imply, for example, that *Queen Anne < Fremont < Wallingford*, or even that *Wallingford - Queen Anne = Fremont*, which (niche demographic jokes aside) does not make much sense.\n",
88
88
"\n",
89
-
"In this case, one proven technique is to use *Ooe-hot encoding*, which effectively creates extra columns indicating the presence or absence of a category with a value of 1 or 0, respectively.\n",
89
+
"In this case, one proven technique is to use *one-hot encoding*, which effectively creates extra columns indicating the presence or absence of a category with a value of 1 or 0, respectively.\n",
90
90
"When your data comes as a list of dictionaries, Scikit-Learn's ``DictVectorizer`` will do this for you:"
0 commit comments