Submitted by YamEnvironmental4720 t3_zaifed in MachineLearning
MrHumanist t1_iymei70 wrote
Groupby -> entropy is used as a feature to estimate uncertainty for a group.
YamEnvironmental4720 OP t1_iymfub5 wrote
Do you mean that uncertainty means almost equality of probabilities? For entropy, we usually group the space into halfspaces separated by a coordinate hyperplane. But any hypersurface, such as the zero level of a function f, also does this. A classifier function f whose zero level hypersurface yields a splitting of the full space that does not significantly reduce the entropy would probably be a bad classifier by other metrics also.
Oceanboi t1_iypx661 wrote
Half spaces, hyper planes, hmm. It seems as though my current understanding of entropy is very limited. Could you link me some relevant material so I can understand what a "zero level hypersurface" is? I only have ever seen simple examples of entropy / gini impurity for splitting random forest so I'm interested in learning more.
YamEnvironmental4720 OP t1_iyrgm9g wrote
I could recommend the lecture on tree classifiers by Nando de Freitas in his course on Machine Learning. It's all on YouTube.
Viewing a single comment thread. View all comments