Viewing a single comment thread. View all comments

MrHumanist t1_iymei70 wrote

Groupby -> entropy is used as a feature to estimate uncertainty for a group.

2

YamEnvironmental4720 OP t1_iymfub5 wrote

Do you mean that uncertainty means almost equality of probabilities? For entropy, we usually group the space into halfspaces separated by a coordinate hyperplane. But any hypersurface, such as the zero level of a function f, also does this. A classifier function f whose zero level hypersurface yields a splitting of the full space that does not significantly reduce the entropy would probably be a bad classifier by other metrics also.

1

Oceanboi t1_iypx661 wrote

Half spaces, hyper planes, hmm. It seems as though my current understanding of entropy is very limited. Could you link me some relevant material so I can understand what a "zero level hypersurface" is? I only have ever seen simple examples of entropy / gini impurity for splitting random forest so I'm interested in learning more.

1

YamEnvironmental4720 OP t1_iyrgm9g wrote

I could recommend the lecture on tree classifiers by Nando de Freitas in his course on Machine Learning. It's all on YouTube.

1