Submitted by Constant-Cranberry29 t3_11mokqu in deeplearning
neuralbeans t1_jbiu3io wrote
Yes, if the features include the model's target output. Then, the overfitting would result in the model outputting that feature as is. Of course this is a useless solution, but the more similar the features are to the output, the less overfitting will be a problem and the less data you would need to generalise.
BamaDane t1_jbjhitr wrote
I’m not sure I understand what your method does. If Y is the output, then you say I should also include Y as an input? And if I manage to design my model so it doesn’t just select the Y input, then I’m not overfitting? This makes sense that it doesn’t overfit, but doesn’t it also mean I am dumbing-down my model? Don’t I want my model to preferentially select features that are most similar to the output?
neuralbeans t1_jbjizpw wrote
It's a degenerate case, not something anyone should do. If you include Y in your input, then overfitting will lead to the best generalisation. This shows that the input does affect overfitting. In fact, the more similar the input is to the output, the simpler the model can be and thus the less it can overfit.
Constant-Cranberry29 OP t1_jbiutcs wrote
Can you provide a reference that states that feature engineering can address overfitting?
neuralbeans t1_jbixife wrote
Constant-Cranberry29 OP t1_jbixwf1 wrote
I think feature selection and feature engineering are different
neuralbeans t1_jbiygo0 wrote
Well selection is part of engineering, is it not?
Constant-Cranberry29 OP t1_jbiyw8d wrote
because I've read from some paper, they saying FS and FE is different
Viewing a single comment thread. View all comments