[D] Is 20 single layer Neural networks equals to a single 20 layer neural network? Submitted by eternal-abyss-77 t3_z9hff3 on December 1, 2022 at 7:14 AM in MachineLearning [removed] 6 comments 0
ML4Bratwurst t1_iygsjef wrote on December 1, 2022 at 7:26 AM #838,839 No because stacking layers is basically what gives neural network their ability to extract high level features 5
bacon_boat t1_iygvgdf wrote on December 1, 2022 at 8:07 AM #838,855 f(f(f(f(x)))) =/= f(x)+f(x)+f(x)+f(x) 9
eternal-abyss-77 OP t1_iygwbbr wrote on December 1, 2022 at 8:20 AM #838,861 Replying to bacon_boat (#838,855) Got it bro, thanks 0
eternal-abyss-77 OP t1_iygx65a wrote on December 1, 2022 at 8:32 AM #838,866 Replying to bacon_boat (#838,855) Bro, but let me ask you one more question, please bear with me. If the result [ f(x)+f(x)+f(x)+f(x) ] >= result [ f(f(f(f(x)))) ] (Result is feature map, features retained or extracted ) Can I conclude that both are same? −2−
bacon_boat t1_iygxmve wrote on December 1, 2022 at 8:39 AM #838,869 Replying to eternal-abyss-77 (#838,866) I think you need to check if you have a specific case in mind. They are obviously not the same in general. 5
Crafty_Primary_2776 t1_iyhy4dp wrote on December 1, 2022 at 3:19 PM #839,841 That’s a good point. Actually slightly change your question leads to the problem of neural network width vs depth. Check these materials. Do Wide and Deep Networks Learn the Same Things? Universal approximation theorem. 2
ML4Bratwurst t1_iygsjef wrote
No because stacking layers is basically what gives neural network their ability to extract high level features