Submitted by PleaseKillMeNowOkay t3_xtadfd in deeplearning
PleaseKillMeNowOkay OP t1_iqqxw6h wrote
Reply to comment by thebear96 in Neural network that models a probability distribution by PleaseKillMeNowOkay
I wouldn't call it a bigger network necessarily. The second network has two more output neurons compared to the first. Rest are the same. How much difference that makes. Idk
thebear96 t1_iqqykoz wrote
That shouldn't create a lot of difference but yes the performance should be worse than the first network in that case. It's far easier to predict two outputs than four. You can try increasing linear layers and using a slower learning rate to see if the model improves.
PleaseKillMeNowOkay OP t1_iqqz3lp wrote
I could add more linear layers and based on my experiments it would probably help but my intention is to compare my new model with the old one for which I presume the architecture should be as close as possible.
thebear96 t1_iqr04o9 wrote
Ideally it should. In that case you will have a worse performance for the second architecture. When you compare you'll have to say that. But it's pretty expected that the second architecture will not perform as well as the first one, so I'm not sure if there's much use comparing. But it's definitely doable.
Viewing a single comment thread. View all comments