emmytau

emmytau t1_j2r7cbc wrote

Is it problematic that my Bart summarization model's training loss drop below validation loss? I could for example stop the training already after 2 epochs. However, it would be nice to train more epochs but maybe it would just require more data - or do you have any training argument suggestions?

See graph of training- and validation loss https://imgur.com/mF7Frfd

Model here: https://huggingface.co/emmyapi/distilbart-podimo-data-eval-2

2