CrazyCrab

CrazyCrab t1_izgcu6i wrote

Ok, so my annotated data consists of about 50 images of size 10000x5000 pixels on average. The task is binary segmentation. Positives constitute approximately 8% of all pixels. 38 images are in the training part, 12 images are in the test part (I divided them randomly).

The batch cross entropy plot and the validation cross entropy plot were crazy unstable during training. After a little bit of training there mostly wasn't any stable trend in either going up or down. However, as the time went on, the best validation cross entropy over all checkpoints went down and went down...

So I think my checkpoint-selecting method gave me a model overfit to the validation dataset. I.e., I expect that on future samples the performance will be more like on the training dataset than on the validation dataset. The only other likely explanation I can think of is that I got unlucky and my validation dataset turned out to be significantly easier than my training dataset.

1

CrazyCrab t1_iuvhe4r wrote

The US has a long term visa, I forgot its name. The idea is that it's for exceptional people, including exceptional scientists. And to prove that you're exceptional, there's a list of like 9 items, of which you need to have 3 or more or something like that. One of them is "has published an article in a serious publication venue". Another is "has been a reviewer of the works of others at a serious venue" or something like that. I wonder if other countries have this.

2