Is training a function of epochs or iterations? Ie., are these two statements correct?
Given a small sample of A and B images (say, 100), training will be fast... but likely poor.
Given a large sample of A and B images (say, 5000), training will be slow... but likely of much higher quality.
I ask because I am training a model now that has 1500 A images and 4000 B images (all good variance and good quality) but the training is taking forever (Dfl Sae at 256). The last time I trained this model was with 1200 images on each side and it was much further along at this point than I am now. Is the large number of B images in the current model (4000) the reason this model is training slower?