Does VRAM Requirements Fall
Do VRAM requirements fall as the model grows?
I'm sorry if this is a newbie question but I can't understand why one model with the same settings fails while another model is exceptional. I started to wonder if I had my previous model on Central Storage distribution and later switched over, because when I started training my new model it kept crashing until I switched over to central storage. After several thousand "its" of training can I switch back to default distribution?