Hi Friends,
I have been successfully training deepfake models for a good few months now, and I am taking some time to research and plan the next iteration of my process.
I was reading that the Apple M processors have unified memory, which would be a god send for large resolution and batch training.
However there isn't a lot of information on the subject as NVidia is the king on the subject.
I am training on 4090s, which gives me 20gb ish of headroom for data on both my machines. And while they are really fast compared to everything else I've tested, I wonder if say using an m4 Ultra with 256gb of memory would unlock a much higher resolution. I know it would be a lot slower. But would it work that way?
Would the 256gb of unified memory give me room to train say a 1024x model at a batch of 4 or even 8, even though much slower? Time isn't the problem, I am chasing quality.
Any personal experience or feedback would be great.
Cheers,
A