Okay, so let me say this faceswap was probably going to be the first and last one I ever do. I have been training for about 340 hours. I thought that I could end at about 40 - 50 thousand iters, but no. It doesn't look good enough. At first I kept training till 30,000 iterations and then converted it to a video. The lip sync was not bad, but the problem was that it was really blurry. I thought "No problem! I only have 30,000 iterations!". So I trained some more. 40000 iterations - still blurry. In fact, I couldn't even tell they were different. Now I am at 54,000 iteration with not even the slightest improvement. And for that matter, check out this video:
She did 70,000 iterations and got crystal clear results. I only have about 15,000 iterations less then them. But it still look really blurry. And I don't think getting to 70,000 iters will make a big difference looking at the 30,000 video and the 54,000 one. What am I doing wrong?
I don't have the appropriate specs to run faceswap but I managed anyways. Here they are below:
OS: Windows 10 Home 64 Bit Build 2004
Faceswap Mode: CPU
CPU: Intel Core i5 inside 3.20 GHz
RAM: 16 GB
Training Batch Size: 64
Trainer: Original
Conversion Writer: Used Ffmpeg for 30,000 iters, Pillow for 40,000, and OpenCV for 54,000
Conversion Scaling: None
Conversion Mask Type: Extended
Conversion Color Adjustment: Avg-Color for 40,000 iters and 50,000 iters, used Match-Hist for 30,000
Just saying, I am not on a VM.