Strange Training Graph

Want to understand the training process better? Got tips for which model to use and when? This is the place for you


Forum rules

Read the FAQs and search the forum before posting a new topic.

This forum is for discussing tips and understanding the process involved with Training a Faceswap model.

If you have found a bug are having issues with the Training process not working, then you should post in the Training Support forum.

Please mark any answers that fixed your problems so others can find the solutions.

Locked
User avatar
esmkevi
Posts: 2
Joined: Wed Oct 07, 2020 10:57 pm

Strange Training Graph

Post by esmkevi »

I've spent 74hrs training my set and my Loss / Iteration graph looks somewhat strange.

Image

There is a significant drop at around 75k iterations, this correlates with the same time I restarted the training to test the footage. Similarly upon my most resent reactivation my loss jumped insanely.

I can't seem to grasp why restarting the training would make it so much worse when previously it had the opposite effect.

Any ideas?

User avatar
torzdf
Posts: 2681
Joined: Fri Jul 12, 2019 12:53 am
Answers: 159
Has thanked: 133 times
Been thanked: 625 times

Re: Strange Training Graph

Post by torzdf »

When did you start training (what date approx.)?

Which model?

Did you change any of the loss multipliers for any of the sessions?

My word is final

User avatar
esmkevi
Posts: 2
Joined: Wed Oct 07, 2020 10:57 pm

Re: Strange Training Graph

Post by esmkevi »

The training begun on Oct 4th, I'm using the Dfaker model. I haven't changed any settings for the training except for the batch size. That being said, it's irrelevant as those changes occurred only during the first few hrs.

Thank you for your help!

Yup.
User avatar
abigflea
Posts: 182
Joined: Sat Feb 22, 2020 10:59 pm
Answers: 2
Has thanked: 20 times
Been thanked: 62 times

Re: Strange Training Graph

Post by abigflea »

I can't imagine what could cause that other than turning on warp to landmarks, and then turning it back off and it jumped back up

:o I dunno what I'm doing :shock:
2X RTX 3090 : RTX 3080 : RTX: 2060 : 2x RTX 2080 Super : Ghetto 1060

User avatar
torzdf
Posts: 2681
Joined: Fri Jul 12, 2019 12:53 am
Answers: 159
Has thanked: 133 times
Been thanked: 625 times

Re: Strange Training Graph

Post by torzdf »

Ok. my advice would be "don't worry about it too much".....

We don't save optimizer weights in the model (as it will double the size of the model file), so it effectively means the model "starts again" every time you stop and start it. However, as the weights have already been updated from previous sessions, it usually quickly recovers to where it was before.

The reason I asked about the date was there was a commit a while back, where I screwed up the loss calculations, and it made the graph look a bit like yours, but this won't be what has impacted you.

Ultimately Loss (as we use it) isn't really a measure of anything useful. It just shows a direction of travel, so as long as it is trending downwards, you should be fine.

My word is final

Locked