Search found 790 matches

by bryanlyon
Fri Dec 02, 2022 5:37 am
Forum: Training Support
Topic: Bicubic vs. Nearest Neighbor
Replies: 4
Views: 1546

Re: Bicubic vs. Nearest Neighbor

by bryanlyon
Fri Nov 25, 2022 10:51 pm
Forum: Training Discussion
Topic: Nan Question I haven't seen yet
Replies: 6
Views: 1376

Re: Nan Question I haven't seen yet

The short answer is "probably not but maybe". One thing that we do with our training process is we don't save out the optimizer weights. The Optimizer is a part of the model, but to save it's weights can triple the size of the model. Because of that, each time we start the model the weight...
by bryanlyon
Sun Nov 20, 2022 8:25 pm
Forum: Training Discussion
Topic: Thoughts on Auto-Training
Replies: 3
Views: 958

Re: Thoughts on Auto-Training

We welcome PRs, as like I said, the mechanics are easy but this is unlikely to be done by us since we don't know how we'd set up the options.

by bryanlyon
Sun Nov 20, 2022 6:37 pm
Forum: Training Discussion
Topic: Training freezes after 1100 iterations
Replies: 2
Views: 892

Re: Training freezes after 1100 iterations

Hmmm, can't be sure, but it looks like a storage access problem. ml.p3.2xlarge doesn't have local storage and it's synchronized to the network. Not sure why the hangup is happening though. Try a much higher save interval, not a lower one. That sets how many iterations between saving out your models.
by bryanlyon
Sun Nov 20, 2022 9:07 am
Forum: Training Discussion
Topic: Thoughts on Auto-Training
Replies: 3
Views: 958

Re: Thoughts on Auto-Training

Yes, this is possible, and the mechanics of it are quite easy. The hard part is the heuristics. When you roll back, how much you roll back, how much you lower LR and all of that are completely unknowable and so we leave it as an exercise to the user.

by bryanlyon
Sat Nov 19, 2022 10:48 am
Forum: General Discussion
Topic: Where are the biases?
Replies: 3
Views: 1459

Re: Where are the biases?

Biases exist intrinsically in dense (Fully connected) layers but are generally not included in convolutional layers because convolutions work in a different way and don't really need them. HOWEVER, "weights" are generally taken to mean all variables used in the model. For example, convolut...
by bryanlyon
Sat Nov 19, 2022 12:42 am
Forum: Extract Discussion
Topic: Is it a good idea to extract from your faces folder to consolidate the alignments file?
Replies: 1
Views: 5521

Re: Is it a good idea to extract from your faces folder to consolidate the alignments file?

No, you don't need to combine multiple extracts for one alignment file. The alignment file is only needed for the convert step, and since each time you extract a video it creates a matching alignment file you don't need to combine them at all anymore.

by bryanlyon
Sat Nov 12, 2022 6:40 pm
Forum: Extract Discussion
Topic: Cleaning Alignments - Error
Replies: 3
Views: 5809

Re: Cleaning Alignments - Error

Check the file at 'C:\Users\chris\Desktop\NIKBIT\spiderman0 (40).png' first. That filename doesn't look like an extracted file to me. If that's not it, I'd suggest starting first with the Tools -> Alignments -> Rename to make sure they're all named with the original name and check if they all match....
by bryanlyon
Sat Nov 12, 2022 6:31 pm
Forum: Training Discussion
Topic: Experiments in pretraining
Replies: 16
Views: 5022

Re: Experiments in pretraining

I tried raising the learning rate and it appears to destroy the model and progress. I'm assuming that is because it is now broadening the learning landscape and "resetting." But again I'm going to let it run for fun as an experiment. Either way it turns out it's going to teach me a good l...
by bryanlyon
Sat Nov 12, 2022 11:48 am
Forum: Training Discussion
Topic: Experiments in pretraining
Replies: 16
Views: 5022

Re: Perfect 0.0 Loss value achieved !

Identity leakage is a subtle thing. It is easier to see in some contexts over others but it's there when you re-use a decoder. If you're training with FFHQ or similar dataset, the recommended method is to train a model with those, then copy the encoder over using the "Load Weights" when cr...
by bryanlyon
Sat Nov 12, 2022 10:01 am
Forum: Training Discussion
Topic: Experiments in pretraining
Replies: 16
Views: 5022

Re: Perfect 0.0 Loss value achieved !

I am not blasting you. Just pointing out Dunning Kruger effect. The simple matter is that I don't have to train to 100 billion to know that FaceSwap can never become an "instant swapper" for the same reason that I know that a car cannot fly if you give it high enough octane fuel. It's not ...
by bryanlyon
Sat Nov 12, 2022 8:39 am
Forum: Training Discussion
Topic: Experiments in pretraining
Replies: 16
Views: 5022

Re: Perfect 0.0 Loss value achieved !

Perfect loss is mathematically impossible. It's not just an infeasible thing, it's complete nonsense. Like claiming that some invention creates more power than it takes in -- It's simply impossible. That said, what you're doing is even worse in many ways. Effectively you're massively overtraining yo...
by bryanlyon
Sat Nov 12, 2022 2:48 am
Forum: Extract Discussion
Topic: Cleaning Alignments - Error
Replies: 3
Views: 5809

Re: Cleaning Alignments - Error

I've never seen this. Are you sure that only images from that fsa are in the folder?

by bryanlyon
Thu Oct 27, 2022 1:21 am
Forum: General Discussion
Topic: PCIe 2.0 and 4.0
Replies: 2
Views: 1321

Re: PCIe 2.0 and 4.0

On a single GPU your link speed is unlikely to be a major bottleneck.

by bryanlyon
Thu Oct 20, 2022 4:00 am
Forum: Training Discussion
Topic: Let me see if I have this right regarding training time...
Replies: 3
Views: 1020

Re: Let me see if I have this right regarding training time...

No. Or more accurately mostly no. The model will train at the same speed if the data is good. That's not the whole story though. If you have a very few images then you'll "overtrain" before the model reaches it's maximum quality. That may show you what looks like high quality faces but it'...
by bryanlyon
Mon Oct 17, 2022 5:07 pm
Forum: Training Discussion
Topic: okay this totally threw me...
Replies: 2
Views: 799

Re: okay this totally threw me...

Check if your GPU is overclocked (factory overclocks too) and remove them. Often overclocking leads to slight instability that you wont notice in a video game but can be catastrophic to machine learning. If that still doesn't fix it, try power cycling your device. Turn it off. Unplug it, and leave i...
by bryanlyon
Fri Oct 14, 2022 5:17 pm
Forum: Training Discussion
Topic: Avoiding NaN Errors
Replies: 14
Views: 2841

Re: Avoiding NaN Errors

Nobody has all the answers regarding NaNs. If you're running into them a lot with a given dataset, then you may want to try lowering the LR, reducing the Batch size, making sure not to use any multipliers, turning off mixed precision, and maybe even changing the loss function. Any or all of these ca...
by bryanlyon
Thu Oct 13, 2022 6:13 pm
Forum: Training Discussion
Topic: [Discussion] Notes on Loss functions
Replies: 13
Views: 31200

Re: [Discussion] Notes on Loss functions

Things outside the mask are just noise. Don't worry about them at all.

by bryanlyon
Sun Oct 02, 2022 11:02 pm
Forum: Extract Discussion
Topic: Where does the eye direction get stored for the training/conversion?
Replies: 2
Views: 1781

Re: Where does the eye direction get stored for the training/conversion?

It's not provided explicitly to the model. It learns to find it by itself from the data.

by bryanlyon
Fri Aug 05, 2022 7:33 pm
Forum: Training Discussion
Topic: Potential VRAM Saving techniques
Replies: 34
Views: 8673

Re: Potential VRAM Saving techniques

Aside, have any of you guys played with one of these?: https://www.nvidia.com/en-us/design-visualization/rtx-a6000/ ...The 48GB VRAM is tempting. I would LOVE to play with 48gb cards, sadly Open Source doesn't pay well enough to justify such large purchases (and nobody has volunteered to let us use...