Code: Select all
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Added defaults: model.unbalanced
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Adding defaults: (filename: villain_defaults.py, module_path: plugins.train.model, plugin_type: model
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Importing defaults module: plugins.train.model.villain_defaults
02/15/2021 11:03:46 MainProcess _training_0 config add_section DEBUG Add section: (title: 'model.villain', info: 'A Higher resolution version of the Original Model by VillainGuy.\nExtremely VRAM heavy. Don't try to run this if you have a small GPU.\n\nNB: Unless specifically stated, values changed here will only take effect when creating a new model.')
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'model.villain', title: 'lowmem', datatype: '<class 'bool'>', default: 'False', info: 'Lower memory mode. Set to 'True' if having issues with VRAM useage.\nNB: Models with a changed lowmem mode are not compatible with each other.', rounding: 'None', min_max: None, choices: [], gui_radio: False, fixed: True, group: settings)
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Added defaults: model.villain
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Adding defaults: (filename: original_defaults.py, module_path: plugins.train.trainer, plugin_type: trainer
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Importing defaults module: plugins.train.trainer.original_defaults
02/15/2021 11:03:46 MainProcess _training_0 config add_section DEBUG Add section: (title: 'trainer.original', info: 'Original Trainer Options.\nWARNING: The defaults for augmentation will be fine for 99.9% of use cases. Only change them if you absolutely know what you are doing!')
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'preview_images', datatype: '<class 'int'>', default: '14', info: 'Number of sample faces to display for each side in the preview when training.', rounding: '2', min_max: (2, 16), choices: None, gui_radio: False, fixed: True, group: evaluation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'zoom_amount', datatype: '<class 'int'>', default: '5', info: 'Percentage amount to randomly zoom each training image in and out.', rounding: '1', min_max: (0, 25), choices: None, gui_radio: False, fixed: True, group: image augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'rotation_range', datatype: '<class 'int'>', default: '10', info: 'Percentage amount to randomly rotate each training image.', rounding: '1', min_max: (0, 25), choices: None, gui_radio: False, fixed: True, group: image augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'shift_range', datatype: '<class 'int'>', default: '5', info: 'Percentage amount to randomly shift each training image horizontally and vertically.', rounding: '1', min_max: (0, 25), choices: None, gui_radio: False, fixed: True, group: image augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'flip_chance', datatype: '<class 'int'>', default: '50', info: 'Percentage chance to randomly flip each training image horizontally.\nNB: This is ignored if the 'no-flip' option is enabled', rounding: '1', min_max: (0, 75), choices: None, gui_radio: False, fixed: True, group: image augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'color_lightness', datatype: '<class 'int'>', default: '30', info: 'Percentage amount to randomly alter the lightness of each training image.\nNB: This is ignored if the 'no-flip' option is enabled', rounding: '1', min_max: (0, 75), choices: None, gui_radio: False, fixed: True, group: color augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'color_ab', datatype: '<class 'int'>', default: '8', info: 'Percentage amount to randomly alter the 'a' and 'b' colors of the L*a*b* color space of each training image.\nNB: This is ignored if the 'no-flip' option is enabled', rounding: '1', min_max: (0, 50), choices: None, gui_radio: False, fixed: True, group: color augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'color_clahe_chance', datatype: '<class 'int'>', default: '50', info: 'Percentage chance to perform Contrast Limited Adaptive Histogram Equalization on each training image.\nNB: This is ignored if the 'no-augment-color' option is enabled', rounding: '1', min_max: (0, 75), choices: None, gui_radio: False, fixed: False, group: color augmentation)
02/15/2021 11:03:46 MainProcess _training_0 config add_item DEBUG Add item: (section: 'trainer.original', title: 'color_clahe_max_size', datatype: '<class 'int'>', default: '4', info: 'The grid size dictates how much Contrast Limited Adaptive Histogram Equalization is performed on any training image selected for clahe. Contrast will be applied randomly with a gridsize of 0 up to the maximum. This value is a multiplier calculated from the training image size.\nNB: This is ignored if the 'no-augment-color' option is enabled', rounding: '1', min_max: (1, 8), choices: None, gui_radio: False, fixed: True, group: color augmentation)
02/15/2021 11:03:46 MainProcess _training_0 _config load_module DEBUG Added defaults: trainer.original
02/15/2021 11:03:46 MainProcess _training_0 config handle_config DEBUG Handling config
02/15/2021 11:03:46 MainProcess _training_0 config check_exists DEBUG Config file exists: 'C:\Users\USER\faceswap\config\train.ini'
02/15/2021 11:03:46 MainProcess _training_0 config load_config VERBOSE Loading config: 'C:\Users\USER\faceswap\config\train.ini'
02/15/2021 11:03:46 MainProcess _training_0 config validate_config DEBUG Validating config
02/15/2021 11:03:46 MainProcess _training_0 config check_config_change DEBUG Default config has not changed
02/15/2021 11:03:46 MainProcess _training_0 config check_config_choices DEBUG Checking config choices
02/15/2021 11:03:46 MainProcess _training_0 config check_config_choices DEBUG Checked config choices
02/15/2021 11:03:46 MainProcess _training_0 config validate_config DEBUG Validated config
02/15/2021 11:03:46 MainProcess _training_0 config handle_config DEBUG Handled config
02/15/2021 11:03:46 MainProcess _training_0 config __init__ DEBUG Initialized: Config
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'centering')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'str'>, value: face)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'coverage')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'float'>, value: 95.0)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'icnr_init')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: False)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'conv_aware_init')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: True)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'optimizer')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'str'>, value: adam)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'learning_rate')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'float'>, value: 5e-05)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'reflect_padding')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: True)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'allow_growth')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: False)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'mixed_precision')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: True)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global', option: 'convert_batchsize')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 16)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'loss_function')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'str'>, value: ssim)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'mask_loss_function')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'str'>, value: mse)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'l2_reg_term')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 100)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'eye_multiplier')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 3)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'mouth_multiplier')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 2)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'penalized_mask_loss')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: True)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'mask_type')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'str'>, value: extended)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'mask_blur_kernel')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 3)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'mask_threshold')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 4)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'global.loss', option: 'learn_mask')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'bool'>, value: False)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'preview_images')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 14)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'zoom_amount')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 5)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'rotation_range')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 10)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'shift_range')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 5)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'flip_chance')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 50)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'color_lightness')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 30)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'color_ab')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 8)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'color_clahe_chance')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 50)
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Getting config item: (section: 'trainer.original', option: 'color_clahe_max_size')
02/15/2021 11:03:46 MainProcess _training_0 config get DEBUG Returning item: (type: <class 'int'>, value: 4)
02/15/2021 11:03:46 MainProcess _training_0 _base add_session_batchsize DEBUG Adding session batch size: 13
02/15/2021 11:03:46 MainProcess _training_0 _base __init__ DEBUG Initializing _TrainingAlignments: (model: <plugins.train.model.dfl_sae.Model object at 0x000002124BDE5DC0>, image counts: {'a': 3497, 'b': 1513})
02/15/2021 11:03:46 MainProcess _training_0 _base _load_detected_faces DEBUG side: a, file count: 3497
02/15/2021 11:03:46 MainProcess _training_0 image read_image_meta_batch DEBUG Submitting 3497 items to executor
02/15/2021 11:03:46 MainProcess _training_0 image read_image_meta_batch DEBUG Succesfully submitted 3497 items to executor
02/15/2021 11:03:46 MainProcess _training_0 multithreading run DEBUG Error in thread (_training_0): 'width'
02/15/2021 11:03:47 MainProcess MainThread train _monitor DEBUG Thread error detected
02/15/2021 11:03:47 MainProcess MainThread train _monitor DEBUG Closed Monitor
02/15/2021 11:03:47 MainProcess MainThread train _end_thread DEBUG Ending Training thread
02/15/2021 11:03:47 MainProcess MainThread train _end_thread CRITICAL Error caught! Exiting...
02/15/2021 11:03:47 MainProcess MainThread multithreading join DEBUG Joining Threads: '_training'
02/15/2021 11:03:47 MainProcess MainThread multithreading join DEBUG Joining Thread: '_training_0'
02/15/2021 11:03:47 MainProcess MainThread multithreading join ERROR Caught exception in thread: '_training_0'
Traceback (most recent call last):
File "C:\Users\USER\faceswap\lib\cli\launcher.py", line 182, in execute_script
process.process()
File "C:\Users\USER\faceswap\scripts\train.py", line 170, in process
self._end_thread(thread, err)
File "C:\Users\USER\faceswap\scripts\train.py", line 210, in _end_thread
thread.join()
File "C:\Users\USER\faceswap\lib\multithreading.py", line 121, in join
raise thread.err[1].with_traceback(thread.err[2])
File "C:\Users\USER\faceswap\lib\multithreading.py", line 37, in run
self._target(*self._args, **self._kwargs)
File "C:\Users\USER\faceswap\scripts\train.py", line 232, in _training
raise err
File "C:\Users\USER\faceswap\scripts\train.py", line 221, in _training
trainer = self._load_trainer(model)
File "C:\Users\USER\faceswap\scripts\train.py", line 267, in _load_trainer
trainer = trainer(model,
File "C:\Users\USER\faceswap\plugins\train\trainer\original.py", line 10, in __init__
super().__init__(*args, **kwargs)
File "C:\Users\USER\faceswap\plugins\train\trainer\_base.py", line 83, in __init__
alignment_data = self._get_alignments_data()
File "C:\Users\USER\faceswap\plugins\train\trainer\_base.py", line 143, in _get_alignments_data
alignments = _TrainingAlignments(self._model, self._images)
File "C:\Users\USER\faceswap\plugins\train\trainer\_base.py", line 1090, in __init__
self._detected_faces = self._load_detected_faces(image_list)
File "C:\Users\USER\faceswap\plugins\train\trainer\_base.py", line 1140, in _load_detected_faces
self._validate_image_size(side, filename, meta["width"], meta["height"])
KeyError: 'width'
============ System Information ============
encoding: cp1252
git_branch: master
git_commits: fcf6d82 Bugfixes: - lib.image Don't raise error if legacy non-png is found when reading header data - plugins.train.trainer._base - Correctly pass legacy alignments through to DetectedFace. b1cfbe4 Update extracted faces to use PNG EXIF data (#1123). a84462f Training - Move disable warp option out of config. ec3fb8b gui - bugfix: Clear recent files on data corruption. 57b29b0 preview tool: bugfix ttk.PanedWindow
gpu_cuda: No global version found. Check Conda packages for Conda Cuda
gpu_cudnn: No global version found. Check Conda packages for Conda cuDNN
gpu_devices: GPU_0: GeForce RTX 2080 Ti
gpu_devices_active: GPU_0
gpu_driver: 457.30
gpu_vram: GPU_0: 11264MB
os_machine: AMD64
os_platform: Windows-10-10.0.19041-SP0
os_release: 10
py_command: C:\Users\USER\faceswap\faceswap.py train -A F:/FSWAP/K6N/Source/Training Set -ala F:/FSWAP/K6N/Source/Training Set/alignments.fsa -B F:/FSWAP/New K/Experimental Training Set -alb F:/FSWAP/New K/Experimental Training Set/alignments.fsa -m F:/FSWAP/K6N/Model DFL SAE DF -t dfl-sae -bs 13 -it 1000000 -s 500 -ss 25000 -ps 75 -nw -L INFO -gui
py_conda_version: conda 4.9.2
py_implementation: CPython
py_version: 3.8.5
py_virtual_env: True
sys_cores: 8
sys_processor: Intel64 Family 6 Model 62 Stepping 4, GenuineIntel
sys_ram: Total: 32712MB, Available: 27525MB, Used: 5186MB, Free: 27525MB
=============== Pip Packages ===============
============== Conda Packages ==============
# packages in environment at C:\Users\USER\MiniConda3\envs\394400:
#
# Name Version Build Channel
_tflow_select 2.3.0 gpu
absl-py 0.11.0 pyhd3eb1b0_1
aiohttp 3.7.3 py38h2bbff1b_1
astunparse 1.6.3 py_0
async-timeout 3.0.1 py38_0
attrs 20.3.0 pyhd3eb1b0_0
blas 1.0 mkl
blinker 1.4 py38_0
brotlipy 0.7.0 py38h2bbff1b_1003
ca-certificates 2020.12.8 haa95532_0
cachetools 4.2.0 pyhd3eb1b0_0
certifi 2020.12.5 py38haa95532_0
cffi 1.14.4 py38hcd4344a_0
chardet 3.0.4 py38haa95532_1003
click 7.1.2 py_0
cryptography 2.9.2 py38h7a1dbc1_0
cudatoolkit 10.1.243 h74a9793_0
cudnn 7.6.5 cuda10.1_0
cycler 0.10.0 py38_0
fastcluster 1.1.26 py38h251f6bf_2 conda-forge
ffmpeg 4.3.1 ha925a31_0 conda-forge
ffmpy 0.2.3 pypi_0 pypi
freetype 2.10.4 hd328e21_0
gast 0.4.0 py_0
git 2.23.0 h6bb4b03_0
google-auth 1.24.0 pyhd3eb1b0_0
google-auth-oauthlib 0.4.2 pyhd3eb1b0_2
google-pasta 0.2.0 py_0
grpcio 1.31.0 py38he7da953_0
h5py 2.10.0 py38h5e291fa_0
hdf5 1.10.4 h7ebc959_0
icc_rt 2019.0.0 h0cc432a_1
icu 58.2 ha925a31_3
idna 2.10 py_0
imageio 2.9.0 py_0
imageio-ffmpeg 0.4.2 py_0 conda-forge
importlib-metadata 2.0.0 py_1
intel-openmp 2020.2 254
joblib 1.0.0 pyhd3eb1b0_0
jpeg 9b hb83a4c4_2
keras-applications 1.0.8 py_1
keras-preprocessing 1.1.0 py_1
kiwisolver 1.3.0 py38hd77b12b_0
libpng 1.6.37 h2a8f88b_0
libprotobuf 3.13.0.1 h200bbdf_0
libtiff 4.1.0 h56a325e_1
lz4-c 1.9.2 hf4a77e7_3
markdown 3.3.3 py38haa95532_0
matplotlib 3.2.2 0
matplotlib-base 3.2.2 py38h64f37c6_0
mkl 2020.2 256
mkl-service 2.3.0 py38h196d8e1_0
mkl_fft 1.2.0 py38h45dec08_0
mkl_random 1.1.1 py38h47e9c7a_0
multidict 4.7.6 py38he774522_1
numpy 1.19.2 py38hadc3359_0
numpy-base 1.19.2 py38ha3acd2a_0
nvidia-ml-py3 7.352.1 pypi_0 pypi
oauthlib 3.1.0 py_0
olefile 0.46 py_0
opencv-python 4.4.0.46 pypi_0 pypi
openssl 1.1.1i h2bbff1b_0
opt_einsum 3.1.0 py_0
pathlib 1.0.1 py_1
pillow 8.0.1 py38h4fa10fc_0
pip 20.3.3 py38haa95532_0
protobuf 3.13.0.1 py38ha925a31_1
psutil 5.7.2 py38he774522_0
pyasn1 0.4.8 py_0
pyasn1-modules 0.2.8 py_0
pycparser 2.20 py_2
pyjwt 2.0.0 py38haa95532_0
pyopenssl 20.0.1 pyhd3eb1b0_1
pyparsing 2.4.7 py_0
pyqt 5.9.2 py38ha925a31_4
pyreadline 2.1 py38_1
pysocks 1.7.1 py38haa95532_0
python 3.8.5 h5fd99cc_1
python-dateutil 2.8.1 py_0
python_abi 3.8 1_cp38 conda-forge
pywin32 227 py38he774522_1
qt 5.9.7 vc14h73c81de_0
requests 2.25.1 pyhd3eb1b0_0
requests-oauthlib 1.3.0 py_0
rsa 4.6 py_0
scikit-learn 0.23.2 py38h47e9c7a_0
scipy 1.5.2 py38h14eb087_0
setuptools 51.0.0 py38haa95532_2
sip 4.19.13 py38ha925a31_0
six 1.15.0 py38haa95532_0
sqlite 3.33.0 h2a8f88b_0
tensorboard 2.3.0 pyh4dce500_0
tensorboard-plugin-wit 1.6.0 py_0
tensorflow 2.3.0 mkl_py38h1fcfbd6_0
tensorflow-base 2.3.0 gpu_py38h7339f5a_0
tensorflow-estimator 2.3.0 pyheb71bc4_0
tensorflow-gpu 2.3.0 he13fc11_0
termcolor 1.1.0 py38_1
threadpoolctl 2.1.0 pyh5ca1d4c_0
tk 8.6.10 he774522_0
tornado 6.1 py38h2bbff1b_0
tqdm 4.54.1 pyhd3eb1b0_0
typing-extensions 3.7.4.3 0
typing_extensions 3.7.4.3 py_0
urllib3 1.26.2 pyhd3eb1b0_0
vc 14.2 h21ff451_1
vs2015_runtime 14.27.29016 h5e58377_2
werkzeug 1.0.1 py_0
wheel 0.36.2 pyhd3eb1b0_0
win_inet_pton 1.1.0 py38haa95532_0
wincertstore 0.2 py38_0
wrapt 1.12.1 py38he774522_1
xz 5.2.5 h62dcd97_0
yarl 1.5.1 py38he774522_0
zipp 3.4.0 pyhd3eb1b0_0
zlib 1.2.11 h62dcd97_4
zstd 1.4.5 h04227a9_0
================= Configs ==================
--------- .faceswap ---------
backend: nvidia
--------- convert.ini ---------
[color.color_transfer]
clip: True
preserve_paper: True
[color.manual_balance]
colorspace: HSV
balance_1: 0.0
balance_2: 0.0
balance_3: 0.0
contrast: 0.0
brightness: 0.0
[color.match_hist]
threshold: 99.0
[mask.box_blend]
type: gaussian
distance: 11.0
radius: 5.0
passes: 1
[mask.mask_blend]
type: gaussian
kernel_size: 3
passes: 4
threshold: 4
erosion: 0.0
[scaling.sharpen]
method: none
amount: 150
radius: 0.3
threshold: 5.0
[writer.ffmpeg]
container: mp4
codec: libx264
crf: 23
preset: medium
tune: None
profile: auto
level: auto
skip_mux: False
[writer.gif]
fps: 25
loop: 0
palettesize: 256
subrectangles: False
[writer.opencv]
format: jpg
draw_transparent: False
jpg_quality: 75
png_compress_level: 3
[writer.pillow]
format: jpg
draw_transparent: False
optimize: False
gif_interlace: True
jpg_quality: 75
png_compress_level: 3
tif_compression: tiff_deflate
--------- extract.ini ---------
[global]
allow_growth: False
[align.fan]
batch-size: 12
[detect.cv2_dnn]
confidence: 50
[detect.mtcnn]
minsize: 20
threshold_1: 0.6
threshold_2: 0.7
threshold_3: 0.7
scalefactor: 0.709
batch-size: 8
[detect.s3fd]
confidence: 70
batch-size: 4
[mask.unet_dfl]
batch-size: 8
[mask.vgg_clear]
batch-size: 6
[mask.vgg_obstructed]
batch-size: 2
--------- gui.ini ---------
[global]
fullscreen: False
tab: extract
options_panel_width: 30
console_panel_height: 20
icon_size: 14
font: default
font_size: 9
autosave_last_session: prompt
timeout: 120
auto_load_model_stats: True
--------- train.ini ---------
[global]
centering: face
coverage: 95.0
icnr_init: False
conv_aware_init: True
optimizer: adam
learning_rate: 5e-05
reflect_padding: True
allow_growth: False
mixed_precision: True
convert_batchsize: 16
[global.loss]
loss_function: ssim
mask_loss_function: mse
l2_reg_term: 100
eye_multiplier: 3
mouth_multiplier: 2
penalized_mask_loss: True
mask_type: extended
mask_blur_kernel: 3
mask_threshold: 4
learn_mask: False
[model.dfaker]
output_size: 128
[model.dfl_h128]
lowmem: False
[model.dfl_sae]
input_size: 192
clipnorm: False
architecture: df
autoencoder_dims: 0
encoder_dims: 42
decoder_dims: 21
multiscale_decoder: True
[model.dlight]
features: best
details: good
output_size: 256
[model.original]
lowmem: False
[model.realface]
input_size: 128
output_size: 256
dense_nodes: 1536
complexity_encoder: 128
complexity_decoder: 512
[model.unbalanced]
input_size: 128
lowmem: False
clipnorm: True
nodes: 1024
complexity_encoder: 128
complexity_decoder_a: 384
complexity_decoder_b: 512
[model.villain]
lowmem: False
[trainer.original]
preview_images: 14
zoom_amount: 5
rotation_range: 10
shift_range: 5
flip_chance: 50
color_lightness: 30
color_ab: 8
color_clahe_chance: 50
color_clahe_max_size: 4