12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Found 15 ops to copy within [, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ], seed [], stop_at [, , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG ops_to_copy = [, , , , , , , , , , , , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/reshape_1/Reshape 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/upscale_4_0_conv2d/convolution 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/dense_1/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/flatten_1/Reshape 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/flatten_1/strided_slice 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/reshape_1/strided_slice 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/reshape_1/Shape 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/dense_2/MatMul 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/flatten_1/stack 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/flatten_1/Shape 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/reshape_1/Reshape/shape 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/dense_2/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/dense_1/MatMul 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/upscale_4_0_conv2d/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/flatten_1/Prod 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Copied [, , , , , , , , , , , , , , ] to dict_values([, , , , , , , , , , , , , , ]): [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Rewired %s in place of %s restricted to %s: [['training/Adam/decoder_a/upscale_16_0_pixelshuffler/transpose_sg:0', 'training/Adam/decoder_a/upscale_32_0_pixelshuffler/Reshape_1_sg:0', 'training/Adam/decoder_a/upscale_8_0_leakyrelu/LeakyRelu_sg:0', 'training/Adam/encoder/conv_16_0_leakyrelu/LeakyRelu_sg:0', 'training/Adam/loss/decoder_a_loss/truediv_sg:0'], ['decoder_a/upscale_16_0_pixelshuffler/transpose:0', 'decoder_a/upscale_32_0_pixelshuffler/Reshape_1:0', 'decoder_a/upscale_8_0_leakyrelu/LeakyRelu:0', 'encoder/conv_16_0_leakyrelu/LeakyRelu:0', 'loss/decoder_a_loss/truediv:0'], ['training/Adam/encoder/dense_1/BiasAdd', 'training/Adam/encoder/dense_1/MatMul', 'training/Adam/encoder/dense_2/BiasAdd', 'training/Adam/encoder/dense_2/MatMul', 'training/Adam/encoder/flatten_1/Prod', 'training/Adam/encoder/flatten_1/Reshape', 'training/Adam/encoder/flatten_1/Shape', 'training/Adam/encoder/flatten_1/stack', 'training/Adam/encoder/flatten_1/strided_slice', 'training/Adam/encoder/reshape_1/Reshape', 'training/Adam/encoder/reshape_1/Reshape/shape', 'training/Adam/encoder/reshape_1/Shape', 'training/Adam/encoder/reshape_1/strided_slice', 'training/Adam/encoder/upscale_4_0_conv2d/BiasAdd', 'training/Adam/encoder/upscale_4_0_conv2d/convolution']] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Got gradients [None, None, None, None, , None, None, None, None, None, None, , , , , , , None, None, None, None, None, None, None, None]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG for []: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG with respect to [, , , , , , , , , , , , , , , , , , , , , , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG with boundary backprop substitutions []: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Processing list []: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Found 9 ops to copy within [, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ], seed [], stop_at [, , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG ops_to_copy = [, , , , , , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_64_0_leakyrelu/LeakyRelu 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_16_0_conv2d/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_64_0_conv2d/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_16_0_conv2d/convolution 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_64_0_conv2d/convolution 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_32_0_leakyrelu/LeakyRelu 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_16_0_leakyrelu/LeakyRelu 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_32_0_conv2d/BiasAdd 12/20/2019 22:02:35 MainProcess _training_0 transform _connect_control_inputs DEBUG Connecting control inputs of op: encoder/conv_32_0_conv2d/convolution 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Copied [, , , , , , , , ] to dict_values([, , , , , , , , ]): [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Rewired %s in place of %s restricted to %s: [['training/Adam/decoder_a/upscale_16_0_pixelshuffler/transpose_sg:0', 'training/Adam/decoder_a/upscale_32_0_pixelshuffler/Reshape_1_sg:0', 'training/Adam/decoder_a/upscale_8_0_leakyrelu/LeakyRelu_sg:0', 'training/Adam/encoder/upscale_4_0_conv2d/BiasAdd_sg:0', 'training/Adam/loss/decoder_a_loss/truediv_sg:0'], ['decoder_a/upscale_16_0_pixelshuffler/transpose:0', 'decoder_a/upscale_32_0_pixelshuffler/Reshape_1:0', 'decoder_a/upscale_8_0_leakyrelu/LeakyRelu:0', 'encoder/upscale_4_0_conv2d/BiasAdd:0', 'loss/decoder_a_loss/truediv:0'], ['training/Adam/encoder/conv_16_0_conv2d/BiasAdd', 'training/Adam/encoder/conv_16_0_conv2d/convolution', 'training/Adam/encoder/conv_16_0_leakyrelu/LeakyRelu', 'training/Adam/encoder/conv_32_0_conv2d/BiasAdd', 'training/Adam/encoder/conv_32_0_conv2d/convolution', 'training/Adam/encoder/conv_32_0_leakyrelu/LeakyRelu', 'training/Adam/encoder/conv_64_0_conv2d/BiasAdd', 'training/Adam/encoder/conv_64_0_conv2d/convolution', 'training/Adam/encoder/conv_64_0_leakyrelu/LeakyRelu']] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG Got gradients [None, None, None, None, None, , , , , , , None, None, None, None, None, None, None, None, None, None, None, None, None, None]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG for []: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG with respect to [, , , , , , , , , , , , , , , , , , , , , , , , ]: [] 12/20/2019 22:02:35 MainProcess _training_0 memory_saving_gradients debug_print DEBUG with boundary backprop substitutions []: [] 12/20/2019 22:02:41 MainProcess _training_0 _base generate_preview DEBUG Generating preview 12/20/2019 22:02:41 MainProcess _training_0 _base largest_face_index DEBUG 0 12/20/2019 22:02:41 MainProcess _training_0 _base compile_sample DEBUG Compiling samples: (side: 'a', samples: 14) 12/20/2019 22:02:41 MainProcess _training_0 _base get_sample DEBUG Getting time-lapse samples: 'a' 12/20/2019 22:02:41 MainProcess _training_0 _base _setup DEBUG Setting up time-lapse 12/20/2019 22:02:41 MainProcess _training_0 _base _setup DEBUG Time-lapse output set to 'C:\Users\thiny\Project\faceswap\sample\output\timelapse_output' 12/20/2019 22:02:41 MainProcess _training_0 utils get_image_paths DEBUG Scanned Folder contains 0 files 12/20/2019 22:02:41 MainProcess _training_0 utils get_image_paths DEBUG Returning 0 images 12/20/2019 22:02:41 MainProcess _training_0 utils get_image_paths DEBUG Scanned Folder contains 0 files 12/20/2019 22:02:41 MainProcess _training_0 utils get_image_paths DEBUG Returning 0 images 12/20/2019 22:02:41 MainProcess _training_0 _base set_timelapse_feed DEBUG Setting time-lapse feed: (side: 'a', input_images: '[]', batch_size: 0) 12/20/2019 22:02:41 MainProcess _training_0 _base _load_generator DEBUG Loading generator: a 12/20/2019 22:02:41 MainProcess _training_0 _base _load_generator DEBUG input_size: 64, output_shapes: [(64, 64, 3)] 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initializing TrainingDataGenerator: (model_input_size: 64, model_output_shapes: [(64, 64, 3)], training_opts: {'alignments': {'a': 'C:\\Users\\thiny\\Project\\faceswap\\sample\\input\\alina\\alina_alignments.fsa', 'b': 'C:\\Users\\thiny\\Project\\faceswap\\sample\\input\\so\\vids\\faceswap_alignments.fsa'}, 'preview_scaling': 0.25, 'warp_to_landmarks': False, 'augment_color': True, 'no_flip': False, 'pingpong': True, 'snapshot_interval': 25000, 'training_size': 256, 'no_logs': False, 'coverage_ratio': 0.6875, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'penalized_mask_loss': True}, landmarks: {}, masks: {'a': 4685, 'b': 5690}, config: {'coverage': 68.75, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'icnr_init': False, 'conv_aware_init': False, 'subpixel_upscaling': False, 'reflect_padding': False, 'penalized_mask_loss': True, 'loss_function': 'ssim', 'learning_rate': 5e-05, 'preview_images': 14, 'zoom_amount': 5, 'rotation_range': 10, 'shift_range': 5, 'flip_chance': 50, 'color_lightness': 30, 'color_ab': 8, 'color_clahe_chance': 50, 'color_clahe_max_size': 4}) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initialized TrainingDataGenerator 12/20/2019 22:02:41 MainProcess _training_0 training_data minibatch_ab DEBUG Queue batches: (image_count: 0, batchsize: 0, side: 'a', do_shuffle: False, is_preview, False, is_timelapse: True) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initializing ImageAugmentation: (batchsize: 0, is_display: True, input_size: 64, output_shapes: [(64, 64, 3)], coverage_ratio: 0.6875, config: {'coverage': 68.75, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'icnr_init': False, 'conv_aware_init': False, 'subpixel_upscaling': False, 'reflect_padding': False, 'penalized_mask_loss': True, 'loss_function': 'ssim', 'learning_rate': 5e-05, 'preview_images': 14, 'zoom_amount': 5, 'rotation_range': 10, 'shift_range': 5, 'flip_chance': 50, 'color_lightness': 30, 'color_ab': 8, 'color_clahe_chance': 50, 'color_clahe_max_size': 4}) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Output sizes: [64] 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initialized ImageAugmentation 12/20/2019 22:02:41 MainProcess _training_0 multithreading __init__ DEBUG Initializing BackgroundGenerator: (target: '_run', thread_count: 2) 12/20/2019 22:02:41 MainProcess _training_0 multithreading __init__ DEBUG Initialized BackgroundGenerator: '_run' 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread(s): '_run' 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread 1 of 2: '_run_0' 12/20/2019 22:02:41 MainProcess _run_0 training_data _minibatch DEBUG Loading minibatch generator: (image_count: 0, side: 'a', do_shuffle: False) 12/20/2019 22:02:41 MainProcess _run_0 multithreading run DEBUG Error in thread (_run_0): tuple index out of range 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread 2 of 2: '_run_1' 12/20/2019 22:02:41 MainProcess _run_1 training_data _minibatch DEBUG Loading minibatch generator: (image_count: 0, side: 'a', do_shuffle: False) 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Started all threads '_run': 2 12/20/2019 22:02:41 MainProcess _training_0 _base set_timelapse_feed DEBUG Set time-lapse feed 12/20/2019 22:02:41 MainProcess _training_0 _base set_timelapse_feed DEBUG Setting time-lapse feed: (side: 'b', input_images: '[]', batch_size: 0) 12/20/2019 22:02:41 MainProcess _training_0 _base _load_generator DEBUG Loading generator: b 12/20/2019 22:02:41 MainProcess _training_0 _base _load_generator DEBUG input_size: 64, output_shapes: [(64, 64, 3)] 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initializing TrainingDataGenerator: (model_input_size: 64, model_output_shapes: [(64, 64, 3)], training_opts: {'alignments': {'a': 'C:\\Users\\thiny\\Project\\faceswap\\sample\\input\\alina\\alina_alignments.fsa', 'b': 'C:\\Users\\thiny\\Project\\faceswap\\sample\\input\\so\\vids\\faceswap_alignments.fsa'}, 'preview_scaling': 0.25, 'warp_to_landmarks': False, 'augment_color': True, 'no_flip': False, 'pingpong': True, 'snapshot_interval': 25000, 'training_size': 256, 'no_logs': False, 'coverage_ratio': 0.6875, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'penalized_mask_loss': True}, landmarks: {}, masks: {'a': 4685, 'b': 5690}, config: {'coverage': 68.75, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'icnr_init': False, 'conv_aware_init': False, 'subpixel_upscaling': False, 'reflect_padding': False, 'penalized_mask_loss': True, 'loss_function': 'ssim', 'learning_rate': 5e-05, 'preview_images': 14, 'zoom_amount': 5, 'rotation_range': 10, 'shift_range': 5, 'flip_chance': 50, 'color_lightness': 30, 'color_ab': 8, 'color_clahe_chance': 50, 'color_clahe_max_size': 4}) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initialized TrainingDataGenerator 12/20/2019 22:02:41 MainProcess _training_0 training_data minibatch_ab DEBUG Queue batches: (image_count: 0, batchsize: 0, side: 'b', do_shuffle: False, is_preview, False, is_timelapse: True) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initializing ImageAugmentation: (batchsize: 0, is_display: True, input_size: 64, output_shapes: [(64, 64, 3)], coverage_ratio: 0.6875, config: {'coverage': 68.75, 'mask_type': 'extended', 'mask_blur_kernel': 3, 'mask_threshold': 4, 'learn_mask': False, 'icnr_init': False, 'conv_aware_init': False, 'subpixel_upscaling': False, 'reflect_padding': False, 'penalized_mask_loss': True, 'loss_function': 'ssim', 'learning_rate': 5e-05, 'preview_images': 14, 'zoom_amount': 5, 'rotation_range': 10, 'shift_range': 5, 'flip_chance': 50, 'color_lightness': 30, 'color_ab': 8, 'color_clahe_chance': 50, 'color_clahe_max_size': 4}) 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Output sizes: [64] 12/20/2019 22:02:41 MainProcess _training_0 training_data __init__ DEBUG Initialized ImageAugmentation 12/20/2019 22:02:41 MainProcess _training_0 multithreading __init__ DEBUG Initializing BackgroundGenerator: (target: '_run', thread_count: 2) 12/20/2019 22:02:41 MainProcess _training_0 multithreading __init__ DEBUG Initialized BackgroundGenerator: '_run' 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread(s): '_run' 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread 1 of 2: '_run_0' 12/20/2019 22:02:41 MainProcess _run_0 training_data _minibatch DEBUG Loading minibatch generator: (image_count: 0, side: 'b', do_shuffle: False) 12/20/2019 22:02:41 MainProcess _run_0 multithreading run DEBUG Error in thread (_run_0): tuple index out of range 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Starting thread 2 of 2: '_run_1' 12/20/2019 22:02:41 MainProcess _run_1 training_data _minibatch DEBUG Loading minibatch generator: (image_count: 0, side: 'b', do_shuffle: False) 12/20/2019 22:02:41 MainProcess _training_0 multithreading start DEBUG Started all threads '_run': 2 12/20/2019 22:02:41 MainProcess _training_0 _base set_timelapse_feed DEBUG Set time-lapse feed 12/20/2019 22:02:41 MainProcess _training_0 _base _setup DEBUG Set up time-lapse 12/20/2019 22:02:41 MainProcess _training_0 multithreading check_and_raise_error DEBUG Thread error caught: [(, IndexError('tuple index out of range',), )] 12/20/2019 22:02:41 MainProcess _training_0 multithreading run DEBUG Error in thread (_training_0): tuple index out of range 12/20/2019 22:02:41 MainProcess _run_1 multithreading run DEBUG Error in thread (_run_1): tuple index out of range 12/20/2019 22:02:42 MainProcess MainThread train _monitor DEBUG Thread error detected 12/20/2019 22:02:42 MainProcess MainThread train _monitor DEBUG Closed Monitor 12/20/2019 22:02:42 MainProcess MainThread train _end_thread DEBUG Ending Training thread 12/20/2019 22:02:42 MainProcess MainThread train _end_thread CRITICAL Error caught! Exiting... 12/20/2019 22:02:42 MainProcess MainThread multithreading join DEBUG Joining Threads: '_training' 12/20/2019 22:02:42 MainProcess MainThread multithreading join DEBUG Joining Thread: '_training_0' 12/20/2019 22:02:42 MainProcess MainThread multithreading join ERROR Caught exception in thread: '_training_0' Traceback (most recent call last): File "C:\Users\thiny\Project\faceswap\faceswap2\lib\cli.py", line 129, in execute_script process.process() File "C:\Users\thiny\Project\faceswap\faceswap2\scripts\train.py", line 159, in process self._end_thread(thread, err) File "C:\Users\thiny\Project\faceswap\faceswap2\scripts\train.py", line 199, in _end_thread thread.join() File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 121, in join raise thread.err[1].with_traceback(thread.err[2]) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 37, in run self._target(*self._args, **self._kwargs) File "C:\Users\thiny\Project\faceswap\faceswap2\scripts\train.py", line 224, in _training raise err File "C:\Users\thiny\Project\faceswap\faceswap2\scripts\train.py", line 214, in _training self._run_training_cycle(model, trainer) File "C:\Users\thiny\Project\faceswap\faceswap2\scripts\train.py", line 303, in _run_training_cycle trainer.train_one_step(viewer, timelapse) File "C:\Users\thiny\Project\faceswap\faceswap2\plugins\train\trainer\_base.py", line 316, in train_one_step raise err File "C:\Users\thiny\Project\faceswap\faceswap2\plugins\train\trainer\_base.py", line 290, in train_one_step self._timelapse.get_sample(side, timelapse_kwargs) File "C:\Users\thiny\Project\faceswap\faceswap2\plugins\train\trainer\_base.py", line 947, in get_sample self._samples.images[side] = self._batchers[side].compile_timelapse_sample() File "C:\Users\thiny\Project\faceswap\faceswap2\plugins\train\trainer\_base.py", line 539, in compile_timelapse_sample batch = next(self._timelapse_feed) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 156, in iterator self.check_and_raise_error() File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 84, in check_and_raise_error raise error[1].with_traceback(error[2]) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 37, in run self._target(*self._args, **self._kwargs) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\multithreading.py", line 145, in _run for item in self.generator(*self._gen_args, **self._gen_kwargs): File "C:\Users\thiny\Project\faceswap\faceswap2\lib\training_data.py", line 189, in _minibatch yield self._process_batch(img_paths, side) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\training_data.py", line 198, in _process_batch batch = self._apply_mask(filenames, batch, side) File "C:\Users\thiny\Project\faceswap\faceswap2\lib\training_data.py", line 254, in _apply_mask masks = self._resize_masks(batch.shape[1], masks) IndexError: tuple index out of range ============ System Information ============ encoding: cp936 git_branch: Not Found git_commits: Not Found gpu_cuda: No global version found. Check Conda packages for Conda Cuda gpu_cudnn: No global version found. Check Conda packages for Conda cuDNN gpu_devices: GPU_0: GeForce GTX 1050 gpu_devices_active: GPU_0 gpu_driver: 441.66 gpu_vram: GPU_0: 4096MB os_machine: AMD64 os_platform: Windows-10-10.0.18362-SP0 os_release: 10 py_command: C:\Users\thiny\Project\faceswap\faceswap2\faceswap.py train -A C:/Users/thiny/Project/faceswap/sample/output/extract_alina -ala C:/Users/thiny/Project/faceswap/sample/input/alina/alina_alignments.fsa -B C:/Users/thiny/Project/faceswap/sample/output/extract_so -alb C:/Users/thiny/Project/faceswap/sample/input/so/vids/faceswap_alignments.fsa -m C:/Users/thiny/Project/faceswap/sample/output/model -t lightweight -bs 8 -it 1000000 -g 1 -msg -o -pp -s 100 -ss 25000 -tia C:/Users/thiny/Project/faceswap/sample/output/timelapse_alina -tib C:/Users/thiny/Project/faceswap/sample/output/timelapse_so -to C:/Users/thiny/Project/faceswap/sample/output/timelapse_output -ps 25 -L INFO -gui py_conda_version: conda 4.8.0 py_implementation: CPython py_version: 3.6.9 py_virtual_env: True sys_cores: 8 sys_processor: Intel64 Family 6 Model 158 Stepping 9, GenuineIntel sys_ram: Total: 32619MB, Available: 17914MB, Used: 14705MB, Free: 17914MB =============== Pip Packages =============== absl-py==0.8.1 astor==0.8.0 astroid==2.3.0 certifi==2019.11.28 cloudpickle==1.2.2 colorama==0.4.1 cycler==0.10.0 cytoolz==0.10.1 dask==2.8.0 decorator==4.4.1 fastcluster==1.1.25 ffmpy==0.2.2 gast==0.3.2 grpcio==1.16.1 h5py==2.9.0 imageio==2.6.1 imageio-ffmpeg==0.3.0 isort==4.3.21 joblib==0.14.0 Keras==2.2.4 Keras-Applications==1.0.8 Keras-Preprocessing==1.1.0 kiwisolver==1.1.0 lazy-object-proxy==1.4.2 Markdown==3.1.1 matplotlib==3.1.1 mccabe==0.6.1 mkl-fft==1.0.15 mkl-random==1.1.0 mkl-service==2.3.0 networkx==2.4 numpy==1.17.4 nvidia-ml-py3==7.352.1 olefile==0.46 opencv-python==4.1.2.30 pathlib==1.0.1 Pillow==6.2.1 protobuf==3.10.1 psutil==5.6.5 pylint==2.4.4 pyparsing==2.4.5 pyreadline==2.1 python-dateutil==2.8.1 pytz==2019.3 PyWavelets==1.1.1 pywin32==223 PyYAML==5.1.2 scikit-image==0.15.0 scikit-learn==0.21.3 scipy==1.3.1 six==1.12.0 tensorboard==1.14.0 tensorflow==1.14.0 tensorflow-estimator==1.14.0 termcolor==1.1.0 toolz==0.10.0 toposort==1.5 tornado==6.0.3 tqdm==4.38.0 typed-ast==1.2.0 Werkzeug==0.16.0 wincertstore==0.2 wrapt==1.11.2 ============== Conda Packages ============== # packages in environment at C:\Users\thiny\.conda\envs\faceswap: # # Name Version Build Channel _tflow_select 2.1.0 gpu absl-py 0.8.1 py36_0 astor 0.8.0 py36_0 blas 1.0 mkl ca-certificates 2019.11.27 0 certifi 2019.11.28 py36_0 cloudpickle 1.2.2 py_0 cudatoolkit 10.0.130 0 cudnn 7.6.4 cuda10.0_0 cycler 0.10.0 py36h009560c_0 cytoolz 0.10.1 py36he774522_0 dask-core 2.8.0 py_0 decorator 4.4.1 py_0 fastcluster 1.1.25 py36he350917_1000 conda-forge ffmpeg 4.2 h6538335_0 conda-forge ffmpy 0.2.2 pypi_0 pypi freetype 2.9.1 ha9979f8_1 gast 0.3.2 py_0 grpcio 1.16.1 py36h351948d_1 h5py 2.9.0 py36h5e291fa_0 hdf5 1.10.4 h7ebc959_0 icc_rt 2019.0.0 h0cc432a_1 icu 58.2 ha66f8fd_1 imageio 2.6.1 py36_0 imageio-ffmpeg 0.3.0 py_0 conda-forge intel-openmp 2019.4 245 joblib 0.14.0 py_0 jpeg 9b hb83a4c4_2 keras 2.2.4 0 keras-applications 1.0.8 py_0 keras-base 2.2.4 py36_0 keras-preprocessing 1.1.0 py_1 kiwisolver 1.1.0 py36ha925a31_0 libmklml 2019.0.5 0 libpng 1.6.37 h2a8f88b_0 libprotobuf 3.10.1 h7bd577a_0 libtiff 4.1.0 h56a325e_0 markdown 3.1.1 py36_0 matplotlib 3.1.1 py36hc8f65d3_0 mkl 2019.4 245 mkl-service 2.3.0 py36hb782905_0 mkl_fft 1.0.15 py36h14836fe_0 mkl_random 1.1.0 py36h675688f_0 networkx 2.4 py_0 numpy 1.17.4 py36h4320e6b_0 numpy-base 1.17.4 py36hc3f5095_0 nvidia-ml-py3 7.352.1 pypi_0 pypi olefile 0.46 py36_0 opencv-python 4.1.1.26 pypi_0 pypi openssl 1.1.1d he774522_3 pathlib 1.0.1 py36_1 pillow 6.2.1 py36hdc69c19_0 pip 19.3.1 py36_0 protobuf 3.10.1 py36h33f27b4_0 psutil 5.6.5 py36he774522_0 pyparsing 2.4.5 py_0 pyqt 5.9.2 py36h6538335_2 pyreadline 2.1 py36_1 python 3.6.9 h5500b2f_0 python-dateutil 2.8.1 py_0 pytz 2019.3 py_0 pywavelets 1.1.1 py36he774522_0 pywin32 223 py36hfa6e2cd_1 pyyaml 5.1.2 py36he774522_0 qt 5.9.7 vc14h73c81de_0 scikit-image 0.15.0 py36ha925a31_0 scikit-learn 0.21.3 py36h6288b17_0 scipy 1.3.1 py36h29ff71c_0 setuptools 41.6.0 py36_0 sip 4.19.8 py36h6538335_0 six 1.13.0 py36_0 sqlite 3.30.1 he774522_0 tensorboard 1.14.0 py36he3c9ec2_0 tensorflow 1.14.0 gpu_py36h305fd99_0 tensorflow-base 1.14.0 gpu_py36h55fc52a_0 tensorflow-estimator 1.14.0 py_0 tensorflow-gpu 1.14.0 h0d30ee6_0 termcolor 1.1.0 py36_1 tk 8.6.8 hfa6e2cd_0 toolz 0.10.0 py_0 toposort 1.5 py_3 conda-forge tornado 6.0.3 py36he774522_0 tqdm 4.38.0 py_0 vc 14.1 h0510ff6_4 vs2015_runtime 14.16.27012 hf0eaf9b_0 werkzeug 0.16.0 py_0 wheel 0.33.6 py36_0 wincertstore 0.2 py36h7fe50ca_0 wrapt 1.11.2 py36he774522_0 xz 5.2.4 h2fa13f4_4 yaml 0.1.7 hc54c509_2 zlib 1.2.11 h62dcd97_3 zstd 1.3.7 h508b16e_0 ================= Configs ================== --------- .faceswap --------- backend: nvidia --------- convert.ini --------- [color.color_transfer] clip: True preserve_paper: True [color.manual_balance] colorspace: HSV balance_1: 0.0 balance_2: 0.0 balance_3: 0.0 contrast: 0.0 brightness: 0.0 [color.match_hist] threshold: 99.0 [mask.box_blend] type: gaussian distance: 11.0 radius: 5.0 passes: 1 [mask.mask_blend] type: normalized radius: 3.0 passes: 4 erosion: 0.0 [scaling.sharpen] method: unsharp_mask amount: 150 radius: 0.3 threshold: 5.0 [writer.ffmpeg] container: mp4 codec: libx264 crf: 23 preset: medium tune: none profile: auto level: auto [writer.gif] fps: 25 loop: 0 palettesize: 256 subrectangles: False [writer.opencv] format: png draw_transparent: False jpg_quality: 75 png_compress_level: 3 [writer.pillow] format: png draw_transparent: False optimize: False gif_interlace: True jpg_quality: 75 png_compress_level: 3 tif_compression: tiff_deflate --------- extract.ini --------- [global] allow_growth: False [align.fan] batch-size: 12 [detect.cv2_dnn] confidence: 50 [detect.mtcnn] minsize: 20 threshold_1: 0.6 threshold_2: 0.7 threshold_3: 0.7 scalefactor: 0.709 batch-size: 8 [detect.s3fd] confidence: 70 batch-size: 4 [mask.unet_dfl] batch-size: 8 [mask.vgg_clear] batch-size: 6 [mask.vgg_obstructed] batch-size: 2 --------- gui.ini --------- [global] fullscreen: False tab: extract options_panel_width: 30 console_panel_height: 20 icon_size: 14 font: default font_size: 9 autosave_last_session: prompt timeout: 120 auto_load_model_stats: True --------- train.ini --------- [global] coverage: 68.75 mask_type: extended mask_blur_kernel: 3 mask_threshold: 4 learn_mask: False icnr_init: False conv_aware_init: False subpixel_upscaling: False reflect_padding: False penalized_mask_loss: True loss_function: ssim learning_rate: 5e-05 [model.dfl_h128] lowmem: True [model.dfl_sae] input_size: 128 clipnorm: True architecture: df autoencoder_dims: 0 encoder_dims: 42 decoder_dims: 21 multiscale_decoder: False [model.dlight] features: best details: good output_size: 256 [model.original] lowmem: False [model.realface] input_size: 64 output_size: 128 dense_nodes: 1536 complexity_encoder: 128 complexity_decoder: 512 [model.unbalanced] input_size: 128 lowmem: False clipnorm: True nodes: 1024 complexity_encoder: 128 complexity_decoder_a: 384 complexity_decoder_b: 512 [model.villain] lowmem: False [trainer.original] preview_images: 14 zoom_amount: 5 rotation_range: 10 shift_range: 5 flip_chance: 50 color_lightness: 30 color_ab: 8 color_clahe_chance: 50 color_clahe_max_size: 4