Code Monkey home page Code Monkey logo

srflow's Introduction

SRFlow

Official SRFlow training code: Super-Resolution using Normalizing Flow in PyTorch


News: Unified Image Super-Resolution and Rescaling [code]

SRFlow


Setup: Data, Environment, PyTorch Demo


git clone https://github.com/andreas128/SRFlow.git && cd SRFlow && ./setup.sh

This oneliner will:

  • Clone SRFlow
  • Setup a python3 virtual env
  • Install the packages from requirements.txt
  • Download the pretrained models
  • Download the DIV2K validation data
  • Run the Demo Jupyter Notebook

If you want to install it manually, read the setup.sh file. (Links to data/models, pip packages)



Demo: Try Normalizing Flow in PyTorch

./run_jupyter.sh

This notebook lets you:

  • Load the pretrained models.
  • Super-resolve images.
  • Measure PSNR/SSIM/LPIPS.
  • Infer the Normalizing Flow latent space.



Testing: Apply the included pretrained models

source myenv/bin/activate                      # Use the env you created using setup.sh
cd code
CUDA_VISIBLE_DEVICES=-1 python test.py ./confs/SRFlow_DF2K_4X.yml      # Diverse Images 4X (Dataset Included)
CUDA_VISIBLE_DEVICES=-1 python test.py ./confs/SRFlow_DF2K_8X.yml      # Diverse Images 8X (Dataset Included)
CUDA_VISIBLE_DEVICES=-1 python test.py ./confs/SRFlow_CelebA_8X.yml    # Faces 8X

For testing, we apply SRFlow to the full images on CPU.



Training: Reproduce or train on your Data

The following commands train the Super-Resolution network using Normalizing Flow in PyTorch:

source myenv/bin/activate                      # Use the env you created using setup.sh
cd code
python train.py -opt ./confs/SRFlow_DF2K_4X.yml      # Diverse Images 4X (Dataset Included)
python train.py -opt ./confs/SRFlow_DF2K_8X.yml      # Diverse Images 8X (Dataset Included)
python train.py -opt ./confs/SRFlow_CelebA_8X.yml    # Faces 8X
  • To reduce the GPU memory, reduce the batch size in the yml file.
  • CelebA does not allow us to host the dataset. A script will follow.

How to prepare CelebA?

1. Get HD-CelebA-Cropper

git clone https://github.com/LynnHo/HD-CelebA-Cropper

2. Download the dataset

img_celeba.7z and annotations.zip as desribed in the Readme.

3. Run the crop align

python3 align.py --img_dir ./data/data --crop_size_h 640 --crop_size_w 640 --order 3 --face_factor 0.6 --n_worker 8

4. Downsample for GT

Use the matlablike kernel to downscale to 160x160 for the GT images.

5. Downsample for LR

Downscale the GT using the Matlab kernel to the LR size (40x40 or 20x20)

6. Train/Validation

For training and validation, we use the corresponding sets defined by CelebA (Train: 000001-162770, Validation: 162771-182637)

7. Pack to pickle for training

cd code && python prepare_data.py /path/to/img_dir



Dataset: How to train on your own data

The following command creates the pickel files that you can use in the yaml config file:

cd code
python prepare_data.py /path/to/img_dir

The precomputed DF2K dataset gets downloaded using setup.sh. You can reproduce it or prepare your own dataset.



Our paper explains

  • How to train Conditional Normalizing Flow
    We designed an architecture that archives state-of-the-art super-resolution quality.
  • How to train Normalizing Flow on a single GPU
    We based our network on GLOW, which uses up to 40 GPUs to train for image generation. SRFlow only needs a single GPU for training conditional image generation.
  • How to use Normalizing Flow for image manipulation
    How to exploit the latent space for Normalizing Flow for controlled image manipulations
  • See many Visual Results
    Compare GAN vs Normalizing Flow yourself. We've included a lot of visuals results in our [Paper].



GAN vs Normalizing Flow - Blog

  • Sampling: SRFlow outputs many different images for a single input.
  • Stable Training: SRFlow has much fewer hyperparameters than GAN approaches, and we did not encounter training stability issues.
  • Convergence: While GANs cannot converge, conditional Normalizing Flows converge monotonic and stable.
  • Higher Consistency: When downsampling the super-resolution, one obtains almost the exact input.

Get a quick introduction to Normalizing Flow in our [Blog].




Wanna help to improve the code?

If you found a bug or improved the code, please do the following:

  • Fork this repo.
  • Push the changes to your repo.
  • Create a pull request.



Paper

[Paper] ECCV 2020 Spotlight

@inproceedings{lugmayr2020srflow,
  title={SRFlow: Learning the Super-Resolution Space with Normalizing Flow},
  author={Lugmayr, Andreas and Danelljan, Martin and Van Gool, Luc and Timofte, Radu},
  booktitle={ECCV},
  year={2020}
}



srflow's People

Contributors

andreas128 avatar hwong557 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

srflow's Issues

Bus error (core dumped)

Thank you very much for your work. When I reproduce the X4 super-resolution completely with your code, there will be a Bus error (core dumped) after training for a while. What should I do?

Is GT_size 160 (size of HR image) hard-coded in the code

  1. Is GT_size (size of HR image) hard-coded in the following place:
  1. If yes, does it mean that one can not use any other HR image size during training?
  2. Does SRFlow impose any limit on what this hardcarded value could be? In other words, can I use for example 180?

Is the RRDBNet correct?

In the forward function of the RRDBNet

def forward(self, x, get_steps=False):
    fea = self.conv_first(x)
    block_idxs = opt_get(self.opt, ['network_G', 'flow', 'stackRRDB', 'blocks']) or []
    block_results = {}
    for idx, m in enumerate(self.RRDB_trunk.children()):            
        fea = m(fea)                                               
        for b in block_idxs:
            if b == idx:
                block_results["block_{}".format(idx)] = fea         
    trunk = self.trunk_conv(fea)
    last_lr_fea = fea + trunk             

It seems that fea in the last_lr_fea=fea+trunk line is the feature of the last RRDB block before self.turnk_conv. But to my knowledge, the skip connection should be last_lr_fea = x+ trunk. Is my understanding right?

How to generate a SR image with only 1 channel?

Thanks for your code and contributions!
There is a task that generate a SR image, but I just need a grayscale generation(that means the generation data has only 1 channel), the origin code can only generate a 3 channel's RGB images.
So how to generate the 1 channel data?

x8 RuntimeError: Error(s) in loading state_dict for SRFlowNet:

#14
I run command again. (python train.py -opt ./confs/SRFlow_DF2K_4X.yml # Diverse Images 4X (Dataset Included)) and the result is
(myenv) (base) ubuntu@ubuntu:~/Desktop/data1024/SRFlow-master/code$ python train.py -opt ./confs/SRFlow_DF2K_4X.yml
OrderedDict([('manual_seed', 10), ('lr_G', 0.00025), ('weight_decay_G', 0), ('beta1', 0.9), ('beta2', 0.99), ('lr_scheme', 'MultiStepLR'), ('warmup_iter', -1), ('lr_steps_rel', [0.5, 0.75, 0.9, 0.95]), ('lr_gamma', 0.5), ('niter', 200000), ('val_freq', 40000), ('lr_steps', [100000, 150000, 180000, 190000])])
Disabled distributed training.
pretrain_model path will be ignored when resuming training.
21-02-14 23:52:33.972 - INFO: name: train
use_tb_logger:
model: SRFlow
distortion: sr
scale: 4
gpu_ids: [0]
datasets:[
train:[
name: CelebA_160_tr
mode: LRHR_PKL
dataroot_GT: ../datasets/DF2K-tr.pklv4
dataroot_LQ: ../datasets/DF2K-tr_X4.pklv4
quant: 32
use_shuffle: True
n_workers: 3
batch_size: 12
GT_size: 160
use_flip: True
color: RGB
phase: train
scale: 4
data_type: img
]
val:[
name: CelebA_160_va
mode: LRHR_PKL
dataroot_GT: ../datasets/DIV2K-va.pklv4
dataroot_LQ: ../datasets/DIV2K-va_X4.pklv4
quant: 32
n_max: 20
phase: val
scale: 4
data_type: img
]
]
dataroot_GT: ../datasets/div2k-validation-modcrop8-gt
dataroot_LR: ../datasets/div2k-validation-modcrop8-x4
model_path: ../pretrained_models/SRFlow_DF2K_4X.pth
heat: 0.9
network_G:[
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 23
upscale: 4
train_RRDB: False
train_RRDB_delay: 0.5
flow:[
K: 16
L: 3
noInitialInj: True
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:[
enable: True
]
fea_up0: True
stackRRDB:[
blocks: [1, 8, 15, 22]
concat: True
]
]
scale: 4
]
path:[
pretrain_model_G: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models/200000_G.pth
strict_load: True
resume_state: auto
root: /home/ubuntu/Desktop/data1024/SRFlow-master
experiments_root: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
models: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models
training_state: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/training_state
log: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
val_images: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/val_images
]
train:[
manual_seed: 10
lr_G: 0.00025
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1
lr_steps_rel: [0.5, 0.75, 0.9, 0.95]
lr_gamma: 0.5
niter: 200000
val_freq: 40000
lr_steps: [100000, 150000, 180000, 190000]
]
val:[
heats: [0.0, 0.5, 0.75, 1.0]
n_sample: 3
]
logger:[
print_freq: 100
save_checkpoint_freq: 1000.0
]
is_train: True
dist: False

21-02-14 23:52:34.059 - INFO: Random seed: 10
{'name': 'CelebA_160_tr', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DF2K-tr.pklv4', 'dataroot_LQ': '../datasets/DF2K-tr_X4.pklv4', 'quant': 32, 'use_shuffle': True, 'n_workers': 3, 'batch_size': 12, 'GT_size': 160, 'use_flip': True, 'color': 'RGB', 'phase': 'train', 'scale': 4, 'data_type': 'img'}
Loaded 162150 HR images with [0.00, 255.00] in 15.40s from ../datasets/DF2K-tr.pklv4
Loaded 162150 LR images with [0.00, 255.00] in 15.40s from ../datasets/DF2K-tr_X4.pklv4
21-02-14 23:52:49.469 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_tr] is created.
Dataset created
21-02-14 23:52:49.480 - INFO: Number of train images: 162,150, iters: 13,513
21-02-14 23:52:49.480 - INFO: Total epochs needed: 15 for iters 200,000
{'name': 'CelebA_160_va', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DIV2K-va.pklv4', 'dataroot_LQ': '../datasets/DIV2K-va_X4.pklv4', 'quant': 32, 'n_max': 20, 'phase': 'val', 'scale': 4, 'data_type': 'img'}
Loaded 20 HR images with [0.00, 255.00] in 0.79s from ../datasets/DIV2K-va.pklv4
Loaded 20 LR images with [0.00, 255.00] in 0.79s from ../datasets/DIV2K-va_X4.pklv4
21-02-14 23:52:50.273 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_va] is created.
21-02-14 23:52:50.273 - INFO: Number of val images in [CelebA_160_va]: 20
21-02-14 23:52:52.479 - INFO: Network G structure: DataParallel - SRFlowNet, with parameters: 39,541,811

module.RRDB.conv_first.weight True
opt module.RRDB.conv_first.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.conv_first.weight] will not optimize.
module.RRDB.conv_first.bias True
opt module.RRDB.conv_first.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.conv_first.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv1.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv1.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv2.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv2.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv3.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv3.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv4.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv4.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv5.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv5.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv1.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv1.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv2.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv2.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv3.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv3.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv4.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv4.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv5.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv5.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv1.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv1.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv2.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv2.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv3.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv3.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv4.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv4.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv5.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv5.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv1.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv1.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv2.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv2.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv3.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv3.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv4.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv4.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv5.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv5.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv1.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv1.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv2.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv2.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv3.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv3.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv4.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv4.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv5.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv5.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv1.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv1.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv2.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv2.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv3.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv3.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv4.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv4.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv5.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv5.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv1.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv1.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv2.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv2.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv3.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv3.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv4.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv4.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv5.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv5.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv1.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv1.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv2.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv2.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv3.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv3.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv4.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv4.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv5.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv5.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv1.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv1.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv2.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv2.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv3.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv3.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv4.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv4.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv5.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv5.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv1.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv1.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv2.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv2.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv3.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv3.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv4.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv4.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv5.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv5.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv1.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv1.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv2.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv2.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv3.weight
21-02-14 23:52:53.184 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv3.bias
21-02-14 23:52:53.184 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv4.weight
·························································································································································
arams [module.flowUpsamplerNet.layers.12.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.actnorm.logs True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.invconv.weight True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.weight True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.13.actnorm.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.invconv.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.14.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.invconv.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.15.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.invconv.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.16.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.16.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.invconv.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.17.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.17.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.17.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.invconv.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.18.actnorm.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.invconv.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.19.conv.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.weight] will not optimize.
module.flowUpsamplerNet.layers.19.conv.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.bias] will not optimize.
module.flowUpsamplerNet.layers.19.conv.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.logs] will not optimize.
module.flowUpsamplerNet.layers.21.actnorm.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.21.actnorm.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.21.invconv.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.22.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.22.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.22.invconv.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.23.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.invconv.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.24.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.invconv.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.25.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.invconv.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.26.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.26.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.26.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.invconv.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.27.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.invconv.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.28.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.actnorm.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.invconv.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.29.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.invconv.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.30.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.invconv.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.31.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.invconv.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.32.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.32.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.invconv.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.33.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.33.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.33.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.invconv.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.34.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.invconv.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.35.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.actnorm.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.invconv.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.36.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.invconv.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.37.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.invconv.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.38.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.invconv.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.40.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.40.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.40.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.40.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.40.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.40.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.41.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.41.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.41.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.42.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.43.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.invconv.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.44.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.invconv.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.45.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.45.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.invconv.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.46.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.46.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.46.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.invconv.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.47.actnorm.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.invconv.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.48.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.invconv.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.49.actnorm.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.invconv.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.50.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.invconv.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.51.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.invconv.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.52.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.52.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.invconv.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.53.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.invconv.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.54.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.actnorm.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.invconv.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.55.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.invconv.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.56.actnorm.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.invconv.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.57.actnorm.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.actnorm.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.invconv.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.f.0.weight True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.f.0.weight] will not optimize.
module.flowUpsamplerNet.f.0.bias True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.f.0.bias] will not optimize.
rrdb params 702
21-02-14 23:52:53.341 - INFO: Model [SRFlowModel] is created.
21-02-14 23:52:53.341 - INFO: Resuming training from epoch: 14, iter: 200000.
21-02-14 23:52:53.685 - INFO: Start training from epoch: 14, iter: 200000
21-02-14 23:52:55.919 - INFO: Saving the final model.
21-02-14 23:52:56.557 - INFO: End of training.

LPIPS version

Hi, which LPIPS version did you used in your evaluation script? v0.0 or v0.1?

Bus error (core dumped)

First of all thank you very much for your work .
when I train x4 super-resolution completely with your code, but after a period of training,it will occur "Bus error (core dumped)".when I use "python -X faulthandler train.py -opt
./confs/SRFlow_DF2K_4X.yml"

it will output:
"
21-02-15 21:58:11.131 - INFO: Model [SRFlowModel] is created.
21-02-15 21:58:11.131 - INFO: Resuming training from epoch: 2, iter: 51000.
21-02-15 21:58:11.450 - INFO: Start training from epoch: 2, iter: 51000
/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/functional.py:3103: UserWarning: The default behavior for interpolate/upsample with float scale_factor changed in 1.6.0 to align with other frameworks/libraries, and now uses scale_factor directly, instead of relying on the computed output size. If you wish to restore the old behavior, please set recompute_scale_factor=True. See the documentation of nn.Upsample for details.
warnings.warn("The default behavior for interpolate/upsample with float scale_factor changed "
<epoch: 2, iter: 51,001, lr:2.500e-04, t:-1.00e+00, td:9.45e-01, eta:-4.14e+01, nll:-1.566e+01>
<epoch: 2, iter: 51,002, lr:2.500e-04, t:-1.00e+00, td:8.32e-04, eta:-4.14e+01, nll:-1.597e+01>
<epoch: 2, iter: 51,003, lr:2.500e-04, t:1.94e+00, td:2.45e-03, eta:8.01e+01, nll:-1.660e+01>
<epoch: 2, iter: 51,004, lr:2.500e-04, t:1.78e+00, td:3.97e-03, eta:7.37e+01, nll:-1.757e+01>
<epoch: 2, iter: 51,005, lr:2.500e-04, t:1.77e+00, td:8.54e-04, eta:7.32e+01, nll:-1.686e+01>
<epoch: 2, iter: 51,006, lr:2.500e-04, t:2.06e+00, td:6.81e-04, eta:8.52e+01, nll:-1.774e+01>
<epoch: 2, iter: 51,007, lr:2.500e-04, t:1.71e+00, td:1.89e-03, eta:7.06e+01, nll:-1.683e+01>
<epoch: 2, iter: 51,008, lr:2.500e-04, t:1.93e+00, td:2.01e-03, eta:7.98e+01, nll:-1.652e+01>
<epoch: 2, iter: 51,009, lr:2.500e-04, t:1.97e+00, td:2.18e-03, eta:8.16e+01, nll:-1.687e+01>
<epoch: 2, iter: 51,010, lr:2.500e-04, t:1.87e+00, td:2.10e-03, eta:7.72e+01, nll:-1.748e+01>
<epoch: 2, iter: 51,011, lr:2.500e-04, t:1.78e+00, td:3.10e-03, eta:7.36e+01, nll:-1.672e+01>
<epoch: 2, iter: 51,012, lr:2.500e-04, t:2.06e+00, td:3.12e-03, eta:8.51e+01, nll:-1.859e+01>
<epoch: 2, iter: 51,013, lr:2.500e-04, t:1.83e+00, td:2.23e-03, eta:7.57e+01, nll:-1.672e+01>
<epoch: 2, iter: 51,014, lr:2.500e-04, t:1.81e+00, td:2.39e-03, eta:7.50e+01, nll:-1.772e+01>
<epoch: 2, iter: 51,015, lr:2.500e-04, t:1.84e+00, td:1.94e-03, eta:7.60e+01, nll:-1.877e+01>
<epoch: 2, iter: 51,016, lr:2.500e-04, t:1.73e+00, td:3.45e-03, eta:7.17e+01, nll:-1.696e+01>
<epoch: 2, iter: 51,017, lr:2.500e-04, t:1.84e+00, td:2.32e-03, eta:7.62e+01, nll:-1.874e+01>
<epoch: 2, iter: 51,018, lr:2.500e-04, t:2.22e+00, td:2.27e-03, eta:9.18e+01, nll:-1.709e+01>
<epoch: 2, iter: 51,019, lr:2.500e-04, t:1.90e+00, td:1.72e-03, eta:7.87e+01, nll:-1.638e+01>
<epoch: 2, iter: 51,020, lr:2.500e-04, t:1.77e+00, td:2.30e-03, eta:7.31e+01, nll:-1.529e+01>
<epoch: 2, iter: 51,021, lr:2.500e-04, t:1.86e+00, td:3.02e-03, eta:7.70e+01, nll:-1.642e+01>
<epoch: 2, iter: 51,022, lr:2.500e-04, t:1.81e+00, td:2.15e-03, eta:7.48e+01, nll:-1.789e+01>
<epoch: 2, iter: 51,023, lr:2.500e-04, t:1.85e+00, td:2.35e-03, eta:7.65e+01, nll:-1.866e+01>
<epoch: 2, iter: 51,024, lr:2.500e-04, t:1.83e+00, td:2.18e-03, eta:7.57e+01, nll:-1.676e+01>
<epoch: 2, iter: 51,100, lr:2.500e-04, t:1.88e+00, td:2.37e-03, eta:7.78e+01, nll:-1.536e+01>
<epoch: 2, iter: 51,200, lr:2.500e-04, t:1.90e+00, td:2.51e-03, eta:7.86e+01, nll:-1.572e+01>
<epoch: 2, iter: 51,300, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.75e+01, nll:-1.708e+01>
<epoch: 2, iter: 51,400, lr:2.500e-04, t:1.86e+00, td:2.42e-03, eta:7.68e+01, nll:-1.943e+01>
<epoch: 2, iter: 51,500, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.76e+01, nll:-1.640e+01>
<epoch: 2, iter: 51,600, lr:2.500e-04, t:1.87e+00, td:2.39e-03, eta:7.71e+01, nll:-1.571e+01>
<epoch: 2, iter: 51,700, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.74e+01, nll:-1.633e+01>
<epoch: 2, iter: 51,800, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.73e+01, nll:-1.499e+01>
<epoch: 2, iter: 51,900, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.71e+01, nll:-1.538e+01>
<epoch: 2, iter: 52,000, lr:2.500e-04, t:1.87e+00, td:2.40e-03, eta:7.70e+01, nll:-1.629e+01>
21-02-15 22:29:40.137 - INFO: Saving models and training states.
<epoch: 2, iter: 52,100, lr:2.500e-04, t:1.90e+00, td:2.42e-03, eta:7.79e+01, nll:-1.673e+01>
<epoch: 2, iter: 52,200, lr:2.500e-04, t:1.89e+00, td:2.46e-03, eta:7.77e+01, nll:-1.898e+01>
<epoch: 2, iter: 52,300, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.77e+01, nll:-1.815e+01>
<epoch: 2, iter: 52,400, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.64e+01, nll:-1.801e+01>
<epoch: 2, iter: 52,500, lr:2.500e-04, t:1.89e+00, td:2.48e-03, eta:7.76e+01, nll:-1.746e+01>
<epoch: 2, iter: 52,600, lr:2.500e-04, t:1.88e+00, td:2.53e-03, eta:7.70e+01, nll:-1.614e+01>
<epoch: 2, iter: 52,700, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.66e+01, nll:-1.496e+01>
<epoch: 2, iter: 52,800, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.71e+01, nll:-1.682e+01>
<epoch: 2, iter: 52,900, lr:2.500e-04, t:1.87e+00, td:2.48e-03, eta:7.66e+01, nll:-1.676e+01>
<epoch: 2, iter: 53,000, lr:2.500e-04, t:1.87e+00, td:2.42e-03, eta:7.62e+01, nll:-1.719e+01>
21-02-15 23:01:01.845 - INFO: Saving models and training states.
<epoch: 2, iter: 53,100, lr:2.500e-04, t:1.87e+00, td:2.37e-03, eta:7.62e+01, nll:-1.640e+01>
<epoch: 2, iter: 53,200, lr:2.500e-04, t:1.87e+00, td:2.41e-03, eta:7.64e+01, nll:-1.765e+01>
<epoch: 2, iter: 53,300, lr:2.500e-04, t:1.89e+00, td:2.45e-03, eta:7.69e+01, nll:-1.725e+01>
<epoch: 2, iter: 53,400, lr:2.500e-04, t:1.89e+00, td:2.45e-03, eta:7.70e+01, nll:-1.702e+01>
<epoch: 2, iter: 53,500, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.64e+01, nll:-1.803e+01>
<epoch: 2, iter: 53,600, lr:2.500e-04, t:1.88e+00, td:2.42e-03, eta:7.65e+01, nll:-1.760e+01>
<epoch: 2, iter: 53,700, lr:2.500e-04, t:1.86e+00, td:2.40e-03, eta:7.57e+01, nll:-1.747e+01>
<epoch: 2, iter: 53,800, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.66e+01, nll:-2.144e+01>
<epoch: 2, iter: 53,900, lr:2.500e-04, t:1.90e+00, td:2.43e-03, eta:7.72e+01, nll:-1.826e+01>
<epoch: 2, iter: 54,000, lr:2.500e-04, t:1.88e+00, td:2.40e-03, eta:7.64e+01, nll:-1.700e+01>
21-02-15 23:32:23.089 - INFO: Saving models and training states.
<epoch: 2, iter: 54,100, lr:2.500e-04, t:1.90e+00, td:2.55e-03, eta:7.70e+01, nll:-1.809e+01>
<epoch: 2, iter: 54,200, lr:2.500e-04, t:1.89e+00, td:2.46e-03, eta:7.64e+01, nll:-1.832e+01>
<epoch: 2, iter: 54,300, lr:2.500e-04, t:1.90e+00, td:2.47e-03, eta:7.67e+01, nll:-1.641e+01>
<epoch: 2, iter: 54,400, lr:2.500e-04, t:1.89e+00, td:2.47e-03, eta:7.63e+01, nll:-1.669e+01>
<epoch: 2, iter: 54,500, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.60e+01, nll:-1.491e+01>
<epoch: 2, iter: 54,600, lr:2.500e-04, t:1.88e+00, td:2.46e-03, eta:7.58e+01, nll:-1.798e+01>
<epoch: 2, iter: 54,700, lr:2.500e-04, t:1.90e+00, td:2.41e-03, eta:7.65e+01, nll:-1.596e+01>
<epoch: 2, iter: 54,800, lr:2.500e-04, t:1.88e+00, td:2.53e-03, eta:7.59e+01, nll:-1.580e+01>
<epoch: 2, iter: 54,900, lr:2.500e-04, t:1.88e+00, td:2.40e-03, eta:7.58e+01, nll:-1.713e+01>
<epoch: 2, iter: 55,000, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.62e+01, nll:-1.874e+01>
21-02-16 00:03:50.708 - INFO: Saving models and training states.
<epoch: 2, iter: 55,100, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.62e+01, nll:-1.506e+01>
<epoch: 2, iter: 55,200, lr:2.500e-04, t:1.88e+00, td:2.39e-03, eta:7.55e+01, nll:-1.786e+01>
<epoch: 2, iter: 55,300, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.58e+01, nll:-1.834e+01>
<epoch: 2, iter: 55,400, lr:2.500e-04, t:1.88e+00, td:2.39e-03, eta:7.54e+01, nll:-1.841e+01>
<epoch: 2, iter: 55,500, lr:2.500e-04, t:1.89e+00, td:2.41e-03, eta:7.58e+01, nll:-1.820e+01>
<epoch: 2, iter: 55,600, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.51e+01, nll:-1.633e+01>
<epoch: 2, iter: 55,700, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.57e+01, nll:-1.660e+01>
<epoch: 2, iter: 55,800, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.59e+01, nll:-1.856e+01>
<epoch: 2, iter: 55,900, lr:2.500e-04, t:1.90e+00, td:2.45e-03, eta:7.59e+01, nll:-1.613e+01>
CUBLAS error: out of memory (3) in magma_sgetrf_gpu_expert at /opt/conda/conda-bld/magma-cuda102_1583546904148/work/src/sgetrf_gpu.cpp:126
CUBLAS error: not initialized (1) in magma_sgetrf_gpu_expert at /opt/conda/conda-bld/magma-cuda102_1583546904148/work/src/sgetrf_gpu.cpp:126
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
Caught RuntimeError in replica 1 on device 1.
Original Traceback (most recent call last):
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/parallel/parallel_apply.py", line 61, in _worker
output = module(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/SRFlowNet_arch.py", line 65, in forward
y_onehot=y_label)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/SRFlowNet_arch.py", line 101, in normal_flow
y_onehot=y_onehot)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowUpsamplerNet.py", line 213, in forward
z, logdet = self.encode(gt, rrdbResults, logdet=logdet, epses=epses, y_onehot=y_onehot)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowUpsamplerNet.py", line 238, in encode
fl_fea, logdet = layer(fl_fea, logdet, reverse=reverse, rrdbResults=level_conditionals[level])
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 84, in forward
return self.normal_flow(input, logdet, rrdbResults)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 103, in normal_flow
self, z, logdet, False)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 35, in
"invconv": lambda obj, z, logdet, rev: obj.invconv(z, logdet, rev),
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 48, in forward
weight, dlogdet = self.get_weight(input, reverse)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 37, in get_weight
dlogdet = torch.slogdet(self.weight)[1] * pixels
RuntimeError: CUDA error: resource already mapped

<epoch: 2, iter: 56,000, lr:2.500e-04, t:1.86e+00, td:2.41e-03, eta:7.44e+01, nll:-1.589e+01>
21-02-16 00:35:13.687 - INFO: Saving models and training states.
<epoch: 2, iter: 56,100, lr:2.500e-04, t:1.88e+00, td:2.41e-03, eta:7.51e+01, nll:-1.545e+01>
<epoch: 2, iter: 56,200, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.48e+01, nll:-1.524e+01>
<epoch: 2, iter: 56,300, lr:2.500e-04, t:1.88e+00, td:2.50e-03, eta:7.49e+01, nll:-1.727e+01>
<epoch: 2, iter: 56,400, lr:2.500e-04, t:1.85e+00, td:2.40e-03, eta:7.40e+01, nll:-1.717e+01>
<epoch: 2, iter: 56,500, lr:2.500e-04, t:1.88e+00, td:2.48e-03, eta:7.48e+01, nll:-1.548e+01>
<epoch: 2, iter: 56,600, lr:2.500e-04, t:1.86e+00, td:2.48e-03, eta:7.42e+01, nll:-1.752e+01>
<epoch: 2, iter: 56,700, lr:2.500e-04, t:1.88e+00, td:2.48e-03, eta:7.47e+01, nll:-1.669e+01>
<epoch: 2, iter: 56,800, lr:2.500e-04, t:1.86e+00, td:2.43e-03, eta:7.40e+01, nll:-1.632e+01>
<epoch: 2, iter: 56,900, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.40e+01, nll:-1.778e+01>
<epoch: 2, iter: 57,000, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.47e+01, nll:-1.696e+01>
21-02-16 01:06:23.673 - INFO: Saving models and training states.
<epoch: 2, iter: 57,100, lr:2.500e-04, t:1.90e+00, td:2.48e-03, eta:7.54e+01, nll:-1.575e+01>
<epoch: 2, iter: 57,200, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.39e+01, nll:-1.667e+01>
<epoch: 2, iter: 57,300, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.47e+01, nll:-1.871e+01>
<epoch: 2, iter: 57,400, lr:2.500e-04, t:1.88e+00, td:2.50e-03, eta:7.44e+01, nll:-1.781e+01>
<epoch: 2, iter: 57,500, lr:2.500e-04, t:1.86e+00, td:2.43e-03, eta:7.36e+01, nll:-1.881e+01>
<epoch: 2, iter: 57,600, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.38e+01, nll:-1.742e+01>
<epoch: 2, iter: 57,700, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.38e+01, nll:-1.726e+01>
<epoch: 2, iter: 57,800, lr:2.500e-04, t:1.86e+00, td:2.42e-03, eta:7.34e+01, nll:-1.844e+01>
<epoch: 2, iter: 57,900, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.36e+01, nll:-1.622e+01>
<epoch: 2, iter: 58,000, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.38e+01, nll:-1.635e+01>
21-02-16 01:37:34.238 - INFO: Saving models and training states.
<epoch: 2, iter: 58,100, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.44e+01, nll:-1.692e+01>
<epoch: 2, iter: 58,200, lr:2.500e-04, t:1.84e+00, td:2.40e-03, eta:7.25e+01, nll:-1.594e+01>
<epoch: 2, iter: 58,300, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.36e+01, nll:-1.747e+01>
<epoch: 2, iter: 58,400, lr:2.500e-04, t:1.87e+00, td:2.49e-03, eta:7.34e+01, nll:-1.949e+01>
<epoch: 2, iter: 58,500, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.32e+01, nll:-1.595e+01>
<epoch: 2, iter: 58,600, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.32e+01, nll:-1.600e+01>
<epoch: 2, iter: 58,700, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.39e+01, nll:-1.668e+01>
<epoch: 2, iter: 58,800, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.33e+01, nll:-1.868e+01>
<epoch: 2, iter: 58,900, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.28e+01, nll:-1.802e+01>
<epoch: 2, iter: 59,000, lr:2.500e-04, t:1.86e+00, td:2.47e-03, eta:7.27e+01, nll:-1.569e+01>
21-02-16 02:08:39.673 - INFO: Saving models and training states.
<epoch: 2, iter: 59,100, lr:2.500e-04, t:1.87e+00, td:2.42e-03, eta:7.34e+01, nll:-1.721e+01>
<epoch: 2, iter: 59,200, lr:2.500e-04, t:1.84e+00, td:2.39e-03, eta:7.21e+01, nll:-1.866e+01>
<epoch: 2, iter: 59,300, lr:2.500e-04, t:1.85e+00, td:2.47e-03, eta:7.22e+01, nll:-1.685e+01>
<epoch: 2, iter: 59,400, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.35e+01, nll:-1.809e+01>
<epoch: 2, iter: 59,500, lr:2.500e-04, t:1.88e+00, td:2.42e-03, eta:7.35e+01, nll:-1.618e+01>
Fatal Python error: Bus error

Thread 0x00007f2e50c69700 (most recent call first):
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 37 in get_weight
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 48 in fFatal Python error: orwarSegmentation faultd

File "/run/meSegmentation fault (core dumped)"
(myenv) (python37) [root@master code]#

what should i do ?

Image manipulations

Hi, the work use latent space normalization(global, local, spatial) for image manipulations , but I can not find code of this part in the project, could you provide the code for image manipulations, thanks.

Why NLL is negative during the training?

Thanks for your impressive work. During the training process, we found that the output NLL is negative. But theoretically, NLL should be positive. Is there any explanation for this?

Installation error

I am sorry, I'm not very knowledgeable about python to say the least. I tried installing your package by following the setup script on my Windows 10 environment, but I'm getting an error during the requirements installation, namely with torch.

Collecting torch==1.7.0
  Downloading torch-1.7.0-cp37-cp37m-win_amd64.whl (184.0 MB)
     |████████████████████████████████| 184.0 MB 1.3 MB/s
ERROR: torch has an invalid wheel, .dist-info directory not found

I understand this is not directly related to your code, but I have nowhere else to turn. I did upgrade pip and run the install in a fresh venv.

NLL Loss Down and Up During Training

Hi authors,

Thanks for sharing your work. I am trying to extend this model into other datasets. During training, the nll loss first decreases from 1 to -4.5 in the first 3k steps, and then goes back to -0.5 in the later 3k steps. A 10 times larger or smaller learning rate does not prevent this phenomenon.

Have you observed similar results, or any hints available?

Thanks for your help

train error: svd_cuda

Hello, thank you for your work.When I tried to train with my own datasets, I used the SRFlow_DF2K_4x model, but during the training process, "svd_cuda: the updating process of SBDSDC did not converage (error:11)" error appeared.Have you ever encountered the same reason in your training?Is there a solution?
image

Thanks for your advice ! I have tried those experiments:

Thanks for your advice ! I have tried those experiments:

  1. I have decoded the z-LR pair with z of standard deviation [0, 0.1, 0.2, ... 1.0, 1.1]. When deviation is less than 0.5, The black reigions disappear. However, the image is blurry and lacks details.

2.I also have added a small number to the division to protect against divide by zero . But tn the training stage, "RuntimeError: 'DivBackWard0' return nan values in its 1th output." still exists.

For Adding Nosie :
I have added noise to the HR image that you mentioned . And The pixel value in the input HR ranges [0, 255] , [-0.5, 0.5] or [0, 1] ?

Originally posted by @Alan-xw in #2 (comment)

How to improve the results

Hi @andreas128
Thanks for your great work and sharing the code.
I tried to train SRFlow using REDS dataset. After training the RRDBN network on REDS (with PSNR around 29 dB), I used this config to train SRFlow. However, the resulted SR images are not good and the PSNR is around 17 dB.
I did not change the code by the way. Could you please give me a hint on how can I improve the training?

network_G:
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 25
upscale: 4
train_RRDB: false
train_RRDB_delay: 0.5

flow:
K: 10
L: 3
noInitialInj: true
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:
enable: true
fea_up0: true
stackRRDB:
blocks: [ 1,5, 15, 22 ]
concat: true

train:
manual_seed: 1386
lr_G: !!float 1e-5
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1 # no warm up
lr_steps_rel: [ 0.5, 0.75, 0.9, 0.95 ]
lr_gamma: 0.8
weight_l1: 0.1

niter: 300000
val_freq: !!float 5e3

validation settings

val:
heats: [ 0.0, 0.5]
n_sample: 1

How to optimize the jacobian matrix?

Hi, I'm confused about how to optimize the Jacobian. I've read several papers related to invertible networks, and they usually use the determinant of Jacobian as loss. But I found there is only nll loss in your code , it seems Jacobian is useless?

meet NAN when I use realSR dataset to train SrFlow

At the early beginning, the loss like this
image
But after 3 k epoch, the loss meet NaN problem ,like this:
image
I use pretrain model "SRFlow_DF2K_4X.pth" as initial weights. Could you give me some suggestion to solve this problem?

Is there something wrong in original code?

Skipping ERROR caught in nll = model.optimize_parameters(current_step):
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias True
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs True
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs] will not optimize.

And I don't know what caused this problem。

Why add noise and how to calculate the logdet of this operation

if add_gt_noise:
# Setup
noiseQuant = opt_get(self.opt, ['network_G', 'flow', 'augmentation', 'noiseQuant'], True)
if noiseQuant:
z = z + ((torch.rand(z.shape, device=z.device) - 0.5) / self.quant)
logdet = logdet + float(-np.log(self.quant) * pixels)

According to the lines, uniform noise between [-0.5,0.5) is added into the input image, is the operation necessary? And why the logdet should be updated like this? When I read the paper, I noticed the description in the trainning details section as below, but i am not sure whether it's implemented by the mentioned lines or not.

adding slight random noise to the target image helps the training process and leads to better visual results. We therefore add Gaussian noise with a standard deviation of $\sigma = 4/ \sqrt 3$ to the high-resolution image. In contrast to [21], we do not employ 5-bit quantization.

how to calculate loss

Are we assumed to use final training output to calculate the loss as in eq (3)?
In GLOW, (-loss / (log(2) * n_pixel)).mean() is implemented as well. It is not mentioned in SRFLOW, do we need similar processing steps?

RuntimeError: Error(s) in loading state_dict for SRFlowNet:

Thx author again!I have train x4 nettwork and then train x8 network.
source myenv/bin/activate
cd code
python train.py -opt ./confs/SRFlow_DF2K_4X.yml
python train.py -opt ./confs/SRFlow_DF2K_8X.yml
The wrong is that:

(myenv) (base) ubuntu@ubuntu:~/Desktop/data1024/SRFlow-master/code$ python train.py -opt ./confs/SRFlow_DF2K_8X.yml
OrderedDict([('manual_seed', 10), ('lr_G', 0.0005), ('weight_decay_G', 0), ('beta1', 0.9), ('beta2', 0.99), ('lr_scheme', 'MultiStepLR'), ('warmup_iter', -1), ('lr_steps_rel', [0.5, 0.75, 0.9, 0.95]), ('lr_gamma', 0.5), ('niter', 200000), ('val_freq', 40000), ('lr_steps', [100000, 150000, 180000, 190000])])
Disabled distributed training.
pretrain_model path will be ignored when resuming training.
21-01-27 01:59:25.817 - INFO: name: train
use_tb_logger: True
model: SRFlow
distortion: sr
scale: 8
gpu_ids: [0]
datasets:[
train:[
name: CelebA_160_tr
mode: LRHR_PKL
dataroot_GT: ../datasets/DF2K-tr.pklv4
dataroot_LQ: ../datasets/DF2K-tr_X8.pklv4
quant: 32
use_shuffle: True
n_workers: 3
batch_size: 16
GT_size: 160
use_flip: True
color: RGB
phase: train
scale: 8
data_type: img
]
val:[
name: CelebA_160_va
mode: LRHR_PKL
dataroot_GT: ../datasets/DIV2K-va.pklv4
dataroot_LQ: ../datasets/DIV2K-va_X8.pklv4
quant: 32
n_max: 20
phase: val
scale: 8
data_type: img
]
]
dataroot_GT: ../datasets/div2k-validation-modcrop8-gt
dataroot_LR: ../datasets/div2k-validation-modcrop8-x8
model_path: ../pretrained_models/SRFlow_DF2K_8X.pth
heat: 0.9
network_G:[
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 23
upscale: 8
train_RRDB: False
train_RRDB_delay: 0.5
flow:[
K: 16
L: 4
noInitialInj: True
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:[
enable: True
]
fea_up0: True
stackRRDB:[
blocks: [1, 3, 5, 7]
concat: True
]
]
scale: 8
]
path:[
pretrain_model_G: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models/200000_G.pth
strict_load: True
resume_state: auto
root: /home/ubuntu/Desktop/data1024/SRFlow-master
experiments_root: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
models: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models
training_state: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/training_state
log: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
val_images: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/val_images
]
train:[
manual_seed: 10
lr_G: 0.0005
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1
lr_steps_rel: [0.5, 0.75, 0.9, 0.95]
lr_gamma: 0.5
niter: 200000
val_freq: 40000
lr_steps: [100000, 150000, 180000, 190000]
]
val:[
heats: [0.0, 0.5, 0.75, 1.0]
n_sample: 3
]
test:[
heats: [0.0, 0.7, 0.8, 0.9]
]
logger:[
print_freq: 100
save_checkpoint_freq: 1000.0
]
is_train: True
dist: False

21-01-27 01:59:25.867 - INFO: Random seed: 10
{'name': 'CelebA_160_tr', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DF2K-tr.pklv4', 'dataroot_LQ': '../datasets/DF2K-tr_X8.pklv4', 'quant': 32, 'use_shuffle': True, 'n_workers': 3, 'batch_size': 16, 'GT_size': 160, 'use_flip': True, 'color': 'RGB', 'phase': 'train', 'scale': 8, 'data_type': 'img'}
Loaded 162150 HR images with [0.00, 255.00] in 15.97s from ../datasets/DF2K-tr.pklv4
Loaded 162150 LR images with [0.00, 255.00] in 15.97s from ../datasets/DF2K-tr_X8.pklv4
21-01-27 01:59:41.838 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_tr] is created.
Dataset created
21-01-27 01:59:41.845 - INFO: Number of train images: 162,150, iters: 10,135
21-01-27 01:59:41.845 - INFO: Total epochs needed: 20 for iters 200,000
{'name': 'CelebA_160_va', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DIV2K-va.pklv4', 'dataroot_LQ': '../datasets/DIV2K-va_X8.pklv4', 'quant': 32, 'n_max': 20, 'phase': 'val', 'scale': 8, 'data_type': 'img'}
Loaded 20 HR images with [0.00, 255.00] in 0.87s from ../datasets/DIV2K-va.pklv4
Loaded 20 LR images with [0.00, 255.00] in 0.87s from ../datasets/DIV2K-va_X8.pklv4
21-01-27 01:59:42.711 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_va] is created.
21-01-27 01:59:42.711 - INFO: Number of val images in [CelebA_160_va]: 20
21-01-27 01:59:45.179 - INFO: Network G structure: DataParallel - SRFlowNet, with parameters: 50,821,891
21-01-27 01:59:45.180 - INFO: SRFlowNet(
(RRDB): RRDBNet(
(conv_first): Conv2d(3, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(RRDB_trunk): Sequential(
(0): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(1): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(2): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(3): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(4): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(5): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(6): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(7): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(8): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(9): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(10): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(11): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(12): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(13): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(14): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(15): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(16): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(17): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(18): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(19): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(20): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(21): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(22): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
)
(trunk_conv): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv1): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv2): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv3): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(HRconv): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv_last): Conv2d(64, 3, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(flowUpsamplerNet): FlowUpsamplerNet(
(layers): ModuleList(
(0): SqueezeLayer()
(1): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(2): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(3): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(4): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(5): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(6): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(7): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(8): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(9): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(10): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(11): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(12): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(13): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(14): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(15): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(16): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(17): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(18): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(19): Split2d(
(conv): Conv2dZeros(6, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(20): SqueezeLayer()
(21): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(22): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(23): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(24): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(25): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(26): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(27): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(28): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(29): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(30): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(31): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(32): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(33): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(34): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(35): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(36): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(37): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(38): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(39): Split2d(
(conv): Conv2dZeros(12, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(40): SqueezeLayer()
(41): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(42): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(43): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(44): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(45): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(46): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(47): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(48): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(49): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(50): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(51): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(52): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(53): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(54): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(55): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(56): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(57): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(58): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(59): SqueezeLayer()
(60): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(61): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(62): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(63): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(64): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(65): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(66): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(67): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(68): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(69): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(70): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(71): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(72): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(73): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(74): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(75): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(76): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(77): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
)
(f): Sequential(
(0): Conv2d(320, 96, kernel_size=[3, 3], stride=(1, 1), padding=(1, 1))
)
)
)
Traceback (most recent call last):
File "train.py", line 324, in
main()
File "train.py", line 158, in main
model = create_model(opt, current_step)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/init.py", line 50, in create_model
m = M(opt, step)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/SRFlow_model.py", line 58, in init
self.load()
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/SRFlow_model.py", line 267, in load
self.load_network(get_resume_model_path, self.netG, strict=True, submodule=None)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/base_model.py", line 124, in load_network
network.load_state_dict(load_net_clean, strict=strict)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/myenv/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in load_state_dict
raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(
RuntimeError: Error(s) in loading state_dict for SRFlowNet:
Missing key(s) in state_dict: "RRDB.upconv3.weight", "RRDB.upconv3.bias", "flowUpsamplerNet.layers.39.conv.weight", "flowUpsamplerNet.layers.39.conv.bias", "flowUpsamplerNet.layers.39.conv.logs", "flowUpsamplerNet.layers.58.actnorm.bias", "flowUpsamplerNet.layers.58.actnorm.logs", "flowUpsamplerNet.layers.58.invconv.weight", "flowUpsamplerNet.layers.58.affine.fAffine.0.weight", "flowUpsamplerNet.layers.58.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fAffine.2.weight", "flowUpsamplerNet.layers.58.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fAffine.4.weight", "flowUpsamplerNet.layers.58.affine.fAffine.4.bias", "flowUpsamplerNet.layers.58.affine.fAffine.4.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.60.actnorm.bias", "flowUpsamplerNet.layers.60.actnorm.logs", "flowUpsamplerNet.layers.60.invconv.weight", "flowUpsamplerNet.layers.61.actnorm.bias", "flowUpsamplerNet.layers.61.actnorm.logs", "flowUpsamplerNet.layers.61.invconv.weight", "flowUpsamplerNet.layers.62.actnorm.bias", "flowUpsamplerNet.layers.62.actnorm.logs", "flowUpsamplerNet.layers.62.invconv.weight", "flowUpsamplerNet.layers.62.affine.fAffine.0.weight", "flowUpsamplerNet.layers.62.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fAffine.2.weight", "flowUpsamplerNet.layers.62.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fAffine.4.weight", "flowUpsamplerNet.layers.62.affine.fAffine.4.bias", "flowUpsamplerNet.layers.62.affine.fAffine.4.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.63.actnorm.bias", "flowUpsamplerNet.layers.63.actnorm.logs", "flowUpsamplerNet.layers.63.invconv.weight", "flowUpsamplerNet.layers.63.affine.fAffine.0.weight", "flowUpsamplerNet.layers.63.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fAffine.2.weight", "flowUpsamplerNet.layers.63.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fAffine.4.weight", "flowUpsamplerNet.layers.63.affine.fAffine.4.bias", "flowUpsamplerNet.layers.63.affine.fAffine.4.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.64.actnorm.bias", "flowUpsamplerNet.layers.64.actnorm.logs", "flowUpsamplerNet.layers.64.invconv.weight", "flowUpsamplerNet.layers.64.affine.fAffine.0.weight", "flowUpsamplerNet.layers.64.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fAffine.2.weight", "flowUpsamplerNet.layers.64.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fAffine.4.weight", "flowUpsamplerNet.layers.64.affine.fAffine.4.bias", "flowUpsamplerNet.layers.64.affine.fAffine.4.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.65.actnorm.bias", "flowUpsamplerNet.layers.65.actnorm.logs", "flowUpsamplerNet.layers.65.invconv.weight", "flowUpsamplerNet.layers.65.affine.fAffine.0.weight", "flowUpsamplerNet.layers.65.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fAffine.2.weight", "flowUpsamplerNet.layers.65.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fAffine.4.weight", "flowUpsamplerNet.layers.65.affine.fAffine.4.bias", "flowUpsamplerNet.layers.65.affine.fAffine.4.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.66.actnorm.bias", "flowUpsamplerNet.layers.66.actnorm.logs", "flowUpsamplerNet.layers.66.invconv.weight", "flowUpsamplerNet.layers.66.affine.fAffine.0.weight", "flowUpsamplerNet.layers.66.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fAffine.2.weight", "flowUpsamplerNet.layers.66.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fAffine.4.weight", "flowUpsamplerNet.layers.66.affine.fAffine.4.bias", "flowUpsamplerNet.layers.66.affine.fAffine.4.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.67.actnorm.bias", "flowUpsamplerNet.layers.67.actnorm.logs", "flowUpsamplerNet.layers.67.invconv.weight", "flowUpsamplerNet.layers.67.affine.fAffine.0.weight", "flowUpsamplerNet.layers.67.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fAffine.2.weight", "flowUpsamplerNet.layers.67.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fAffine.4.weight", "flowUpsamplerNet.layers.67.affine.fAffine.4.bias", "flowUpsamplerNet.layers.67.affine.fAffine.4.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.68.actnorm.bias", "flowUpsamplerNet.layers.68.actnorm.logs", "flowUpsamplerNet.layers.68.invconv.weight", "flowUpsamplerNet.layers.68.affine.fAffine.0.weight", "flowUpsamplerNet.layers.68.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fAffine.2.weight", "flowUpsamplerNet.layers.68.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fAffine.4.weight", "flowUpsamplerNet.layers.68.affine.fAffine.4.bias", "flowUpsamplerNet.layers.68.affine.fAffine.4.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.69.actnorm.bias", "flowUpsamplerNet.layers.69.actnorm.logs", "flowUpsamplerNet.layers.69.invconv.weight", "flowUpsamplerNet.layers.69.affine.fAffine.0.weight", "flowUpsamplerNet.layers.69.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fAffine.2.weight", "flowUpsamplerNet.layers.69.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fAffine.4.weight", "flowUpsamplerNet.layers.69.affine.fAffine.4.bias", "flowUpsamplerNet.layers.69.affine.fAffine.4.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.70.actnorm.bias", "flowUpsamplerNet.layers.70.actnorm.logs", "flowUpsamplerNet.layers.70.invconv.weight", "flowUpsamplerNet.layers.70.affine.fAffine.0.weight", "flowUpsamplerNet.layers.70.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fAffine.2.weight", "flowUpsamplerNet.layers.70.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fAffine.4.weight", "flowUpsamplerNet.layers.70.affine.fAffine.4.bias", "flowUpsamplerNet.layers.70.affine.fAffine.4.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.71.actnorm.bias", "flowUpsamplerNet.layers.71.actnorm.logs", "flowUpsamplerNet.layers.71.invconv.weight", "flowUpsamplerNet.layers.71.affine.fAffine.0.weight", "flowUpsamplerNet.layers.71.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fAffine.2.weight", "flowUpsamplerNet.layers.71.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fAffine.4.weight", "flowUpsamplerNet.layers.71.affine.fAffine.4.bias", "flowUpsamplerNet.layers.71.affine.fAffine.4.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.72.actnorm.bias", "flowUpsamplerNet.layers.72.actnorm.logs", "flowUpsamplerNet.layers.72.invconv.weight", "flowUpsamplerNet.layers.72.affine.fAffine.0.weight", "flowUpsamplerNet.layers.72.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fAffine.2.weight", "flowUpsamplerNet.layers.72.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fAffine.4.weight", "flowUpsamplerNet.layers.72.affine.fAffine.4.bias", "flowUpsamplerNet.layers.72.affine.fAffine.4.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.73.actnorm.bias", "flowUpsamplerNet.layers.73.actnorm.logs", "flowUpsamplerNet.layers.73.invconv.weight", "flowUpsamplerNet.layers.73.affine.fAffine.0.weight", "flowUpsamplerNet.layers.73.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fAffine.2.weight", "flowUpsamplerNet.layers.73.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fAffine.4.weight", "flowUpsamplerNet.layers.73.affine.fAffine.4.bias", "flowUpsamplerNet.layers.73.affine.fAffine.4.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.74.actnorm.bias", "flowUpsamplerNet.layers.74.actnorm.logs", "flowUpsamplerNet.layers.74.invconv.weight", "flowUpsamplerNet.layers.74.affine.fAffine.0.weight", "flowUpsamplerNet.layers.74.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fAffine.2.weight", "flowUpsamplerNet.layers.74.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fAffine.4.weight", "flowUpsamplerNet.layers.74.affine.fAffine.4.bias", "flowUpsamplerNet.layers.74.affine.fAffine.4.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.75.actnorm.bias", "flowUpsamplerNet.layers.75.actnorm.logs", "flowUpsamplerNet.layers.75.invconv.weight", "flowUpsamplerNet.layers.75.affine.fAffine.0.weight", "flowUpsamplerNet.layers.75.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fAffine.2.weight", "flowUpsamplerNet.layers.75.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fAffine.4.weight", "flowUpsamplerNet.layers.75.affine.fAffine.4.bias", "flowUpsamplerNet.layers.75.affine.fAffine.4.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.76.actnorm.bias", "flowUpsamplerNet.layers.76.actnorm.logs", "flowUpsamplerNet.layers.76.invconv.weight", "flowUpsamplerNet.layers.76.affine.fAffine.0.weight", "flowUpsamplerNet.layers.76.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fAffine.2.weight", "flowUpsamplerNet.layers.76.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fAffine.4.weight", "flowUpsamplerNet.layers.76.affine.fAffine.4.bias", "flowUpsamplerNet.layers.76.affine.fAffine.4.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.77.actnorm.bias", "flowUpsamplerNet.layers.77.actnorm.logs", "flowUpsamplerNet.layers.77.invconv.weight", "flowUpsamplerNet.layers.77.affine.fAffine.0.weight", "flowUpsamplerNet.layers.77.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fAffine.2.weight", "flowUpsamplerNet.layers.77.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fAffine.4.weight", "flowUpsamplerNet.layers.77.affine.fAffine.4.bias", "flowUpsamplerNet.layers.77.affine.fAffine.4.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.4.logs".
Unexpected key(s) in state_dict: "flowUpsamplerNet.layers.40.actnorm.bias", "flowUpsamplerNet.layers.40.actnorm.logs", "flowUpsamplerNet.layers.40.invconv.weight", "flowUpsamplerNet.layers.42.affine.fAffine.0.weight", "flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fAffine.2.weight", "flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fAffine.4.weight", "flowUpsamplerNet.layers.42.affine.fAffine.4.bias", "flowUpsamplerNet.layers.42.affine.fAffine.4.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.4.logs".
size mismatch for flowUpsamplerNet.layers.41.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.41.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.41.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.42.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.42.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.42.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.43.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).

How to crop CelebA dataset for training

Thanks for your great work!
I have a question about training using CelebA dataset
In SRFlow paper,
we pre-process the dataset by cropping aligned patches, which are
resized to the HR resolution of 160 × 160.

I am wondering how to crop celeba aligned dataset!
Thank you!

Meet "NaN" Problem.

Hey,
I implement the SRFlow based on the paper and Glow source code with Pytorch. But I encounter the "NaN" problem during testing. I put the testing LR image and generated z_samples into SRflow reverse net. But It will generate "NaN" value . Have you ever met this issue?

Looking forward to your reply.!

how to solve it?

Skipping ERROR caught in nll = model.optimize_parameters(current_step):
Given groups=1, weight if size [64,3,3,3], expected input[1,4,40,40] to have 3 channels, but got 4 channels instead

How does the Normalizing Flow loss in SRFlow behave?

I followed this repo, paper and glow paper. My loss looks like this

    _logp =  GaussianDiag.logp(None, None, (z))
    obj = logdets + logp
    loss = -obj / (tf.math.log(2.) * tf.cast(pixels, tf.float32))_

There are 2 terms added together : one is logp which is always negative, and logdets which starts with a negative number during training, because of the minus sign in -obj the loss is positive, but after training for around 100k steps, both the logp and logdets increases. The logp term increases and becomes positive number and hence the final loss becomes negative around -3.xxx. Just wanted to know whether it is an expected behaviour ?

Skipping ERROR caught in nll = model.optimize_parameters(current_step): svd_cuda: the updating process of SBDSDC did not converge (error: 23)

Thx author!I train x4 is ok!
But when I train x8:
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
svd_cuda: the updating process of SBDSDC did not converge (error: 23)

Python 3.8.5 (default, Sep 4 2020, 07:30:14)
[GCC 7.3.0] :: Anaconda, Inc. on linux
Type "help", "copyright", "credits" or "license" for more information.

import torch
print(torch.version)
1.7.1+cu110
print(torch.version.cuda)
11.0

print(torch.backends.cudnn.version())
8005

············································································································
My GPU is 3090.I run setup code and find the cuda version is not compare.After that I reinstall with (pip install torch==1.7.1+cu110 torchvision==0.8.2+cu110 torchaudio===0.7.2

About iters is 10000.

Super Resolution Video

Hi, may I know if it's possible to get super resolution videos with SRFlow? Kindly let me know if you have any resources to share. Thank you!

Quantitative evaluation

Hi,

Thanks for your great and inspiring work!
Just a question about how you do the evaluation for your model. Since it is possible to sample different SR images for a given LR image, what is your methodology to decide which sample is chosen as the output SR image for evaluation? Do you simply draw one random sample or do you draw multiple samples and choose the one with the best result?

please upload the training and test source code

Dear author:
You SRFlow work is really insightful. Lots of researchers just like me are looking forward to try your models. Please kindly share the source code asap. Thank you very much!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.