andreas128 / srflow Goto Github PK
View Code? Open in Web Editor NEWOfficial SRFlow training code: Super-Resolution using Normalizing Flow in PyTorch
License: Other
Official SRFlow training code: Super-Resolution using Normalizing Flow in PyTorch
License: Other
Hi, may I know if it's possible to get super resolution videos with SRFlow? Kindly let me know if you have any resources to share. Thank you!
Hi, the work use latent space normalization(global, local, spatial) for image manipulations , but I can not find code of this part in the project, could you provide the code for image manipulations, thanks.
Any ETA on that?
Hi, I'm confused about how to optimize the Jacobian. I've read several papers related to invertible networks, and they usually use the determinant of Jacobian as loss. But I found there is only nll loss in your code , it seems Jacobian is useless?
Thx author again!I have train x4 nettwork and then train x8 network.
source myenv/bin/activate
cd code
python train.py -opt ./confs/SRFlow_DF2K_4X.yml
python train.py -opt ./confs/SRFlow_DF2K_8X.yml
The wrong is that:
(myenv) (base) ubuntu@ubuntu:~/Desktop/data1024/SRFlow-master/code$ python train.py -opt ./confs/SRFlow_DF2K_8X.yml
OrderedDict([('manual_seed', 10), ('lr_G', 0.0005), ('weight_decay_G', 0), ('beta1', 0.9), ('beta2', 0.99), ('lr_scheme', 'MultiStepLR'), ('warmup_iter', -1), ('lr_steps_rel', [0.5, 0.75, 0.9, 0.95]), ('lr_gamma', 0.5), ('niter', 200000), ('val_freq', 40000), ('lr_steps', [100000, 150000, 180000, 190000])])
Disabled distributed training.
pretrain_model path will be ignored when resuming training.
21-01-27 01:59:25.817 - INFO: name: train
use_tb_logger: True
model: SRFlow
distortion: sr
scale: 8
gpu_ids: [0]
datasets:[
train:[
name: CelebA_160_tr
mode: LRHR_PKL
dataroot_GT: ../datasets/DF2K-tr.pklv4
dataroot_LQ: ../datasets/DF2K-tr_X8.pklv4
quant: 32
use_shuffle: True
n_workers: 3
batch_size: 16
GT_size: 160
use_flip: True
color: RGB
phase: train
scale: 8
data_type: img
]
val:[
name: CelebA_160_va
mode: LRHR_PKL
dataroot_GT: ../datasets/DIV2K-va.pklv4
dataroot_LQ: ../datasets/DIV2K-va_X8.pklv4
quant: 32
n_max: 20
phase: val
scale: 8
data_type: img
]
]
dataroot_GT: ../datasets/div2k-validation-modcrop8-gt
dataroot_LR: ../datasets/div2k-validation-modcrop8-x8
model_path: ../pretrained_models/SRFlow_DF2K_8X.pth
heat: 0.9
network_G:[
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 23
upscale: 8
train_RRDB: False
train_RRDB_delay: 0.5
flow:[
K: 16
L: 4
noInitialInj: True
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:[
enable: True
]
fea_up0: True
stackRRDB:[
blocks: [1, 3, 5, 7]
concat: True
]
]
scale: 8
]
path:[
pretrain_model_G: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models/200000_G.pth
strict_load: True
resume_state: auto
root: /home/ubuntu/Desktop/data1024/SRFlow-master
experiments_root: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
models: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models
training_state: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/training_state
log: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
val_images: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/val_images
]
train:[
manual_seed: 10
lr_G: 0.0005
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1
lr_steps_rel: [0.5, 0.75, 0.9, 0.95]
lr_gamma: 0.5
niter: 200000
val_freq: 40000
lr_steps: [100000, 150000, 180000, 190000]
]
val:[
heats: [0.0, 0.5, 0.75, 1.0]
n_sample: 3
]
test:[
heats: [0.0, 0.7, 0.8, 0.9]
]
logger:[
print_freq: 100
save_checkpoint_freq: 1000.0
]
is_train: True
dist: False
21-01-27 01:59:25.867 - INFO: Random seed: 10
{'name': 'CelebA_160_tr', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DF2K-tr.pklv4', 'dataroot_LQ': '../datasets/DF2K-tr_X8.pklv4', 'quant': 32, 'use_shuffle': True, 'n_workers': 3, 'batch_size': 16, 'GT_size': 160, 'use_flip': True, 'color': 'RGB', 'phase': 'train', 'scale': 8, 'data_type': 'img'}
Loaded 162150 HR images with [0.00, 255.00] in 15.97s from ../datasets/DF2K-tr.pklv4
Loaded 162150 LR images with [0.00, 255.00] in 15.97s from ../datasets/DF2K-tr_X8.pklv4
21-01-27 01:59:41.838 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_tr] is created.
Dataset created
21-01-27 01:59:41.845 - INFO: Number of train images: 162,150, iters: 10,135
21-01-27 01:59:41.845 - INFO: Total epochs needed: 20 for iters 200,000
{'name': 'CelebA_160_va', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DIV2K-va.pklv4', 'dataroot_LQ': '../datasets/DIV2K-va_X8.pklv4', 'quant': 32, 'n_max': 20, 'phase': 'val', 'scale': 8, 'data_type': 'img'}
Loaded 20 HR images with [0.00, 255.00] in 0.87s from ../datasets/DIV2K-va.pklv4
Loaded 20 LR images with [0.00, 255.00] in 0.87s from ../datasets/DIV2K-va_X8.pklv4
21-01-27 01:59:42.711 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_va] is created.
21-01-27 01:59:42.711 - INFO: Number of val images in [CelebA_160_va]: 20
21-01-27 01:59:45.179 - INFO: Network G structure: DataParallel - SRFlowNet, with parameters: 50,821,891
21-01-27 01:59:45.180 - INFO: SRFlowNet(
(RRDB): RRDBNet(
(conv_first): Conv2d(3, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(RRDB_trunk): Sequential(
(0): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(1): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(2): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(3): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(4): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(5): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(6): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(7): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(8): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(9): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(10): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(11): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(12): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(13): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(14): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(15): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(16): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(17): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(18): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(19): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(20): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(21): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
(22): RRDB(
(RDB1): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB2): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(RDB3): ResidualDenseBlock_5C(
(conv1): Conv2d(64, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv2): Conv2d(96, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv3): Conv2d(128, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv4): Conv2d(160, 32, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv5): Conv2d(192, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
)
)
(trunk_conv): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv1): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv2): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(upconv3): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(HRconv): Conv2d(64, 64, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(conv_last): Conv2d(64, 3, kernel_size=(3, 3), stride=(1, 1), padding=(1, 1))
(lrelu): LeakyReLU(negative_slope=0.2, inplace=True)
)
(flowUpsamplerNet): FlowUpsamplerNet(
(layers): ModuleList(
(0): SqueezeLayer()
(1): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(2): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(3): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(4): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(5): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(6): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(7): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(8): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(9): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(10): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(11): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(12): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(13): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(14): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(15): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(16): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(17): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(18): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
326, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(19): Split2d(
(conv): Conv2dZeros(6, 12, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(20): SqueezeLayer()
(21): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(22): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(23): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(24): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(25): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(26): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(27): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(28): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(29): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(30): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(31): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(32): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(33): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(34): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(35): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(36): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(37): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(38): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
332, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(39): Split2d(
(conv): Conv2dZeros(12, 24, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(40): SqueezeLayer()
(41): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(42): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(43): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(44): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(45): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(46): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(47): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(48): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(49): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(50): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(51): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(52): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(53): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(54): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(55): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(56): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(57): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(58): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
344, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 48, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 96, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(59): SqueezeLayer()
(60): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(61): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
)
(62): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(63): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(64): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(65): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(66): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(67): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(68): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(69): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(70): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(71): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(72): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(73): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(74): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(75): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(76): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
(77): FlowStep(
(actnorm): ActNorm2d()
(invconv): InvertibleConv1x1()
(affine): CondAffineSeparatedAndCond(
(fAffine): Sequential(
(0): Conv2d(
416, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 192, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
(fFeatures): Sequential(
(0): Conv2d(
320, 64, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1], bias=False
(actnorm): ActNorm2d()
)
(1): ReLU()
(2): Conv2d(
64, 64, kernel_size=[1, 1], stride=[1, 1], padding=[0, 0], bias=False
(actnorm): ActNorm2d()
)
(3): ReLU()
(4): Conv2dZeros(64, 384, kernel_size=[3, 3], stride=[1, 1], padding=[1, 1])
)
)
)
)
(f): Sequential(
(0): Conv2d(320, 96, kernel_size=[3, 3], stride=(1, 1), padding=(1, 1))
)
)
)
Traceback (most recent call last):
File "train.py", line 324, in
main()
File "train.py", line 158, in main
model = create_model(opt, current_step)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/init.py", line 50, in create_model
m = M(opt, step)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/SRFlow_model.py", line 58, in init
self.load()
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/SRFlow_model.py", line 267, in load
self.load_network(get_resume_model_path, self.netG, strict=True, submodule=None)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/code/models/base_model.py", line 124, in load_network
network.load_state_dict(load_net_clean, strict=strict)
File "/home/ubuntu/Desktop/data1024/SRFlow-master/myenv/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in load_state_dict
raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(
RuntimeError: Error(s) in loading state_dict for SRFlowNet:
Missing key(s) in state_dict: "RRDB.upconv3.weight", "RRDB.upconv3.bias", "flowUpsamplerNet.layers.39.conv.weight", "flowUpsamplerNet.layers.39.conv.bias", "flowUpsamplerNet.layers.39.conv.logs", "flowUpsamplerNet.layers.58.actnorm.bias", "flowUpsamplerNet.layers.58.actnorm.logs", "flowUpsamplerNet.layers.58.invconv.weight", "flowUpsamplerNet.layers.58.affine.fAffine.0.weight", "flowUpsamplerNet.layers.58.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fAffine.2.weight", "flowUpsamplerNet.layers.58.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fAffine.4.weight", "flowUpsamplerNet.layers.58.affine.fAffine.4.bias", "flowUpsamplerNet.layers.58.affine.fAffine.4.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.58.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.58.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.58.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.60.actnorm.bias", "flowUpsamplerNet.layers.60.actnorm.logs", "flowUpsamplerNet.layers.60.invconv.weight", "flowUpsamplerNet.layers.61.actnorm.bias", "flowUpsamplerNet.layers.61.actnorm.logs", "flowUpsamplerNet.layers.61.invconv.weight", "flowUpsamplerNet.layers.62.actnorm.bias", "flowUpsamplerNet.layers.62.actnorm.logs", "flowUpsamplerNet.layers.62.invconv.weight", "flowUpsamplerNet.layers.62.affine.fAffine.0.weight", "flowUpsamplerNet.layers.62.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fAffine.2.weight", "flowUpsamplerNet.layers.62.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fAffine.4.weight", "flowUpsamplerNet.layers.62.affine.fAffine.4.bias", "flowUpsamplerNet.layers.62.affine.fAffine.4.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.62.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.62.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.62.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.63.actnorm.bias", "flowUpsamplerNet.layers.63.actnorm.logs", "flowUpsamplerNet.layers.63.invconv.weight", "flowUpsamplerNet.layers.63.affine.fAffine.0.weight", "flowUpsamplerNet.layers.63.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fAffine.2.weight", "flowUpsamplerNet.layers.63.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fAffine.4.weight", "flowUpsamplerNet.layers.63.affine.fAffine.4.bias", "flowUpsamplerNet.layers.63.affine.fAffine.4.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.63.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.63.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.63.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.64.actnorm.bias", "flowUpsamplerNet.layers.64.actnorm.logs", "flowUpsamplerNet.layers.64.invconv.weight", "flowUpsamplerNet.layers.64.affine.fAffine.0.weight", "flowUpsamplerNet.layers.64.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fAffine.2.weight", "flowUpsamplerNet.layers.64.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fAffine.4.weight", "flowUpsamplerNet.layers.64.affine.fAffine.4.bias", "flowUpsamplerNet.layers.64.affine.fAffine.4.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.64.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.64.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.64.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.65.actnorm.bias", "flowUpsamplerNet.layers.65.actnorm.logs", "flowUpsamplerNet.layers.65.invconv.weight", "flowUpsamplerNet.layers.65.affine.fAffine.0.weight", "flowUpsamplerNet.layers.65.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fAffine.2.weight", "flowUpsamplerNet.layers.65.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fAffine.4.weight", "flowUpsamplerNet.layers.65.affine.fAffine.4.bias", "flowUpsamplerNet.layers.65.affine.fAffine.4.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.65.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.65.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.65.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.66.actnorm.bias", "flowUpsamplerNet.layers.66.actnorm.logs", "flowUpsamplerNet.layers.66.invconv.weight", "flowUpsamplerNet.layers.66.affine.fAffine.0.weight", "flowUpsamplerNet.layers.66.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fAffine.2.weight", "flowUpsamplerNet.layers.66.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fAffine.4.weight", "flowUpsamplerNet.layers.66.affine.fAffine.4.bias", "flowUpsamplerNet.layers.66.affine.fAffine.4.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.66.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.66.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.66.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.67.actnorm.bias", "flowUpsamplerNet.layers.67.actnorm.logs", "flowUpsamplerNet.layers.67.invconv.weight", "flowUpsamplerNet.layers.67.affine.fAffine.0.weight", "flowUpsamplerNet.layers.67.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fAffine.2.weight", "flowUpsamplerNet.layers.67.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fAffine.4.weight", "flowUpsamplerNet.layers.67.affine.fAffine.4.bias", "flowUpsamplerNet.layers.67.affine.fAffine.4.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.67.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.67.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.67.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.68.actnorm.bias", "flowUpsamplerNet.layers.68.actnorm.logs", "flowUpsamplerNet.layers.68.invconv.weight", "flowUpsamplerNet.layers.68.affine.fAffine.0.weight", "flowUpsamplerNet.layers.68.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fAffine.2.weight", "flowUpsamplerNet.layers.68.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fAffine.4.weight", "flowUpsamplerNet.layers.68.affine.fAffine.4.bias", "flowUpsamplerNet.layers.68.affine.fAffine.4.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.68.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.68.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.68.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.69.actnorm.bias", "flowUpsamplerNet.layers.69.actnorm.logs", "flowUpsamplerNet.layers.69.invconv.weight", "flowUpsamplerNet.layers.69.affine.fAffine.0.weight", "flowUpsamplerNet.layers.69.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fAffine.2.weight", "flowUpsamplerNet.layers.69.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fAffine.4.weight", "flowUpsamplerNet.layers.69.affine.fAffine.4.bias", "flowUpsamplerNet.layers.69.affine.fAffine.4.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.69.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.69.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.69.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.70.actnorm.bias", "flowUpsamplerNet.layers.70.actnorm.logs", "flowUpsamplerNet.layers.70.invconv.weight", "flowUpsamplerNet.layers.70.affine.fAffine.0.weight", "flowUpsamplerNet.layers.70.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fAffine.2.weight", "flowUpsamplerNet.layers.70.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fAffine.4.weight", "flowUpsamplerNet.layers.70.affine.fAffine.4.bias", "flowUpsamplerNet.layers.70.affine.fAffine.4.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.70.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.70.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.70.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.71.actnorm.bias", "flowUpsamplerNet.layers.71.actnorm.logs", "flowUpsamplerNet.layers.71.invconv.weight", "flowUpsamplerNet.layers.71.affine.fAffine.0.weight", "flowUpsamplerNet.layers.71.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fAffine.2.weight", "flowUpsamplerNet.layers.71.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fAffine.4.weight", "flowUpsamplerNet.layers.71.affine.fAffine.4.bias", "flowUpsamplerNet.layers.71.affine.fAffine.4.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.71.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.71.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.71.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.72.actnorm.bias", "flowUpsamplerNet.layers.72.actnorm.logs", "flowUpsamplerNet.layers.72.invconv.weight", "flowUpsamplerNet.layers.72.affine.fAffine.0.weight", "flowUpsamplerNet.layers.72.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fAffine.2.weight", "flowUpsamplerNet.layers.72.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fAffine.4.weight", "flowUpsamplerNet.layers.72.affine.fAffine.4.bias", "flowUpsamplerNet.layers.72.affine.fAffine.4.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.72.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.72.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.72.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.73.actnorm.bias", "flowUpsamplerNet.layers.73.actnorm.logs", "flowUpsamplerNet.layers.73.invconv.weight", "flowUpsamplerNet.layers.73.affine.fAffine.0.weight", "flowUpsamplerNet.layers.73.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fAffine.2.weight", "flowUpsamplerNet.layers.73.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fAffine.4.weight", "flowUpsamplerNet.layers.73.affine.fAffine.4.bias", "flowUpsamplerNet.layers.73.affine.fAffine.4.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.73.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.73.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.73.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.74.actnorm.bias", "flowUpsamplerNet.layers.74.actnorm.logs", "flowUpsamplerNet.layers.74.invconv.weight", "flowUpsamplerNet.layers.74.affine.fAffine.0.weight", "flowUpsamplerNet.layers.74.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fAffine.2.weight", "flowUpsamplerNet.layers.74.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fAffine.4.weight", "flowUpsamplerNet.layers.74.affine.fAffine.4.bias", "flowUpsamplerNet.layers.74.affine.fAffine.4.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.74.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.74.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.74.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.75.actnorm.bias", "flowUpsamplerNet.layers.75.actnorm.logs", "flowUpsamplerNet.layers.75.invconv.weight", "flowUpsamplerNet.layers.75.affine.fAffine.0.weight", "flowUpsamplerNet.layers.75.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fAffine.2.weight", "flowUpsamplerNet.layers.75.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fAffine.4.weight", "flowUpsamplerNet.layers.75.affine.fAffine.4.bias", "flowUpsamplerNet.layers.75.affine.fAffine.4.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.75.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.75.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.75.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.76.actnorm.bias", "flowUpsamplerNet.layers.76.actnorm.logs", "flowUpsamplerNet.layers.76.invconv.weight", "flowUpsamplerNet.layers.76.affine.fAffine.0.weight", "flowUpsamplerNet.layers.76.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fAffine.2.weight", "flowUpsamplerNet.layers.76.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fAffine.4.weight", "flowUpsamplerNet.layers.76.affine.fAffine.4.bias", "flowUpsamplerNet.layers.76.affine.fAffine.4.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.76.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.76.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.76.affine.fFeatures.4.logs", "flowUpsamplerNet.layers.77.actnorm.bias", "flowUpsamplerNet.layers.77.actnorm.logs", "flowUpsamplerNet.layers.77.invconv.weight", "flowUpsamplerNet.layers.77.affine.fAffine.0.weight", "flowUpsamplerNet.layers.77.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fAffine.2.weight", "flowUpsamplerNet.layers.77.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fAffine.4.weight", "flowUpsamplerNet.layers.77.affine.fAffine.4.bias", "flowUpsamplerNet.layers.77.affine.fAffine.4.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.77.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.77.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.77.affine.fFeatures.4.logs".
Unexpected key(s) in state_dict: "flowUpsamplerNet.layers.40.actnorm.bias", "flowUpsamplerNet.layers.40.actnorm.logs", "flowUpsamplerNet.layers.40.invconv.weight", "flowUpsamplerNet.layers.42.affine.fAffine.0.weight", "flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fAffine.2.weight", "flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fAffine.4.weight", "flowUpsamplerNet.layers.42.affine.fAffine.4.bias", "flowUpsamplerNet.layers.42.affine.fAffine.4.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.0.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.2.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs", "flowUpsamplerNet.layers.42.affine.fFeatures.4.weight", "flowUpsamplerNet.layers.42.affine.fFeatures.4.bias", "flowUpsamplerNet.layers.42.affine.fFeatures.4.logs".
size mismatch for flowUpsamplerNet.layers.41.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.41.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.41.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.42.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.42.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.42.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.43.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.43.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.43.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.44.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.44.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.45.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.45.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.46.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.46.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.47.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.47.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.48.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.48.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.49.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.49.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.50.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.50.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.51.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.51.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.52.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.52.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.53.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.53.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.54.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.54.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.55.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.55.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.56.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.56.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.actnorm.bias: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.actnorm.logs: copying a param with shape torch.Size([1, 96, 1, 1]) from checkpoint, the shape in current model is torch.Size([1, 48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.invconv.weight: copying a param with shape torch.Size([96, 96]) from checkpoint, the shape in current model is torch.Size([48, 48]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.0.weight: copying a param with shape torch.Size([64, 368, 3, 3]) from checkpoint, the shape in current model is torch.Size([64, 344, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.weight: copying a param with shape torch.Size([96, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([48, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.bias: copying a param with shape torch.Size([96]) from checkpoint, the shape in current model is torch.Size([48]).
size mismatch for flowUpsamplerNet.layers.57.affine.fAffine.4.logs: copying a param with shape torch.Size([96, 1, 1]) from checkpoint, the shape in current model is torch.Size([48, 1, 1]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.weight: copying a param with shape torch.Size([192, 64, 3, 3]) from checkpoint, the shape in current model is torch.Size([96, 64, 3, 3]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.bias: copying a param with shape torch.Size([192]) from checkpoint, the shape in current model is torch.Size([96]).
size mismatch for flowUpsamplerNet.layers.57.affine.fFeatures.4.logs: copying a param with shape torch.Size([192, 1, 1]) from checkpoint, the shape in current model is torch.Size([96, 1, 1]).
First of all thank you very much for your work .
when I train x4 super-resolution completely with your code, but after a period of training,it will occur "Bus error (core dumped)".when I use "python -X faulthandler train.py -opt
./confs/SRFlow_DF2K_4X.yml"
it will output:
"
21-02-15 21:58:11.131 - INFO: Model [SRFlowModel] is created.
21-02-15 21:58:11.131 - INFO: Resuming training from epoch: 2, iter: 51000.
21-02-15 21:58:11.450 - INFO: Start training from epoch: 2, iter: 51000
/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/functional.py:3103: UserWarning: The default behavior for interpolate/upsample with float scale_factor changed in 1.6.0 to align with other frameworks/libraries, and now uses scale_factor directly, instead of relying on the computed output size. If you wish to restore the old behavior, please set recompute_scale_factor=True. See the documentation of nn.Upsample for details.
warnings.warn("The default behavior for interpolate/upsample with float scale_factor changed "
<epoch: 2, iter: 51,001, lr:2.500e-04, t:-1.00e+00, td:9.45e-01, eta:-4.14e+01, nll:-1.566e+01>
<epoch: 2, iter: 51,002, lr:2.500e-04, t:-1.00e+00, td:8.32e-04, eta:-4.14e+01, nll:-1.597e+01>
<epoch: 2, iter: 51,003, lr:2.500e-04, t:1.94e+00, td:2.45e-03, eta:8.01e+01, nll:-1.660e+01>
<epoch: 2, iter: 51,004, lr:2.500e-04, t:1.78e+00, td:3.97e-03, eta:7.37e+01, nll:-1.757e+01>
<epoch: 2, iter: 51,005, lr:2.500e-04, t:1.77e+00, td:8.54e-04, eta:7.32e+01, nll:-1.686e+01>
<epoch: 2, iter: 51,006, lr:2.500e-04, t:2.06e+00, td:6.81e-04, eta:8.52e+01, nll:-1.774e+01>
<epoch: 2, iter: 51,007, lr:2.500e-04, t:1.71e+00, td:1.89e-03, eta:7.06e+01, nll:-1.683e+01>
<epoch: 2, iter: 51,008, lr:2.500e-04, t:1.93e+00, td:2.01e-03, eta:7.98e+01, nll:-1.652e+01>
<epoch: 2, iter: 51,009, lr:2.500e-04, t:1.97e+00, td:2.18e-03, eta:8.16e+01, nll:-1.687e+01>
<epoch: 2, iter: 51,010, lr:2.500e-04, t:1.87e+00, td:2.10e-03, eta:7.72e+01, nll:-1.748e+01>
<epoch: 2, iter: 51,011, lr:2.500e-04, t:1.78e+00, td:3.10e-03, eta:7.36e+01, nll:-1.672e+01>
<epoch: 2, iter: 51,012, lr:2.500e-04, t:2.06e+00, td:3.12e-03, eta:8.51e+01, nll:-1.859e+01>
<epoch: 2, iter: 51,013, lr:2.500e-04, t:1.83e+00, td:2.23e-03, eta:7.57e+01, nll:-1.672e+01>
<epoch: 2, iter: 51,014, lr:2.500e-04, t:1.81e+00, td:2.39e-03, eta:7.50e+01, nll:-1.772e+01>
<epoch: 2, iter: 51,015, lr:2.500e-04, t:1.84e+00, td:1.94e-03, eta:7.60e+01, nll:-1.877e+01>
<epoch: 2, iter: 51,016, lr:2.500e-04, t:1.73e+00, td:3.45e-03, eta:7.17e+01, nll:-1.696e+01>
<epoch: 2, iter: 51,017, lr:2.500e-04, t:1.84e+00, td:2.32e-03, eta:7.62e+01, nll:-1.874e+01>
<epoch: 2, iter: 51,018, lr:2.500e-04, t:2.22e+00, td:2.27e-03, eta:9.18e+01, nll:-1.709e+01>
<epoch: 2, iter: 51,019, lr:2.500e-04, t:1.90e+00, td:1.72e-03, eta:7.87e+01, nll:-1.638e+01>
<epoch: 2, iter: 51,020, lr:2.500e-04, t:1.77e+00, td:2.30e-03, eta:7.31e+01, nll:-1.529e+01>
<epoch: 2, iter: 51,021, lr:2.500e-04, t:1.86e+00, td:3.02e-03, eta:7.70e+01, nll:-1.642e+01>
<epoch: 2, iter: 51,022, lr:2.500e-04, t:1.81e+00, td:2.15e-03, eta:7.48e+01, nll:-1.789e+01>
<epoch: 2, iter: 51,023, lr:2.500e-04, t:1.85e+00, td:2.35e-03, eta:7.65e+01, nll:-1.866e+01>
<epoch: 2, iter: 51,024, lr:2.500e-04, t:1.83e+00, td:2.18e-03, eta:7.57e+01, nll:-1.676e+01>
<epoch: 2, iter: 51,100, lr:2.500e-04, t:1.88e+00, td:2.37e-03, eta:7.78e+01, nll:-1.536e+01>
<epoch: 2, iter: 51,200, lr:2.500e-04, t:1.90e+00, td:2.51e-03, eta:7.86e+01, nll:-1.572e+01>
<epoch: 2, iter: 51,300, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.75e+01, nll:-1.708e+01>
<epoch: 2, iter: 51,400, lr:2.500e-04, t:1.86e+00, td:2.42e-03, eta:7.68e+01, nll:-1.943e+01>
<epoch: 2, iter: 51,500, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.76e+01, nll:-1.640e+01>
<epoch: 2, iter: 51,600, lr:2.500e-04, t:1.87e+00, td:2.39e-03, eta:7.71e+01, nll:-1.571e+01>
<epoch: 2, iter: 51,700, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.74e+01, nll:-1.633e+01>
<epoch: 2, iter: 51,800, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.73e+01, nll:-1.499e+01>
<epoch: 2, iter: 51,900, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.71e+01, nll:-1.538e+01>
<epoch: 2, iter: 52,000, lr:2.500e-04, t:1.87e+00, td:2.40e-03, eta:7.70e+01, nll:-1.629e+01>
21-02-15 22:29:40.137 - INFO: Saving models and training states.
<epoch: 2, iter: 52,100, lr:2.500e-04, t:1.90e+00, td:2.42e-03, eta:7.79e+01, nll:-1.673e+01>
<epoch: 2, iter: 52,200, lr:2.500e-04, t:1.89e+00, td:2.46e-03, eta:7.77e+01, nll:-1.898e+01>
<epoch: 2, iter: 52,300, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.77e+01, nll:-1.815e+01>
<epoch: 2, iter: 52,400, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.64e+01, nll:-1.801e+01>
<epoch: 2, iter: 52,500, lr:2.500e-04, t:1.89e+00, td:2.48e-03, eta:7.76e+01, nll:-1.746e+01>
<epoch: 2, iter: 52,600, lr:2.500e-04, t:1.88e+00, td:2.53e-03, eta:7.70e+01, nll:-1.614e+01>
<epoch: 2, iter: 52,700, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.66e+01, nll:-1.496e+01>
<epoch: 2, iter: 52,800, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.71e+01, nll:-1.682e+01>
<epoch: 2, iter: 52,900, lr:2.500e-04, t:1.87e+00, td:2.48e-03, eta:7.66e+01, nll:-1.676e+01>
<epoch: 2, iter: 53,000, lr:2.500e-04, t:1.87e+00, td:2.42e-03, eta:7.62e+01, nll:-1.719e+01>
21-02-15 23:01:01.845 - INFO: Saving models and training states.
<epoch: 2, iter: 53,100, lr:2.500e-04, t:1.87e+00, td:2.37e-03, eta:7.62e+01, nll:-1.640e+01>
<epoch: 2, iter: 53,200, lr:2.500e-04, t:1.87e+00, td:2.41e-03, eta:7.64e+01, nll:-1.765e+01>
<epoch: 2, iter: 53,300, lr:2.500e-04, t:1.89e+00, td:2.45e-03, eta:7.69e+01, nll:-1.725e+01>
<epoch: 2, iter: 53,400, lr:2.500e-04, t:1.89e+00, td:2.45e-03, eta:7.70e+01, nll:-1.702e+01>
<epoch: 2, iter: 53,500, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.64e+01, nll:-1.803e+01>
<epoch: 2, iter: 53,600, lr:2.500e-04, t:1.88e+00, td:2.42e-03, eta:7.65e+01, nll:-1.760e+01>
<epoch: 2, iter: 53,700, lr:2.500e-04, t:1.86e+00, td:2.40e-03, eta:7.57e+01, nll:-1.747e+01>
<epoch: 2, iter: 53,800, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.66e+01, nll:-2.144e+01>
<epoch: 2, iter: 53,900, lr:2.500e-04, t:1.90e+00, td:2.43e-03, eta:7.72e+01, nll:-1.826e+01>
<epoch: 2, iter: 54,000, lr:2.500e-04, t:1.88e+00, td:2.40e-03, eta:7.64e+01, nll:-1.700e+01>
21-02-15 23:32:23.089 - INFO: Saving models and training states.
<epoch: 2, iter: 54,100, lr:2.500e-04, t:1.90e+00, td:2.55e-03, eta:7.70e+01, nll:-1.809e+01>
<epoch: 2, iter: 54,200, lr:2.500e-04, t:1.89e+00, td:2.46e-03, eta:7.64e+01, nll:-1.832e+01>
<epoch: 2, iter: 54,300, lr:2.500e-04, t:1.90e+00, td:2.47e-03, eta:7.67e+01, nll:-1.641e+01>
<epoch: 2, iter: 54,400, lr:2.500e-04, t:1.89e+00, td:2.47e-03, eta:7.63e+01, nll:-1.669e+01>
<epoch: 2, iter: 54,500, lr:2.500e-04, t:1.88e+00, td:2.43e-03, eta:7.60e+01, nll:-1.491e+01>
<epoch: 2, iter: 54,600, lr:2.500e-04, t:1.88e+00, td:2.46e-03, eta:7.58e+01, nll:-1.798e+01>
<epoch: 2, iter: 54,700, lr:2.500e-04, t:1.90e+00, td:2.41e-03, eta:7.65e+01, nll:-1.596e+01>
<epoch: 2, iter: 54,800, lr:2.500e-04, t:1.88e+00, td:2.53e-03, eta:7.59e+01, nll:-1.580e+01>
<epoch: 2, iter: 54,900, lr:2.500e-04, t:1.88e+00, td:2.40e-03, eta:7.58e+01, nll:-1.713e+01>
<epoch: 2, iter: 55,000, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.62e+01, nll:-1.874e+01>
21-02-16 00:03:50.708 - INFO: Saving models and training states.
<epoch: 2, iter: 55,100, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.62e+01, nll:-1.506e+01>
<epoch: 2, iter: 55,200, lr:2.500e-04, t:1.88e+00, td:2.39e-03, eta:7.55e+01, nll:-1.786e+01>
<epoch: 2, iter: 55,300, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.58e+01, nll:-1.834e+01>
<epoch: 2, iter: 55,400, lr:2.500e-04, t:1.88e+00, td:2.39e-03, eta:7.54e+01, nll:-1.841e+01>
<epoch: 2, iter: 55,500, lr:2.500e-04, t:1.89e+00, td:2.41e-03, eta:7.58e+01, nll:-1.820e+01>
<epoch: 2, iter: 55,600, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.51e+01, nll:-1.633e+01>
<epoch: 2, iter: 55,700, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.57e+01, nll:-1.660e+01>
<epoch: 2, iter: 55,800, lr:2.500e-04, t:1.89e+00, td:2.44e-03, eta:7.59e+01, nll:-1.856e+01>
<epoch: 2, iter: 55,900, lr:2.500e-04, t:1.90e+00, td:2.45e-03, eta:7.59e+01, nll:-1.613e+01>
CUBLAS error: out of memory (3) in magma_sgetrf_gpu_expert at /opt/conda/conda-bld/magma-cuda102_1583546904148/work/src/sgetrf_gpu.cpp:126
CUBLAS error: not initialized (1) in magma_sgetrf_gpu_expert at /opt/conda/conda-bld/magma-cuda102_1583546904148/work/src/sgetrf_gpu.cpp:126
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
Caught RuntimeError in replica 1 on device 1.
Original Traceback (most recent call last):
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/parallel/parallel_apply.py", line 61, in _worker
output = module(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/SRFlowNet_arch.py", line 65, in forward
y_onehot=y_label)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/SRFlowNet_arch.py", line 101, in normal_flow
y_onehot=y_onehot)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowUpsamplerNet.py", line 213, in forward
z, logdet = self.encode(gt, rrdbResults, logdet=logdet, epses=epses, y_onehot=y_onehot)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowUpsamplerNet.py", line 238, in encode
fl_fea, logdet = layer(fl_fea, logdet, reverse=reverse, rrdbResults=level_conditionals[level])
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 84, in forward
return self.normal_flow(input, logdet, rrdbResults)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 103, in normal_flow
self, z, logdet, False)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/FlowStep.py", line 35, in
"invconv": lambda obj, z, logdet, rev: obj.invconv(z, logdet, rev),
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/myenv/lib/python3.7/site-packages/torch/nn/modules/module.py", line 727, in _call_impl
result = self.forward(*input, **kwargs)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 48, in forward
weight, dlogdet = self.get_weight(input, reverse)
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 37, in get_weight
dlogdet = torch.slogdet(self.weight)[1] * pixels
RuntimeError: CUDA error: resource already mapped
<epoch: 2, iter: 56,000, lr:2.500e-04, t:1.86e+00, td:2.41e-03, eta:7.44e+01, nll:-1.589e+01>
21-02-16 00:35:13.687 - INFO: Saving models and training states.
<epoch: 2, iter: 56,100, lr:2.500e-04, t:1.88e+00, td:2.41e-03, eta:7.51e+01, nll:-1.545e+01>
<epoch: 2, iter: 56,200, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.48e+01, nll:-1.524e+01>
<epoch: 2, iter: 56,300, lr:2.500e-04, t:1.88e+00, td:2.50e-03, eta:7.49e+01, nll:-1.727e+01>
<epoch: 2, iter: 56,400, lr:2.500e-04, t:1.85e+00, td:2.40e-03, eta:7.40e+01, nll:-1.717e+01>
<epoch: 2, iter: 56,500, lr:2.500e-04, t:1.88e+00, td:2.48e-03, eta:7.48e+01, nll:-1.548e+01>
<epoch: 2, iter: 56,600, lr:2.500e-04, t:1.86e+00, td:2.48e-03, eta:7.42e+01, nll:-1.752e+01>
<epoch: 2, iter: 56,700, lr:2.500e-04, t:1.88e+00, td:2.48e-03, eta:7.47e+01, nll:-1.669e+01>
<epoch: 2, iter: 56,800, lr:2.500e-04, t:1.86e+00, td:2.43e-03, eta:7.40e+01, nll:-1.632e+01>
<epoch: 2, iter: 56,900, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.40e+01, nll:-1.778e+01>
<epoch: 2, iter: 57,000, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.47e+01, nll:-1.696e+01>
21-02-16 01:06:23.673 - INFO: Saving models and training states.
<epoch: 2, iter: 57,100, lr:2.500e-04, t:1.90e+00, td:2.48e-03, eta:7.54e+01, nll:-1.575e+01>
<epoch: 2, iter: 57,200, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.39e+01, nll:-1.667e+01>
<epoch: 2, iter: 57,300, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.47e+01, nll:-1.871e+01>
<epoch: 2, iter: 57,400, lr:2.500e-04, t:1.88e+00, td:2.50e-03, eta:7.44e+01, nll:-1.781e+01>
<epoch: 2, iter: 57,500, lr:2.500e-04, t:1.86e+00, td:2.43e-03, eta:7.36e+01, nll:-1.881e+01>
<epoch: 2, iter: 57,600, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.38e+01, nll:-1.742e+01>
<epoch: 2, iter: 57,700, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.38e+01, nll:-1.726e+01>
<epoch: 2, iter: 57,800, lr:2.500e-04, t:1.86e+00, td:2.42e-03, eta:7.34e+01, nll:-1.844e+01>
<epoch: 2, iter: 57,900, lr:2.500e-04, t:1.87e+00, td:2.44e-03, eta:7.36e+01, nll:-1.622e+01>
<epoch: 2, iter: 58,000, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.38e+01, nll:-1.635e+01>
21-02-16 01:37:34.238 - INFO: Saving models and training states.
<epoch: 2, iter: 58,100, lr:2.500e-04, t:1.89e+00, td:2.43e-03, eta:7.44e+01, nll:-1.692e+01>
<epoch: 2, iter: 58,200, lr:2.500e-04, t:1.84e+00, td:2.40e-03, eta:7.25e+01, nll:-1.594e+01>
<epoch: 2, iter: 58,300, lr:2.500e-04, t:1.87e+00, td:2.43e-03, eta:7.36e+01, nll:-1.747e+01>
<epoch: 2, iter: 58,400, lr:2.500e-04, t:1.87e+00, td:2.49e-03, eta:7.34e+01, nll:-1.949e+01>
<epoch: 2, iter: 58,500, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.32e+01, nll:-1.595e+01>
<epoch: 2, iter: 58,600, lr:2.500e-04, t:1.86e+00, td:2.44e-03, eta:7.32e+01, nll:-1.600e+01>
<epoch: 2, iter: 58,700, lr:2.500e-04, t:1.88e+00, td:2.47e-03, eta:7.39e+01, nll:-1.668e+01>
<epoch: 2, iter: 58,800, lr:2.500e-04, t:1.87e+00, td:2.46e-03, eta:7.33e+01, nll:-1.868e+01>
<epoch: 2, iter: 58,900, lr:2.500e-04, t:1.86e+00, td:2.46e-03, eta:7.28e+01, nll:-1.802e+01>
<epoch: 2, iter: 59,000, lr:2.500e-04, t:1.86e+00, td:2.47e-03, eta:7.27e+01, nll:-1.569e+01>
21-02-16 02:08:39.673 - INFO: Saving models and training states.
<epoch: 2, iter: 59,100, lr:2.500e-04, t:1.87e+00, td:2.42e-03, eta:7.34e+01, nll:-1.721e+01>
<epoch: 2, iter: 59,200, lr:2.500e-04, t:1.84e+00, td:2.39e-03, eta:7.21e+01, nll:-1.866e+01>
<epoch: 2, iter: 59,300, lr:2.500e-04, t:1.85e+00, td:2.47e-03, eta:7.22e+01, nll:-1.685e+01>
<epoch: 2, iter: 59,400, lr:2.500e-04, t:1.88e+00, td:2.45e-03, eta:7.35e+01, nll:-1.809e+01>
<epoch: 2, iter: 59,500, lr:2.500e-04, t:1.88e+00, td:2.42e-03, eta:7.35e+01, nll:-1.618e+01>
Fatal Python error: Bus error
Thread 0x00007f2e50c69700 (most recent call first):
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 37 in get_weight
File "/run/media/root/7de46b27-ca07-4d98-8955-0d77387c5764/test/SRFlow/code/models/modules/Permutations.py", line 48 in fFatal Python error: orwarSegmentation faultd
File "/run/meSegmentation fault (core dumped)"
(myenv) (python37) [root@master code]#
what should i do ?
Thx author!This is a great work!
But I train x4(python train.py -opt ./confs/SRFlow_DF2K_4X.yml) with problem.And with origin code.
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
svd_cuda: the updating process of SBDSDC did not converge (error: 11)
In the forward function of the RRDBNet
def forward(self, x, get_steps=False):
fea = self.conv_first(x)
block_idxs = opt_get(self.opt, ['network_G', 'flow', 'stackRRDB', 'blocks']) or []
block_results = {}
for idx, m in enumerate(self.RRDB_trunk.children()):
fea = m(fea)
for b in block_idxs:
if b == idx:
block_results["block_{}".format(idx)] = fea
trunk = self.trunk_conv(fea)
last_lr_fea = fea + trunk
It seems that fea
in the last_lr_fea=fea+trunk
line is the feature of the last RRDB block before self.turnk_conv
. But to my knowledge, the skip connection should be last_lr_fea = x+ trunk
. Is my understanding right?
Congrats and excellent work!
Maybe a dumb question, what's the difference between this and VAE, and how are they related?
just like title, how to how to train get SRFlow latent?
thank you very much
Thx author!I train x4 is ok!
But when I train x8:
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
svd_cuda: the updating process of SBDSDC did not converge (error: 23)
Python 3.8.5 (default, Sep 4 2020, 07:30:14)
[GCC 7.3.0] :: Anaconda, Inc. on linux
Type "help", "copyright", "credits" or "license" for more information.
import torch
print(torch.version)
1.7.1+cu110
print(torch.version.cuda)
11.0
print(torch.backends.cudnn.version())
8005
············································································································
My GPU is 3090.I run setup code and find the cuda version is not compare.After that I reinstall with (pip install torch==1.7.1+cu110 torchvision==0.8.2+cu110 torchaudio===0.7.2
About iters is 10000.
Hello, thank you for your work.When I tried to train with my own datasets, I used the SRFlow_DF2K_4x model, but during the training process, "svd_cuda: the updating process of SBDSDC did not converage (error:11)" error appeared.Have you ever encountered the same reason in your training?Is there a solution?
I had check the SRFlow and RRDB, except DDRB_DF2K_4X all the model's result is right, so can you check it ?
#14
I run command again. (python train.py -opt ./confs/SRFlow_DF2K_4X.yml # Diverse Images 4X (Dataset Included)) and the result is
(myenv) (base) ubuntu@ubuntu:~/Desktop/data1024/SRFlow-master/code$ python train.py -opt ./confs/SRFlow_DF2K_4X.yml
OrderedDict([('manual_seed', 10), ('lr_G', 0.00025), ('weight_decay_G', 0), ('beta1', 0.9), ('beta2', 0.99), ('lr_scheme', 'MultiStepLR'), ('warmup_iter', -1), ('lr_steps_rel', [0.5, 0.75, 0.9, 0.95]), ('lr_gamma', 0.5), ('niter', 200000), ('val_freq', 40000), ('lr_steps', [100000, 150000, 180000, 190000])])
Disabled distributed training.
pretrain_model path will be ignored when resuming training.
21-02-14 23:52:33.972 - INFO: name: train
use_tb_logger:
model: SRFlow
distortion: sr
scale: 4
gpu_ids: [0]
datasets:[
train:[
name: CelebA_160_tr
mode: LRHR_PKL
dataroot_GT: ../datasets/DF2K-tr.pklv4
dataroot_LQ: ../datasets/DF2K-tr_X4.pklv4
quant: 32
use_shuffle: True
n_workers: 3
batch_size: 12
GT_size: 160
use_flip: True
color: RGB
phase: train
scale: 4
data_type: img
]
val:[
name: CelebA_160_va
mode: LRHR_PKL
dataroot_GT: ../datasets/DIV2K-va.pklv4
dataroot_LQ: ../datasets/DIV2K-va_X4.pklv4
quant: 32
n_max: 20
phase: val
scale: 4
data_type: img
]
]
dataroot_GT: ../datasets/div2k-validation-modcrop8-gt
dataroot_LR: ../datasets/div2k-validation-modcrop8-x4
model_path: ../pretrained_models/SRFlow_DF2K_4X.pth
heat: 0.9
network_G:[
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 23
upscale: 4
train_RRDB: False
train_RRDB_delay: 0.5
flow:[
K: 16
L: 3
noInitialInj: True
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:[
enable: True
]
fea_up0: True
stackRRDB:[
blocks: [1, 8, 15, 22]
concat: True
]
]
scale: 4
]
path:[
pretrain_model_G: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models/200000_G.pth
strict_load: True
resume_state: auto
root: /home/ubuntu/Desktop/data1024/SRFlow-master
experiments_root: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
models: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/models
training_state: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/training_state
log: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train
val_images: /home/ubuntu/Desktop/data1024/SRFlow-master/experiments/train/val_images
]
train:[
manual_seed: 10
lr_G: 0.00025
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1
lr_steps_rel: [0.5, 0.75, 0.9, 0.95]
lr_gamma: 0.5
niter: 200000
val_freq: 40000
lr_steps: [100000, 150000, 180000, 190000]
]
val:[
heats: [0.0, 0.5, 0.75, 1.0]
n_sample: 3
]
logger:[
print_freq: 100
save_checkpoint_freq: 1000.0
]
is_train: True
dist: False
21-02-14 23:52:34.059 - INFO: Random seed: 10
{'name': 'CelebA_160_tr', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DF2K-tr.pklv4', 'dataroot_LQ': '../datasets/DF2K-tr_X4.pklv4', 'quant': 32, 'use_shuffle': True, 'n_workers': 3, 'batch_size': 12, 'GT_size': 160, 'use_flip': True, 'color': 'RGB', 'phase': 'train', 'scale': 4, 'data_type': 'img'}
Loaded 162150 HR images with [0.00, 255.00] in 15.40s from ../datasets/DF2K-tr.pklv4
Loaded 162150 LR images with [0.00, 255.00] in 15.40s from ../datasets/DF2K-tr_X4.pklv4
21-02-14 23:52:49.469 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_tr] is created.
Dataset created
21-02-14 23:52:49.480 - INFO: Number of train images: 162,150, iters: 13,513
21-02-14 23:52:49.480 - INFO: Total epochs needed: 15 for iters 200,000
{'name': 'CelebA_160_va', 'mode': 'LRHR_PKL', 'dataroot_GT': '../datasets/DIV2K-va.pklv4', 'dataroot_LQ': '../datasets/DIV2K-va_X4.pklv4', 'quant': 32, 'n_max': 20, 'phase': 'val', 'scale': 4, 'data_type': 'img'}
Loaded 20 HR images with [0.00, 255.00] in 0.79s from ../datasets/DIV2K-va.pklv4
Loaded 20 LR images with [0.00, 255.00] in 0.79s from ../datasets/DIV2K-va_X4.pklv4
21-02-14 23:52:50.273 - INFO: Dataset [LRHR_PKLDataset - CelebA_160_va] is created.
21-02-14 23:52:50.273 - INFO: Number of val images in [CelebA_160_va]: 20
21-02-14 23:52:52.479 - INFO: Network G structure: DataParallel - SRFlowNet, with parameters: 39,541,811
module.RRDB.conv_first.weight True
opt module.RRDB.conv_first.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.conv_first.weight] will not optimize.
module.RRDB.conv_first.bias True
opt module.RRDB.conv_first.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.conv_first.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv1.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv1.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv2.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv2.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv3.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv3.bias
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv4.weight
21-02-14 23:52:53.174 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv4.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB1.conv5.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB1.conv5.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv1.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv1.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv2.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv2.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv3.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv3.bias
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv4.weight
21-02-14 23:52:53.175 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv4.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB2.conv5.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB2.conv5.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv1.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv1.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv2.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv2.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv3.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv3.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv4.weight
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv4.bias
21-02-14 23:52:53.176 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.0.RDB3.conv5.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.0.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.0.RDB3.conv5.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.0.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv1.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv1.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv2.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv2.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv3.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv3.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv4.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv4.bias
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB1.conv5.weight
21-02-14 23:52:53.177 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB1.conv5.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv1.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv1.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv2.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv2.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv3.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv3.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv4.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv4.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB2.conv5.weight
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB2.conv5.bias
21-02-14 23:52:53.178 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv1.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv1.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv2.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv2.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv3.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv3.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv4.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv4.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.1.RDB3.conv5.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.1.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.1.RDB3.conv5.bias
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.1.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv1.weight
21-02-14 23:52:53.179 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv1.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv2.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv2.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv3.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv3.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv4.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv4.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB1.conv5.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB1.conv5.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv1.weight
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv1.bias
21-02-14 23:52:53.180 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv2.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv2.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv3.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv3.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv4.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv4.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB2.conv5.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB2.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB2.conv5.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB2.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv1.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv1.weight
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv1.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv1.bias
21-02-14 23:52:53.181 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv2.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv2.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv2.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv2.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv3.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv3.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv3.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv3.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv4.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv4.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv4.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv4.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv5.weight True
opt module.RRDB.RRDB_trunk.2.RDB3.conv5.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.2.RDB3.conv5.bias True
opt module.RRDB.RRDB_trunk.2.RDB3.conv5.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.2.RDB3.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv1.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv1.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv1.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv1.bias
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv2.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv2.weight
21-02-14 23:52:53.182 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv2.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv2.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv3.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv3.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv3.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv3.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv4.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv4.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv4.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv4.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv4.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv4.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv5.weight True
opt module.RRDB.RRDB_trunk.3.RDB1.conv5.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv5.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB1.conv5.bias True
opt module.RRDB.RRDB_trunk.3.RDB1.conv5.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB1.conv5.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv1.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv1.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv1.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv1.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv1.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv1.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv2.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv2.weight
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv2.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv2.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv2.bias
21-02-14 23:52:53.183 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv2.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv3.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv3.weight
21-02-14 23:52:53.184 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv3.weight] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv3.bias True
opt module.RRDB.RRDB_trunk.3.RDB2.conv3.bias
21-02-14 23:52:53.184 - WARNING: Params [module.RRDB.RRDB_trunk.3.RDB2.conv3.bias] will not optimize.
module.RRDB.RRDB_trunk.3.RDB2.conv4.weight True
opt module.RRDB.RRDB_trunk.3.RDB2.conv4.weight
·························································································································································
arams [module.flowUpsamplerNet.layers.12.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.actnorm.logs True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.invconv.weight True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.weight True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.257 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fAffine.4.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.weight True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.258 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.12.affine.fFeatures.4.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.12.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.13.actnorm.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.invconv.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.weight True
21-02-14 23:52:53.259 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fAffine.4.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.weight True
21-02-14 23:52:53.260 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.13.affine.fFeatures.4.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.13.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.14.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.invconv.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.weight True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.261 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fAffine.4.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.weight True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.bias True
21-02-14 23:52:53.262 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.14.affine.fFeatures.4.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.14.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.15.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.invconv.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.weight True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.bias True
21-02-14 23:52:53.263 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fAffine.4.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.weight True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.15.affine.fFeatures.4.logs True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.15.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.16.actnorm.bias True
21-02-14 23:52:53.264 - WARNING: Params [module.flowUpsamplerNet.layers.16.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.invconv.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.weight True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.bias True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fAffine.4.logs True
21-02-14 23:52:53.265 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.weight True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.16.affine.fFeatures.4.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.16.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.17.actnorm.bias True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.17.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.actnorm.logs True
21-02-14 23:52:53.266 - WARNING: Params [module.flowUpsamplerNet.layers.17.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.invconv.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fAffine.4.logs True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.weight True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.267 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.17.affine.fFeatures.4.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.17.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.18.actnorm.bias True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.actnorm.logs True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.invconv.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.weight True
21-02-14 23:52:53.268 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fAffine.4.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.weight True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.269 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.18.affine.fFeatures.4.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.18.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.19.conv.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.weight] will not optimize.
module.flowUpsamplerNet.layers.19.conv.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.bias] will not optimize.
module.flowUpsamplerNet.layers.19.conv.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.19.conv.logs] will not optimize.
module.flowUpsamplerNet.layers.21.actnorm.bias True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.21.actnorm.logs True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.21.invconv.weight True
21-02-14 23:52:53.270 - WARNING: Params [module.flowUpsamplerNet.layers.21.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.22.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.22.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.22.invconv.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.22.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.23.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.invconv.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.weight True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.271 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fAffine.4.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.weight True
21-02-14 23:52:53.272 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.23.affine.fFeatures.4.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.23.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.24.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.invconv.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.weight True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.273 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fAffine.4.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.weight True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.bias True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.24.affine.fFeatures.4.logs True
21-02-14 23:52:53.274 - WARNING: Params [module.flowUpsamplerNet.layers.24.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.25.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.invconv.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.weight True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.bias True
21-02-14 23:52:53.275 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fAffine.4.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.weight True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.25.affine.fFeatures.4.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.25.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.26.actnorm.bias True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.26.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.actnorm.logs True
21-02-14 23:52:53.276 - WARNING: Params [module.flowUpsamplerNet.layers.26.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.invconv.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.bias True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fAffine.4.logs True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.weight True
21-02-14 23:52:53.277 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.26.affine.fFeatures.4.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.26.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.27.actnorm.bias True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.actnorm.logs True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.invconv.weight True
21-02-14 23:52:53.278 - WARNING: Params [module.flowUpsamplerNet.layers.27.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fAffine.4.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.weight True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.279 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.27.affine.fFeatures.4.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.27.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.28.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.actnorm.logs True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.invconv.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.weight True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.280 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fAffine.4.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.weight True
21-02-14 23:52:53.281 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.28.affine.fFeatures.4.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.28.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.29.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.invconv.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.weight True
21-02-14 23:52:53.282 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fAffine.4.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.weight True
21-02-14 23:52:53.283 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.29.affine.fFeatures.4.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.29.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.30.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.invconv.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.weight True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.284 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fAffine.4.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.weight True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.bias True
21-02-14 23:52:53.285 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.30.affine.fFeatures.4.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.30.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.31.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.invconv.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.weight True
21-02-14 23:52:53.286 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fAffine.4.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.weight True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.31.affine.fFeatures.4.logs True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.31.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.32.actnorm.bias True
21-02-14 23:52:53.287 - WARNING: Params [module.flowUpsamplerNet.layers.32.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.invconv.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.weight True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.bias True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fAffine.4.logs True
21-02-14 23:52:53.288 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.weight True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.32.affine.fFeatures.4.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.32.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.33.actnorm.bias True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.33.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.actnorm.logs True
21-02-14 23:52:53.289 - WARNING: Params [module.flowUpsamplerNet.layers.33.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.invconv.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.bias True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fAffine.4.logs True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.weight True
21-02-14 23:52:53.290 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.33.affine.fFeatures.4.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.33.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.34.actnorm.bias True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.actnorm.logs True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.invconv.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.weight True
21-02-14 23:52:53.291 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fAffine.4.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.weight True
21-02-14 23:52:53.292 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.34.affine.fFeatures.4.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.34.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.35.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.actnorm.logs True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.invconv.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.weight True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.293 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fAffine.4.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.weight True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.294 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.35.affine.fFeatures.4.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.35.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.36.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.invconv.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.weight True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.295 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fAffine.4.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.weight True
21-02-14 23:52:53.296 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.36.affine.fFeatures.4.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.36.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.37.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.invconv.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.weight True
21-02-14 23:52:53.297 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fAffine.4.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.weight True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.bias True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.37.affine.fFeatures.4.logs True
21-02-14 23:52:53.298 - WARNING: Params [module.flowUpsamplerNet.layers.37.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.38.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.invconv.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.weight True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.bias True
21-02-14 23:52:53.299 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fAffine.4.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.weight True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.38.affine.fFeatures.4.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.38.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.40.actnorm.bias True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.40.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.40.actnorm.logs True
21-02-14 23:52:53.300 - WARNING: Params [module.flowUpsamplerNet.layers.40.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.40.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.40.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.41.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.41.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.41.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.41.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.42.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.invconv.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.weight True
21-02-14 23:52:53.301 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fAffine.4.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.weight True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.302 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.42.affine.fFeatures.4.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.42.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.43.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.invconv.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.weight True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.303 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fAffine.4.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.weight True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.bias True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.43.affine.fFeatures.4.logs True
21-02-14 23:52:53.304 - WARNING: Params [module.flowUpsamplerNet.layers.43.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.44.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.invconv.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.weight True
21-02-14 23:52:53.305 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fAffine.4.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.weight True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.44.affine.fFeatures.4.logs True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.44.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.45.actnorm.bias True
21-02-14 23:52:53.306 - WARNING: Params [module.flowUpsamplerNet.layers.45.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.invconv.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.weight True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.bias True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fAffine.4.logs True
21-02-14 23:52:53.307 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.weight True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.45.affine.fFeatures.4.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.45.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.46.actnorm.bias True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.46.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.actnorm.logs True
21-02-14 23:52:53.308 - WARNING: Params [module.flowUpsamplerNet.layers.46.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.invconv.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fAffine.4.logs True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.weight True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.309 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.46.affine.fFeatures.4.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.46.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.47.actnorm.bias True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.actnorm.logs True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.invconv.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.weight True
21-02-14 23:52:53.310 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fAffine.4.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.weight True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.311 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.47.affine.fFeatures.4.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.47.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.48.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.invconv.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.weight True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.312 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fAffine.4.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.weight True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.313 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.48.affine.fFeatures.4.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.48.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.49.actnorm.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.invconv.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.weight True
21-02-14 23:52:53.314 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fAffine.4.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.weight True
21-02-14 23:52:53.315 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.49.affine.fFeatures.4.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.49.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.50.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.invconv.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.weight True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.316 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fAffine.4.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.weight True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.bias True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.50.affine.fFeatures.4.logs True
21-02-14 23:52:53.317 - WARNING: Params [module.flowUpsamplerNet.layers.50.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.51.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.invconv.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.weight True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.bias True
21-02-14 23:52:53.318 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fAffine.4.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.weight True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.51.affine.fFeatures.4.logs True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.51.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.52.actnorm.bias True
21-02-14 23:52:53.319 - WARNING: Params [module.flowUpsamplerNet.layers.52.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.invconv.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.bias True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fAffine.4.logs True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.weight True
21-02-14 23:52:53.320 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.52.affine.fFeatures.4.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.52.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.53.actnorm.bias True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.actnorm.logs True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.invconv.weight True
21-02-14 23:52:53.321 - WARNING: Params [module.flowUpsamplerNet.layers.53.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fAffine.4.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.weight True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.322 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.53.affine.fFeatures.4.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.53.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.54.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.actnorm.logs True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.invconv.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.weight True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.323 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fAffine.4.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.weight True
21-02-14 23:52:53.324 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.54.affine.fFeatures.4.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.54.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.55.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.invconv.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.weight True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.325 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fAffine.4.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.weight True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.326 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.55.affine.fFeatures.4.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.55.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.56.actnorm.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.invconv.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.weight True
21-02-14 23:52:53.327 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fAffine.4.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.weight True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.328 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.56.affine.fFeatures.4.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.56.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.layers.57.actnorm.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.actnorm.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.invconv.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.invconv.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.bias True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.logs True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.weight True
21-02-14 23:52:53.329 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fAffine.4.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fAffine.4.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.0.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.weight True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.bias True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.logs True
21-02-14 23:52:53.330 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.2.actnorm.logs] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs] will not optimize.
module.flowUpsamplerNet.f.0.weight True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.f.0.weight] will not optimize.
module.flowUpsamplerNet.f.0.bias True
21-02-14 23:52:53.331 - WARNING: Params [module.flowUpsamplerNet.f.0.bias] will not optimize.
rrdb params 702
21-02-14 23:52:53.341 - INFO: Model [SRFlowModel] is created.
21-02-14 23:52:53.341 - INFO: Resuming training from epoch: 14, iter: 200000.
21-02-14 23:52:53.685 - INFO: Start training from epoch: 14, iter: 200000
21-02-14 23:52:55.919 - INFO: Saving the final model.
21-02-14 23:52:56.557 - INFO: End of training.
Hi,
Thanks for your great and inspiring work!
Just a question about how you do the evaluation for your model. Since it is possible to sample different SR images for a given LR image, what is your methodology to decide which sample is chosen as the output SR image for evaluation? Do you simply draw one random sample or do you draw multiple samples and choose the one with the best result?
Thanks for your great work!
I have a question about training using CelebA dataset
In SRFlow paper,
we pre-process the dataset by cropping aligned patches, which are
resized to the HR resolution of 160 × 160.
I am wondering how to crop celeba aligned dataset!
Thank you!
I am sorry, I'm not very knowledgeable about python to say the least. I tried installing your package by following the setup script on my Windows 10 environment, but I'm getting an error during the requirements installation, namely with torch.
Collecting torch==1.7.0
Downloading torch-1.7.0-cp37-cp37m-win_amd64.whl (184.0 MB)
|████████████████████████████████| 184.0 MB 1.3 MB/s
ERROR: torch has an invalid wheel, .dist-info directory not found
I understand this is not directly related to your code, but I have nowhere else to turn. I did upgrade pip and run the install in a fresh venv.
wget --continue http://data.vision.ee.ethz.ch/alugmayr/SRFlow/pretrained_models.zip
Why not directly release the model file?
I am not able to understand the use of sigmoid in the scaling.
Thanks for sharing the code.
I have a confusion that you only maximize the term
However, the image is obtained by
What is the reason?
Looking forward to your reply. Thanks!
Hi @andreas128
Thanks for your great work and sharing the code.
I tried to train SRFlow using REDS dataset. After training the RRDBN network on REDS (with PSNR around 29 dB), I used this config to train SRFlow. However, the resulted SR images are not good and the PSNR is around 17 dB.
I did not change the code by the way. Could you please give me a hint on how can I improve the training?
network_G:
which_model_G: SRFlowNet
in_nc: 3
out_nc: 3
nf: 64
nb: 25
upscale: 4
train_RRDB: false
train_RRDB_delay: 0.5
flow:
K: 10
L: 3
noInitialInj: true
coupling: CondAffineSeparatedAndCond
additionalFlowNoAffine: 2
split:
enable: true
fea_up0: true
stackRRDB:
blocks: [ 1,5, 15, 22 ]
concat: true
train:
manual_seed: 1386
lr_G: !!float 1e-5
weight_decay_G: 0
beta1: 0.9
beta2: 0.99
lr_scheme: MultiStepLR
warmup_iter: -1 # no warm up
lr_steps_rel: [ 0.5, 0.75, 0.9, 0.95 ]
lr_gamma: 0.8
weight_l1: 0.1
niter: 300000
val_freq: !!float 5e3
val:
heats: [ 0.0, 0.5]
n_sample: 1
Dear author:
You SRFlow work is really insightful. Lots of researchers just like me are looking forward to try your models. Please kindly share the source code asap. Thank you very much!
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.weight] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias True
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.bias] will not optimize.
module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs True
23-09-04 05:05:13.371 - WARNING: Params [module.flowUpsamplerNet.layers.57.affine.fFeatures.4.logs] will not optimize.
And I don't know what caused this problem。
Currently, ESRGAN outperforms SRFlow when it comes to uncompressed upscaling.
And SRFlow can't be trained on JPEGs or other artifacts.
Is it possible to use custom LR images that have been augmented?
I followed this repo, paper and glow paper. My loss looks like this
_logp = GaussianDiag.logp(None, None, (z))
obj = logdets + logp
loss = -obj / (tf.math.log(2.) * tf.cast(pixels, tf.float32))_
There are 2 terms added together : one is logp which is always negative, and logdets which starts with a negative number during training, because of the minus sign in -obj the loss is positive, but after training for around 100k steps, both the logp and logdets increases. The logp term increases and becomes positive number and hence the final loss becomes negative around -3.xxx. Just wanted to know whether it is an expected behaviour ?
Thanks for your code and contributions!
There is a task that generate a SR image, but I just need a grayscale generation(that means the generation data has only 1 channel), the origin code can only generate a 3 channel's RGB images.
So how to generate the 1 channel data?
Hi authors,
Thanks for sharing your work. I am trying to extend this model into other datasets. During training, the nll loss first decreases from 1 to -4.5 in the first 3k steps, and then goes back to -0.5 in the later 3k steps. A 10 times larger or smaller learning rate does not prevent this phenomenon.
Have you observed similar results, or any hints available?
Thanks for your help
Thank you very much for your work. When I reproduce the X4 super-resolution completely with your code, there will be a Bus error (core dumped) after training for a while. What should I do?
Are we assumed to use final training output to calculate the loss as in eq (3)?
In GLOW, (-loss / (log(2) * n_pixel)).mean()
is implemented as well. It is not mentioned in SRFLOW, do we need similar processing steps?
Hi, which LPIPS version did you used in your evaluation script? v0.0 or v0.1?
Thanks for your advice ! I have tried those experiments:
2.I also have added a small number to the division to protect against divide by zero . But tn the training stage, "RuntimeError: 'DivBackWard0' return nan values in its 1th output." still exists.
For Adding Nosie :
I have added noise to the HR image that you mentioned . And The pixel value in the input HR ranges [0, 255] , [-0.5, 0.5] or [0, 1] ?
Originally posted by @Alan-xw in #2 (comment)
Thanks for your impressive work. During the training process, we found that the output NLL is negative. But theoretically, NLL should be positive. Is there any explanation for this?
SRFlow/code/models/modules/SRFlowNet_arch.py
Lines 92 to 97 in 5a007ad
According to the lines, uniform noise between [-0.5,0.5) is added into the input image, is the operation necessary? And why the logdet should be updated like this? When I read the paper, I noticed the description in the trainning details section as below, but i am not sure whether it's implemented by the mentioned lines or not.
adding slight random noise to the target image helps the training process and leads to better visual results. We therefore add Gaussian noise with a standard deviation of
$\sigma = 4/ \sqrt 3$ to the high-resolution image. In contrast to [21], we do not employ 5-bit quantization.
Skipping ERROR caught in nll = model.optimize_parameters(current_step):
Given groups=1, weight if size [64,3,3,3], expected input[1,4,40,40] to have 3 channels, but got 4 channels instead
Hey,
I implement the SRFlow based on the paper and Glow source code with Pytorch. But I encounter the "NaN" problem during testing. I put the testing LR image and generated z_samples into SRflow reverse net. But It will generate "NaN" value . Have you ever met this issue?
Looking forward to your reply.!
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.