Comments (5)
Can I know what is the batch size you used during training? E.g., what is the number of GPUs you used?
And which checkpoint you used for evaluation, e.g., how many iterations you trained?
from ml-gmpi.
I have used these hyperparameters
curriculum: {'res_dict': {256: {'batch_size': 8, 'num_steps': 32, 'img_size': 256, 'tex_size': 256, 'batch_split': 1, 'gen_lr': 0.002, 'disc_lr': 0.002}, 512: {'batch_size': 4, 'num_steps': 32, 'img_size': 512, 'tex_size': 512, 'batch_split': 1, 'gen_lr': 0.002, 'disc_lr': 0.002}, 1024: {'batch_size': 4, 'num_steps': 32, 'img_size': 1024, 'tex_size': 1024, 'batch_split': 2, 'gen_lr': 0.002, 'disc_lr': 0.002}}, 'res_dict_learnable_param': {256: {'batch_size': 4, 'num_steps': 32, 'img_size': 256, 'tex_size': 256, 'batch_split': 1, 'gen_lr': 0.002, 'disc_lr': 0.002}, 512: {'batch_size': 4, 'num_steps': 32, 'img_size': 512, 'tex_size': 512, 'batch_split': 2, 'gen_lr': 0.002, 'disc_lr': 0.002}, 1024: {'batch_size': 4, 'num_steps': 32, 'img_size': 1024, 'tex_size': 1024, 'batch_split': 2, 'gen_lr': 0.002, 'disc_lr': 0.002}}, 0: {'batch_size': 8, 'num_steps': 32, 'img_size': 256, 'tex_size': 256, 'batch_split': 1, 'gen_lr': 0.002, 'disc_lr': 0.002}, 200000: {}, 'dataset_path': '/proj/cvl/users/x_fahkh/mn/ml-gmpi/runtime_dataset/ffhq256x256.zip', 'pose_data_path': '/proj/cvl/users/x_fahkh/mn/ml-gmpi/runtime_dataset/ffhq256_deep3dface_coeffs', 'fov': 12.6, 'ray_start': 0.95, 'ray_end': 1.12, 'h_stddev': 0.289, 'v_stddev': 0.127, 'h_mean': 0.0, 'v_mean': 0.0, 'latent_dim': 512, 'stylegan2_w_dim': 512, 'generator_label_dim': 0, 'fade_steps': 10000, 'betas': (0, 0.9), 'unique_lr': False, 'weight_decay': 0, 'r1_lambda': 10.0, 'grad_clip': 10, 'dataset': 'FFHQ', 'z_dist': 'gaussian', 'raw_img_size': 256, 'eval_img_size': 256}
Used 8 GPUs for training...
Trained the model for the 17000 iterations
from ml-gmpi.
Thanks. Do you mind evaluating the checkpoint at 5k iterations, which is what we used for evaluation? I think the code should save all checkpoints during training so 5k-iteration checkpoint should be there.
from ml-gmpi.
When I used the 5k-iteration checkpoint the reported number comes to be for FFHQ256
depth: 0.5485293 0.36907703
angle: 0.004393761830653509 0.006464534215649069
from ml-gmpi.
Thanks a lot for checking this. I think the current number looks reasonable to me. Since the depth and angle scores are computed on renderings from randomly-sampled camera poses, I would say the difference of ~0.01 or 0.02 is within expectation.
Regarding the reason about why the evaluation from the checkpoint at 17k iteration differs a lot, my gut is that the discriminator may be too strong after that long training and the generator cannot beat it. The curve of loss_d
and loss_g_fake
in the tensorboard may not be stable at 17k iterations.
It is an interesting finding though. I think some techniques for stabilizing GAN training might be helpful here.
from ml-gmpi.
Related Issues (20)
- environment.yml error HOT 13
- Problem In training the FFHQ1024 model HOT 3
- Great Work!!!!!! Few Queries HOT 7
- "RuntimeError: output is too large" when extracting meshes with 1024x1024 models HOT 2
- Problem in Importing upfirdn2d_plugin.so HOT 2
- RuntimeError: Unable to find a valid cuDNN algorithm to run convolution HOT 10
- RuntimeError: "slow_conv_transpose2d_out_cpu" not implemented for 'Half' HOT 1
- Great Work!!!!! HOT 9
- I got error when I start to train HOT 9
- download FFHQ tfrecord dataset related issue HOT 1
- Problem in reproducing fid score HOT 9
- Maximum range of yaw and pitches HOT 1
- 感觉真是充满坑点 HOT 1
- pose of AFHQ HOT 2
- Great Work!!!! Few queries regarding the evaluation code HOT 3
- Query Regarding the evaluation code for FFHOQ512 HOT 6
- Problem in calculating the Identity metric
- Query in calculating the depth and pose metric HOT 1
- Code for estimating pose for preprocessing (used in "Deep3DFaceRecon" directory) HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from ml-gmpi.