Giter Site home page Giter Site logo

Comments (5)

mmderakhshani avatar mmderakhshani commented on May 31, 2024

@daib13. Hi Bin, I would really appreciate it if you could tell me the commands you really used in order to generate table 1 results.

from twostagevae.

daib13 avatar daib13 commented on May 31, 2024

Hi @mmderakhshani, sorry for the late reply. For celeba, we just use the default setting in the repository. You can run python demo.py --dataset celeba using this code. For cifar10, we use 1000 epochs for the first stage and 2000 epochs for the second stage. But note that the FID for cifar10 is very weird. Saving the real images in jpg files and then reading the jpg files will give different FID than just reading from the original files. Also, according to my experience, using pytorch will also produce different FID than using Tensorflow framework.

from twostagevae.

mmderakhshani avatar mmderakhshani commented on May 31, 2024

@daib13 Thanks for your reply. Regarding table 2, did you calculate those scores with Resnet architecture or it is similar to table one and calculated using Infogan architecture?

from twostagevae.

daib13 avatar daib13 commented on May 31, 2024

@mmderakhshani Table 2 is applied on WAE network which is defined in

class Wae(TwoStageVaeModel):

We exactly follow the training protocol of the WAE paper. You can reproduce the result using the command

python dome.py --dataset celeba --epochs 70 --lr-epochs 30 --epochs2 70 --lr-epochs2 30 --network-structure Wae

To calculate the FID score, we use the standard inception feature for both table 1 and table 2, which is also consistent to most of the previous works. The model is defined in https://github.com/openai/improved-gan/blob/master/inception_score/model.py. You can check how we calculate the FID score in

def evaluate_fid_score(fake_images, dataset, root_folder, norm=True):

from twostagevae.

saehoonkim avatar saehoonkim commented on May 31, 2024

Hello, I really enjoy reading your article. It contains many interesting observations both theoretically and empirically.
But, I’m struggling with reproducing FID scores with CIFAR-10. After running the command below, I’ve obtained FID scores on CIFAR-10: 86.1316 (reconstruction), 105.9603 (first stage), and 101.6009 (second stage). After saving the numpy array in JPEG format and reloading them to calculate FID scores, the numbers are 77.3854 (reconstruction), 89.8473 (first stage), 89.1814 (second stage).

python demo.py --dataset cifar10 --epochs 1000 --lr-epochs 300 --epochs2 2000 --lr-epochs2 600
--network-structure Resnet --num-scale 4 --base-dim 32 --latent-dim 64
--gpu 0 --exp-name [EXP-NAME]

I believe that the above configuration is exactly the same with Appendix D, but even after saving and reloading them, the numbers are higher than the ones reported in Table 1.

And, I’ve found that there exists a slight difference between Figure 16 in the arxiv version and the implementation. In the implementation, global averaging pooling was used instead of a flatten layer, which seems to be a minor difference.

Thanks in advance!

from twostagevae.

Related Issues (15)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.