Giter Site home page Giter Site logo

advcl's People

Contributors

lijiefan avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

advcl's Issues

About pretraining model

Could you please share us your pretraining model?
Actually, I am very interested about your work, but I met some difficulties.
So, I sincerely hope that the pretraining model can be shared. Thanks.

No support for CIFAR100 dataset

Hi Lijie, current code only supports CIFAR10, can you kindly share the code of the ClusterFit component or the pseudo-labels of CIFAR100? Thanks.

Details about the clustering process

Could you shed some light on the clustering process? In particular,

  • Which latent represents did you use to train KMeans?
  • If the encoder is being trained using AdvCL then which encoder was used to compute the latent representations that were used during KMeans training?
  • What did you use to build the clustering index?

Doubt in the SupCon loss function

Hi there. In the end of SupCon loss fn, you multiply the final loss by temp / base_temp value. What's the purpose for that? Thanks

Training command

It seems that there is no argument --syncBN and --decay in pretraining_advCL.py.
syncBN is set by default in code, and maybe decay is --weight_decay.

Unable to reproduce the results

Dear Authors,

Thanks for your interesting paper and the code! I encountered the issue when I reproduced your results.

I conducted AdvCL on CIFAR-10 by using your pre-training code and your finetuning code to conduct SLF on CIFAR-10. I got the following results copied from the generated log of your finetuning code:

best accuracy: 10.43
best accuracy clean: 13.14

In addition, I repeated the experiments including the pre-training and the finetuning procedure again. The following results are copied from the generated log:

best accuracy: 10.92
best accuracy clean: 14.12

Compared with your reported results on CIFAR-10 (standard test accuracy is 80.85% and PGD-20 robust test accuracy is 50.45%), my reproduced results are significantly lower than what you reported.

Could you help me figure out the reasons and guide me to reproduce your results? Thank you!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.