Giter Site home page Giter Site logo

residual_adapters's People

Contributors

hbilen avatar srebuffi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

residual_adapters's Issues

validation accuracy not match the published results in the paper

Hi,

I have ran the code and found some surprising results. For training from scratch, fine-tuning and parallel adapter training, the validation accuracy for the aircraft data set I get is 38%, 50%, 44%, however the published test accuracy are 57.10%, 60.34%, 64.21%. The results are lower than the published results.
For the UCF101 data set, the training from scratch, fine-tuning and parallel adapter training validation accuracy I get is 72%, 81%, 75%, however the published test accuracy are 43.27%, 51.20%, 50.94%. The results are higher than the published results.
Finally, the validation results I got for the cifar100 data set match the published results.
I simply ran the code in this github and did not make any changes. Although I am comparing the validation accuracy with the test accuracy, however the difference are quite large. Is there something wrong with the published results? Thanks.

Batch Normalization layers inconsistently placed in regards to paper

According to "Efficient parametrization of multi-domain deep neural networks" the Batch Normalization (BN) layers are not needed for the parallel configuration but the model it appears that you still use task specific BN layers for the parallel configuration:

elif config_task.mode == 'parallel_adapters' and is_proj: self.parallel_conv = nn.ModuleList([conv1x1(in_planes, planes, stride) for i in range(nb_tasks)]) self.bns = nn.ModuleList([nn.BatchNorm2d(planes) for i in range(nb_tasks)])

Additionally in the series configurations, the BN layer is stated to come before the convolution yet that is not the case in the provided code:

if config_task.mode == 'series_adapters' and is_proj: self.bns = nn.ModuleList([nn.Sequential(conv1x1(planes), nn.BatchNorm2d(planes)) for i in range(nb_tasks)])

question about parallel SVD

Hi Sylvestre Rebuffi,
It's inspiring to read your work Efficient parametrization of multi-domain deep neural networks , but I have a problem understanding section 3.4. Cross-domain adapter compression.
As I understand, you used decomposition to compress the number of parameters, and thus shared information among different datasets. But I am confused about how you realize the perspective of sharing parameters to allow target tasks to communicate. Is this part (SVD) also included in your code? Or is there any important information I've missed? It would be very kind of you to reply.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.