Comments (2)
Ok makes total sense!
Problem is that even with the original implementation I cannot retrieve the numbers stated in the paper.
Will let you know if I get to anything with my own implementation.
from hypertorch.
Hi @zaccharieramzi,
thanks a lot for your interest in the code :).
I did not try to reproduce the exact numbers in the iMAML paper, but just to implement the main components of the experimental setup. The purpose of the iMAML example is just to showcase what it can be achieved with the hypergradient approximation methods implemented here and possibly be used as a starting point for further research or to apply the method to different data.
As you noticed there are some differences in the implementation which happened mostly by chance. You can try to modify the implementation to match the one described by the original paper and code to see if this improves accuracy. I'd try first with larger batch size and using the same network (with correct placement of batch norm), possibly tweaking the learning rate. I don't think initialising the final linear layer weights randomly instead of to zero would make much difference.
The approximation I'm referring to here is the hypergradient approximation that they use, which relies on the conjugate gradient method to solve the linear system.
from hypertorch.
Related Issues (9)
- Could you put a license in this repo? HOT 1
- imaml.py issue regarding inner_loop solver HOT 3
- Add suport for unused variables in the graph
- support for custom datasets HOT 1
- Modification for the feature-head setting of meta-learning? HOT 2
- What does "stochastic" mean here? HOT 2
- Hyperparameter Optimization on MLP HOT 1
- Optimizers of inner problem and fixed point maps in approximate implicit hypergradient computations do not necessarily match HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from hypertorch.