Comments (3)
Hi @DC95,
First of all, sorry for my slow response time, I had a lot on my plate.
Now it's better :) I will have a look at it !
Could you provide me with:
- the command line you are using?
- the dictionary structure of the checkpoint you are trying to evaluate?
For the second point, what I mean by dictionary structure is something like this:
{
"mode": {
"state_dict": [name of the weights]
}
}
Or if you have a public checkpoint that I can use, I can try that as well !
Thank you,
Quentin
from vissl.
Thanks for your response :)
The link for DC_v2 checkpoint is (https://gigamove.rwth-aachen.de/en/download/859065e3c72eabd19c45578abfd17ba0)
The link for simclr checkpoint is (https://gigamove.rwth-aachen.de/en/download/cb9771810632728d4f5885f72ac4e551)
The checkpoint structure looks like this:
phase_idx iteration loss iteration_num train_phase_idx classy_state_dict type
The classy_state_dict looks like this:
dict_keys(['train', 'base_model', 'meters', 'optimizer', 'phase_idx', 'train_phase_idx', 'num_updates', 'losses', 'hooks', 'loss', 'train_dataset_iterator', 'amp'])
The trunk keys looks like this : '_feature_blocks.conv1.weight', '_feature_blocks.bn1.weight'
The command line: python run_distributed_engines.py \config=pretrain/linear_classifier/linear_classifier_k7_g128
Thanks for taking time out,
DC
from vissl.
Wondering if it was in the right direction. Or was I doing something wrong?
from vissl.
Related Issues (20)
- Example of using ZeRO and LARC
- Incosistent results when running inference on trained resnet model HOT 1
- Low performance of Supervised training
- MLP_DIM in ViT is not adjustable through the config. HOT 2
- [Feature] Online linear probing
- "on_forward"-hook operations and back-propagation
- EMA does not work on fp16 and does not copy weights?
- First-class support for timm models
- Can't load pretrained seer/swav checkpoints (RG_Y_128GF) HOT 1
- Is it possible to get full imagenet pretrained weights of MoCo v2 ? HOT 2
- NPID + MoCoV2 weights are the same? HOT 1
- Potential confusion over licensing comment
- Does the MoCo implementation do shuffled Batch Norm when run on a single GPU? HOT 3
- Loading TRAINED models with VISSL HOT 9
- Use pretrain outside of vissl
- can someone share/has anyone shared a checkpoint ( with the class head) of a ssl model on imagenet?
- What config relate to learning rate warm up, weight decay, and momentum in 1 node n GPUs (n > 1 && n < 8) config?
- AttributeError: module 'importlib_resources' has no attribute 'is_resource'
- all vissl.ai/tutorials links broken
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from vissl.