Giter Site home page Giter Site logo

Comments (4)

larjung avatar larjung commented on July 17, 2024 1

Hi Jean,
Ok you are right, it is my mistake. Probably used some older version of the _likelihood() function without the scale_bound (as in the original tf_compression code, the scale lower bound does not exist in the likelihood function).
Thanks!

from compressai.

jbegaint avatar jbegaint commented on July 17, 2024

Hi, good point. Thanks for the report!

The _biases, _factors, _matrices are actually updated by the auxiliary optimizer, but their gradients are computed via the "main" loss, not the auxiliary ones.

You are right to highlight it would be more correct to only update the quantiles with the aux_optimizer, and update all the other parameters via the main optimizer. I'm actually working on a cleaner implementation, which would also not require to override .parameters() (which is a major issue do support multi-gpus training right now) and I'll address then.

In the meantime, the current version should still allow you to train and reproduce results. That said, please let me know if you have troubles reproducing results or training. Thanks!

from compressai.

larjung avatar larjung commented on July 17, 2024

Hi Jean,
Thank you for acknowledging this point.
Regarding being able to reproduce the training results, one interesting issue that I've encountered and had to solve, concerns MeanScaleHyperprior and JointAutoregressiveHierarchicalPriors models. In both of them, the estimated bpp on the latent entropy model (i.e. the gaussian conditional) started very high when training (around 20-30 bpp) and hasn't gone down to the the proper values (e.g. 1-2bpp). So although D loss term gone down, R remained high resulting in a poor model.
After investigation I found out that the root cause of this, is the fact that some "scale_hat" values (as predicted by the hyper-synthesis subnetwork) were negative (e.g -0.3) at the beginning of training, as there is no ReLU at the end of h_s (or context_prediction in case of JAHP). Those negative scales disrupted the entropy model in a way that the optimizer couldn't get out of.
My solution is to lower bound the predicted "scale_hat" to zero using LowerBound class. After this, training behaves normally and bpp values are as expected.
It's probably not the most elegant or correct solution, but it worked for me. One other option is to add a final ReLU to h_s (or context_prediction) but apply it only on the chunk that belongs to scale.
(Note that the ScaleHyperprior model doesn't suffer from this issue, since it has a ReLU at the end of h_s.)
If you managed to train successfully without encountering this issue at all, I'd be curious to understand what makes the difference. Maybe some other training hyper-params, since we are probably using different training scripts...

Anyway, it is a pleasure to work with your great framework. Thanks again.

Danny

from compressai.

jbegaint avatar jbegaint commented on July 17, 2024

Hi Danny, mmh this is weird.

So I've seen some papers and implementations where authors used ReLU or an exponential function to prevent the scales values going below zero. I've experimented a bit with torch.exp for the scales but then I had convergence issues.

Anyway, regarding the GaussianConditional class, there's already a scale_bound parameter that uses the LowerBound op to clip the scales to a minimum values (https://github.com/InterDigitalInc/CompressAI/blob/master/compressai/entropy_models/entropy_models.py#L496 and https://github.com/InterDigitalInc/CompressAI/blob/master/compressai/entropy_models/entropy_models.py#L567).
How do you initialize the GaussianConditional class? Can you share your lambda and hyper-parameters values?

from compressai.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.