Giter Site home page Giter Site logo

Comments (12)

rasenganai avatar rasenganai commented on July 24, 2024 1

@Optimox I think the problem here is how to train the weak learners .

As in boosted trees this is done by gini index (for training a weak tree) etc. And the cross entropy was used on whole algorithm that is to find out residuals and another tree is trained on those residuals.

But here each step requires some gradients to train not as in tree (gini index only).

A solution could be to train each step on cross-entropy (for 1 or 2 epochs or gradient steps (weak learners)) predicitng classes probab and using that probab calculating the residuals for the next step to train using cross entropy in a same way and so on ?

from tabnet.

AlexisMignon avatar AlexisMignon commented on July 24, 2024 1

@Optimox
You may want to have a look at the Friedman paper about Gradient Boosting :
https://statweb.stanford.edu/~jhf/ftp/trebst.pdf
You'll see what I meant by using regressors only.

@Jaskaran170599

In case of Gradient Boostin technique the output of each step will be multiplied by a Learning rate and will be sum to get the log odd on which we can apply sigmoid to get probability (0/1)?

Exactly. The idea is to fit the decision function before the sigmoid is applied. And compute the gradient with respect to this decision function values. So at each step, the weak learner is trained to fit the gradient (hence the regressor), the result is added (with a weight) to the the previous decision function. And class probability can be computed by applying the sigmoid function for binary problems or softmax for multi-class problems.

from tabnet.

bibhabasumohapatra avatar bibhabasumohapatra commented on July 24, 2024 1

Thanks @bibhabasumohapatra, looks promising. Is there a research paper related to the repo?

Yes.
https://arxiv.org/abs/2106.05239

from tabnet.

AlexisMignon avatar AlexisMignon commented on July 24, 2024

Why not try a mere application of gradient boosting ? Each step fits the gradient of the loss function (as computed so far) and adds it (using line search) to the previous result. Only regression is needed internally (to fit the gradient) and it allows for regression and classification.

from tabnet.

rasenganai avatar rasenganai commented on July 24, 2024

Interesting ,
For classification , i think we can try the same way as gradient boosting algorithm and or adaboost as mentioned in ther paper using cross_entropy loss function ?

In case of Gradient Boostin technique the output of each step will be multiplied by a Learning rate and will be sum to get the log odd on which we can apply sigmoid to get probability (0/1)?

In case of adaboost we can maybe use the same weightage formula as mentioned in the paper.

interesting would be to some how use MASK weights to give "IMPORTANCE WEIGHT" to each step to contribute to the final prediction as MASK heatmap shows us that some MASK weights are not that activated as others , It may improve decision making.

Different would be the training as in case of boosted algorithms they trin one tree then use it in boosting but here all the weak learners would be learning simultaneously.

I would like to do some research and contribution to this.

#Abhishek-eBook

from tabnet.

Optimox avatar Optimox commented on July 24, 2024

@AlexisMignon approaching classification problems with regression could be a solution but I feel like it's not satisfying and especially for multi class classification...

@JaskaranSingh-Precily tabnet is using cross entropy already, but you need to have integers as targets to apply cross entropy, so I don't see how a boosted version could use cross entropy at every step. Could you explain and/or give some links to literature? I probably just need to dig a bit deeper on how XGBoost deals with multi class classification.

@Jaskaran170599 Not sure you'll double your chance of winning Abhishek's ebook that way to be honnest! :)

from tabnet.

rasenganai avatar rasenganai commented on July 24, 2024

@Optimox actually commented with the company account that was not my personal account

from tabnet.

rasenganai avatar rasenganai commented on July 24, 2024

@AlexisMignon Yeah and i think here in tabnet case that weak learner is one block of the architecture and the main task that is different than Boosting algos is to train that block .

from tabnet.

bibhabasumohapatra avatar bibhabasumohapatra commented on July 24, 2024

https://github.com/tusharsarkar3/XBNet

from tabnet.

Optimox avatar Optimox commented on July 24, 2024

Thanks @bibhabasumohapatra, looks promising. Is there a research paper related to the repo?

from tabnet.

bibhabasumohapatra avatar bibhabasumohapatra commented on July 24, 2024

Thanks @bibhabasumohapatra, looks promising. Is there a research paper related to the repo?

Yes.

from tabnet.

ShuyangenFrance avatar ShuyangenFrance commented on July 24, 2024

https://github.com/tusharsarkar3/XBNet

This is a good job, but rather a completely design from my point of view.

from tabnet.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.