Comments (6)
@firmai Thanks for your interest in the project.
As of now, it is implemented for regression tasks only, since classification tasks are probabilistic already, assigning probabilities to each class.
from lightgbmlss.
Agreed. What I'd personally be interested in is a classifier which can output calibrated probabilities in one go. Then you don't have to fit a calibrator (e.g. sklearn CalibratedClassifier
) on top of the already trained model. I am not sure if this is even possible by the way. Just throwing it out here because @StatMixedML you really seem to know what you are doing :)
from lightgbmlss.
@ihopethiswillfi Thanks for the clarification.
I definitely do see the value of probability calibration and for it being a valuable tool for evaluating any classifier, but I am no sure I how you would get that out from the specified likelihood.
Both XGBoostLSS and LightGBMLSS are trained to learn the parameters of a specified likelihood. So for the binary-case it would be a Bernoulli, and for the multi-class classification a multi-nomial distribution. For the Bernoulli, there is only one parameter p (q=1-p)
that the model would learn. I don't see how to extend that to allow for a calibration, but very happy to hear your comments on this. Would be a nice extension.
from lightgbmlss.
Unfortunately I have no idea either. Thanks for the quick response.
(by the way I'm not OP)
from lightgbmlss.
@firmai Can we close the issue?
from lightgbmlss.
I was thinking about (probabilistic) classification problems in the context of a meta-model for the distribution of Bernoulii p itself (perhaps a Beta). Yes, when you get a probability from a classifier, it's Bernoulii and you can sample from it. But what if the probability is 1 (or 0)? Your samples will all be 1 (or 0) and they would have 0 variance. What I think would be interesting would be the ability to sample from the distribution of p rather than the (Bernoulli) distribution parameterized by p. That would, I think, let you examine questions about what your model might have predicted, within some credibility bounds. You'd be able to talk about the confidence of the model's predictions rather than the confidence in the model's outcomes given that the predictions are correct. Am I thinking about that correctly?
from lightgbmlss.
Related Issues (20)
- General Discussion HOT 3
- Multi-parameter optimization with custom loss function for probabilistic forecasting HOT 6
- ETA for code HOT 2
- Out of sample prediction and overall questions HOT 5
- [Feature request] latest lightgbm support HOT 2
- Monotonicity of expectiles HOT 2
- Can't install on Ubuntu 20.04 HOT 2
- Userdefine CV Class?
- Lower python version support
- error when using weights in lgb.Dataset HOT 3
- Models with init_score HOT 8
- Silent freezes HOT 3
- Zero-and-One-Adjusted Beta HOT 1
- kwargs from the lightgbm predict are not available with the LightGBMLSS.predict HOT 5
- lightgbm v4.0.0? HOT 3
- Multi-task learning and ONNX support
- IndexError: LightGBMLSS.train(valid_sets=[dataset_val]) calls set_valid_margin, which seems to expect both train+val HOT 7
- Prediction of quantiles for parametric distributions HOT 5
- SHAP interpretations from zero-adjusted gamma model HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from lightgbmlss.