Coder Social home page Coder Social logo

y0ast / due Goto Github PK

View Code? Open in Web Editor NEW
107.0 107.0 22.0 22 KB

Code for "On Feature Collapse and Deep Kernel Learning for Single Forward Pass Uncertainty".

Home Page: https://arxiv.org/abs/2102.11409

License: MIT License

Python 79.59% Jupyter Notebook 20.41%

due's People

Contributors

blackhc avatar y0ast avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

due's Issues

Is there MC sampling in the training phase?

Dear authors,

I really like the idea of this paper. Thank you very much! However, while reading the code, I am a little confused by the implementation of DUE.

In the paper (Appendix A.2, right after Equation 9), it is mentioned that MC sampling is used for training the classification model. But according to the implementation, I couldn't find the location where MC sampling is implemented. Based on my observation of the code, the output of the model is a Gaussian distribution. Then this distribution is fed into the VariationalElbo directly to compute the loss. Therefore, could you please link me to the place where MC sampling is implemented?

Another question about MC sampling during training is more fundamental. I think MC sampling is non-differentiable. If MC sampling exists during training, how the gradient could be propagated to the feature extractor part?

Please let me apologize if I am asking wrong questions. (I am new to Gaussian process.)

Thank you very much

__

I have deleted the issue. sorry for bothering you.

load trained model

Could you add a example about how to load trained modle and likelyhood model for prediction in toymodel_regression script?

Loss become negative

When I apply DUE regression to my task, the expected_log_prob became negative gradually?
Does it represent overfitting of my model?

Why the predicted mean is a same value for each input?

I have implemented your DUE to my 6D pose regression task, but after training, in the evaluation, the predicted mean is a constant value, could you give me some inspirations? The input data is 128 diemensions. output is 6 diem.

error in num_outputs > 1 in toy_regression when batch_size >1

When I try to extend to num_outputs > 1 based on the toy_regression script

  • No problem when num_outputs >1 with batch_size =1.
  • But, when batch_size>1, error happens.

if you can modify toy regression script to support num_outputs>1, I will be very very appreciated!
Thanks ahead!

initial values selection

About def initial_values(train_dataset, feature_extractor, n_inducing_points):
In toymodel regression script, the train_dataset is the whole training datset, could it be the mini-batch instead of whole training dataset?

About the accuracy of the baseline (SNGP)

Hi, I have successfully run this code and it performed well without "--sngp". However, when I added "sngp", the accuracy fell to 0.1. It might make sense when it was 0.7 or 0.8, but it seemed too low to be 0.1.
Is it because I did something wrong, or the performance of GP layer is highly correlated with the feature extractor? Is such optimizer applicable for some bigger feature extractors, such as BERT for text classification for instant?
And by the way, is the performance significantly correlated with the optimizer? Is adam also okay?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.