Coder Social home page Coder Social logo

convergence of in-the-loop training about spin HOT 8 CLOSED

nkolot avatar nkolot commented on August 17, 2024
convergence of in-the-loop training

from spin.

Comments (8)

iammarvelous avatar iammarvelous commented on August 17, 2024 1

OK. I will try that. I tried static training before and everything looks good except that loss_regr_betas has some extreme values once a while. I am curious about your training details. Did you train without run_smplify at the beginning? Also, did you use the default loss weight here? It seems that shape_loss_weight is 0? Also about openpose_train_weight and gt_train_weight, any suggest on this? Do you also suggest openpose_train_weight as in the fitting stage?

from spin.

nkolot avatar nkolot commented on August 17, 2024 1

So just to be clear, the previous version of the code could still be using the extreme betas from the static fits. What I do is I check which fit has the lowest reprojection error and in the beginning of training it is almost certain that the static fits have lower error because the network has not started producing accurate shapes, i.e. good initializations for the optimization. This is one reason we enable the in-the-loop training later on.

from spin.

nkolot avatar nkolot commented on August 17, 2024

Hmm it looks like a learning rate issue. Can you try and train with a 10 times lower learning rate? Unfortunately Adam does not behave like SGD, so scaling the learning rate when increasing the batch size can have unexpected behavior.
Also try not to use --run_smplify in the beginning.

from spin.

nkolot avatar nkolot commented on August 17, 2024

The parameters that you mentioned are the ones used in the paper. I also inspected the code and I noticed that there is a small bug in the released version. Essentially I am not filtering out extreme betas in the static fits, only for the in the loop fits. I will do the modification now and I will push the changes.

from spin.

iammarvelous avatar iammarvelous commented on August 17, 2024

I think this should be the reason for the bumps in loss_regr_betas. I recalled in the paper you mentioned about the per vertices loss, i.e., shape loss in the code. Did you ever run the experiment on the shape loss? If so, did you disable it because of inferior performance?

from spin.

nkolot avatar nkolot commented on August 17, 2024

We mention in the supplementary material that the weight we chose for this loss is 0. The reason that we ended up not using this loss is that it is highly correlated with the other losses.

from spin.

iammarvelous avatar iammarvelous commented on August 17, 2024

Well, I just take a look at your commit b4c70d5. The above figures are run under run_smplify. So I feel there may be other reasons for the bumps?

from spin.

ikvision avatar ikvision commented on August 17, 2024

@iammarvelous were you able to avoid this problem by reducing the learning_rate or enlarging the batch_size?

from spin.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.