Coder Social home page Coder Social logo

Bug in forward? about glom-pytorch HOT 9 OPEN

lucidrains avatar lucidrains commented on August 17, 2024 1
Bug in forward?

from glom-pytorch.

Comments (9)

A7ocin avatar A7ocin commented on August 17, 2024 1

Yes I have, it works. I'm currently trying to figure out how to improve the model

from glom-pytorch.

lucidrains avatar lucidrains commented on August 17, 2024

@A7ocin Hi! There wasn't strictly a bug because the variables never got used, but I removed it to avoid any confusion!

from glom-pytorch.

A7ocin avatar A7ocin commented on August 17, 2024

Oh, I see, my bad! That's better anyway :)

from glom-pytorch.

A7ocin avatar A7ocin commented on August 17, 2024

Hi, I'm reopening this again because there is something that is not clear to me.
In the README, you give some examples, in particular, when using 6 levels:

top_level = all_levels[7, :, :, -1] # get the top-level embeddings after iteration 6

which sounds good. However, in the network, you initialise the number of iterations as double the number of levels, for the top-down and bottom-up information to propagate. So my question is: why do we get the top-level embeddings at iteration 7 and not, let's say, 12?
Thank you for your patience

from glom-pytorch.

lucidrains avatar lucidrains commented on August 17, 2024

@A7ocin Hi Nicola again! I may not be thinking clearly, but my reasoning was that that is how long it would take for the data to propagate all the way up the layers and back down to the start. So say you were doing some segmentation task akin to https://arxiv.org/abs/2103.13413 , then you could attach losses to the bottom layer as well. If you are only doing losses on the top layer, 6 iterations is enough!

from glom-pytorch.

lucidrains avatar lucidrains commented on August 17, 2024

@A7ocin I see some connections between GLOM and the feature pyramids commonly used in vision nets

from glom-pytorch.

pfabreu avatar pfabreu commented on August 17, 2024

@lucidrains sorry to hijack this issue, but have you (or anyone else) tried training this as a denoising autoencoder on MNIST or CIFAR?

from glom-pytorch.

pfabreu avatar pfabreu commented on August 17, 2024

Could you please share your code/training loop? I think an interesting idea would be to make the autoencoders VAEs (the embeddings would be distribution parameters and sampled from at inference time), for a purely generative model.

from glom-pytorch.

MrChill avatar MrChill commented on August 17, 2024

Yes I have, it works. I'm currently trying to figure out how to improve the model

@A7ocin @lucidrains What did you change? what patch size did you use?

from glom-pytorch.

Related Issues (7)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.