Coder Social home page Coder Social logo

vae-celeba's People

Contributors

dekhtiarjonathan avatar yzwxx avatar zsdonghao avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

vae-celeba's Issues

Reproduce the plain VAE result

HI,

Thank you for sharing your code.

However, when I run the plain VAE code using the default parameters, I cannot reproduce the result you post on the REAMED page.

The reconstruction image and random image I generated using your code is:

train_29_3014

train_29_3014_random

Can you provide some suggestions on how to reproduce your result?

NaN losses

I'm training the model with different datasets and image sizes. I have noticed that if I use images larger than 64x64, all the losses become NaN after a certain point. It seems like it happens earlier as the image size increase (for instance, it happens after 77 epochs with 128x128 images and after 3 epochs with 1024x1024 images). Do you happen to know why this is happening and do you have any advice to address it?

Thanks!

Paper and code difference

Why does the implementation differ from the paper? Is this per intention? The file model_vae.py for example has a final filter depth of 512, and uses a 5x5 kernel size. What produces the best results? The paper implementation or the code?

train_size is not effective

I understand that the author may not have time to maintain the code.

For those who use the code, please be aware that the param train_size is not effective. It is only used for logging purpose.

I want to use the encoder part only as a CNN, what layers should I use and after which one I put the softmax layer?

If I want to use the encoder only as a cnn, what layer is the one containing the embedded coding that I need to use before the soft max? that is, I want to use the encoder with the trained parameters as a pretrained CNN, But I am not sure what will be the layer that has to be connected to the softmax. This one is the equivalent the last layer of a CNN before the softmax layer, correct?

UpSampling2dLayer+Conv2d or DeConv2d?

Thanks for the good code!
I have two questions:

  1. Is UpSampling2dLayer+Conv2d better than DeConv2d? How about the results of DeConv2d?
  2. How about the results of using the original VAE loss without perceptual loss?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.