Coder Social home page Coder Social logo

clipnerf's People

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

clipnerf's Issues

Code Release

Hey Cassie!
Will there be a code release for the ClipNerf. I am interested specifically in part 1 training.

Shape editing support

If I am not missing anything, there currently is no shape editing support as the paper suggests. How would shape editing fit in this pipeline?

Thanks.

The input image of clip

Hi, according to your code the original nerf training process, one would randomly choose a batch of rays (say 1024) in the training process and compare this to the ground truth pixel-wise values of the ground truth image sampled by the same set of rays. So the image sent to compute the clip loss is just a batch of random pixels, without any semantic information. Is my understanding correct? And if so, why would it be possible to compare this 'image' to the input prompt?

This is the image sent to the clip loss during your training process.
fdc3d5653ab7244e5512b5eb7b42afc

The code does not match the description in the paper

Hi~ Thanks for your excellent work!

But when I read the open source code, I find it does not match the description in the paper, e.g. first train the disentangled conditional NeRF including the conditional NeRF generator and the deformation network; then we fix the weights of the generator and train the CLIP manipulation parts

I think that it just finetunes pretrained NeRF using clip. Could you please open-source the code described in the paper? Thank you very much.

Code share

Thanks for the nice work!
Is there a plan to release the code?
I'm really looking forward to it!

About the training data.

Hi, in your paper you described the training data of your work are Photoshapes and Carla. However, these dataset is not originally designed to be used in nerf, I was wondering if you could show us the data preprocessing code of Carla so that we could make some comparsion with your work on that. Thank you.

How to get the text in training?

Thanks for the great work. I have a question, since it is a feed-forward framework, I think you may need the text in training. How to get it? I think there are no texts in both of the two datasets. Thanks!

RuntimeError: shape '[60, 60, -1]' is invalid for input of size 3072 in llff dataset

Hi cassie,
thank you so much for your amazing works! i have encountered some problems when running the llff data set. It said that RuntimeError: shape '[60, 60, -1]' is invalid for input of size 3072. how can i fix this problem? Look forward for your replying. thanks again!
Traceback (most recent call last):
File "run_nerf_clip.py", line 885, in
train()
File "run_nerf_clip.py", line 805, in train
rgb_img = rgb.view(sample_scale, sample_scale, -1)
RuntimeError: shape '[60, 60, -1]' is invalid for input of size 3072

RuntimeError:CUDA out of memory.

I follow the default setting of ClipNerf on one RTX TiTian X with cuda out of memory.

I run successfully standard nerf on TiTian X with nerf-pytorch, which is similar to ColorNerf.

Then I change mse loss to gray mse loss and add Clip loss on nerf-pytorch in the way described in ClipNerf, which also can successfully run on one Titian X.

Is there anything wrong?

Thank you for your time.

Can i get entire code descripted on paper?

Thank you for nice work of your paper and i have huge interest about your work!

In released code, i can not find your idea of Training strategy, like two stage training with Discriminator
Could you please release open-source of the paper??

The settings of training parameters for red excavator

Hello!First of all,congratulations on such an amazing paper and thank you very much for making the code public.I have a question regarding the training parameters settings for the red excavator.I set use_alpha,use_feature and use_view parameters. sample_scale=45. I loaded the model you gave with the trained model and trained to 250000 iterations with a single 2080Ti GPU.But the results are not ideal, could you please share the training parameters of red excavator. Thanks you for your attention.
图片
图片

No module named 'clip'

First of all, I'd like to thank you for sharing this amazing work to all!

I'm pretty interested in using your code. However, when I try to run your code, it shows no module named 'clip'.
image

Here are the files that clip is used.
Screenshot from 2022-06-22 13-34-51

I was wondering if there's some extra software to install, or you haven't finished cleaning and uploading the code for CLIPNeRF? Thanks a lot!

Code release

Thanks for your excellent work on language guided NeRF editting! The results are really amazing!
Since two months have passed since you last answered this quetion, I'm wondering if you are now ready to release the main part of your code.
I'm quite excited to try it out! Looking forward to your reply!

new code release

Hi, Cassie,

Thank you so much for your amazing job! I'm wondering if that possible for you to release the full code. I'm really looking forward to trying it! Thanks again!

cannot get reasonable results!

Clip-NeRF is a great work, I tried to reproduce some results recently, but can not get reasoable results. I can only get the map like this:
image

Very slow training

Hi, I have encountered a very slow training, it took me about 10 seconds to finish an iteration. The clip process costs the most of the time. My training script is directly copied from your README.md without any changes, and I checked the CLIP model, input image and text are all on cuda. Is this speed normal? Is there any way to speed up this process?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.