Coder Social home page Coder Social logo

deltas's People

Contributors

asinha87 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

deltas's Issues

How to obtain ground truth 2D point matches?

Hi, thanks for your excellent work. In the original paper you write:

3.4 Overall training objective
(b) a smooth-L1 loss between the 2D points output after soft argmax and ground truth 2D point matches...

I am a bit confused by the term "ground truth 2D point matches". Are they obtained by transforming the 2D points in a reference image to a target image using the ground truth pose?

The image size used for depth evaluation

Hello!
I found that the evaluation results of DPSNET on SUN3D in your paper have obvious differences with the original paper. I check your code, and I found that you downsample the GT depth to (320,240) to evaluate with pred_depth while prior works usually upsample the pred_depth to the original size to do the evaluation. Although this minor difference won't influence your paper's conclusion, maybe you could clarify this difference in the READE.ME to prevent other readers from misusing your numbers reported in your paper. If possible, may you report your tables with the original size like other works and readers could easily cite your tables?

Thank you.

The inference time

Hello!
First of all, DELTAS is a great research and an easy-to-work-with code, so thank you.
I see in the paper that you report the forward pass time as ~90 milliseconds on a Titan RTX. I am using image size 320 x 240 for all of the upcoming values that I write here. Also I am using the faster SVD implementation that you share. I am trying to measure the inference time on a GTX 1060 and a GTX 1080Ti, but I can only measure ~530 milliseconds on 1060, and ~500 milliseconds on 1080Ti per prediction with batch size = 1. If I increase the batch size to 8, I can get around 200 milliseconds on 1080Ti per prediction. It might be possible to get 90 milliseconds per prediction with batch size 16 on a Titan RTX, unfortunately I can not test it. But I am not sure if it is meaningful to measure the inference time with batch size 16 anyway, in fact anything over 1 is pretty much always memory dependent and does not reflect the real frames per second value. The problem is, I am able to measure ~90 ms for MVDepthNet (this is close to what you report as ~60 ms), and ~380 ms for DPSNet with batch size = 1. And you clearly state that DPSNet is 2.5 times slower than DELTAS.
Maybe I am making a mistake, could you please share the code or share some details on how you measure the inference time?

About the function "patch_for_depth_guided_range" in triangulation.py

Thank you for your remarkable work!

The implementation of learnable triangulation is a great inspiration.

But I'm confused about the function "patch_for_depth_guided_range" in triangulation.py. The calculation of the fundamental matrices and epipolar lines make sense to me. But the meaning of variables like x2ord, y2ord, x2, and y2 isn't clear to me.

Would you please help to explain this function? Thank you.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.