zhang-tao-whu / e2ec Goto Github PK
View Code? Open in Web Editor NEWE2EC: An End-to-End Contour-based Method for High-Quality High-Speed Instance Segmentation
License: Other
E2EC: An End-to-End Contour-based Method for High-Quality High-Speed Instance Segmentation
License: Other
sir, 2 days ago, I saw the project related to DVIS, but now, I cannot found that repository.. :( Do you have any plan to reopen that project?
您好,SBD数据集中官方的训练集和测试集是8498和2857,想知道论文中5623和5732的划分方式
Thanks for your excellent work!
I wonder when the multi GPU training codes will be released, which is very important for reproduction and consequent research.
I really appreciate your generosity and kindness.
Hi, thanks for your excellent work!
I tried to train this work on COCO dataset, but it converges very slowly and needs to train more than 100 epochs.
I notice that you mention it is because of CenterNet(detector) in previous issue, and you rebuild this network in FCOS version.
I wonder when will the FCOS version be released?
Or can you explain how do I change the detector? Which files and what functions should I change?
Thanks! :)
hi,你好,我将cfg.model.use_dcn 设置为False,但是训练程序运行起来还是会import dcn导致出错,这块该如何处理呢?
Thanks for the excellent paper and code.
But my experiments (DCN disabled) on KINS dataset always deadlock at epoch 20 with GPU and CPU busy.
I use command python -m torch.distributed.launch --nproc_per_node 2 train_net_ddp.py --config_file kitti --bs 4 --gpus 2
.
Is there any advice to check? Thanks advance.
BTW, what are the minimum bs and epoch in KINS dataset? Bs 64 and epoch 150 seem too huge.
Hello! Thank you for your amazing work.
I used your network (with the coco config) to train on a dataset of custom imagery with coco style object instance annotations. Now I wish to generate predictions for the test split of this dataset for which I have no annotation file. I notice you are using an empty annotation file to perform evaluations on the coco test dev split. Could you kindly guide me on how I can perform single image predictions (object polygons and seg masks for each image) on my custom dataset using the model I trained? Looking forward to your reply.
Thank you!
Hello, when we train on the KITTI dataset, the input size is (384,896) or (512,896)?
我使用了mmcv中的dcn,并按照您的说明和问题23修改了代码,除此以外没有修改。运行时有一些警告但没有报错,但准确率一直是0,可视化图像也是错误的,coco和kitti两个数据集都是一样的情况,想请教一下是不是哪里操作不当?
测试时使用的代码:
python test.py kitti --checkpoint model/model_kitti.pth --stage coarse
python visualize.py coco data/coco --checkpoint model/model_coco.pth --with_nms True --output_dir data/result/coco
得到的可视化图像是一团混乱的线条,能看到线条下方有缩小的原图
运行时的警告:
load model: model/model_kitti.pth
loading annotations into memory...
Done (t=3.27s)
creating index...
index created!
命令语法不正确。
loading annotations into memory...
Done (t=3.30s)
creating index...
index created!
0%|
UserWarning: Named tensors and all their associated APIs are an experimental feature and subject to change. Please do not use them for anything important until they are released as stable. (Triggered internally at ..\c10/core/TensorImpl.h:1156.)
return torch.max_pool2d(input, kernel_size, stride, padding, dilation, ceil_mode)
UserWarning: Default grid_sample and affine_grid behavior has changed to align_corners=False since 1.3.0. Please specify align_corners=True if the old behavior is desired. See the documentation of grid_sample for details.
"Default grid_sample and affine_grid behavior has changed "
when i run test.py show the error:
python=3.7
and i think the same one like CharlesShang/DCNv2#24.
can u help me to solve this problem?
Thanks!
Hello, thank you for your awesome work.
I am currently training a model on the Cityscapes dataset. Apart from setting the batch size to 8 in the config, everything else matches your config file. However, the final result I obtained is only around 0.27 AP, which is significantly lower than the 0.34 provided with your pretrained model. I would like to ask if there are any differences between the config used for your best model and the one you provided?
作者您好!非常欣赏您的工作,在我的数据集上表现良好。请问您有尝试过将模型部署到c++上吗?不知可否提供相关的参考和指导,非常感谢!
Thanks for your work!
But I am confused about the Multi-direction alignment (MDA) part, can you explain in more detail?
Which part of the code does the MDA correspond to? If I want to try to change the value of M, where should I adjust it?
Thanks very much for your work, I have some questions about get_gcn_feature function. In this function, we have input 'cnn_feature','img_poly' . Our main focus is to extract the features of img_polys‘ points from the feature map, is this correct? If so, we need to use torch.nn.functional.grid_sample to finish this task. So why in your code, you didn't normalize img_polys between [-1,1] to target the location in cnn_feature instead
img_poly[..., 0] = img_poly[..., 0] / (w / 2.) - 1
img_poly[..., 1] = img_poly[..., 1] / (h / 2.) - 1
I dont understand how this works, I hope you can help me out.
In torch.nn.functional.grid_sample(input, grid, mode='bilinear', padding_mode='zeros', align_corners=None)
grid specifies the sampling pixel locations normalized by the input spatial dimensions. Therefore, it should have most values in the range of [-1, 1]
Hi, thank you very much for your work. Is the pre-trained model (eg. model_sbd.pth) you provided the training result parameters of the 149th epoch? Or the epoch with the best performance among all the training epochs? Because no relevant information was found in the model.pth for you provided.
您好,我是新手,请问您的代码是基于Linux的吗
你好,我看到您的论文实验中用到了KINS数据集,我想训练自己的amodal数据集,请问如何把它做成KINS格式的,谢谢!
When there are more than 100 instances in each image, which parameter should I modify?
SBD数据集的格式跟其他数据集不太像
所以套用readme中的格式会报错
可以提供下SBD数据集的visualize方式吗
万分感谢!
author您好,请问一下有没有在Windows系统下的安装方法呢?
Hi, thank you for your work! I was wondering if it is possible to keep the org resolution during visualization, to create example Images like in your paper.
Thank you!
How to train ous dataset?
when i train in my dataset it get
pos_loss = torch.log(pred) * torch.pow(1 - pred, 2) * pos_inds
RuntimeError: The size of tensor a (80) must match the size of tensor b (2) at non-singleton dimension 1
Is the classes number problem?
I can not find the classes config
Due to my GPU device limit, I change the batchsize from 32 to 8 in cityscapes training, but it got an unexpected error that in the computation of loss, the tensor missmatch the size. I am comfused what's wrong with this error if I lower the batchsize. Or this code only can train cityscapes in batchsize of 32?
Hi, I tried the multi-gpu training code but the program always got stuck after a few iterations.
Environment:
Reproduce the bug:
CUDA_VISIBLE_DEVICES=0,1,2,3 python -m torch.distributed.launch --nproc_per_node 4 train_net_ddp.py --config_file coco --gpus 4
Output:
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/aa/anaconda3/envs/e2ec/lib/python3.7/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/site-packages/torch/distributed/launch.py", line 260, in <module>
main()
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/site-packages/torch/distributed/launch.py", line 253, in main
process.wait()
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/subprocess.py", line 1019, in wait
return self._wait(timeout=timeout)
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/subprocess.py", line 1653, in _wait
(pid, sts) = self._try_wait(0)
File "/home/a/anaconda3/envs/e2ec/lib/python3.7/subprocess.py", line 1611, in _try_wait
(pid, sts) = os.waitpid(self.pid, wait_flags)
KeyboardInterrupt```
Hello, if the number of instances in one batch is zero, the smooth L1 loss(Linit, Lcoarse, Liter) will be 'nan'. How can I address this problem?
Have you done some comparative experiments about the model with the same backbone?For example, replacing your backbone to resnet50?
The model actually produces five polygons: init, coarse, final1, final2, and final3.
Coarse is refined using the proposed global deformation module, while final* are refined using the circular convolution module.
What if we use all global deformation modules or circular convolution modules for all the polygon refinement?
dear author:
我看到您论文里写用dla-34作为backbone,dla-34是什么呢,有没有参考资料?
Q1: Is it possible to effectively fit the regression points to the real points in discrete spatial dimensions with a simple l1 loss?
dear author:
1、return self.setloss(ini_pred_poly, pred_polys_, gt_polys, keyPointsMask) 为什么不直接用pred_polys_去计算最近距离,而是采用ini_pred_poly去计算最近距离呢?
2、ground truth的轮廓坐标如何获得?
Hi, thank you for your work! Could you update the download link of cityscapes_anno.tar.gz file? Now click on it to show that the link has expired.
Thank you!
I set use_dcn = False,dowload the pretrained model and the dataset to test .The code run successfully, but the result is 0.
Hi, thanks for your wonderful work. But I found a bug in your README file. In the H3 title, Testing on COCO
, the pretrained model URL should be gpcv.whu.edu.cn/member/ZhangTao/model.zip
. However, Github converts it into a relative link in your repo, resulting in a 404 page. You can add https://
ahead of your pretrained model URL to avoid it.
Thank you very much for your work, do you know why this method needs to train more than 100 epochs, the convergence speed is very slow, which makes training a lot of inconvenience, some instance segmentation methods only need more than 20 epochs, is it because of CenterNet?
In the file dataset/data_loader.py, you have set num_workers=batch_size
in the function make_ddp_train_loader
. Is there any specific reason for this? You mentioned in issue #13 that you've done this for convenience. Can you please elaborate on this?
This looks like it should instead be num_workers=train.cfg.num_workers
. Please let me know if this is correct.
Thank you!
您好,十分想学习您的工作成果,但是很多数据集无法访问了,可以麻烦您更新一下吗?感谢!
请问为什么检测框和分割的轮廓差别如此大?
检测框可视化的代码使用的是snake仓库中的代码,
box = output['detection'][:, :4].detach().cpu().numpy()
for i in range(len(ex)):
color = next(colors).tolist()
poly = ex[i]
poly = np.append(poly, [poly[0]], axis=0)
ax.plot(poly[:, 0], poly[:, 1], color=color, linewidth=5)
x_min, y_min, x_max, y_max = box[i]
ax.plot([x_min, x_min, x_max, x_max, x_min], [y_min, y_max, y_max, y_min, y_min], color='w', linewidth=0.5)
希望可以收到回复,非常感谢
I have several questions about MDA and DML.
key point
is needed to calculate the loss. How to obtain the key point
in the training? Does it comes from annotation information?Segment-wise Matching Scheme
in DANSE
. Am I correct?A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.