aim-uofa / matcher Goto Github PK
View Code? Open in Web Editor NEW[ICLR'24] Matcher: Segment Anything with One Shot Using All-Purpose Feature Matching
Home Page: https://arxiv.org/abs/2305.13310
License: MIT License
[ICLR'24] Matcher: Segment Anything with One Shot Using All-Purpose Feature Matching
Home Page: https://arxiv.org/abs/2305.13310
License: MIT License
Congratulation for your great work!
I have several questions about bidirectional matching:
1、How do you perform bipartite matching between the points on the reference mask Pr and the patch-level features of target image Zt. As the dimension of the reference mask is 3256192 and the dimension of the patch feature is 7681612.
2、How to solve the problem of having more points than patch features(192 patch) using bipartite matching in forward matching? The number of Pr is much greater than the number of patch feature, how to evaluate the similarity of the Pr and Zt?
3、Could you elaborate on the implementation process and implementation details of forward matching or the bidirectional matching?
When will you release the code for training?
I want to know it. Thanks a lot!
In forward matching, what is the dimesion of the points on the reference mask Pr? And what does the L means?
Hi, thanks for the great article
May I ask you what time are you going to release the source code? Will you release the pretrained models too?
Hello,
Congratulation for your great and interesting work!
I have several questions to see if this model match my use case:
1- If I run the model on a target image without the reference object, will it still predict something or will it be able to say (with a given confidence) that the image does not have the queried object ?
2- I am interested to run this model with several categories as inputs. Is there a mecanism to run the inference on several categories at the same time or will I have to run distinct predictions for each categories ?
3- Can the model be extended to do few-shot with several reference masks for one same object ?
Thank you in advance!
作者您好,想请教几个问题。
1.论文里似乎只提到对于三种level有不同的采样的策略,但具体是怎么采样的似乎没有提及。请问可以展开说说吗?
2.关于kmeans++的cluster center数量似乎也没有明确,请问可以说说相关的选择策略吗?
3.图1中,输入point+box prompt后SAM似乎有多个mask proposals,sam中默认mult_mask_outputs的数量是3,请问你们是设置了多少?
Thank you for very inspiring work! Our research group especially thinks that creation of new, demanding few shot dataset is a great idea. As we would like to use this dataset for evaluation we have a few questions about lvis 92-i dataset:
Thank you very much for your response in advance!
Thank you for your outstanding work!
Can you please describe how patch-level features are generated and how they are sized?
Also, I'd like to ask what the center prompt means and how the model generates it.
Your excellent will be a great help to my research!
Hi, thanks for the great work and code!
I want to try it on VOS tasks, but that part of code is not released for now.
Could we expect the release of VOS code in the near future, or some estimated date? Thank you!
Thank you for your outstanding work!
The paper mentions that you use SAM as the class-agnostic segmentation model, does this mean that Matcher does not have the ability to recognize semantic information while segmenting?
In the meantime, I'm curious as to when the source code will be released.
Your excellent will be a great help to my research!
Hi, thanks for this great work!
I tried to test Matcher on COCO-20i by following the command here but I got following errors.
FileNotFoundError: [Errno 2] No such file or directory: 'datasets/COCO2014/annotations/val2014/COCO_val2014_000000507081.png'
I believe this is because the code requires the COCO-20i mask annotation to be in image format, but the official annotation is in json format.
Could you please check it and provide the converted image format of the mask annotation as well, thank you so much!
Have you try directly use SAM encoder to extract feature instead use other pretrained model?
作者您好!请问可以在Windows系统上用Anaconda部署吗,我看安装说明里面只说了Linux和mac。谢谢!
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.