Coder Social home page Coder Social logo

aioz-ai / miccai19-medvqa Goto Github PK

View Code? Open in Web Editor NEW
43.0 7.0 25.0 140 KB

AIOZ AI - Overcoming Data Limitation in Medical Visual Question Answering (MICCAI 2019)

Home Page: https://blog.ai.aioz.io/research/vqa-mevf/

License: MIT License

Python 99.58% Shell 0.42%
ai medical medical-image-processing deep-learning vqa visual-question-answering medvqa aioz aioz-ai miccai

miccai19-medvqa's People

Contributors

gazeal avatar quangduytran avatar xuanbinh-nguyen96 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar

miccai19-medvqa's Issues

Request for the splited data for training MAML

Hello,
The paper describes that all the images are categorized into 9 classes: head normal, head abnormal present, head abnormal organ, chest normal, chest abnormal organ, chest abnormal present,abdominal normal, abdominal abnormal organ, and abdominal abnormal present. Could you please provide the split data?
Thanks!

problem

Hello, excuse me, I reproduced it with the provided code, found that the accuracy was a little worse than the paper, and then I evaluated it with the provided weight file, found that the accuracy was even worse, and wanted to ask if this was caused by the use of different data sets, did you use the data set to make some fine-tuning? Thank you and bother.

not able to understand MAML in code files

As per mentioned in paper, for MAML: 5 tasks are considered for 1 iteration. Each task contains 3 classes out of 9 subcategories and 6 images (3 for train and 3 for validation). However inside the code sections, I am not finding this part. The validation set is not seen in the code, as only training is carried out. Also the meta update, train parts are not understandable. Can you please help to know in which file these things are considered and also a high level briefing on the code sections for MAML (meta learning).

FileNotFoundError: [Errno 2] No such file or directory: 'data_RAD/imgid2idx.json'

I downloaded the dataset from the link, then split the json file into trainset.json and test.json, but still can not find imgid2idx.json. When I run
python3 main.py --model SAN --use_RAD --RAD_dir data_RAD --maml --autoencoder --output saved_models/SAN_MEVF
FileNotFoundError: [Errno 2] No such file or directory: 'data_RAD/imgid2idx.json'

Could you please show where I can get imgid2idx.json?

code for generating the files in data_RAD

Hello, I am interested in VQA in medical field and your work is amazaing! But I am confused that how to generate the files such as trainset.json, testset.json images84x84, images128x128 in data_RAD directory with the original data. Could you plz share the code for this? Thank you!

Non-deterministic results in different runs

Thanks for publishing the code-base.

I noticed that in different runs, I get different results. It seems to be because of pytorch itself and setting torch.backends.cudnn.benchmark = False gives deterministic results in different runs. However, the performance drops by a few percentage.
I was wondering whether you had noticed this and maybe report the average or max of different runs in the paper? If that is the case, how many time you ran the experiments?

Thanks in advance.

Get the model output based on a given image

Hi, thanks for providing the autoencoder model and the checkpoints of two models. I'm wondering whether there is a quick way to encode the given new image by the pretrained CDAE, feed into the BAN model and get the results based on the original question set. Just want to have a play with this :)

The accuracy of the model

Dear author,
I followed the training step strictly and used the data_RAD/ file from the link you offered. But I can only get around 59% accuracy when testing BAN_MEVF model.
I could reproduce the accuracy of 62% using the pretrained BAN_MEVF model.
I wonder if it’s because the pretrained ae model or maml model in data_RAD/ is not the best?

How to get pretrained MAML weights

Hello,
Thank you for your kind sharing. I'm trying to reproduce this paper but having some troubles.
In base_model.py line 164, the code maml_v_emb.load_state_dict(torch.load(weight_path)) loads pretrained weights for MAML, but I can't find the corresponding .pth file.
I tried to load pretrained_maml.weights, but the weights from this file are inconsistent with the structure of MAML model.
Here are my questions:

  • How can I obtain the .pth file to initialize MAML correctly, or am I expected to train it from scratch?
  • Since pretrained_maml.weights doesn't seem to be used to initialize MAML, what's the use of it?

Thanks a lot.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.