Coder Social home page Coder Social logo

kristiyanvachev / leaf-question-generation Goto Github PK

View Code? Open in Web Editor NEW
119.0 4.0 27.0 189 KB

Easy to use and understand multiple-choice question generation algorithm using T5 Transformers.

License: MIT License

Python 3.54% Jupyter Notebook 96.46%
question-generation multiple-choice transformers mcq distractors t5 ml neural-networks quiz test

leaf-question-generation's People

Contributors

kristiyanvachev avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

leaf-question-generation's Issues

Some questions regarding training

Hi

Thank you very much for your implementation. It's been very helpful to me.
Regarding your example script to generate questions and answers (google collab), I would like you to clarify some doubts for me, if possible.

1 - Do you know if it is possible to get generated tokens in training_step? Instead of getting them just at the end, via generate() inference method.

2 - Do you have a particular reason to do encoding using tokenizer(answer + <sep> + context, ...) instead of using tokenizer(answer, context, ...)?

3 - Have you encountered overfitting throughout your experiments? Unfortunately, my dev loss only improves up to the second epoch (cross-entropy loss of 1.35) and then increases. From what I observer from your experiments, you can reach the 4th epoch with a loss of 1.17374.
Note: I am using the same SQUAD v1.1 splits, model: t5-base, batch size: 32, optimizer: AdamW (eps = 1e-6).

4 - Related to the previous question. Do you have any idea what a "good loss" is for this QG task? In any case, I haven't seen the dev loss reach a value lower than 1.

Thanks in advance.
Bernardo

Inference time of the distractor generation module

Hi Kristian,

Thanks for the code, it is really helpful.
I'm wondering about one thing: how do you ensure that your distractor module will output a sequence with 2 tokens in it? Because it appears to me that training with three distractors doesn't necessarily make you 100% certain that the model will generate outputs of this form. Or do you do multiple samples and just concatenate them?
Thanks in advance.

Python Version

Hello
Could you please tell what python version did you use ?

Clarification on how to run the script

Hello,
First of all, amazing work! I have so much to learn!

I apologize if the question is too basic: I followed all the steps of installation but I'm not sure how I can actually "run" the code in a way that the Angular app can send requests to it.

Thank you very much!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.