salesforce / glad Goto Github PK
View Code? Open in Web Editor NEWGlobal-Locally Self-Attentive Dialogue State Tracker
License: BSD 3-Clause "New" or "Revised" License
Global-Locally Self-Attentive Dialogue State Tracker
License: BSD 3-Clause "New" or "Revised" License
Could you please provide the cleaned DSTC2 data? I found this link (mi.eng.cam.ac.uk/~nm480/dstc2-clean.zip) is invalid.
Same issue with stanfordnlp/stanza#16. Any solution?
I noticed that they had the same input
I am trying to run the train.py and the embedding I am using is Wikipedia embedding with 50 dimension size in python 3. I am getting this following error.
Namespace(batch_size=50, demb=400, dexp='exp', dhid=200, dout='exp/glad/default', dropout={'emb': 0.2, 'local': 0.2, 'global': 0.2}, epoch=50, gpu=0, lr=0.001, model='glad', nick='default', resume=None, seed=42, stop='joint_goal', test=False)
WARNING:root:loading split train
WARNING:root:loading split dev
WARNING:root:loading split test
INFO:root:dataset sizes: {'dev': 200, 'test': 400, 'train': 600}
INFO:root:loaded model <class 'models.glad.Model'>
INFO:root:saving config to exp/glad/default/config.json
Traceback (most recent call last):
File "train.py", line 66, in
run(args)
File "train.py", line 24, in run
model.load_emb(Eword)
File "/mount/studenten/SpokenLanguageProcessing/2019/WokeSpoke/SLU/glad_master/models/glad.py", line 151, in load_emb
self.emb_fixed.weight.data.copy_(new(Eword))
RuntimeError: The expanded size of the tensor (400) must match the existing size (150) at non-singleton dimension 1. Target sizes: [950, 400]. Tensor sizes: [950, 150]
On my local machine, when I run python preprocess_data.py
and the script computes word embeddings, it dies with a MemoryError. I've killed all other processes and have about 6GB free RAM, but that doesn't seem to be enough. Is this expected? Anything I can do against it? Perhaps download and use precomputed embeddings from somewhere?
@vzhong Thank you!
Hi.
Since the local RNN for a slot type is a bidirectional RNN, should the local self-attention component be initialized as
SelfAttention(2 * dhid, dropout=self.dropout.get('selfattn', 0.))
instead of
SelfAttention(din, dropout=self.dropout.get('selfattn', 0.))
as in the line 101 of glad.py?
I have downloaded the clean dstc-2 dataset and trained the system but unable to reproduce the results. Please share the hyper-parameters for the dstc-2 dataset.
It seems that one batch contains batch_size of one-turn-utterance.
@vzhong Thank you!!
I didn't find any code of dstc2.0
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.