Coder Social home page Coder Social logo

mickeysjm / synsetmine-pytorch Goto Github PK

View Code? Open in Web Editor NEW
65.0 2.0 14.0 71.7 MB

PyTorch implementation of paper "Mining Entity Synonyms with Efficient Neural Set Generation" in AAAI 2019

Python 98.10% Shell 1.90%
pytorch supervised-clustering synonym-discovery synonym-dataset

synsetmine-pytorch's Introduction

Mining Entity Synonyms with Efficient Neural Set Generation

Documentation Status

This repo includes datasets, model training scripts, and model evaluation scripts used in paper -- Mining Entity Synonyms with Efficient Neural Set Generation.

Details about SynSetMine model can be accessed here, and this implementation is based on the PyTorch library.

The documents would be available here.

Installation

Simply clone this repository via

git clone https://github.com/mickeystroller/SynSetMine-pytorch.git
cd SynSetMine-pytorch

Check whether the below dependencies are satisfied. If not, simply install them via

pip install -r requirements_full.txt

Training Model

You can train SynSetMine model and test its performance using commands in run.sh

chmod +x run.sh
./run.sh

By default, we will run on NYT dataset. You can uncomment the code in run.sh to run on the other two datasets.

Model snapshots will be saved in ./snapshots/ directory. Logs will be saved in ./runs/ directory, and final results will be stored in ./results/ directory.

Loading Pre-trained Model for Prediction

We save three pre-trained models, one for each dataset in ./snapshots/ directory. You can load them directly for prediction via:

chmod +x predict.sh
./predict.sh

Dependencies

  • Python 3 with NumPy
  • PyTorch > 0.4.0
  • sklearn
  • tensorboardX (to display/log information while model running)
  • gensim (to load embedding files)
  • tqdm (to display information while model running)
  • networkx (to calculate one particular evaluation metric)

Screenshot

References

If you find this code useful for your research, please cite the following paper in your publication:

@inproceedings{Shen2019SynSetMine,
  title={Mining Entity Synonyms with Efficient Neural Set Generation},
  author={Jiaming Shen and Ruiilang Lv and Xiang Ren and Michelle Vanni and Brian Sadler and Jiawei Han},
  booktitle={AAAI},
  year={2019}
}

synsetmine-pytorch's People

Contributors

mickeysjm avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

synsetmine-pytorch's Issues

About modify element_set to Dataset

Hey, I transfer the element_set.py to Dataset, so I need apply element_set.ElementSet.get_train_batch to __getitem__.

I have already done the above work. But my Dataset can't achieve the sample performance with element_set.ElementSet.

my code as follow:

class SynsetDataset(Dataset):

    def __init__(self,
                 ....
                 ):
        

    def __len__(self):
        return len(self.train_labels)

    ...

    def __getitem__(self, index):
        sets = self.train_sets[index]
        # word_sets = np.zeros(self.max_set_lengths)
        # word_sets[:len(sets)] = sets
        word_sets = sets
        word_inst = self.train_insts[index]
        label = self.train_labels[index]
        return torch.tensor(word_sets).long(), torch.tensor(word_inst).long(), torch.tensor(label).long()

    def load_word_synset(self, synset_path):
        """
        读取文本同义词集合,c

        :synset_path: the path of synset
        :type synset_path: str
        :return :None
        :rtype: None
        """
        text = read_synset(synset_path)
        random.shuffle(text)
        for line in tqdm(text, desc='loading word synset...'):
            words = sorted([self.word2index[word]for word in line])
            self.positive_sets.append(words)
            self.vocab.extend(words)

        self.vocab = sorted(self.vocab)
        # for sets in self.positive_sets:
        #     self.vocab.extend(sets)


    def sample_pos_neg_data(self, raw_sets, max_length, neg_sample_size):
        """
        same with synsetmine
        """
        sip_triplets = []
        pos_sip_cnt_sum = 0
        neg_sip_cnt_sum = 0

        sample_sets = []
        sample_insts = []
        sample_labels = []

        for subset_size in range(1, max_length + 1):
            for raw_set in raw_sets:  # 每一个都是同义词集
                if len(raw_set) < subset_size:
                    continue
                raw_set_new = raw_set.copy()
                random.shuffle(raw_set_new)
                batch_set = []
                batch_pos = []
                if len(raw_set) == subset_size:  # put the entire full set
                    for _ in range(neg_sample_size + 1):
                        batch_set.append(raw_set)
                        batch_pos.append(random.choice(raw_set))
                else:
                    for _ in range(neg_sample_size + 1):
                        '''
                        subset_size = 1
                        raw_set_new = [a,b,c,d]

                        '''
                        for start_idx in range(0, len(raw_set_new) - subset_size, subset_size + 1):
                            # slide window= subset_size
                            subset = raw_set_new[start_idx:start_idx + subset_size]
                            # 取subset后一个词
                            pos_inst = raw_set_new[start_idx + subset_size]
                            batch_set.append(subset)
                            batch_pos.append(pos_inst)
                        random.shuffle(raw_set_new)

                pos_sip_cnt = int(len(batch_set) / (neg_sample_size + 1))
                pos_sip_cnt_sum += pos_sip_cnt
                neg_sip_cnt = int(pos_sip_cnt * neg_sample_size)
                neg_sip_cnt_sum += neg_sip_cnt

                negative_pool = [
                    ele for ele in self.vocab if ele not in raw_set]
                sample_size = math.gcd(
                    neg_sip_cnt, len(negative_pool))  # 最大公约数
                sample_times = int(neg_sip_cnt / sample_size)

                batch_neg = []
                for _ in range(sample_times):
                    batch_neg.extend(random.sample(
                        negative_pool, sample_size))
                    # 每次抽出相同大小的neg
                for idx, subset in enumerate(batch_set):

    
                    if idx < pos_sip_cnt:
                        pos = batch_pos[idx]

                        # 设置随机
                        # random.shuffle(subset)
                        sample_sets.append(subset)
                        sample_insts.append(pos)
                        sample_labels.append(1)
                        sip_triplets.append((subset, pos, 1))
                    else:
                        neg = batch_neg[idx - pos_sip_cnt]
                        # random.shuffle(subset)
                        sip_triplets.append((subset, neg, 0))
                        sample_sets.append(subset)
                        sample_insts.append(neg)
                        sample_labels.append(0)
                    

        print(len(sip_triplets))
        return sample_sets, sample_insts, sample_labels

  
    def generate_data(self,sample_iter=15,strategy='paper'):
 
        print(len(self.positive_sets))
        for i in range(sample_iter):
            if strategy=='paper':
                # paper strategy
                # best performance 0.27
                sample_sets, sample_insts, sample_labels = self.sample_pos_neg_data(
                    self.positive_sets, self.max_set_lengths, self.neg_sample_size)

            self.train_sets.extend(sample_sets)
            self.train_insts.extend(sample_insts)
            self.train_labels.extend(sample_labels)
        return self.train_sets, self.train_insts, self.train_labels

the difference of my code is that my code is seem to offline sample? , ElementSet is online sample .

Could you give my some advice?

thanks! :D

Share Token method to sample negative items

Hi !
In your paper, you proposed a method to sample negative items named share token right?
I can not understand what it means!
However,
in your source code, I only find complete-random method to sample negative items.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.