Giter VIP home page Giter VIP logo

ldsgm's Introduction

LDSGM

A Label Dependence-aware Sequence Generation Model for Multi-level Implicit Discourse Relation Recognition (AAAI 2022)

Main Dependecies and installation

pytorch 1.3.1

transformer 4.12.4

pytorch_pretrained_bert 0.6.2

DataSet

  1. Download the PDTB2.0 , put it under /raw/
  2. python3 preprocess.py
  3. Download the pretrained model roberta base,put it under /pretrained/roberta-base/

Run

python3 run.py

ldsgm's People

Contributors

galbya avatar nlpersecjtu avatar

Stargazers

Shahana Mogal avatar Dongqi avatar Yuxin Jiang 姜宇心 avatar  avatar Liu tianrui avatar SEV avatar  avatar

Watchers

 avatar  avatar

ldsgm's Issues

Question for the data.

Hi,

(1) The data size in this paper is "the training set with 12,775 instances (Section 2-20), the validation set with 1,183 instances (Section 0-1), and the test set with 1,046 instances (Section 21-22)", Is this before or after processing?

*Processing means: "Further, there exist 16 second-level labels, five of which with few training instances and no validation and test instance are removed. Therefore, we conduct an 11-way classification on the second-level labels. " in the paper.

(2) And in paper "On the Importance of Word and Sentence Representation Learning in Implicit Discourse Relation Classification", the data size is Train/dev/test: 12362, 1183, 1046. Which one is right and If there any code to process the dataset.

Thank you!

questions about data preprocessing and graph construction

Hi,

Thanks for your interesting work. I am confused about two parts of your implementation.

The first one is the data proprecessing. In the preprocess function, I see you maintain several "other" arrays, such as arg1_train_other and arg2_train_other and so on. If a sample doesn't have a second-level sence label, it will be assigned with a default label and added to those arrays. When evaluating, you calculate results based on the predictions on both normal samples and samples with default labels. This is fine for top-level evaluation, but not the case for second-level. Because previous works usually consider only samples with gold second level labels.

The second one is the construction of the graph. You mentioned in your paper that nodes will have a self-loop edges. But in the implementation, there is not self-loop in the adjacent matrix your provided. This contradicts your description.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.