Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Hierarchical Inter-Attention Network for Document Classification with Multi-Task Learning

Authors: Bing Tian, Yong Zhang, Jin Wang, Chunxiao Xing

IJCAI 2019 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experimental results on 15 public datasets demonstrate the benefits of our proposed model.
Researcher Affiliation Academia 1RIIT, TNList, Dept. of Computer Science and Technology, Tsinghua University, Beijing, China. 2Computer Science Department, University of California, Los Angeles
Pseudocode No The paper describes its methods using text and mathematical equations and provides a network architecture diagram (Figure 2, Figure 3, Figure 4), but does not include any explicit pseudocode or algorithm blocks.
Open Source Code No The paper does not include any explicit statement about open-sourcing the code for the methodology or provide a link to a code repository.
Open Datasets Yes The first 14 datasets are Amazon product reviews coming from different domains such as Books, Music, Baby, etc. These datasets are collected based on the dataset 1provided by Blitzer et al. [2007]. The last IMDB dataset contains movie reviews with binary classes [Maas et al., 2011]. 1https://www.cs.jhu.edu/ mdredze/datasets/sentiment/
Dataset Splits Yes Following previous studies, we randomly split these datasets into training sets, development sets and testing sets with the proportion of 70%, 10% and 20% respectively.
Hardware Specification No The paper does not explicitly describe the hardware specifications (e.g., specific GPU or CPU models, memory details) used for running the experiments.
Software Dependencies No The paper mentions using GloVe vectors and the Adam optimizer, but it does not specify any software dependencies with version numbers (e.g., Python, TensorFlow/PyTorch versions, or specific library versions).
Experiment Setup Yes The detailed settings of hyper-parameters are shown in Table 2. Word embedding size d = 200 Size of word-level basic-LSTM layer hw = 50 ... Initial learning rate 0.001 Regularization 1E 5