Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Exploring Effective Inter-Encoder Semantic Interaction for Document-Level Relation Extraction

Authors: Liang Zhang, Zijun Min, Jinsong Su, Pei Yu, Ante Wang, Yidong Chen

IJCAI 2023 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experimental results on four benchmark datasets prove the effectiveness of our model.
Researcher Affiliation Academia School of Informatics, Xiamen University, China Key Laboratory of Digital Protection and Intelligent Processing of Intangible Cultural Heritage of Fujian and Taiwan (Xiamen University), Ministry of Culture and Tourism, China
Pseudocode No The paper describes methods and equations but does not include any explicit pseudocode or algorithm blocks.
Open Source Code Yes Our source code is available at https://github.com/ Deep Learn XMU/Doc RE-BSI.
Open Datasets Yes We evaluate our model on four commonly-used datasets: Doc RED [Yao et al., 2019], Revisit-Doc RED [Huang et al., 2022], Re-Doc RED [Tan et al., 2022b], DWIE [Zaporojets et al., 2021].
Dataset Splits Yes Doc RED [Yao et al., 2019] is a large-scale document-level RE dataset with 96 predefined relations... It contains 5,053 documents, which is divided into 3,053 documents for training, 1,000 for development, and 1,000 for test.
Hardware Specification No The paper does not provide specific details about the hardware (e.g., GPU, CPU models, memory) used for running the experiments.
Software Dependencies No The paper mentions software like 'Huggingface’s Transformers' and 'PyTorch' but does not specify their version numbers, which are required for a reproducible description of ancillary software.
Experiment Setup Yes We use AdamW [Loshchilov and Hutter, 2019] as our optimizer, which is equipped with a weight decay of 1e-4 and a linear warmup [Goyal et al., 2017] for the first 6% training steps. ... α and β are hyper-parameters, which are empirically set to 0.1 and 0.01, respectively.