Modeling Selective Feature Attention for Lightweight Text Matching

Authors: Jianxiang Zang, Hui Liu

IJCAI 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experimental evaluations conducted across diverse text matching baselines and benchmarks underscore the indispensability of modeling feature attention and the superiority of the selection mechanism. Table 1 reports the evaluation accuracies of six lightweight text matching baselines, as well as their performances following the integration of FA and SFA blocks.
Researcher Affiliation Academia Jianxiang Zang , Hui Liu School of Statistics and Information, Shanghai University of International Business and Economics {21349110, liuh}@suibe.edu.cn,
Pseudocode No The paper includes equations and figures illustrating the architecture and processes, but it does not contain a dedicated pseudocode or algorithm block.
Open Source Code Yes 1Codes available:https://github.com/hggzjx/SFA
Open Datasets Yes evaluated their performance on following benchmarks: QQP [Iyer et al., 2017], MRPC [Dolan and Brockett, 2005], Bool Q [Clark et al., 2019], SNLI[Bowman et al., 2015], MNLI [Williams et al., 2018](matched&mismatched), QNLI [Wang et al., 2018], and Scitail [Khot et al., 2018].
Dataset Splits No The paper evaluates performance on various text matching benchmarks but does not explicitly provide the specific percentages or sample counts for training, validation, and test splits used to reproduce the data partitioning. While Figure 3 refers to 'dev.' for loss curves, it doesn't detail the split ratios.
Hardware Specification No The paper does not provide specific hardware details (e.g., GPU/CPU models, memory) used for running its experiments.
Software Dependencies No The paper does not provide specific ancillary software details with version numbers (e.g., library or solver names with versions) needed to replicate the experiment.
Experiment Setup No The paper mentions controlling 'r' and 'N' hyperparameters to manage parameter increment but does not provide specific experimental setup details such as learning rates, batch sizes, number of epochs, or optimizer settings for training.