GLPocket: A Multi-Scale Representation Learning Approach for Protein Binding Site Prediction

Authors: Peiying Li, Yongchang Liu, Shikui Tu, Lei Xu

IJCAI 2023 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experiments show that GLPocket improves by 0.5% 4% on DCA Top-n prediction compared with previous state-of-the-art methods on four datasets. Our code has been released in https://github.com/CMACH508/GLPocket.
Researcher Affiliation Academia 1Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China 2Guangdong Institute of Intelligence Science and Technology, Zhuhai, Guangdong, China {lipeiying, liuyongchang, tushikui, leixu}@sjtu.edu.cn
Pseudocode No The paper does not contain any structured pseudocode or algorithm blocks.
Open Source Code Yes Our code has been released in https://github.com/CMACH508/GLPocket.
Open Datasets Yes We use sc PDB [Desaphy et al., 2015] as training set, COACH420, HOLO4k, PDBbind [Wang et al., 2005], SC6K as testing sets.
Dataset Splits Yes We divide the dataset into ten parts and use one of them as validation dataset.
Hardware Specification Yes GLPocket is implemented in Py Torch and trained for 30 epochs with a batch size of 12 on 3 A100 GPUs.
Software Dependencies No GLPocket is implemented in Py Torch and trained for 30 epochs with a batch size of 12 on 3 A100 GPUs. SGD optimizer was applied to train the model. The learning rate is set to 0.001 and remains the same. The binary cross entropy loss is employed to optimize our network.
Experiment Setup Yes GLPocket is implemented in Py Torch and trained for 30 epochs with a batch size of 12 on 3 A100 GPUs. SGD optimizer was applied to train the model. The learning rate is set to 0.001 and remains the same. The binary cross entropy loss is employed to optimize our network.