ToDD: Topological Compound Fingerprinting in Computer-Aided Drug Discovery

Authors: AndaƧ Demir, Baris Coskunuzer, Yulia Gel, Ignacio Segovia-Dominguez, Yuzhou Chen, Bulent Kiziltan

NeurIPS 2022 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We perform extensive numerical experiments in VS, showing that our To DD models outperform all state-of-the-art methods by a wide margin (See Figure 1).
Researcher Affiliation Collaboration Andac Demir Novartis andac.demir@novartis.com Baris Coskunuzer University of Texas at Dallas coskunuz@utdallas.edu Ignacio Segovia-Dominguez University of Texas at Dallas Jet Propulsion Laboratory, Caltech Yuzhou Chen Temple University Yulia Gel University of Texas at Dallas National Science Foundation Bulent Kiziltan Novartis bulent.kiziltan@novartis.com
Pseudocode No The paper describes the process in numbered steps but does not provide a formal pseudocode block or algorithm listing.
Open Source Code No The paper does not contain an explicit statement about releasing the source code for the described methodology or a direct link to a code repository for their work.
Open Datasets Yes Cleves-Jain: This is a relatively small dataset [26] that has 1149 compounds.* There are 22 different drug targets, and for each one of them the dataset provides only 2-3 template active compounds dedicated for model training, which presents a few-shot learning task. ... *Cleves-Jain dataset: https://www.jainlab.org/Public/SF-Test-Data-Drug Space-2006.zip DUD-E Diverse: DUD-E (Directory of Useful Decoys, Enhanced) dataset [67] is a comprehensive ligand dataset with 102 targets and approximately 1.5 million compounds.* ... *DUD-E Diverse dataset: http://dude.docking.org/subsets/diverse
Dataset Splits Yes The performance of all models was assessed by 5-fold cross-validation (CV).
Hardware Specification Yes Training time of To DD-Vi T and To DD-Conv Ne Xt for each individual drug target takes less than 1 hour on a single GPU (NVIDIA RTX 2080 Ti).
Software Dependencies No The paper mentions software components like 'Adam optimizer' and 'Conv Ne Xt_tiny models' but does not specify their version numbers or the versions of underlying libraries like Python or PyTorch.
Experiment Setup Yes Transfer learning via fine-tuning Vi T_b_16 and Conv Ne Xt_tiny models using Adam optimizer with a learning rate of 5e-4, no warmup or layerwise learning rate decay, cosine annealing schedule for 5 epochs, stochastic weight averaging for 5 epochs, weight decay of 1e-4, and a batch size of 64 for 10 epochs in total led to significantly better performance in Enrichment Factor and ROC-AUC scores compared to training from scratch.