Learning Multi-Way Relations via Tensor Decomposition With Neural Networks
Authors: Koji Maruhashi, Masaru Todoriki, Takuya Ohwa, Keisuke Goto, Yu Hasegawa, Hiroya Inakoshi, Hirokazu Anai
AAAI 2018 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Our experiments on three different domains show that our method is highly accurate, especially on higher order data. It also enables us to interpret the classification results along with the matrices calculated with the novel tensor decomposition. |
| Researcher Affiliation | Industry | Koji Maruhashi, Masaru Todoriki, Takuya Ohwa, Keisuke Goto, Yu Hasegawa, Hiroya Inakoshi, Hirokazu Anai Fujitsu Laboratories Ltd. 4-1-1, Kamikodanaka, Nakahara-ku, Kawasaki, Kanagawa, Japan {maruhashi.koji, todoriki.masaru, takuyaohwa, goto.keisuke, latente, inakoshi.hiroya, anai}@jp.fujitsu.com |
| Pseudocode | Yes | Algorithm 1: Deep Tensor |
| Open Source Code | No | The paper mentions that their method is implemented in Python using Chainer, but does not provide a link or statement about their own source code being publicly available. |
| Open Datasets | Yes | We use DARPA Intrusion Detection Data Sets1 and use a 1998 dataset... 1http://www.ll.mit.edu/ideval/data/; transactions from the Show Me the Money website2... 2http://smtm.labs.theodi.org/; ENZYMES, NCI1, and NCI109, downloaded from the ETH z urich website 3... 3https://www.bsse.ethz.ch/mlcb/research/machinelearning/graph-kernels/weisfeiler-lehman-graph-kernels.html; CHEM is provided by the Pub Chem Bio Assay 4... 4https://www.ncbi.nlm.nih.gov/pcassay |
| Dataset Splits | Yes | We choose the best parameters by using the validation datasets... On BIO, we evaluate the methods based on 10-fold cross validation repeated five times and take their average accuracy. |
| Hardware Specification | Yes | All the experiments were conducted on an Intel(R) Xeon(R) CPU E5-2630 v3 2.40GHz with 132GB of memory and a Tesla P100 GPU, running Linux. |
| Software Dependencies | Yes | Deep Tensor and Tucker NN are implemented in Python 2.7, using Chainer 1.24 (http://chainer.org/). |
| Experiment Setup | Yes | The best size of the core tensor is mainly 10 10 10 10 for IDS, 10 10 10 for LEND, and 50 50 for BIO, chosen out of {10, 25, 50} for each mode. The best number of hidden layers is mainly 4 for IDS, 3 for LEND, 4 for ENZYMES, and 4 for NCI1 and NCI109, chosen out of various numbers of hidden layers; up to 7, with 1, 000 neurons in each layer. The number of epochs for training is 200 and the mini-batch size is 100. We use ReLU (Glorot, Bordes, and Bengio 2011) as the activation function and use it with or without dropout (Srivastava et al. 2014) and batch normalization (Ioffe and Szegedy 2015) techniques. |