Fast Learning from Distributed Datasets without Entity Matching
Authors: Giorgio Patrini, Richard Nock, Stephen Hardy, Tiberio Caetano
IJCAI 2016 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | 5 Experiments, We have evaluated the leverage that DRL provides compared to the peers..., We ran experiments on a dozen UCI domains., Table 3: Results on domain ionosphere..., Table 4: Results on domain musk... |
| Researcher Affiliation | Collaboration | Australian National University1, NICTA2, Ambiata3, University of New South Wales4 |
| Pseudocode | Yes | Algorithm 1 RADOCRAFT(P1, P2, ..., Pp), Algorithm 2 DRL(P1, P2, ..., Pp; Γ) |
| Open Source Code | No | The paper does not include an unambiguous statement about releasing the source code for the described methodology, nor does it provide a direct link to a source-code repository. |
| Open Datasets | Yes | We ran experiments on a dozen UCI domains. |
| Dataset Splits | Yes | Each peer Pj estimates learns through a ten-folds stratified cross-validation (CV) minimization of sql(Sj, ; γ Iddj)..., CV is performed on rados as follows: first, RB is split in 10 folds, RB, , for = 1, 2, ..., 10. |
| Hardware Specification | No | The paper does not provide specific hardware details (exact GPU/CPU models, processor types with speeds, memory amounts, or detailed computer specifications) used for running its experiments. |
| Software Dependencies | No | The paper does not provide specific ancillary software details (e.g., library or solver names with version numbers like Python 3.8, CPLEX 12.4) needed to replicate the experiment. |
| Experiment Setup | Yes | γ is optimized by a 10-folds CV on I ., We have carried out a very simple optimisation of the regularisation matrix of DRL as a diagonal matrix which weights differently the shared features, Γ .= Diag(lift X(proj J(1)))+ γ Diag(lift X(proj X\J(1))), for γ 2 G. |