Login (DCU Staff Only)
Login (DCU Staff Only)

DORAS | DCU Research Repository

Explore open access research and scholarly works from DCU

Advanced Search

A comparison of distributed machine learning methods for the support of “many labs” collaborations in computational modeling of decision making

Zhang, Lili orcid logoORCID: 0000-0002-2203-2949, Vashisht, Himanshu orcid logoORCID: 0000-0002-2726-946X, Totev, Andrey orcid logoORCID: 0000-0002-9093-5744, Trinh, Nam orcid logoORCID: 0000-0003-0307-3793 and Ward, Tomás E. orcid logoORCID: 0000-0002-6173-6607 (2022) A comparison of distributed machine learning methods for the support of “many labs” collaborations in computational modeling of decision making. Frontiers in Psychology, 13 . ISSN 1664-1078

Abstract
Deep learning models are powerful tools for representing the complex learning processes and decision-making strategies used by humans. Such neural network models make fewer assumptions about the underlying mechanisms thus providing experimental flexibility in terms of applicability. However, this comes at the cost of involving a larger number of parameters requiring significantly more data for effective learning. This presents practical challenges given that most cognitive experiments involve relatively small numbers of subjects. Laboratory collaborations are a natural way to increase overall dataset size. However, data sharing barriers between laboratories as necessitated by data protection regulations encourage the search for alternative methods to enable collaborative data science. Distributed learning, especially federated learning (FL), which supports the preservation of data privacy, is a promising method for addressing this issue. To verify the reliability and feasibility of applying FL to train neural networks models used in the characterization of decision making, we conducted experiments on a real-world, many-labs data pool including experiment data-sets from ten independent studies. The performance of single models trained on single laboratory data-sets was poor. This unsurprising finding supports the need for laboratory collaboration to train more reliable models. To that end we evaluated four collaborative approaches. The first approach represents conventional centralized learning (CL-based) and is the optimal approach but requires complete sharing of data which we wish to avoid. The results however establish a benchmark for the other three approaches, federated learning (FL-based), incremental learning (IL-based), and cyclic incremental learning (CIL-based). We evaluate these approaches in terms of prediction accuracy and capacity to characterize human decision-making strategies. The FL-based model achieves performance most comparable to that of the CL-based model. This indicates that FL has value in scaling data science methods to data collected in computational modeling contexts when data sharing is not convenient, practical or permissible.
Metadata
Item Type:Article (Published)
Refereed:Yes
Uncontrolled Keywords:deep learning; decision-making; distributed learning; federated learning; data privacy
Subjects:UNSPECIFIED
DCU Faculties and Centres:DCU Faculties and Schools > Faculty of Engineering and Computing > School of Computing
Research Institutes and Centres > INSIGHT Centre for Data Analytics
Publisher:Frontiers Media
Official URL:https://doi.org/10.3389/fpsyg.2022.943198
Copyright Information:© 2022 The Authors.
Funders:Allied Irish Banks and Science Foundation Ireland under Grant Number SFI/12/RC/2289_P2.
ID Code:28123
Deposited On:07 Mar 2023 12:54 by Thomas Murtagh . Last Modified 07 Mar 2023 12:58
Documents

Full text available as:

[thumbnail of fpsyg-13-943198.pdf]
Preview
PDF - Requires a PDF viewer such as GSview, Xpdf or Adobe Acrobat Reader
2MB
Downloads

Downloads

Downloads per month over past year

Archive Staff Only: edit this record