Show simple item record

dc.contributor.authorNguyen, Quan
dc.contributor.authorPham, Hieu H.
dc.contributor.authorNguyen, Phi Le
dc.contributor.authorWong, Kok-Seng
dc.contributor.authorNguyen, Truong Thao
dc.contributor.authorDo, Minh N.
dc.date.accessioned2024-11-21T16:26:49Z
dc.date.available2024-11-21T16:26:49Z
dc.date.issued2024-01
dc.identifier.urihttps://vinspace.edu.vn/handle/VIN/433
dc.description.abstractIn Federated Learning (FL), the size of local models matters. On the one hand, it is logical to use large-capacity neural networks in pursuit of high performance. On the other hand, deep convolutional neural networks (CNNs) are exceedingly parameter-hungry, which makes memory a significant bottleneck when training large-scale CNNs on hardware-constrained devices such as smartphones or wearables sensors. Current state-of-the-art (SOTA) FL approaches either only test their convergence properties on tiny CNNs with inferior accuracy or assume clients have the adequate processing power to train large models, which remains a formidable obstacle in actual practice. To overcome these issues, we introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the proposed FedDCT allows a cluster of several clients to collaboratively train a large deep learning model by dividing it into an ensemble of several small sub-models and train them on multiple devices in parallel while maintaining privacy. In this collaborative training process, clients from the same cluster can also learn from each other, further improving their ensemble performance. In the aggregation stage, the server takes a weighted average of all the ensemble models trained by all the clusters. FedDCT reduces the memory requirements and allows low-end devices to participate in FL. We empirically conduct extensive experiments on standardized datasets, including CIFAR-10, CIFAR-100, and two real-world medical datasets HAM10000 and VAIPE. Experimental results show that FedDCT outperforms a set of current SOTA FL methods with interesting convergence behaviors. Furthermore, compared to other existing approaches, FedDCT achieves higher accuracy and substantially reduces the number of communication rounds (with 4–8 times fewer memory requirements) to achieve the desired accuracy on the testing dataset without incurring any extra training cost on the server sideen_US
dc.language.isoenen_US
dc.subjectfederated learningen_US
dc.subjectdeep convolutional neural networksen_US
dc.subjectsplit learningen_US
dc.subjectedge devicesen_US
dc.subjectedge learningen_US
dc.subjectcollaborative trainingen_US
dc.titleFedDCT: Federated Learning of Large Convolutional Neural Networks on Resource-Constrained Devices Using Divide and Collaborative Trainingen_US
dc.typeArticleen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

  • Kok-Seng Wong, PhD [16]
    Associate Professor, Computer Science program, College of Engineering and Computer Science

Show simple item record


Vin University Library
Da Ton, Gia Lam
Vinhomes Oceanpark, Ha Noi, Viet Nam
Phone: +84-2471-089-779 | 1800-8189
Contact: library@vinuni.edu.vn