Loading…

CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning

In a practical setting, how to enable robust Federated Learning (FL) systems, both in terms of generalization and personalization abilities, is one important research question. It is a challenging issue due to the consequences of non-i.i.d. properties of client’s data, often referred to as statistic...

Full description

Saved in:
Bibliographic Details
Published in:Engineering applications of artificial intelligence 2024-07, Vol.133, p.108093, Article 108093
Main Authors: Le, Huy Q., Nguyen, Minh N.H., Pandey, Shashi Raj, Zhang, Chaoning, Hong, Choong Seon
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c307t-dc30f3105b8269da593cfcb7e4c92fe4b6767dac7ab324506a2ac75c15dd89743
container_end_page
container_issue
container_start_page 108093
container_title Engineering applications of artificial intelligence
container_volume 133
creator Le, Huy Q.
Nguyen, Minh N.H.
Pandey, Shashi Raj
Zhang, Chaoning
Hong, Choong Seon
description In a practical setting, how to enable robust Federated Learning (FL) systems, both in terms of generalization and personalization abilities, is one important research question. It is a challenging issue due to the consequences of non-i.i.d. properties of client’s data, often referred to as statistical heterogeneity, and small local data samples from the various data distributions. Therefore, to develop robust generalized global and personalized models, conventional FL methods need to redesign the knowledge aggregation from biased local models while considering huge divergence of learning parameters due to skewed client data. In this work, we demonstrate that the knowledge transfer mechanism achieves these objectives and develop a novel knowledge distillation-based approach to study the extent of knowledge transfer between the global model and local models. Henceforth, our method considers the suitability of transferring the outcome distribution and (or) the embedding vector of representation from trained models during cross-device knowledge transfer using a small proxy dataset in heterogeneous FL. In doing so, we alternatively perform cross-device knowledge transfer following general formulations as (1) global knowledge transfer and (2) on-device knowledge transfer. Through simulations on three federated datasets, we show the proposed method achieves significant speedups and high personalized performance of local models. Furthermore, the proposed approach offers a more stable algorithm than other baselines during the training, with minimal communication data load when exchanging the trained model’s outcomes and representation.
doi_str_mv 10.1016/j.engappai.2024.108093
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_engappai_2024_108093</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0952197624002513</els_id><sourcerecordid>S0952197624002513</sourcerecordid><originalsourceid>FETCH-LOGICAL-c307t-dc30f3105b8269da593cfcb7e4c92fe4b6767dac7ab324506a2ac75c15dd89743</originalsourceid><addsrcrecordid>eNqFkN1KAzEUhIMoWKuvIHmBrfnZTTZeKatVsSBIvQ7Z5GxJrdklWat9e1Oq114N5zAzDB9Cl5TMKKHiaj2DsDLDYPyMEVbmZ00UP0ITWkteCCnUMZoQVbGCKilO0VlKa0IIr0sxQa_N3fOymC-ucRP7lAoHW28Bv4f-awNuBXiMJqQOIv5MPqzwEPvvHXZmNAlG7APuwEE0Izi8ARND9pyjk85sElz86hS9ze-XzWOxeHl4am4XheVEjoXL0nFKqrZmQjlTKW4720oorWIdlG1eLp2x0rSclRURhuWjsrRyrlay5FMkDr12vzxCp4foP0zcaUr0noxe6z8yek9GH8jk4M0hCHnd1kPUyXoIFpyPYEftev9fxQ-rHnCv</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning</title><source>ScienceDirect Journals</source><creator>Le, Huy Q. ; Nguyen, Minh N.H. ; Pandey, Shashi Raj ; Zhang, Chaoning ; Hong, Choong Seon</creator><creatorcontrib>Le, Huy Q. ; Nguyen, Minh N.H. ; Pandey, Shashi Raj ; Zhang, Chaoning ; Hong, Choong Seon</creatorcontrib><description>In a practical setting, how to enable robust Federated Learning (FL) systems, both in terms of generalization and personalization abilities, is one important research question. It is a challenging issue due to the consequences of non-i.i.d. properties of client’s data, often referred to as statistical heterogeneity, and small local data samples from the various data distributions. Therefore, to develop robust generalized global and personalized models, conventional FL methods need to redesign the knowledge aggregation from biased local models while considering huge divergence of learning parameters due to skewed client data. In this work, we demonstrate that the knowledge transfer mechanism achieves these objectives and develop a novel knowledge distillation-based approach to study the extent of knowledge transfer between the global model and local models. Henceforth, our method considers the suitability of transferring the outcome distribution and (or) the embedding vector of representation from trained models during cross-device knowledge transfer using a small proxy dataset in heterogeneous FL. In doing so, we alternatively perform cross-device knowledge transfer following general formulations as (1) global knowledge transfer and (2) on-device knowledge transfer. Through simulations on three federated datasets, we show the proposed method achieves significant speedups and high personalized performance of local models. Furthermore, the proposed approach offers a more stable algorithm than other baselines during the training, with minimal communication data load when exchanging the trained model’s outcomes and representation.</description><identifier>ISSN: 0952-1976</identifier><identifier>EISSN: 1873-6769</identifier><identifier>DOI: 10.1016/j.engappai.2024.108093</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Federated learning ; Knowledge distillation ; Representation learning</subject><ispartof>Engineering applications of artificial intelligence, 2024-07, Vol.133, p.108093, Article 108093</ispartof><rights>2024 Elsevier Ltd</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c307t-dc30f3105b8269da593cfcb7e4c92fe4b6767dac7ab324506a2ac75c15dd89743</cites><orcidid>0000-0002-3035-0816 ; 0009-0007-8342-7614 ; 0000-0002-5781-4131</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27923,27924</link.rule.ids></links><search><creatorcontrib>Le, Huy Q.</creatorcontrib><creatorcontrib>Nguyen, Minh N.H.</creatorcontrib><creatorcontrib>Pandey, Shashi Raj</creatorcontrib><creatorcontrib>Zhang, Chaoning</creatorcontrib><creatorcontrib>Hong, Choong Seon</creatorcontrib><title>CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning</title><title>Engineering applications of artificial intelligence</title><description>In a practical setting, how to enable robust Federated Learning (FL) systems, both in terms of generalization and personalization abilities, is one important research question. It is a challenging issue due to the consequences of non-i.i.d. properties of client’s data, often referred to as statistical heterogeneity, and small local data samples from the various data distributions. Therefore, to develop robust generalized global and personalized models, conventional FL methods need to redesign the knowledge aggregation from biased local models while considering huge divergence of learning parameters due to skewed client data. In this work, we demonstrate that the knowledge transfer mechanism achieves these objectives and develop a novel knowledge distillation-based approach to study the extent of knowledge transfer between the global model and local models. Henceforth, our method considers the suitability of transferring the outcome distribution and (or) the embedding vector of representation from trained models during cross-device knowledge transfer using a small proxy dataset in heterogeneous FL. In doing so, we alternatively perform cross-device knowledge transfer following general formulations as (1) global knowledge transfer and (2) on-device knowledge transfer. Through simulations on three federated datasets, we show the proposed method achieves significant speedups and high personalized performance of local models. Furthermore, the proposed approach offers a more stable algorithm than other baselines during the training, with minimal communication data load when exchanging the trained model’s outcomes and representation.</description><subject>Federated learning</subject><subject>Knowledge distillation</subject><subject>Representation learning</subject><issn>0952-1976</issn><issn>1873-6769</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNqFkN1KAzEUhIMoWKuvIHmBrfnZTTZeKatVsSBIvQ7Z5GxJrdklWat9e1Oq114N5zAzDB9Cl5TMKKHiaj2DsDLDYPyMEVbmZ00UP0ITWkteCCnUMZoQVbGCKilO0VlKa0IIr0sxQa_N3fOymC-ucRP7lAoHW28Bv4f-awNuBXiMJqQOIv5MPqzwEPvvHXZmNAlG7APuwEE0Izi8ARND9pyjk85sElz86hS9ze-XzWOxeHl4am4XheVEjoXL0nFKqrZmQjlTKW4720oorWIdlG1eLp2x0rSclRURhuWjsrRyrlay5FMkDr12vzxCp4foP0zcaUr0noxe6z8yek9GH8jk4M0hCHnd1kPUyXoIFpyPYEftev9fxQ-rHnCv</recordid><startdate>202407</startdate><enddate>202407</enddate><creator>Le, Huy Q.</creator><creator>Nguyen, Minh N.H.</creator><creator>Pandey, Shashi Raj</creator><creator>Zhang, Chaoning</creator><creator>Hong, Choong Seon</creator><general>Elsevier Ltd</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-3035-0816</orcidid><orcidid>https://orcid.org/0009-0007-8342-7614</orcidid><orcidid>https://orcid.org/0000-0002-5781-4131</orcidid></search><sort><creationdate>202407</creationdate><title>CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning</title><author>Le, Huy Q. ; Nguyen, Minh N.H. ; Pandey, Shashi Raj ; Zhang, Chaoning ; Hong, Choong Seon</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c307t-dc30f3105b8269da593cfcb7e4c92fe4b6767dac7ab324506a2ac75c15dd89743</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Federated learning</topic><topic>Knowledge distillation</topic><topic>Representation learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Le, Huy Q.</creatorcontrib><creatorcontrib>Nguyen, Minh N.H.</creatorcontrib><creatorcontrib>Pandey, Shashi Raj</creatorcontrib><creatorcontrib>Zhang, Chaoning</creatorcontrib><creatorcontrib>Hong, Choong Seon</creatorcontrib><collection>CrossRef</collection><jtitle>Engineering applications of artificial intelligence</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Le, Huy Q.</au><au>Nguyen, Minh N.H.</au><au>Pandey, Shashi Raj</au><au>Zhang, Chaoning</au><au>Hong, Choong Seon</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning</atitle><jtitle>Engineering applications of artificial intelligence</jtitle><date>2024-07</date><risdate>2024</risdate><volume>133</volume><spage>108093</spage><pages>108093-</pages><artnum>108093</artnum><issn>0952-1976</issn><eissn>1873-6769</eissn><abstract>In a practical setting, how to enable robust Federated Learning (FL) systems, both in terms of generalization and personalization abilities, is one important research question. It is a challenging issue due to the consequences of non-i.i.d. properties of client’s data, often referred to as statistical heterogeneity, and small local data samples from the various data distributions. Therefore, to develop robust generalized global and personalized models, conventional FL methods need to redesign the knowledge aggregation from biased local models while considering huge divergence of learning parameters due to skewed client data. In this work, we demonstrate that the knowledge transfer mechanism achieves these objectives and develop a novel knowledge distillation-based approach to study the extent of knowledge transfer between the global model and local models. Henceforth, our method considers the suitability of transferring the outcome distribution and (or) the embedding vector of representation from trained models during cross-device knowledge transfer using a small proxy dataset in heterogeneous FL. In doing so, we alternatively perform cross-device knowledge transfer following general formulations as (1) global knowledge transfer and (2) on-device knowledge transfer. Through simulations on three federated datasets, we show the proposed method achieves significant speedups and high personalized performance of local models. Furthermore, the proposed approach offers a more stable algorithm than other baselines during the training, with minimal communication data load when exchanging the trained model’s outcomes and representation.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.engappai.2024.108093</doi><orcidid>https://orcid.org/0000-0002-3035-0816</orcidid><orcidid>https://orcid.org/0009-0007-8342-7614</orcidid><orcidid>https://orcid.org/0000-0002-5781-4131</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0952-1976
ispartof Engineering applications of artificial intelligence, 2024-07, Vol.133, p.108093, Article 108093
issn 0952-1976
1873-6769
language eng
recordid cdi_crossref_primary_10_1016_j_engappai_2024_108093
source ScienceDirect Journals
subjects Federated learning
Knowledge distillation
Representation learning
title CDKT-FL: Cross-device knowledge transfer using proxy dataset in federated learning
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T12%3A56%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=CDKT-FL:%20Cross-device%20knowledge%20transfer%20using%20proxy%20dataset%20in%20federated%20learning&rft.jtitle=Engineering%20applications%20of%20artificial%20intelligence&rft.au=Le,%20Huy%20Q.&rft.date=2024-07&rft.volume=133&rft.spage=108093&rft.pages=108093-&rft.artnum=108093&rft.issn=0952-1976&rft.eissn=1873-6769&rft_id=info:doi/10.1016/j.engappai.2024.108093&rft_dat=%3Celsevier_cross%3ES0952197624002513%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c307t-dc30f3105b8269da593cfcb7e4c92fe4b6767dac7ab324506a2ac75c15dd89743%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true