Loading…

CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance

Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massi...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on computer-aided design of integrated circuits and systems 2024-11, Vol.43 (11), p.4057-4068
Main Authors: Xia, Zeke, Hu, Ming, Yan, Dengke, Xie, Xiaofei, Li, Tianlin, Li, Anran, Zhou, Junlong, Chen, Mingsong
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 4068
container_issue 11
container_start_page 4057
container_title IEEE transactions on computer-aided design of integrated circuits and systems
container_volume 43
creator Xia, Zeke
Hu, Ming
Yan, Dengke
Xie, Xiaofei
Li, Tianlin
Li, Anran
Zhou, Junlong
Chen, Mingsong
description Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.
doi_str_mv 10.1109/TCAD.2024.3446881
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TCAD_2024_3446881</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10745831</ieee_id><sourcerecordid>3124825443</sourcerecordid><originalsourceid>FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853</originalsourceid><addsrcrecordid>eNpNkE1Lw0AQhhdRsFZ_gOBhwXPqfmY33tporRAQpJ6XyWZiU2pSdxOh_96UevA08PK8M8NDyC1nM85Z9rDO508zwYSaSaVSa_kZmfBMmkRxzc_JhAljE8YMuyRXMW4Z40qLbELec1jAsnik83ho_SZ0bTdEusQKA_RY0QIhtE37SX8aoKtmTIPfNB52NAe_QQptNdLQDwHpAnbQerwmFzXsIt78zSn5WD6v81VSvL285vMi8YJnfWIslkLqGk0KJTLNbFmZrFYlqPE5M0YMfSkqmVmjNRe6Tn2qLNTelJJZLafk_rR3H7rvAWPvtt0Q2vGkk1woK7RScqT4ifKhizFg7fah-YJwcJy5ozp3VOeO6tyfurFzd-o0iPiPN0pbyeUvLpto_g</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3124825443</pqid></control><display><type>article</type><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</creator><creatorcontrib>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</creatorcontrib><description>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</description><identifier>ISSN: 0278-0070</identifier><identifier>EISSN: 1937-4151</identifier><identifier>DOI: 10.1109/TCAD.2024.3446881</identifier><identifier>CODEN: ITCSDI</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Accuracy ; Artificial intelligence ; Artificial Intelligence of Things (AIoT) ; asynchronous federated learning (FL) ; Data models ; Data structures ; data/device heterogeneity ; Design automation ; Devices ; feature balance ; Federated learning ; Internet of Things ; Machine learning ; Performance evaluation ; Servers ; Training</subject><ispartof>IEEE transactions on computer-aided design of integrated circuits and systems, 2024-11, Vol.43 (11), p.4057-4068</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>0000-0001-7052-4721 ; 0000-0002-5058-4660 ; 0000-0002-7734-4077 ; 0000-0002-3922-0989 ; 0000-0002-2207-1622 ; 0000-0002-1288-6502 ; 0000-0002-3592-4153</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10745831$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Xia, Zeke</creatorcontrib><creatorcontrib>Hu, Ming</creatorcontrib><creatorcontrib>Yan, Dengke</creatorcontrib><creatorcontrib>Xie, Xiaofei</creatorcontrib><creatorcontrib>Li, Tianlin</creatorcontrib><creatorcontrib>Li, Anran</creatorcontrib><creatorcontrib>Zhou, Junlong</creatorcontrib><creatorcontrib>Chen, Mingsong</creatorcontrib><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><title>IEEE transactions on computer-aided design of integrated circuits and systems</title><addtitle>TCAD</addtitle><description>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</description><subject>Accuracy</subject><subject>Artificial intelligence</subject><subject>Artificial Intelligence of Things (AIoT)</subject><subject>asynchronous federated learning (FL)</subject><subject>Data models</subject><subject>Data structures</subject><subject>data/device heterogeneity</subject><subject>Design automation</subject><subject>Devices</subject><subject>feature balance</subject><subject>Federated learning</subject><subject>Internet of Things</subject><subject>Machine learning</subject><subject>Performance evaluation</subject><subject>Servers</subject><subject>Training</subject><issn>0278-0070</issn><issn>1937-4151</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkE1Lw0AQhhdRsFZ_gOBhwXPqfmY33tporRAQpJ6XyWZiU2pSdxOh_96UevA08PK8M8NDyC1nM85Z9rDO508zwYSaSaVSa_kZmfBMmkRxzc_JhAljE8YMuyRXMW4Z40qLbELec1jAsnik83ho_SZ0bTdEusQKA_RY0QIhtE37SX8aoKtmTIPfNB52NAe_QQptNdLQDwHpAnbQerwmFzXsIt78zSn5WD6v81VSvL285vMi8YJnfWIslkLqGk0KJTLNbFmZrFYlqPE5M0YMfSkqmVmjNRe6Tn2qLNTelJJZLafk_rR3H7rvAWPvtt0Q2vGkk1woK7RScqT4ifKhizFg7fah-YJwcJy5ozp3VOeO6tyfurFzd-o0iPiPN0pbyeUvLpto_g</recordid><startdate>20241101</startdate><enddate>20241101</enddate><creator>Xia, Zeke</creator><creator>Hu, Ming</creator><creator>Yan, Dengke</creator><creator>Xie, Xiaofei</creator><creator>Li, Tianlin</creator><creator>Li, Anran</creator><creator>Zhou, Junlong</creator><creator>Chen, Mingsong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-7052-4721</orcidid><orcidid>https://orcid.org/0000-0002-5058-4660</orcidid><orcidid>https://orcid.org/0000-0002-7734-4077</orcidid><orcidid>https://orcid.org/0000-0002-3922-0989</orcidid><orcidid>https://orcid.org/0000-0002-2207-1622</orcidid><orcidid>https://orcid.org/0000-0002-1288-6502</orcidid><orcidid>https://orcid.org/0000-0002-3592-4153</orcidid></search><sort><creationdate>20241101</creationdate><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><author>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Artificial intelligence</topic><topic>Artificial Intelligence of Things (AIoT)</topic><topic>asynchronous federated learning (FL)</topic><topic>Data models</topic><topic>Data structures</topic><topic>data/device heterogeneity</topic><topic>Design automation</topic><topic>Devices</topic><topic>feature balance</topic><topic>Federated learning</topic><topic>Internet of Things</topic><topic>Machine learning</topic><topic>Performance evaluation</topic><topic>Servers</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xia, Zeke</creatorcontrib><creatorcontrib>Hu, Ming</creatorcontrib><creatorcontrib>Yan, Dengke</creatorcontrib><creatorcontrib>Xie, Xiaofei</creatorcontrib><creatorcontrib>Li, Tianlin</creatorcontrib><creatorcontrib>Li, Anran</creatorcontrib><creatorcontrib>Zhou, Junlong</creatorcontrib><creatorcontrib>Chen, Mingsong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on computer-aided design of integrated circuits and systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Xia, Zeke</au><au>Hu, Ming</au><au>Yan, Dengke</au><au>Xie, Xiaofei</au><au>Li, Tianlin</au><au>Li, Anran</au><au>Zhou, Junlong</au><au>Chen, Mingsong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</atitle><jtitle>IEEE transactions on computer-aided design of integrated circuits and systems</jtitle><stitle>TCAD</stitle><date>2024-11-01</date><risdate>2024</risdate><volume>43</volume><issue>11</issue><spage>4057</spage><epage>4068</epage><pages>4057-4068</pages><issn>0278-0070</issn><eissn>1937-4151</eissn><coden>ITCSDI</coden><abstract>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TCAD.2024.3446881</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0001-7052-4721</orcidid><orcidid>https://orcid.org/0000-0002-5058-4660</orcidid><orcidid>https://orcid.org/0000-0002-7734-4077</orcidid><orcidid>https://orcid.org/0000-0002-3922-0989</orcidid><orcidid>https://orcid.org/0000-0002-2207-1622</orcidid><orcidid>https://orcid.org/0000-0002-1288-6502</orcidid><orcidid>https://orcid.org/0000-0002-3592-4153</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0278-0070
ispartof IEEE transactions on computer-aided design of integrated circuits and systems, 2024-11, Vol.43 (11), p.4057-4068
issn 0278-0070
1937-4151
language eng
recordid cdi_crossref_primary_10_1109_TCAD_2024_3446881
source IEEE Electronic Library (IEL) Journals
subjects Accuracy
Artificial intelligence
Artificial Intelligence of Things (AIoT)
asynchronous federated learning (FL)
Data models
Data structures
data/device heterogeneity
Design automation
Devices
feature balance
Federated learning
Internet of Things
Machine learning
Performance evaluation
Servers
Training
title CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-01T17%3A41%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=CaBaFL:%20Asynchronous%20Federated%20Learning%20via%20Hierarchical%20Cache%20and%20Feature%20Balance&rft.jtitle=IEEE%20transactions%20on%20computer-aided%20design%20of%20integrated%20circuits%20and%20systems&rft.au=Xia,%20Zeke&rft.date=2024-11-01&rft.volume=43&rft.issue=11&rft.spage=4057&rft.epage=4068&rft.pages=4057-4068&rft.issn=0278-0070&rft.eissn=1937-4151&rft.coden=ITCSDI&rft_id=info:doi/10.1109/TCAD.2024.3446881&rft_dat=%3Cproquest_cross%3E3124825443%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3124825443&rft_id=info:pmid/&rft_ieee_id=10745831&rfr_iscdi=true