Loading…
CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance
Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massi...
Saved in:
Published in: | IEEE transactions on computer-aided design of integrated circuits and systems 2024-11, Vol.43 (11), p.4057-4068 |
---|---|
Main Authors: | , , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 4068 |
container_issue | 11 |
container_start_page | 4057 |
container_title | IEEE transactions on computer-aided design of integrated circuits and systems |
container_volume | 43 |
creator | Xia, Zeke Hu, Ming Yan, Dengke Xie, Xiaofei Li, Tianlin Li, Anran Zhou, Junlong Chen, Mingsong |
description | Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements. |
doi_str_mv | 10.1109/TCAD.2024.3446881 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TCAD_2024_3446881</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10745831</ieee_id><sourcerecordid>3124825443</sourcerecordid><originalsourceid>FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853</originalsourceid><addsrcrecordid>eNpNkE1Lw0AQhhdRsFZ_gOBhwXPqfmY33tporRAQpJ6XyWZiU2pSdxOh_96UevA08PK8M8NDyC1nM85Z9rDO508zwYSaSaVSa_kZmfBMmkRxzc_JhAljE8YMuyRXMW4Z40qLbELec1jAsnik83ho_SZ0bTdEusQKA_RY0QIhtE37SX8aoKtmTIPfNB52NAe_QQptNdLQDwHpAnbQerwmFzXsIt78zSn5WD6v81VSvL285vMi8YJnfWIslkLqGk0KJTLNbFmZrFYlqPE5M0YMfSkqmVmjNRe6Tn2qLNTelJJZLafk_rR3H7rvAWPvtt0Q2vGkk1woK7RScqT4ifKhizFg7fah-YJwcJy5ozp3VOeO6tyfurFzd-o0iPiPN0pbyeUvLpto_g</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3124825443</pqid></control><display><type>article</type><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</creator><creatorcontrib>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</creatorcontrib><description>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</description><identifier>ISSN: 0278-0070</identifier><identifier>EISSN: 1937-4151</identifier><identifier>DOI: 10.1109/TCAD.2024.3446881</identifier><identifier>CODEN: ITCSDI</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Accuracy ; Artificial intelligence ; Artificial Intelligence of Things (AIoT) ; asynchronous federated learning (FL) ; Data models ; Data structures ; data/device heterogeneity ; Design automation ; Devices ; feature balance ; Federated learning ; Internet of Things ; Machine learning ; Performance evaluation ; Servers ; Training</subject><ispartof>IEEE transactions on computer-aided design of integrated circuits and systems, 2024-11, Vol.43 (11), p.4057-4068</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>0000-0001-7052-4721 ; 0000-0002-5058-4660 ; 0000-0002-7734-4077 ; 0000-0002-3922-0989 ; 0000-0002-2207-1622 ; 0000-0002-1288-6502 ; 0000-0002-3592-4153</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10745831$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Xia, Zeke</creatorcontrib><creatorcontrib>Hu, Ming</creatorcontrib><creatorcontrib>Yan, Dengke</creatorcontrib><creatorcontrib>Xie, Xiaofei</creatorcontrib><creatorcontrib>Li, Tianlin</creatorcontrib><creatorcontrib>Li, Anran</creatorcontrib><creatorcontrib>Zhou, Junlong</creatorcontrib><creatorcontrib>Chen, Mingsong</creatorcontrib><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><title>IEEE transactions on computer-aided design of integrated circuits and systems</title><addtitle>TCAD</addtitle><description>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</description><subject>Accuracy</subject><subject>Artificial intelligence</subject><subject>Artificial Intelligence of Things (AIoT)</subject><subject>asynchronous federated learning (FL)</subject><subject>Data models</subject><subject>Data structures</subject><subject>data/device heterogeneity</subject><subject>Design automation</subject><subject>Devices</subject><subject>feature balance</subject><subject>Federated learning</subject><subject>Internet of Things</subject><subject>Machine learning</subject><subject>Performance evaluation</subject><subject>Servers</subject><subject>Training</subject><issn>0278-0070</issn><issn>1937-4151</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkE1Lw0AQhhdRsFZ_gOBhwXPqfmY33tporRAQpJ6XyWZiU2pSdxOh_96UevA08PK8M8NDyC1nM85Z9rDO508zwYSaSaVSa_kZmfBMmkRxzc_JhAljE8YMuyRXMW4Z40qLbELec1jAsnik83ho_SZ0bTdEusQKA_RY0QIhtE37SX8aoKtmTIPfNB52NAe_QQptNdLQDwHpAnbQerwmFzXsIt78zSn5WD6v81VSvL285vMi8YJnfWIslkLqGk0KJTLNbFmZrFYlqPE5M0YMfSkqmVmjNRe6Tn2qLNTelJJZLafk_rR3H7rvAWPvtt0Q2vGkk1woK7RScqT4ifKhizFg7fah-YJwcJy5ozp3VOeO6tyfurFzd-o0iPiPN0pbyeUvLpto_g</recordid><startdate>20241101</startdate><enddate>20241101</enddate><creator>Xia, Zeke</creator><creator>Hu, Ming</creator><creator>Yan, Dengke</creator><creator>Xie, Xiaofei</creator><creator>Li, Tianlin</creator><creator>Li, Anran</creator><creator>Zhou, Junlong</creator><creator>Chen, Mingsong</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0001-7052-4721</orcidid><orcidid>https://orcid.org/0000-0002-5058-4660</orcidid><orcidid>https://orcid.org/0000-0002-7734-4077</orcidid><orcidid>https://orcid.org/0000-0002-3922-0989</orcidid><orcidid>https://orcid.org/0000-0002-2207-1622</orcidid><orcidid>https://orcid.org/0000-0002-1288-6502</orcidid><orcidid>https://orcid.org/0000-0002-3592-4153</orcidid></search><sort><creationdate>20241101</creationdate><title>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</title><author>Xia, Zeke ; Hu, Ming ; Yan, Dengke ; Xie, Xiaofei ; Li, Tianlin ; Li, Anran ; Zhou, Junlong ; Chen, Mingsong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accuracy</topic><topic>Artificial intelligence</topic><topic>Artificial Intelligence of Things (AIoT)</topic><topic>asynchronous federated learning (FL)</topic><topic>Data models</topic><topic>Data structures</topic><topic>data/device heterogeneity</topic><topic>Design automation</topic><topic>Devices</topic><topic>feature balance</topic><topic>Federated learning</topic><topic>Internet of Things</topic><topic>Machine learning</topic><topic>Performance evaluation</topic><topic>Servers</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xia, Zeke</creatorcontrib><creatorcontrib>Hu, Ming</creatorcontrib><creatorcontrib>Yan, Dengke</creatorcontrib><creatorcontrib>Xie, Xiaofei</creatorcontrib><creatorcontrib>Li, Tianlin</creatorcontrib><creatorcontrib>Li, Anran</creatorcontrib><creatorcontrib>Zhou, Junlong</creatorcontrib><creatorcontrib>Chen, Mingsong</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on computer-aided design of integrated circuits and systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Xia, Zeke</au><au>Hu, Ming</au><au>Yan, Dengke</au><au>Xie, Xiaofei</au><au>Li, Tianlin</au><au>Li, Anran</au><au>Zhou, Junlong</au><au>Chen, Mingsong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance</atitle><jtitle>IEEE transactions on computer-aided design of integrated circuits and systems</jtitle><stitle>TCAD</stitle><date>2024-11-01</date><risdate>2024</risdate><volume>43</volume><issue>11</issue><spage>4057</spage><epage>4068</epage><pages>4057-4068</pages><issn>0278-0070</issn><eissn>1937-4151</eissn><coden>ITCSDI</coden><abstract>Federated learning (FL) as a promising distributed machine learning paradigm has been widely adopted in Artificial Intelligence of Things (AIoT) applications. However, the efficiency and inference capability of FL is seriously limited due to the presence of stragglers and data imbalance across massive AIoT devices, respectively. To address the above challenges, we present a novel asynchronous FL approach named CaBaFL, which includes a hierarchical cache-based aggregation mechanism and a feature balance-guided device selection strategy. CaBaFL maintains multiple intermediate models simultaneously for local training. The hierarchical cache-based aggregation mechanism enables each intermediate model to be trained on multiple devices to align the training time and mitigate the straggler issue. In specific, each intermediate model is stored in a low-level cache for local training and when it is trained by sufficient local devices, it will be stored in a high-level cache for aggregation. To address the problem of imbalanced data, the feature balance-guided device selection strategy in CaBaFL adopts the activation distribution as a metric, which enables each intermediate model to be trained across devices with totally balanced data distributions before aggregation. Experimental results show that compared to the state-of-the-art FL methods, CaBaFL achieves up to 9.26X training acceleration and 19.71% accuracy improvements.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TCAD.2024.3446881</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0001-7052-4721</orcidid><orcidid>https://orcid.org/0000-0002-5058-4660</orcidid><orcidid>https://orcid.org/0000-0002-7734-4077</orcidid><orcidid>https://orcid.org/0000-0002-3922-0989</orcidid><orcidid>https://orcid.org/0000-0002-2207-1622</orcidid><orcidid>https://orcid.org/0000-0002-1288-6502</orcidid><orcidid>https://orcid.org/0000-0002-3592-4153</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0278-0070 |
ispartof | IEEE transactions on computer-aided design of integrated circuits and systems, 2024-11, Vol.43 (11), p.4057-4068 |
issn | 0278-0070 1937-4151 |
language | eng |
recordid | cdi_crossref_primary_10_1109_TCAD_2024_3446881 |
source | IEEE Electronic Library (IEL) Journals |
subjects | Accuracy Artificial intelligence Artificial Intelligence of Things (AIoT) asynchronous federated learning (FL) Data models Data structures data/device heterogeneity Design automation Devices feature balance Federated learning Internet of Things Machine learning Performance evaluation Servers Training |
title | CaBaFL: Asynchronous Federated Learning via Hierarchical Cache and Feature Balance |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-01T17%3A41%3A13IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=CaBaFL:%20Asynchronous%20Federated%20Learning%20via%20Hierarchical%20Cache%20and%20Feature%20Balance&rft.jtitle=IEEE%20transactions%20on%20computer-aided%20design%20of%20integrated%20circuits%20and%20systems&rft.au=Xia,%20Zeke&rft.date=2024-11-01&rft.volume=43&rft.issue=11&rft.spage=4057&rft.epage=4068&rft.pages=4057-4068&rft.issn=0278-0070&rft.eissn=1937-4151&rft.coden=ITCSDI&rft_id=info:doi/10.1109/TCAD.2024.3446881&rft_dat=%3Cproquest_cross%3E3124825443%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c219t-78eb235fe76abe0508bd79f4ba40147be00ecb2d398755125f6c648afc7b30853%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3124825443&rft_id=info:pmid/&rft_ieee_id=10745831&rfr_iscdi=true |