Loading…
Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays
With the rapid development of deep-learning models, especially the widespread adoption of transformer architectures, the demand for efficient hardware accelerators with field-programmable gate arrays (FPGAs) has increased owing to their flexibility and performance advantages. Although high-level syn...
Saved in:
Published in: | Electronics (Basel) 2024-02, Vol.13 (3), p.552 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c311t-3be5b6ceca53cfc2c3613bdeb02a435b3afb5ae5e209cbc28fb4839f32e900363 |
container_end_page | |
container_issue | 3 |
container_start_page | 552 |
container_title | Electronics (Basel) |
container_volume | 13 |
creator | Jang, Seojin Cho, Yongbeom |
description | With the rapid development of deep-learning models, especially the widespread adoption of transformer architectures, the demand for efficient hardware accelerators with field-programmable gate arrays (FPGAs) has increased owing to their flexibility and performance advantages. Although high-level synthesis can shorten the hardware design cycle, determining the optimal bit-width for various transformer designs remains challenging. Therefore, this paper proposes a novel technique based on a predesigned transformer hardware architecture tailored for various types of FPGAs. The proposed method leverages a reinforcement learning-driven mechanism to automatically adapt and optimize bit-width settings based on user-provided transformer variants during inference on an FPGA, significantly alleviating the challenges related to bit-width optimization. The effect of bit-width settings on resource utilization and performance across different FPGA types was analyzed. The efficacy of the proposed method was demonstrated by optimizing the bit-width settings for users’ transformer-based model inferences on an FPGA. The use of the predesigned hardware architecture significantly enhanced the performance. Overall, the proposed method enables effective and optimized implementations of user-provided transformer-based models on an FPGA, paving the way for edge FPGA-based deep-learning accelerators while reducing the time and effort typically required in fine-tuning bit-width settings. |
doi_str_mv | 10.3390/electronics13030552 |
format | article |
fullrecord | <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_2923905416</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A782089551</galeid><sourcerecordid>A782089551</sourcerecordid><originalsourceid>FETCH-LOGICAL-c311t-3be5b6ceca53cfc2c3613bdeb02a435b3afb5ae5e209cbc28fb4839f32e900363</originalsourceid><addsrcrecordid>eNptUU1rGzEQXUoKDUl-QS-CnjeRNJa9OjpJ8wGGlNalx0WSR2uFXckdKQHnJ_RXV8Y59NCZwwyP92bgvab5LPglgOZXOKIrlGJwWQAHrpT80JxKvtCtllqe_LN_ai5yfua1tIAO-Gnz5zuG6BM5nDAWtkJDMcShvaXwipFdh9L-CpuyZU-7EqbwZkpIkVUBK1tkD2HYtit8xZH92MeK5JBZ8mxNJuZKmpDYbQWHWOHI7gKOm_YbpYHMNBk7Irs3BdmSyOzzefPRmzHjxfs8a37efV3fPLSrp_vHm-WqdSBEacGisnOHzihw3kkHcwF2g5ZLMwNlwXirDCqUXDvrZOftrAPtQaLmHOZw1nw53t1R-v2CufTP6YVifdlXi6qjaiYOrMsjazAj9gePChlXe4NTcCmiDxVfLjrJO62UqAI4ChylnAl9v6MwGdr3gveHoPr_BAV_AV4EjBc</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2923905416</pqid></control><display><type>article</type><title>Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays</title><source>Publicly Available Content Database</source><creator>Jang, Seojin ; Cho, Yongbeom</creator><creatorcontrib>Jang, Seojin ; Cho, Yongbeom</creatorcontrib><description>With the rapid development of deep-learning models, especially the widespread adoption of transformer architectures, the demand for efficient hardware accelerators with field-programmable gate arrays (FPGAs) has increased owing to their flexibility and performance advantages. Although high-level synthesis can shorten the hardware design cycle, determining the optimal bit-width for various transformer designs remains challenging. Therefore, this paper proposes a novel technique based on a predesigned transformer hardware architecture tailored for various types of FPGAs. The proposed method leverages a reinforcement learning-driven mechanism to automatically adapt and optimize bit-width settings based on user-provided transformer variants during inference on an FPGA, significantly alleviating the challenges related to bit-width optimization. The effect of bit-width settings on resource utilization and performance across different FPGA types was analyzed. The efficacy of the proposed method was demonstrated by optimizing the bit-width settings for users’ transformer-based model inferences on an FPGA. The use of the predesigned hardware architecture significantly enhanced the performance. Overall, the proposed method enables effective and optimized implementations of user-provided transformer-based models on an FPGA, paving the way for edge FPGA-based deep-learning accelerators while reducing the time and effort typically required in fine-tuning bit-width settings.</description><identifier>ISSN: 2079-9292</identifier><identifier>EISSN: 2079-9292</identifier><identifier>DOI: 10.3390/electronics13030552</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Accelerators ; Accuracy ; Algorithms ; Circuit design ; Deep learning ; Design ; Digital integrated circuits ; Efficiency ; Fashion models ; Field programmable gate arrays ; Hardware ; High level synthesis ; Language ; Machine learning ; Methods ; Natural language processing ; Neural networks ; Optimization ; Optimization techniques ; Power ; Resource utilization</subject><ispartof>Electronics (Basel), 2024-02, Vol.13 (3), p.552</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c311t-3be5b6ceca53cfc2c3613bdeb02a435b3afb5ae5e209cbc28fb4839f32e900363</cites><orcidid>0000-0002-8443-9373 ; 0000-0002-3264-3575</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/2923905416/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/2923905416?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25753,27924,27925,37012,44590,75126</link.rule.ids></links><search><creatorcontrib>Jang, Seojin</creatorcontrib><creatorcontrib>Cho, Yongbeom</creatorcontrib><title>Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays</title><title>Electronics (Basel)</title><description>With the rapid development of deep-learning models, especially the widespread adoption of transformer architectures, the demand for efficient hardware accelerators with field-programmable gate arrays (FPGAs) has increased owing to their flexibility and performance advantages. Although high-level synthesis can shorten the hardware design cycle, determining the optimal bit-width for various transformer designs remains challenging. Therefore, this paper proposes a novel technique based on a predesigned transformer hardware architecture tailored for various types of FPGAs. The proposed method leverages a reinforcement learning-driven mechanism to automatically adapt and optimize bit-width settings based on user-provided transformer variants during inference on an FPGA, significantly alleviating the challenges related to bit-width optimization. The effect of bit-width settings on resource utilization and performance across different FPGA types was analyzed. The efficacy of the proposed method was demonstrated by optimizing the bit-width settings for users’ transformer-based model inferences on an FPGA. The use of the predesigned hardware architecture significantly enhanced the performance. Overall, the proposed method enables effective and optimized implementations of user-provided transformer-based models on an FPGA, paving the way for edge FPGA-based deep-learning accelerators while reducing the time and effort typically required in fine-tuning bit-width settings.</description><subject>Accelerators</subject><subject>Accuracy</subject><subject>Algorithms</subject><subject>Circuit design</subject><subject>Deep learning</subject><subject>Design</subject><subject>Digital integrated circuits</subject><subject>Efficiency</subject><subject>Fashion models</subject><subject>Field programmable gate arrays</subject><subject>Hardware</subject><subject>High level synthesis</subject><subject>Language</subject><subject>Machine learning</subject><subject>Methods</subject><subject>Natural language processing</subject><subject>Neural networks</subject><subject>Optimization</subject><subject>Optimization techniques</subject><subject>Power</subject><subject>Resource utilization</subject><issn>2079-9292</issn><issn>2079-9292</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNptUU1rGzEQXUoKDUl-QS-CnjeRNJa9OjpJ8wGGlNalx0WSR2uFXckdKQHnJ_RXV8Y59NCZwwyP92bgvab5LPglgOZXOKIrlGJwWQAHrpT80JxKvtCtllqe_LN_ai5yfua1tIAO-Gnz5zuG6BM5nDAWtkJDMcShvaXwipFdh9L-CpuyZU-7EqbwZkpIkVUBK1tkD2HYtit8xZH92MeK5JBZ8mxNJuZKmpDYbQWHWOHI7gKOm_YbpYHMNBk7Irs3BdmSyOzzefPRmzHjxfs8a37efV3fPLSrp_vHm-WqdSBEacGisnOHzihw3kkHcwF2g5ZLMwNlwXirDCqUXDvrZOftrAPtQaLmHOZw1nw53t1R-v2CufTP6YVifdlXi6qjaiYOrMsjazAj9gePChlXe4NTcCmiDxVfLjrJO62UqAI4ChylnAl9v6MwGdr3gveHoPr_BAV_AV4EjBc</recordid><startdate>20240201</startdate><enddate>20240201</enddate><creator>Jang, Seojin</creator><creator>Cho, Yongbeom</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L7M</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><orcidid>https://orcid.org/0000-0002-8443-9373</orcidid><orcidid>https://orcid.org/0000-0002-3264-3575</orcidid></search><sort><creationdate>20240201</creationdate><title>Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays</title><author>Jang, Seojin ; Cho, Yongbeom</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c311t-3be5b6ceca53cfc2c3613bdeb02a435b3afb5ae5e209cbc28fb4839f32e900363</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Accelerators</topic><topic>Accuracy</topic><topic>Algorithms</topic><topic>Circuit design</topic><topic>Deep learning</topic><topic>Design</topic><topic>Digital integrated circuits</topic><topic>Efficiency</topic><topic>Fashion models</topic><topic>Field programmable gate arrays</topic><topic>Hardware</topic><topic>High level synthesis</topic><topic>Language</topic><topic>Machine learning</topic><topic>Methods</topic><topic>Natural language processing</topic><topic>Neural networks</topic><topic>Optimization</topic><topic>Optimization techniques</topic><topic>Power</topic><topic>Resource utilization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Jang, Seojin</creatorcontrib><creatorcontrib>Cho, Yongbeom</creatorcontrib><collection>CrossRef</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Database (1962 - current)</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>ProQuest advanced technologies & aerospace journals</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><jtitle>Electronics (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Jang, Seojin</au><au>Cho, Yongbeom</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays</atitle><jtitle>Electronics (Basel)</jtitle><date>2024-02-01</date><risdate>2024</risdate><volume>13</volume><issue>3</issue><spage>552</spage><pages>552-</pages><issn>2079-9292</issn><eissn>2079-9292</eissn><abstract>With the rapid development of deep-learning models, especially the widespread adoption of transformer architectures, the demand for efficient hardware accelerators with field-programmable gate arrays (FPGAs) has increased owing to their flexibility and performance advantages. Although high-level synthesis can shorten the hardware design cycle, determining the optimal bit-width for various transformer designs remains challenging. Therefore, this paper proposes a novel technique based on a predesigned transformer hardware architecture tailored for various types of FPGAs. The proposed method leverages a reinforcement learning-driven mechanism to automatically adapt and optimize bit-width settings based on user-provided transformer variants during inference on an FPGA, significantly alleviating the challenges related to bit-width optimization. The effect of bit-width settings on resource utilization and performance across different FPGA types was analyzed. The efficacy of the proposed method was demonstrated by optimizing the bit-width settings for users’ transformer-based model inferences on an FPGA. The use of the predesigned hardware architecture significantly enhanced the performance. Overall, the proposed method enables effective and optimized implementations of user-provided transformer-based models on an FPGA, paving the way for edge FPGA-based deep-learning accelerators while reducing the time and effort typically required in fine-tuning bit-width settings.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/electronics13030552</doi><orcidid>https://orcid.org/0000-0002-8443-9373</orcidid><orcidid>https://orcid.org/0000-0002-3264-3575</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2079-9292 |
ispartof | Electronics (Basel), 2024-02, Vol.13 (3), p.552 |
issn | 2079-9292 2079-9292 |
language | eng |
recordid | cdi_proquest_journals_2923905416 |
source | Publicly Available Content Database |
subjects | Accelerators Accuracy Algorithms Circuit design Deep learning Design Digital integrated circuits Efficiency Fashion models Field programmable gate arrays Hardware High level synthesis Language Machine learning Methods Natural language processing Neural networks Optimization Optimization techniques Power Resource utilization |
title | Reinforcement Learning-Driven Bit-Width Optimization for the High-Level Synthesis of Transformer Designs on Field-Programmable Gate Arrays |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-01T10%3A26%3A47IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Reinforcement%20Learning-Driven%20Bit-Width%20Optimization%20for%20the%20High-Level%20Synthesis%20of%20Transformer%20Designs%20on%20Field-Programmable%20Gate%20Arrays&rft.jtitle=Electronics%20(Basel)&rft.au=Jang,%20Seojin&rft.date=2024-02-01&rft.volume=13&rft.issue=3&rft.spage=552&rft.pages=552-&rft.issn=2079-9292&rft.eissn=2079-9292&rft_id=info:doi/10.3390/electronics13030552&rft_dat=%3Cgale_proqu%3EA782089551%3C/gale_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c311t-3be5b6ceca53cfc2c3613bdeb02a435b3afb5ae5e209cbc28fb4839f32e900363%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2923905416&rft_id=info:pmid/&rft_galeid=A782089551&rfr_iscdi=true |