Loading…
Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation
Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains...
Saved in:
Published in: | IEEE journal of solid-state circuits 2020-08, Vol.55 (8), p.2228-2246 |
---|---|
Main Authors: | , , , , , , , , , , , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3 |
---|---|
cites | cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3 |
container_end_page | 2246 |
container_issue | 8 |
container_start_page | 2228 |
container_title | IEEE journal of solid-state circuits |
container_volume | 55 |
creator | Deng, Lei Wang, Guanrui Li, Guoqi Li, Shuangchen Liang, Ling Zhu, Maohua Wu, Yujie Yang, Zheyu Zou, Zhe Pei, Jing Wu, Zhenzhi Hu, Xing Ding, Yufei He, Wei Xie, Yuan Shi, Luping |
description | Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with >610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing. |
doi_str_mv | 10.1109/JSSC.2020.2970709 |
format | article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2426667584</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8998338</ieee_id><sourcerecordid>2426667584</sourcerecordid><originalsourceid>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</originalsourceid><addsrcrecordid>eNo9kE1Lw0AQhhdRsFZ_gHgJeE6d_cruemuDnxQ9tAXBw7LdbOrWNIm7ycF_b0qLp-EdnncGHoSuMUwwBnX3uljkEwIEJkQJEKBO0AhzLlMs6McpGgFgmSoCcI4uYtwOkTGJR-hz6U299fY-mSar2pfeFYmpi2RhTWXWlUvyL98ms-CLja83yaL13y6dmXjE8qbufN03fUzeXB9MNWx2bd-Zzjf1JTorTRXd1XGO0erxYZk_p_P3p5d8Ok8tZbhL5ZoIhzNaEJmVzhaKGQZcGRAZkVZwy4RgAKSUYLnjYHjhLFBYS4IF4yUdo9vD3TY0P72Lnd42faiHl5owkmWZ4JINFD5QNjQxBlfqNvidCb8ag9471HuHeu9QHx0OnZtDxzvn_nmplKRU0j_v52vJ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2426667584</pqid></control><display><type>article</type><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><source>IEEE Xplore (Online service)</source><creator>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</creator><creatorcontrib>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</creatorcontrib><description>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with >610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</description><identifier>ISSN: 0018-9200</identifier><identifier>EISSN: 1558-173X</identifier><identifier>DOI: 10.1109/JSSC.2020.2970709</identifier><identifier>CODEN: IJSCBC</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Bicycles ; Biological neural networks ; Computational modeling ; Computer architecture ; Deep learning accelerator ; hybrid paradigm ; Machine learning ; Mapping ; Neural networks ; neuromorphic chip ; Neuromorphics ; Power efficiency ; unified/scalable architecture</subject><ispartof>IEEE journal of solid-state circuits, 2020-08, Vol.55 (8), p.2228-2246</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</citedby><cites>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</cites><orcidid>0000-0002-8994-431X ; 0000-0002-9829-2202 ; 0000-0002-8534-6494 ; 0000-0002-5172-9411</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8998338$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Deng, Lei</creatorcontrib><creatorcontrib>Wang, Guanrui</creatorcontrib><creatorcontrib>Li, Guoqi</creatorcontrib><creatorcontrib>Li, Shuangchen</creatorcontrib><creatorcontrib>Liang, Ling</creatorcontrib><creatorcontrib>Zhu, Maohua</creatorcontrib><creatorcontrib>Wu, Yujie</creatorcontrib><creatorcontrib>Yang, Zheyu</creatorcontrib><creatorcontrib>Zou, Zhe</creatorcontrib><creatorcontrib>Pei, Jing</creatorcontrib><creatorcontrib>Wu, Zhenzhi</creatorcontrib><creatorcontrib>Hu, Xing</creatorcontrib><creatorcontrib>Ding, Yufei</creatorcontrib><creatorcontrib>He, Wei</creatorcontrib><creatorcontrib>Xie, Yuan</creatorcontrib><creatorcontrib>Shi, Luping</creatorcontrib><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><title>IEEE journal of solid-state circuits</title><addtitle>JSSC</addtitle><description>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with >610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</description><subject>Bicycles</subject><subject>Biological neural networks</subject><subject>Computational modeling</subject><subject>Computer architecture</subject><subject>Deep learning accelerator</subject><subject>hybrid paradigm</subject><subject>Machine learning</subject><subject>Mapping</subject><subject>Neural networks</subject><subject>neuromorphic chip</subject><subject>Neuromorphics</subject><subject>Power efficiency</subject><subject>unified/scalable architecture</subject><issn>0018-9200</issn><issn>1558-173X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNo9kE1Lw0AQhhdRsFZ_gHgJeE6d_cruemuDnxQ9tAXBw7LdbOrWNIm7ycF_b0qLp-EdnncGHoSuMUwwBnX3uljkEwIEJkQJEKBO0AhzLlMs6McpGgFgmSoCcI4uYtwOkTGJR-hz6U299fY-mSar2pfeFYmpi2RhTWXWlUvyL98ms-CLja83yaL13y6dmXjE8qbufN03fUzeXB9MNWx2bd-Zzjf1JTorTRXd1XGO0erxYZk_p_P3p5d8Ok8tZbhL5ZoIhzNaEJmVzhaKGQZcGRAZkVZwy4RgAKSUYLnjYHjhLFBYS4IF4yUdo9vD3TY0P72Lnd42faiHl5owkmWZ4JINFD5QNjQxBlfqNvidCb8ag9471HuHeu9QHx0OnZtDxzvn_nmplKRU0j_v52vJ</recordid><startdate>20200801</startdate><enddate>20200801</enddate><creator>Deng, Lei</creator><creator>Wang, Guanrui</creator><creator>Li, Guoqi</creator><creator>Li, Shuangchen</creator><creator>Liang, Ling</creator><creator>Zhu, Maohua</creator><creator>Wu, Yujie</creator><creator>Yang, Zheyu</creator><creator>Zou, Zhe</creator><creator>Pei, Jing</creator><creator>Wu, Zhenzhi</creator><creator>Hu, Xing</creator><creator>Ding, Yufei</creator><creator>He, Wei</creator><creator>Xie, Yuan</creator><creator>Shi, Luping</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0002-8994-431X</orcidid><orcidid>https://orcid.org/0000-0002-9829-2202</orcidid><orcidid>https://orcid.org/0000-0002-8534-6494</orcidid><orcidid>https://orcid.org/0000-0002-5172-9411</orcidid></search><sort><creationdate>20200801</creationdate><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><author>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Bicycles</topic><topic>Biological neural networks</topic><topic>Computational modeling</topic><topic>Computer architecture</topic><topic>Deep learning accelerator</topic><topic>hybrid paradigm</topic><topic>Machine learning</topic><topic>Mapping</topic><topic>Neural networks</topic><topic>neuromorphic chip</topic><topic>Neuromorphics</topic><topic>Power efficiency</topic><topic>unified/scalable architecture</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Deng, Lei</creatorcontrib><creatorcontrib>Wang, Guanrui</creatorcontrib><creatorcontrib>Li, Guoqi</creatorcontrib><creatorcontrib>Li, Shuangchen</creatorcontrib><creatorcontrib>Liang, Ling</creatorcontrib><creatorcontrib>Zhu, Maohua</creatorcontrib><creatorcontrib>Wu, Yujie</creatorcontrib><creatorcontrib>Yang, Zheyu</creatorcontrib><creatorcontrib>Zou, Zhe</creatorcontrib><creatorcontrib>Pei, Jing</creatorcontrib><creatorcontrib>Wu, Zhenzhi</creatorcontrib><creatorcontrib>Hu, Xing</creatorcontrib><creatorcontrib>Ding, Yufei</creatorcontrib><creatorcontrib>He, Wei</creatorcontrib><creatorcontrib>Xie, Yuan</creatorcontrib><creatorcontrib>Shi, Luping</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE journal of solid-state circuits</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Deng, Lei</au><au>Wang, Guanrui</au><au>Li, Guoqi</au><au>Li, Shuangchen</au><au>Liang, Ling</au><au>Zhu, Maohua</au><au>Wu, Yujie</au><au>Yang, Zheyu</au><au>Zou, Zhe</au><au>Pei, Jing</au><au>Wu, Zhenzhi</au><au>Hu, Xing</au><au>Ding, Yufei</au><au>He, Wei</au><au>Xie, Yuan</au><au>Shi, Luping</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</atitle><jtitle>IEEE journal of solid-state circuits</jtitle><stitle>JSSC</stitle><date>2020-08-01</date><risdate>2020</risdate><volume>55</volume><issue>8</issue><spage>2228</spage><epage>2246</epage><pages>2228-2246</pages><issn>0018-9200</issn><eissn>1558-173X</eissn><coden>IJSCBC</coden><abstract>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with >610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/JSSC.2020.2970709</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-8994-431X</orcidid><orcidid>https://orcid.org/0000-0002-9829-2202</orcidid><orcidid>https://orcid.org/0000-0002-8534-6494</orcidid><orcidid>https://orcid.org/0000-0002-5172-9411</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0018-9200 |
ispartof | IEEE journal of solid-state circuits, 2020-08, Vol.55 (8), p.2228-2246 |
issn | 0018-9200 1558-173X |
language | eng |
recordid | cdi_proquest_journals_2426667584 |
source | IEEE Xplore (Online service) |
subjects | Bicycles Biological neural networks Computational modeling Computer architecture Deep learning accelerator hybrid paradigm Machine learning Mapping Neural networks neuromorphic chip Neuromorphics Power efficiency unified/scalable architecture |
title | Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T05%3A04%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Tianjic:%20A%20Unified%20and%20Scalable%20Chip%20Bridging%20Spike-Based%20and%20Continuous%20Neural%20Computation&rft.jtitle=IEEE%20journal%20of%20solid-state%20circuits&rft.au=Deng,%20Lei&rft.date=2020-08-01&rft.volume=55&rft.issue=8&rft.spage=2228&rft.epage=2246&rft.pages=2228-2246&rft.issn=0018-9200&rft.eissn=1558-173X&rft.coden=IJSCBC&rft_id=info:doi/10.1109/JSSC.2020.2970709&rft_dat=%3Cproquest_ieee_%3E2426667584%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2426667584&rft_id=info:pmid/&rft_ieee_id=8998338&rfr_iscdi=true |