Loading…

Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation

Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains...

Full description

Saved in:
Bibliographic Details
Published in:IEEE journal of solid-state circuits 2020-08, Vol.55 (8), p.2228-2246
Main Authors: Deng, Lei, Wang, Guanrui, Li, Guoqi, Li, Shuangchen, Liang, Ling, Zhu, Maohua, Wu, Yujie, Yang, Zheyu, Zou, Zhe, Pei, Jing, Wu, Zhenzhi, Hu, Xing, Ding, Yufei, He, Wei, Xie, Yuan, Shi, Luping
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3
cites cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3
container_end_page 2246
container_issue 8
container_start_page 2228
container_title IEEE journal of solid-state circuits
container_volume 55
creator Deng, Lei
Wang, Guanrui
Li, Guoqi
Li, Shuangchen
Liang, Ling
Zhu, Maohua
Wu, Yujie
Yang, Zheyu
Zou, Zhe
Pei, Jing
Wu, Zhenzhi
Hu, Xing
Ding, Yufei
He, Wei
Xie, Yuan
Shi, Luping
description Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with >610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.
doi_str_mv 10.1109/JSSC.2020.2970709
format article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2426667584</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>8998338</ieee_id><sourcerecordid>2426667584</sourcerecordid><originalsourceid>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</originalsourceid><addsrcrecordid>eNo9kE1Lw0AQhhdRsFZ_gHgJeE6d_cruemuDnxQ9tAXBw7LdbOrWNIm7ycF_b0qLp-EdnncGHoSuMUwwBnX3uljkEwIEJkQJEKBO0AhzLlMs6McpGgFgmSoCcI4uYtwOkTGJR-hz6U299fY-mSar2pfeFYmpi2RhTWXWlUvyL98ms-CLja83yaL13y6dmXjE8qbufN03fUzeXB9MNWx2bd-Zzjf1JTorTRXd1XGO0erxYZk_p_P3p5d8Ok8tZbhL5ZoIhzNaEJmVzhaKGQZcGRAZkVZwy4RgAKSUYLnjYHjhLFBYS4IF4yUdo9vD3TY0P72Lnd42faiHl5owkmWZ4JINFD5QNjQxBlfqNvidCb8ag9471HuHeu9QHx0OnZtDxzvn_nmplKRU0j_v52vJ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2426667584</pqid></control><display><type>article</type><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><source>IEEE Xplore (Online service)</source><creator>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</creator><creatorcontrib>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</creatorcontrib><description>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with &gt;610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</description><identifier>ISSN: 0018-9200</identifier><identifier>EISSN: 1558-173X</identifier><identifier>DOI: 10.1109/JSSC.2020.2970709</identifier><identifier>CODEN: IJSCBC</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Bicycles ; Biological neural networks ; Computational modeling ; Computer architecture ; Deep learning accelerator ; hybrid paradigm ; Machine learning ; Mapping ; Neural networks ; neuromorphic chip ; Neuromorphics ; Power efficiency ; unified/scalable architecture</subject><ispartof>IEEE journal of solid-state circuits, 2020-08, Vol.55 (8), p.2228-2246</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2020</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</citedby><cites>FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</cites><orcidid>0000-0002-8994-431X ; 0000-0002-9829-2202 ; 0000-0002-8534-6494 ; 0000-0002-5172-9411</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/8998338$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Deng, Lei</creatorcontrib><creatorcontrib>Wang, Guanrui</creatorcontrib><creatorcontrib>Li, Guoqi</creatorcontrib><creatorcontrib>Li, Shuangchen</creatorcontrib><creatorcontrib>Liang, Ling</creatorcontrib><creatorcontrib>Zhu, Maohua</creatorcontrib><creatorcontrib>Wu, Yujie</creatorcontrib><creatorcontrib>Yang, Zheyu</creatorcontrib><creatorcontrib>Zou, Zhe</creatorcontrib><creatorcontrib>Pei, Jing</creatorcontrib><creatorcontrib>Wu, Zhenzhi</creatorcontrib><creatorcontrib>Hu, Xing</creatorcontrib><creatorcontrib>Ding, Yufei</creatorcontrib><creatorcontrib>He, Wei</creatorcontrib><creatorcontrib>Xie, Yuan</creatorcontrib><creatorcontrib>Shi, Luping</creatorcontrib><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><title>IEEE journal of solid-state circuits</title><addtitle>JSSC</addtitle><description>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with &gt;610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</description><subject>Bicycles</subject><subject>Biological neural networks</subject><subject>Computational modeling</subject><subject>Computer architecture</subject><subject>Deep learning accelerator</subject><subject>hybrid paradigm</subject><subject>Machine learning</subject><subject>Mapping</subject><subject>Neural networks</subject><subject>neuromorphic chip</subject><subject>Neuromorphics</subject><subject>Power efficiency</subject><subject>unified/scalable architecture</subject><issn>0018-9200</issn><issn>1558-173X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><recordid>eNo9kE1Lw0AQhhdRsFZ_gHgJeE6d_cruemuDnxQ9tAXBw7LdbOrWNIm7ycF_b0qLp-EdnncGHoSuMUwwBnX3uljkEwIEJkQJEKBO0AhzLlMs6McpGgFgmSoCcI4uYtwOkTGJR-hz6U299fY-mSar2pfeFYmpi2RhTWXWlUvyL98ms-CLja83yaL13y6dmXjE8qbufN03fUzeXB9MNWx2bd-Zzjf1JTorTRXd1XGO0erxYZk_p_P3p5d8Ok8tZbhL5ZoIhzNaEJmVzhaKGQZcGRAZkVZwy4RgAKSUYLnjYHjhLFBYS4IF4yUdo9vD3TY0P72Lnd42faiHl5owkmWZ4JINFD5QNjQxBlfqNvidCb8ag9471HuHeu9QHx0OnZtDxzvn_nmplKRU0j_v52vJ</recordid><startdate>20200801</startdate><enddate>20200801</enddate><creator>Deng, Lei</creator><creator>Wang, Guanrui</creator><creator>Li, Guoqi</creator><creator>Li, Shuangchen</creator><creator>Liang, Ling</creator><creator>Zhu, Maohua</creator><creator>Wu, Yujie</creator><creator>Yang, Zheyu</creator><creator>Zou, Zhe</creator><creator>Pei, Jing</creator><creator>Wu, Zhenzhi</creator><creator>Hu, Xing</creator><creator>Ding, Yufei</creator><creator>He, Wei</creator><creator>Xie, Yuan</creator><creator>Shi, Luping</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0002-8994-431X</orcidid><orcidid>https://orcid.org/0000-0002-9829-2202</orcidid><orcidid>https://orcid.org/0000-0002-8534-6494</orcidid><orcidid>https://orcid.org/0000-0002-5172-9411</orcidid></search><sort><creationdate>20200801</creationdate><title>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</title><author>Deng, Lei ; Wang, Guanrui ; Li, Guoqi ; Li, Shuangchen ; Liang, Ling ; Zhu, Maohua ; Wu, Yujie ; Yang, Zheyu ; Zou, Zhe ; Pei, Jing ; Wu, Zhenzhi ; Hu, Xing ; Ding, Yufei ; He, Wei ; Xie, Yuan ; Shi, Luping</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Bicycles</topic><topic>Biological neural networks</topic><topic>Computational modeling</topic><topic>Computer architecture</topic><topic>Deep learning accelerator</topic><topic>hybrid paradigm</topic><topic>Machine learning</topic><topic>Mapping</topic><topic>Neural networks</topic><topic>neuromorphic chip</topic><topic>Neuromorphics</topic><topic>Power efficiency</topic><topic>unified/scalable architecture</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Deng, Lei</creatorcontrib><creatorcontrib>Wang, Guanrui</creatorcontrib><creatorcontrib>Li, Guoqi</creatorcontrib><creatorcontrib>Li, Shuangchen</creatorcontrib><creatorcontrib>Liang, Ling</creatorcontrib><creatorcontrib>Zhu, Maohua</creatorcontrib><creatorcontrib>Wu, Yujie</creatorcontrib><creatorcontrib>Yang, Zheyu</creatorcontrib><creatorcontrib>Zou, Zhe</creatorcontrib><creatorcontrib>Pei, Jing</creatorcontrib><creatorcontrib>Wu, Zhenzhi</creatorcontrib><creatorcontrib>Hu, Xing</creatorcontrib><creatorcontrib>Ding, Yufei</creatorcontrib><creatorcontrib>He, Wei</creatorcontrib><creatorcontrib>Xie, Yuan</creatorcontrib><creatorcontrib>Shi, Luping</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE journal of solid-state circuits</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Deng, Lei</au><au>Wang, Guanrui</au><au>Li, Guoqi</au><au>Li, Shuangchen</au><au>Liang, Ling</au><au>Zhu, Maohua</au><au>Wu, Yujie</au><au>Yang, Zheyu</au><au>Zou, Zhe</au><au>Pei, Jing</au><au>Wu, Zhenzhi</au><au>Hu, Xing</au><au>Ding, Yufei</au><au>He, Wei</au><au>Xie, Yuan</au><au>Shi, Luping</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation</atitle><jtitle>IEEE journal of solid-state circuits</jtitle><stitle>JSSC</stitle><date>2020-08-01</date><risdate>2020</risdate><volume>55</volume><issue>8</issue><spage>2228</spage><epage>2246</epage><pages>2228-2246</pages><issn>0018-9200</issn><eissn>1558-173X</eissn><coden>IJSCBC</coden><abstract>Toward the long-standing dream of artificial intelligence, two successful solution paths have been paved: 1) neuromorphic computing and 2) deep learning. Recently, they tend to interact for simultaneously achieving biological plausibility and powerful accuracy. However, models from these two domains have to run on distinct substrates, i.e., neuromorphic platforms and deep learning accelerators, respectively. This architectural incompatibility greatly compromises the modeling flexibility and hinders promising interdisciplinary research. To address this issue, we build a unified model description framework and a unified processing architecture (Tianjic), which covers the full stack from software to hardware. By implementing a set of integration and transformation operations, Tianjic is able to support spiking neural networks, biological dynamic neural networks, multilayered perceptron, convolutional neural networks, recurrent neural networks, and so on. A compatible routing infrastructure enables homogeneous and heterogeneous scalability on a decentralized many-core network. Several optimization methods are incorporated, such as resource and data sharing, near-memory processing, compute/access skipping, and intra-/inter-core pipeline, to improve performance and efficiency. We further design streaming mapping schemes for efficient network deployment with a flexible tradeoff between execution throughput and resource overhead. A 28-nm prototype chip is fabricated with &gt;610-GB/s internal memory bandwidth. A variety of benchmarks are evaluated and compared with GPUs and several existing specialized platforms. In summary, the fully unfolded mapping can achieve significantly higher throughput and power efficiency; the semi-folded mapping can save 30x resources while still presenting comparable performance on average. Finally, two hybrid-paradigm examples, a multimodal unmanned bicycle and a hybrid neural network, are demonstrated to show the potential of our unified architecture. This article paves a new way to explore neural computing.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/JSSC.2020.2970709</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-8994-431X</orcidid><orcidid>https://orcid.org/0000-0002-9829-2202</orcidid><orcidid>https://orcid.org/0000-0002-8534-6494</orcidid><orcidid>https://orcid.org/0000-0002-5172-9411</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0018-9200
ispartof IEEE journal of solid-state circuits, 2020-08, Vol.55 (8), p.2228-2246
issn 0018-9200
1558-173X
language eng
recordid cdi_proquest_journals_2426667584
source IEEE Xplore (Online service)
subjects Bicycles
Biological neural networks
Computational modeling
Computer architecture
Deep learning accelerator
hybrid paradigm
Machine learning
Mapping
Neural networks
neuromorphic chip
Neuromorphics
Power efficiency
unified/scalable architecture
title Tianjic: A Unified and Scalable Chip Bridging Spike-Based and Continuous Neural Computation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T05%3A04%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Tianjic:%20A%20Unified%20and%20Scalable%20Chip%20Bridging%20Spike-Based%20and%20Continuous%20Neural%20Computation&rft.jtitle=IEEE%20journal%20of%20solid-state%20circuits&rft.au=Deng,%20Lei&rft.date=2020-08-01&rft.volume=55&rft.issue=8&rft.spage=2228&rft.epage=2246&rft.pages=2228-2246&rft.issn=0018-9200&rft.eissn=1558-173X&rft.coden=IJSCBC&rft_id=info:doi/10.1109/JSSC.2020.2970709&rft_dat=%3Cproquest_ieee_%3E2426667584%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c341t-8b27e163d286fecd94a4059a07628c75c4774002f80c5e50a5dec030b821745f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2426667584&rft_id=info:pmid/&rft_ieee_id=8998338&rfr_iscdi=true