Loading…
High Performance MPI over the Slingshot Interconnect
The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various...
Saved in:
Published in: | Journal of computer science and technology 2023-02, Vol.38 (1), p.128-145 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3 |
container_end_page | 145 |
container_issue | 1 |
container_start_page | 128 |
container_title | Journal of computer science and technology |
container_volume | 38 |
creator | Khorassani, Kawthar Shafie Chen, Chen-Chun Ramesh, Bharath Shafi, Aamir Subramoni, Hari Panda, Dhabaleswar K. |
description | The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect. |
doi_str_mv | 10.1007/s11390-023-2907-5 |
format | article |
fullrecord | <record><control><sourceid>wanfang_jour_proqu</sourceid><recordid>TN_cdi_wanfang_journals_jsjkxjsxb_e202301009</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A745163421</galeid><wanfj_id>jsjkxjsxb_e202301009</wanfj_id><sourcerecordid>jsjkxjsxb_e202301009</sourcerecordid><originalsourceid>FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</originalsourceid><addsrcrecordid>eNp1kF9LwzAUxYsoOKcfwLeCr3be_GvSxzHUDSYO1OeQtrdd65bOpOr89mZU2JMEbsLld27uOVF0TWBCAOSdJ4RlkABlCc1AJuIkGhGVQsIlz07DGwCSLJTz6ML7FoBJ4HwU8XlTr-MVuqpzW2MLjJ9Wi7j7Qhf3a4xfNo2t_brr44Xt0RWdtVj0l9FZZTYer_7ucfT2cP86myfL58fFbLpMCqZkn2AFlGSkQlWWpSGAsqC8oLkQwBQDynNl0pwZkCCoIJSqwNJMpAKMymXOxtHtMPfb2MrYWrfdp7PhR9369n3f-n2ukQbPEOxlAb8Z8J3rPj7R90eeZkQJRUHxQE0GqjYb1I2tut6ZIpwSt00wiFUT-lPJBUkZpyQIyCAoXOe9w0rvXLM17kcT0Ifw9RC-DovoQ_haBA0dND6wtkZ3XOV_0S_uv4P-</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2918582084</pqid></control><display><type>article</type><title>High Performance MPI over the Slingshot Interconnect</title><source>ABI/INFORM Global</source><source>Springer Nature</source><creator>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</creator><creatorcontrib>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</creatorcontrib><description>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</description><identifier>ISSN: 1000-9000</identifier><identifier>EISSN: 1860-4749</identifier><identifier>DOI: 10.1007/s11390-023-2907-5</identifier><language>eng</language><publisher>Singapore: Springer Nature Singapore</publisher><subject>Adaptive control ; Artificial Intelligence ; Central processing units ; Computer Science ; CPUs ; Data Structures and Information Theory ; Evaluation ; Graphics processing units ; Information Systems Applications (incl.Internet) ; Libraries ; Message passing ; Microprocessors ; Nodes ; Rankings ; Regular Paper ; Software Engineering ; Supercomputers ; Theory of Computation</subject><ispartof>Journal of computer science and technology, 2023-02, Vol.38 (1), p.128-145</ispartof><rights>Institute of Computing Technology, Chinese Academy of Sciences 2023</rights><rights>COPYRIGHT 2023 Springer</rights><rights>Institute of Computing Technology, Chinese Academy of Sciences 2023.</rights><rights>Copyright © Wanfang Data Co. Ltd. All Rights Reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Uhttp://www.wanfangdata.com.cn/images/PeriodicalImages/jsjkxjsxb-e/jsjkxjsxb-e.jpg</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2918582084?pq-origsite=primo$$EHTML$$P50$$Gproquest$$H</linktohtml><link.rule.ids>314,780,784,11688,27924,27925,36060,44363</link.rule.ids></links><search><creatorcontrib>Khorassani, Kawthar Shafie</creatorcontrib><creatorcontrib>Chen, Chen-Chun</creatorcontrib><creatorcontrib>Ramesh, Bharath</creatorcontrib><creatorcontrib>Shafi, Aamir</creatorcontrib><creatorcontrib>Subramoni, Hari</creatorcontrib><creatorcontrib>Panda, Dhabaleswar K.</creatorcontrib><title>High Performance MPI over the Slingshot Interconnect</title><title>Journal of computer science and technology</title><addtitle>J. Comput. Sci. Technol</addtitle><description>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</description><subject>Adaptive control</subject><subject>Artificial Intelligence</subject><subject>Central processing units</subject><subject>Computer Science</subject><subject>CPUs</subject><subject>Data Structures and Information Theory</subject><subject>Evaluation</subject><subject>Graphics processing units</subject><subject>Information Systems Applications (incl.Internet)</subject><subject>Libraries</subject><subject>Message passing</subject><subject>Microprocessors</subject><subject>Nodes</subject><subject>Rankings</subject><subject>Regular Paper</subject><subject>Software Engineering</subject><subject>Supercomputers</subject><subject>Theory of Computation</subject><issn>1000-9000</issn><issn>1860-4749</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>M0C</sourceid><recordid>eNp1kF9LwzAUxYsoOKcfwLeCr3be_GvSxzHUDSYO1OeQtrdd65bOpOr89mZU2JMEbsLld27uOVF0TWBCAOSdJ4RlkABlCc1AJuIkGhGVQsIlz07DGwCSLJTz6ML7FoBJ4HwU8XlTr-MVuqpzW2MLjJ9Wi7j7Qhf3a4xfNo2t_brr44Xt0RWdtVj0l9FZZTYer_7ucfT2cP86myfL58fFbLpMCqZkn2AFlGSkQlWWpSGAsqC8oLkQwBQDynNl0pwZkCCoIJSqwNJMpAKMymXOxtHtMPfb2MrYWrfdp7PhR9369n3f-n2ukQbPEOxlAb8Z8J3rPj7R90eeZkQJRUHxQE0GqjYb1I2tut6ZIpwSt00wiFUT-lPJBUkZpyQIyCAoXOe9w0rvXLM17kcT0Ifw9RC-DovoQ_haBA0dND6wtkZ3XOV_0S_uv4P-</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Khorassani, Kawthar Shafie</creator><creator>Chen, Chen-Chun</creator><creator>Ramesh, Bharath</creator><creator>Shafi, Aamir</creator><creator>Subramoni, Hari</creator><creator>Panda, Dhabaleswar K.</creator><general>Springer Nature Singapore</general><general>Springer</general><general>Springer Nature B.V</general><general>Department of Computer Science and Engineering,The Ohio State University,Columbus,OH 43210,U.S.A</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PTHSS</scope><scope>Q9U</scope><scope>2B.</scope><scope>4A8</scope><scope>92I</scope><scope>93N</scope><scope>PSX</scope><scope>TCJ</scope></search><sort><creationdate>20230201</creationdate><title>High Performance MPI over the Slingshot Interconnect</title><author>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Adaptive control</topic><topic>Artificial Intelligence</topic><topic>Central processing units</topic><topic>Computer Science</topic><topic>CPUs</topic><topic>Data Structures and Information Theory</topic><topic>Evaluation</topic><topic>Graphics processing units</topic><topic>Information Systems Applications (incl.Internet)</topic><topic>Libraries</topic><topic>Message passing</topic><topic>Microprocessors</topic><topic>Nodes</topic><topic>Rankings</topic><topic>Regular Paper</topic><topic>Software Engineering</topic><topic>Supercomputers</topic><topic>Theory of Computation</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Khorassani, Kawthar Shafie</creatorcontrib><creatorcontrib>Chen, Chen-Chun</creatorcontrib><creatorcontrib>Ramesh, Bharath</creatorcontrib><creatorcontrib>Shafi, Aamir</creatorcontrib><creatorcontrib>Subramoni, Hari</creatorcontrib><creatorcontrib>Panda, Dhabaleswar K.</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI-INFORM Complete</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Engineering Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>Engineering Collection</collection><collection>ProQuest Central Basic</collection><collection>Wanfang Data Journals - Hong Kong</collection><collection>WANFANG Data Centre</collection><collection>Wanfang Data Journals</collection><collection>万方数据期刊 - 香港版</collection><collection>China Online Journals (COJ)</collection><collection>China Online Journals (COJ)</collection><jtitle>Journal of computer science and technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Khorassani, Kawthar Shafie</au><au>Chen, Chen-Chun</au><au>Ramesh, Bharath</au><au>Shafi, Aamir</au><au>Subramoni, Hari</au><au>Panda, Dhabaleswar K.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>High Performance MPI over the Slingshot Interconnect</atitle><jtitle>Journal of computer science and technology</jtitle><stitle>J. Comput. Sci. Technol</stitle><date>2023-02-01</date><risdate>2023</risdate><volume>38</volume><issue>1</issue><spage>128</spage><epage>145</epage><pages>128-145</pages><issn>1000-9000</issn><eissn>1860-4749</eissn><abstract>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</abstract><cop>Singapore</cop><pub>Springer Nature Singapore</pub><doi>10.1007/s11390-023-2907-5</doi><tpages>18</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1000-9000 |
ispartof | Journal of computer science and technology, 2023-02, Vol.38 (1), p.128-145 |
issn | 1000-9000 1860-4749 |
language | eng |
recordid | cdi_wanfang_journals_jsjkxjsxb_e202301009 |
source | ABI/INFORM Global; Springer Nature |
subjects | Adaptive control Artificial Intelligence Central processing units Computer Science CPUs Data Structures and Information Theory Evaluation Graphics processing units Information Systems Applications (incl.Internet) Libraries Message passing Microprocessors Nodes Rankings Regular Paper Software Engineering Supercomputers Theory of Computation |
title | High Performance MPI over the Slingshot Interconnect |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T02%3A05%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-wanfang_jour_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=High%20Performance%20MPI%20over%20the%20Slingshot%20Interconnect&rft.jtitle=Journal%20of%20computer%20science%20and%20technology&rft.au=Khorassani,%20Kawthar%20Shafie&rft.date=2023-02-01&rft.volume=38&rft.issue=1&rft.spage=128&rft.epage=145&rft.pages=128-145&rft.issn=1000-9000&rft.eissn=1860-4749&rft_id=info:doi/10.1007/s11390-023-2907-5&rft_dat=%3Cwanfang_jour_proqu%3Ejsjkxjsxb_e202301009%3C/wanfang_jour_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2918582084&rft_id=info:pmid/&rft_galeid=A745163421&rft_wanfj_id=jsjkxjsxb_e202301009&rfr_iscdi=true |