Loading…

High Performance MPI over the Slingshot Interconnect

The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various...

Full description

Saved in:
Bibliographic Details
Published in:Journal of computer science and technology 2023-02, Vol.38 (1), p.128-145
Main Authors: Khorassani, Kawthar Shafie, Chen, Chen-Chun, Ramesh, Bharath, Shafi, Aamir, Subramoni, Hari, Panda, Dhabaleswar K.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3
container_end_page 145
container_issue 1
container_start_page 128
container_title Journal of computer science and technology
container_volume 38
creator Khorassani, Kawthar Shafie
Chen, Chen-Chun
Ramesh, Bharath
Shafi, Aamir
Subramoni, Hari
Panda, Dhabaleswar K.
description The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.
doi_str_mv 10.1007/s11390-023-2907-5
format article
fullrecord <record><control><sourceid>wanfang_jour_proqu</sourceid><recordid>TN_cdi_wanfang_journals_jsjkxjsxb_e202301009</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A745163421</galeid><wanfj_id>jsjkxjsxb_e202301009</wanfj_id><sourcerecordid>jsjkxjsxb_e202301009</sourcerecordid><originalsourceid>FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</originalsourceid><addsrcrecordid>eNp1kF9LwzAUxYsoOKcfwLeCr3be_GvSxzHUDSYO1OeQtrdd65bOpOr89mZU2JMEbsLld27uOVF0TWBCAOSdJ4RlkABlCc1AJuIkGhGVQsIlz07DGwCSLJTz6ML7FoBJ4HwU8XlTr-MVuqpzW2MLjJ9Wi7j7Qhf3a4xfNo2t_brr44Xt0RWdtVj0l9FZZTYer_7ucfT2cP86myfL58fFbLpMCqZkn2AFlGSkQlWWpSGAsqC8oLkQwBQDynNl0pwZkCCoIJSqwNJMpAKMymXOxtHtMPfb2MrYWrfdp7PhR9369n3f-n2ukQbPEOxlAb8Z8J3rPj7R90eeZkQJRUHxQE0GqjYb1I2tut6ZIpwSt00wiFUT-lPJBUkZpyQIyCAoXOe9w0rvXLM17kcT0Ifw9RC-DovoQ_haBA0dND6wtkZ3XOV_0S_uv4P-</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2918582084</pqid></control><display><type>article</type><title>High Performance MPI over the Slingshot Interconnect</title><source>ABI/INFORM Global</source><source>Springer Nature</source><creator>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</creator><creatorcontrib>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</creatorcontrib><description>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</description><identifier>ISSN: 1000-9000</identifier><identifier>EISSN: 1860-4749</identifier><identifier>DOI: 10.1007/s11390-023-2907-5</identifier><language>eng</language><publisher>Singapore: Springer Nature Singapore</publisher><subject>Adaptive control ; Artificial Intelligence ; Central processing units ; Computer Science ; CPUs ; Data Structures and Information Theory ; Evaluation ; Graphics processing units ; Information Systems Applications (incl.Internet) ; Libraries ; Message passing ; Microprocessors ; Nodes ; Rankings ; Regular Paper ; Software Engineering ; Supercomputers ; Theory of Computation</subject><ispartof>Journal of computer science and technology, 2023-02, Vol.38 (1), p.128-145</ispartof><rights>Institute of Computing Technology, Chinese Academy of Sciences 2023</rights><rights>COPYRIGHT 2023 Springer</rights><rights>Institute of Computing Technology, Chinese Academy of Sciences 2023.</rights><rights>Copyright © Wanfang Data Co. Ltd. All Rights Reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Uhttp://www.wanfangdata.com.cn/images/PeriodicalImages/jsjkxjsxb-e/jsjkxjsxb-e.jpg</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2918582084?pq-origsite=primo$$EHTML$$P50$$Gproquest$$H</linktohtml><link.rule.ids>314,780,784,11688,27924,27925,36060,44363</link.rule.ids></links><search><creatorcontrib>Khorassani, Kawthar Shafie</creatorcontrib><creatorcontrib>Chen, Chen-Chun</creatorcontrib><creatorcontrib>Ramesh, Bharath</creatorcontrib><creatorcontrib>Shafi, Aamir</creatorcontrib><creatorcontrib>Subramoni, Hari</creatorcontrib><creatorcontrib>Panda, Dhabaleswar K.</creatorcontrib><title>High Performance MPI over the Slingshot Interconnect</title><title>Journal of computer science and technology</title><addtitle>J. Comput. Sci. Technol</addtitle><description>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</description><subject>Adaptive control</subject><subject>Artificial Intelligence</subject><subject>Central processing units</subject><subject>Computer Science</subject><subject>CPUs</subject><subject>Data Structures and Information Theory</subject><subject>Evaluation</subject><subject>Graphics processing units</subject><subject>Information Systems Applications (incl.Internet)</subject><subject>Libraries</subject><subject>Message passing</subject><subject>Microprocessors</subject><subject>Nodes</subject><subject>Rankings</subject><subject>Regular Paper</subject><subject>Software Engineering</subject><subject>Supercomputers</subject><subject>Theory of Computation</subject><issn>1000-9000</issn><issn>1860-4749</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>M0C</sourceid><recordid>eNp1kF9LwzAUxYsoOKcfwLeCr3be_GvSxzHUDSYO1OeQtrdd65bOpOr89mZU2JMEbsLld27uOVF0TWBCAOSdJ4RlkABlCc1AJuIkGhGVQsIlz07DGwCSLJTz6ML7FoBJ4HwU8XlTr-MVuqpzW2MLjJ9Wi7j7Qhf3a4xfNo2t_brr44Xt0RWdtVj0l9FZZTYer_7ucfT2cP86myfL58fFbLpMCqZkn2AFlGSkQlWWpSGAsqC8oLkQwBQDynNl0pwZkCCoIJSqwNJMpAKMymXOxtHtMPfb2MrYWrfdp7PhR9369n3f-n2ukQbPEOxlAb8Z8J3rPj7R90eeZkQJRUHxQE0GqjYb1I2tut6ZIpwSt00wiFUT-lPJBUkZpyQIyCAoXOe9w0rvXLM17kcT0Ifw9RC-DovoQ_haBA0dND6wtkZ3XOV_0S_uv4P-</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Khorassani, Kawthar Shafie</creator><creator>Chen, Chen-Chun</creator><creator>Ramesh, Bharath</creator><creator>Shafi, Aamir</creator><creator>Subramoni, Hari</creator><creator>Panda, Dhabaleswar K.</creator><general>Springer Nature Singapore</general><general>Springer</general><general>Springer Nature B.V</general><general>Department of Computer Science and Engineering,The Ohio State University,Columbus,OH 43210,U.S.A</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L6V</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M7S</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PTHSS</scope><scope>Q9U</scope><scope>2B.</scope><scope>4A8</scope><scope>92I</scope><scope>93N</scope><scope>PSX</scope><scope>TCJ</scope></search><sort><creationdate>20230201</creationdate><title>High Performance MPI over the Slingshot Interconnect</title><author>Khorassani, Kawthar Shafie ; Chen, Chen-Chun ; Ramesh, Bharath ; Shafi, Aamir ; Subramoni, Hari ; Panda, Dhabaleswar K.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Adaptive control</topic><topic>Artificial Intelligence</topic><topic>Central processing units</topic><topic>Computer Science</topic><topic>CPUs</topic><topic>Data Structures and Information Theory</topic><topic>Evaluation</topic><topic>Graphics processing units</topic><topic>Information Systems Applications (incl.Internet)</topic><topic>Libraries</topic><topic>Message passing</topic><topic>Microprocessors</topic><topic>Nodes</topic><topic>Rankings</topic><topic>Regular Paper</topic><topic>Software Engineering</topic><topic>Supercomputers</topic><topic>Theory of Computation</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Khorassani, Kawthar Shafie</creatorcontrib><creatorcontrib>Chen, Chen-Chun</creatorcontrib><creatorcontrib>Ramesh, Bharath</creatorcontrib><creatorcontrib>Shafi, Aamir</creatorcontrib><creatorcontrib>Subramoni, Hari</creatorcontrib><creatorcontrib>Panda, Dhabaleswar K.</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI-INFORM Complete</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>ProQuest Engineering Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Engineering Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>Engineering Collection</collection><collection>ProQuest Central Basic</collection><collection>Wanfang Data Journals - Hong Kong</collection><collection>WANFANG Data Centre</collection><collection>Wanfang Data Journals</collection><collection>万方数据期刊 - 香港版</collection><collection>China Online Journals (COJ)</collection><collection>China Online Journals (COJ)</collection><jtitle>Journal of computer science and technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Khorassani, Kawthar Shafie</au><au>Chen, Chen-Chun</au><au>Ramesh, Bharath</au><au>Shafi, Aamir</au><au>Subramoni, Hari</au><au>Panda, Dhabaleswar K.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>High Performance MPI over the Slingshot Interconnect</atitle><jtitle>Journal of computer science and technology</jtitle><stitle>J. Comput. Sci. Technol</stitle><date>2023-02-01</date><risdate>2023</risdate><volume>38</volume><issue>1</issue><spage>128</spage><epage>145</epage><pages>128-145</pages><issn>1000-9000</issn><eissn>1860-4749</eissn><abstract>The Slingshot interconnect designed by HPE/Cray is becoming more relevant in high-performance computing with its deployment on the upcoming exascale systems. In particular, it is the interconnect empowering the first exascale and highest-ranked supercomputer in the world, Frontier. It offers various features such as adaptive routing, congestion control, and isolated workloads. The deployment of newer interconnects sparks interest related to performance, scalability, and any potential bottlenecks as they are critical elements contributing to the scalability across nodes on these systems. In this paper, we delve into the challenges the Slingshot interconnect poses with current state-of-the-art MPI (message passing interface) libraries. In particular, we look at the scalability performance when using Slingshot across nodes. We present a comprehensive evaluation using various MPI and communication libraries including Cray MPICH, Open- MPI + UCX, RCCL, and MVAPICH2 on CPUs and GPUs on the Spock system, an early access cluster deployed with Slingshot-10, AMD MI100 GPUs and AMD Epyc Rome CPUs to emulate the Frontier system. We also evaluate preliminary CPU-based support of MPI libraries on the Slingshot-11 interconnect.</abstract><cop>Singapore</cop><pub>Springer Nature Singapore</pub><doi>10.1007/s11390-023-2907-5</doi><tpages>18</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1000-9000
ispartof Journal of computer science and technology, 2023-02, Vol.38 (1), p.128-145
issn 1000-9000
1860-4749
language eng
recordid cdi_wanfang_journals_jsjkxjsxb_e202301009
source ABI/INFORM Global; Springer Nature
subjects Adaptive control
Artificial Intelligence
Central processing units
Computer Science
CPUs
Data Structures and Information Theory
Evaluation
Graphics processing units
Information Systems Applications (incl.Internet)
Libraries
Message passing
Microprocessors
Nodes
Rankings
Regular Paper
Software Engineering
Supercomputers
Theory of Computation
title High Performance MPI over the Slingshot Interconnect
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T02%3A05%3A59IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-wanfang_jour_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=High%20Performance%20MPI%20over%20the%20Slingshot%20Interconnect&rft.jtitle=Journal%20of%20computer%20science%20and%20technology&rft.au=Khorassani,%20Kawthar%20Shafie&rft.date=2023-02-01&rft.volume=38&rft.issue=1&rft.spage=128&rft.epage=145&rft.pages=128-145&rft.issn=1000-9000&rft.eissn=1860-4749&rft_id=info:doi/10.1007/s11390-023-2907-5&rft_dat=%3Cwanfang_jour_proqu%3Ejsjkxjsxb_e202301009%3C/wanfang_jour_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c387t-ef02191fe8ddda10e7c24c2b550383024b8a6b3a07052512281fe295650a8b7b3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2918582084&rft_id=info:pmid/&rft_galeid=A745163421&rft_wanfj_id=jsjkxjsxb_e202301009&rfr_iscdi=true