SlideShare uma empresa Scribd logo
1 de 35
Baixar para ler offline
Interconnects
Ken Raffenetti
Software Development Specialist
Programming Models and Runtime Systems Group
Mathematics and Computer Science Division
Argonne National Laboratory
U.S. DOE Potential System Architecture Targets
ATPESC (07/29/2019)
System attributes 2010 2018-2019 2021-2022
System peak 2 Peta 150-200 Petaflop/sec 1 Exaflop/sec
Power 6 MW 15 MW 20 MW
System memory 0.3 PB 5 PB 32-64 PB
Node performance 125 GF 3 TF 30 TF 10 TF 100 TF
Node memory BW 25 GB/s 0.1TB/sec 1 TB/sec 0.4TB/sec 4 TB/sec
Node concurrency 12 O(100) O(1,000) O(1,000) O(10,000)
System size (nodes) 18,700 50,000 5,000 100,000 10,000
Total Node
Interconnect BW
1.5 GB/s 20 GB/sec 200GB/sec
MTTI days O(1day) O(1 day)
Past
production
Current generation
(e.g., CORAL)
Exascale
Goals
[Includes modifications to the DOE Exascale report]
General Trends in System Architecture
§ Number of nodes is increasing, but at a moderate pace
§ Number of cores/threads on a node is increasing rapidly
§ Each core is not increasing in speed (clock frequency)
§ What does this mean for networks?
– More cores will drive the network
– More sharing of the network infrastructure
– The aggregate amount of communication from each node will
increase moderately, but will be divided into many smaller messages
– A single core may not be able fully saturate the NIC
ATPESC (07/29/2019)
A Simplified Network Architecture
§ Hardware components
– Processing cores and
memory subsystem
– I/O bus or links
– Network
adapters/switches
§ Software components
– Communication stack
§ Balanced approach
required to maximize
user-perceived network
performance
ATPESC (07/29/2019)
P0
Core0 Core1
Core2 Core3
P1
Core0 Core1
Core2 Core3
Memory
MemoryI
/
O
B
u
s
Network Adapter
Network
Switch
Processing
Bottlenecks
I/O Interface
Bottlenecks
Network
End-host
Bottlenecks
Network
Fabric
Bottlenecks
Agenda
ATPESC (07/29/2019)
Network Adapters
Network Topologies
Network/Processor/Memory
Interactions
Bottlenecks on Traditional Network Adapters
§ Network speeds plateaued at
around 1Gbps
– Features provided were limited
– Commodity networks were not
considered scalable enough for
very large-scale systems
ATPESC (07/29/2019)
P0
Core0 Core1
Core2 Core3
P1
Core0 Core1
Core2 Core3
Memory
MemoryI
/
O
B
u
s
Network Adapter
Network
Switch
Network
Bottlenecks
Ethernet (1979 - ) 10 Mbit/sec
Fast Ethernet (1993 -) 100 Mbit/sec
Gigabit Ethernet (1995 -) 1000 Mbit /sec
ATM (1995 -) 155/622/1024 Mbit/sec
Myrinet (1993 -) 1 Gbit/sec
Fibre Channel (1994 -) 1 Gbit/sec
End-host Network Interface Speeds
§ HPC network technologies provide high bandwidth links
– InfiniBand EDR gives 100 Gbps per network link
• Will continue to increase (HDR 200 Gbps, etc.)
– Multiple network links becoming a common place
• ORNL Summit and LLNL Sierra machines, Japanese Post T2K machine
• Torus style or other multi-dimensional networks
§ End-host peak network bandwidth is “mostly” no longer
considered a major limitation
§ Network latency still an issue
– That’s a harder problem to solve – limited by physics, not technology
• There is some room to improve it in current technology (trimming the fat)
• Significant effort in making systems denser so as to reduce network latency
§ Other important metrics: message rate, congestion, …
ATPESC (07/29/2019)
Simple Network Architecture (past systems)
§ Processor, memory,
network are all
decoupled
ATPESC (07/29/2019)
P0
Core0 Core1
Core2 Core3
P1
Core0 Core1
Core2 Core3
Memory
MemoryI
/
O
B
u
s
Network Adapter
Network
Switch
Integrated Memory Controllers (current systems)
§ In the past 10 years or so, memory
controllers have been integrated on
to the processor
§ Primary purpose was scalable
memory bandwidth (NUMA)
§ Also helps network communication
– Data transfer to/from network requires
coordination with caches
§ Several network I/O technologies
exist
– PCIe, HTX, NVLink
– Expected to provide higher bandwidth
than what network links will have
ATPESC (07/29/2019)
P1
Core0 Core1
Core2 Core3
Memory
Memory
I
/
O
B
u
s
Network Adapter
Network
Switch
Memory
Controller
P0
Core0 Core1
Core2 Core3
Memory
Controller
Integrated Networks (future systems?)
ATPESC (07/29/2019)
Network
Switch
P0
Core0 Core1
Core2 Core3
Memory
Controller
Network
Interface
Off-chip Memory
§ Several vendors are considering
processor-integrated network
adapters
§ May improve network bandwidth
– Unclear if the I/O bus would be a
bottleneck
§ Improves network latencies
– Control messages between the
processor, network, and memory are
now on-chip
§ Improves network functionality
– Communication is a first-class citizen
and better integrated with processor
features
– E.g., network atomic operations can be
atomic with respect to processor
atomics
In-package
Memory
P1
Core0 Core1
Core2 Core3
Memory
Controller
Network
Interface
Off-chip Memory
In-package
Memory
Processing Bottlenecks in Traditional Protocols
§ Ex: TCP/IP, UDP/IP
§ Generic architecture for all networks
§ Host processor handles almost all
aspects of communication
– Data buffering (copies on sender and
receiver)
– Data integrity (checksum)
– Routing aspects (IP routing)
§ Signaling between different layers
– Hardware interrupt on packet arrival or
transmission
– Software signals between different
layers to handle protocol processing in
different priority levels
ATPESC (07/29/2019)
P0
Core0 Core1
Core2 Core3
P1
Core0 Core1
Core2 Core3
Memory
MemoryI
/
O
B
u
s
Network Adapter
Network
Switch
Processing
Bottlenecks
Network Protocol Stacks: The Offload Era
§ Modern networks are spending more and more network real-estate on
offloading various communication features on hardware
§ Network and transport layers are hardware offloaded for most modern
HPC networks
– Reliability (retransmissions, CRC checks), packetization
– OS-based memory registration, and user-level data transmission
ATPESC (07/29/2019)
Comparing Offloaded Network Stacks with
Traditional Network Stacks
ATPESC (07/29/2019)
Application Layer
MPI, PGAS, File Systems
Transport Layer
Low-level interface
Reliable/unreliable protocols
Link Layer
Flow-control, Error Detection
Physical Layer
Hardware Offload
Copper or Optical
HTTP, FTP, MPI,
File Systems
Routing
Physical Layer
Link Layer
Network Layer
Transport Layer
Application Layer
Traditional Ethernet
Sockets Interface
TCP, UDP
Flow-control and
Error Detection
Copper, Optical or Wireless
Network Layer Routing
Management tools
DNS management tools
Current State for Network APIs
§ A large number of network vendor specific APIs
– InfiniBand verbs, Intel PSM2, IBM PAMI, Cray Gemini/DMAPP, …
§ Recent efforts to standardize these low-level communication
APIs
– Open Fabrics Interface (OFI)
• Effort from Intel, CISCO, etc., to provide a unified low-level communication
layer that exposes features provided by each network
– Unified Communication X (UCX)
• Effort from Mellanox, IBM, ORNL, etc., to provide a unified low-level
communication layer that allows for efficient MPI and PGAS communication
– Portals 4
• Effort from Sandia National Laboratory to provide a network hardware
capability centric API
ATPESC (07/29/2019)
User-level Communication: Memory Registration
ATPESC (07/29/2019)
1. Registration Request
• Send virtual address and length
2. Kernel handles virtual->physical
mapping and pins region into
physical memory
• Process cannot map memory
that it does not own (security !)
3. Network adapter caches the
virtual to physical mapping and
issues a handle
4. Handle is returned to application
Before we do any communication:
All memory used for communication must
be registered
1
3
4
Process
Kernel
Network
2
User-level Communication: OS Bypass
ATPESC (07/29/2019)
Process
Kernel
Network
User-level APIs allow direct interaction with
network adapters § Contrast with traditional network
APIs that trap down to the kernel
§ Eliminates heavyweight context
switch
§ Memory registration caches allow
for fast buffer re-use, further
reducing dependence on the
kernel
Send/Receive Communication
ATPESC (07/29/2019)
Network Interface
Memory Memory
Network Interface
Send Recv
Memory
Segment
Send entry contains information about the
send buffer (multiple non-contiguous
segments)
Processor Processor
Send Recv
Memory
Segment
Receive entry contains information on the receive
buffer (multiple non-contiguous segments);
Incoming messages have to be matched to a
receive entry to know where to place
Hardware ACK
Memory
Segment
Memory
Segment
Memory
Segment
PUT/GET Communication
ATPESC (07/29/2019)
Network Interface
Memory Memory
Network Interface
Send Recv
Memory
Segment
Send entry contains information about the
send buffer (multiple segments) and the
receive buffer (single segment)
Processor Processor
Send Recv
Memory
Segment
Hardware ACK
Memory
Segment
Memory
Segment
Atomic Operations
ATPESC (07/29/2019)
Network Interface
Memory Memory
Network Interface
Send Recv
Memory
Segment
Send entry contains information about the
send buffer and the receive buffer
Processor Processor
Send Recv
Source
Memory
Segment
OP
Destination
Memory
Segment
Network Protocol Stacks: Specialization
§ Increasing network specialization is the focus today
– The next generation of networks plan to have further support for
noncontiguous data movement, and multiple contexts for multithreaded
architectures
§ Some networks, such as the Tofu network, Cray Aries and InfiniBand, are
also offloading some MPI and PGAS features on to hardware
– E.g., PUT/GET communication has hardware support
– Increasing number of atomic operations being offloaded to hardware
• Compare-and-swap, fetch-and-add, swap
– Collective operations (NIC and switch support)
• SHARP collectives
– Hardware tag matching for MPI send/recv
• Cray Seastar, Bull BXI, Mellanox Infiniband (ConnectX-5 and later)
ATPESC (07/29/2019)
Agenda
ATPESC (07/29/2019)
Network Adapters
Network Topologies
Network/Processor/Memory
Interactions
Traditional Network Topologies: Crossbar
§ A network topology describes how different network
adapters and switches are interconnected with each other
§ The ideal network topology (for performance) is a crossbar
– Alltoall connection between network adapters
– Typically done on a single network ASIC
– Current network crossbar ASICs go up to 64 ports; too expensive to
scale to higher port counts
– All communication is nonblocking
ATPESC (07/29/2019)
Traditional Network Topologies: Fat-tree
§ The most common topology for small and medium scale
systems is a fat-tree
– Nonblocking fat-tree switches available in abundance
• Allows for pseudo nonblocking communication
• Between all pairs of processes, there exists a completely nonblocking
path, but not all paths are nonblocking
– More scalable than crossbars, but the number of network links still
increases super-linearly with node count
• Can get very expensive with scale
ATPESC (07/29/2019)
Network Topology Trends
§ Modern topologies are moving towards
more “scalability” (with respect to cost, not
performance)
§ Blue Gene, Cray XE/XK, and K
supercomputers use a torus-network; Cray
XC uses dragonfly
– Linear increase in the number of
links/routers with system size
– Any communication that is more than one
hop away has a possibility of interference –
congestion is not just possible, but common
– Even when there is no congestion, such
topologies increase the network diameter
causing performance loss
§ Take-away: topological locality is important
and its not going to get better
ATPESC (07/29/2019)
Network Congestion Behavior: IBM BG/P
0
500
1000
1500
2000
2500
3000
3500
1
2
4
8
16
32
64
128
256
512
1K
2K
4K
8K
16K
32K
64K
128K
256K
512K
1M
Bandwidth(Mbps)
Message Size (bytes)
P2-P5
P3-P4
No overlap
ATPESC (07/29/2019)
P0 P1 P2 P3 P4 P5 P6 P7
2D Nearest Neighbor: Process Mapping (XYZ)
ATPESC (07/29/2019)
X-Axis
Z-Axis
Y-Axis
Nearest Neighbor Performance: IBM BG/P
ATPESC (07/29/2019)
0
100
200
300
400
500
600
700
800
900
2 4 8 16 32 64 128 256 512 1K
ExecutionTime(us)
Grid Partition (bytes)
System Size : 16K Cores
XYZT
TXYZ
ZYXT
TZYX
0
500
1000
1500
2000
2500
2 4 8 16 32 64 128 256 512 1K
ExecutionTime(us)
Grid Partition (bytes)
System Size : 128K Cores
XYZT
TXYZ
ZYXT
TZYX
2D Halo Exchange
Agenda
ATPESC (07/29/2019)
Network Adapters
Network Topologies
Network/Processor/Memory
Interactions
Network Interactions with Memory/Cache
§ Most network interfaces understand and work with the cache
coherence protocols available on modern systems
– Users do not have to ensure that data is flushed from cache before
communication
– Network and memory controller hardware understand what state the
data is in and communicate appropriately
ATPESC (07/29/2019)
Send-side Network Communication
ATPESC (07/29/2019)
North BridgeCPU
NIC
FSB Memory Bus
I/OBus
Appln.
Buffer
L3 $ Memory
Appln.
Buffer
Receive-side Network Communication
ATPESC (07/29/2019)
North BridgeCPU
NIC
FSB Memory Bus
I/OBus
Appln.
Buffer
L3 $ Memory
Appln.
Buffer
Network/Processor Interoperation Trends
§ Direct cache injection
– Most current networks inject data into memory
• If data is in cache, they flush cache and then inject to memory
– Some networks are investigating direct cache injection
• Data can be injected directly into the last-level cache
• Can be tricky since it can cause cache pollution if the incoming data is not
used immediately
§ Atomic operations
– Current network atomic operations are only atomic with respect to
other network operations and not with respect to processor atomics
• E.g., network fetch-and-add and processor fetch-and-add might corrupt
each other’s data
– With network/processor integration, this is expected to be fixed
ATPESC (07/29/2019)
Network Interactions with Accelerators
§ PCI Express peer-to-peer capabilities enables network
adapters to directly access third-party devices
– Coordination between network adapter and accelerator (GPUs,
FPGAs, …)
– Data does not need to be copied into to/from buffers when going over
the network
– GPUDirect RDMA for NVIDIA GPUs is one example
ATPESC (07/29/2019)
Summary
§ These are interesting times for all components in the overall
system architecture: processor, memory, interconnect
– And interesting times for computational science on these systems
§ Interconnect technology is rapidly advancing
– More hardware integration is the key to removing bottlenecks and
improve functionality
• Processor/memory/network integration is already in progress and will
continue for the foreseeable future
– Offload technologies continue to evolve as we move more
functionality to the network hardware
– Network topologies are becoming more “shared” (cost saving)
ATPESC (07/29/2019)
Thank You!
Email: raffenet@mcs.anl.gov

Mais conteúdo relacionado

Mais procurados

SCFE 2020 OpenCAPI presentation as part of OpenPWOER Tutorial
SCFE 2020 OpenCAPI presentation as part of OpenPWOER TutorialSCFE 2020 OpenCAPI presentation as part of OpenPWOER Tutorial
SCFE 2020 OpenCAPI presentation as part of OpenPWOER TutorialGanesan Narayanasamy
 
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...Ganesan Narayanasamy
 
High-Performance and Scalable Designs of Programming Models for Exascale Systems
High-Performance and Scalable Designs of Programming Models for Exascale SystemsHigh-Performance and Scalable Designs of Programming Models for Exascale Systems
High-Performance and Scalable Designs of Programming Models for Exascale Systemsinside-BigData.com
 
IBM Data Centric Systems & OpenPOWER
IBM Data Centric Systems & OpenPOWERIBM Data Centric Systems & OpenPOWER
IBM Data Centric Systems & OpenPOWERinside-BigData.com
 
TAU E4S ON OpenPOWER /POWER9 platform
TAU E4S ON OpenPOWER /POWER9 platformTAU E4S ON OpenPOWER /POWER9 platform
TAU E4S ON OpenPOWER /POWER9 platformGanesan Narayanasamy
 
FD.io Vector Packet Processing (VPP)
FD.io Vector Packet Processing (VPP)FD.io Vector Packet Processing (VPP)
FD.io Vector Packet Processing (VPP)Kirill Tsym
 
TLDK - FD.io Sept 2016
TLDK - FD.io Sept 2016 TLDK - FD.io Sept 2016
TLDK - FD.io Sept 2016 Benoit Hudzia
 
Trends in Systems and How to Get Efficient Performance
Trends in Systems and How to Get Efficient PerformanceTrends in Systems and How to Get Efficient Performance
Trends in Systems and How to Get Efficient Performanceinside-BigData.com
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Clusterinside-BigData.com
 
Scaling the Container Dataplane
Scaling the Container Dataplane Scaling the Container Dataplane
Scaling the Container Dataplane Michelle Holley
 
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand SolutionsMellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand Solutionsinside-BigData.com
 
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFV
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFVDPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFV
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFVJim St. Leger
 
OpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software StackOpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software Stackinside-BigData.com
 
AI is Impacting HPC Everywhere
AI is Impacting HPC EverywhereAI is Impacting HPC Everywhere
AI is Impacting HPC Everywhereinside-BigData.com
 

Mais procurados (20)

SCFE 2020 OpenCAPI presentation as part of OpenPWOER Tutorial
SCFE 2020 OpenCAPI presentation as part of OpenPWOER TutorialSCFE 2020 OpenCAPI presentation as part of OpenPWOER Tutorial
SCFE 2020 OpenCAPI presentation as part of OpenPWOER Tutorial
 
DOME 64-bit μDataCenter
DOME 64-bit μDataCenterDOME 64-bit μDataCenter
DOME 64-bit μDataCenter
 
ARM HPC Ecosystem
ARM HPC EcosystemARM HPC Ecosystem
ARM HPC Ecosystem
 
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
 
IBM HPC Transformation with AI
IBM HPC Transformation with AI IBM HPC Transformation with AI
IBM HPC Transformation with AI
 
High-Performance and Scalable Designs of Programming Models for Exascale Systems
High-Performance and Scalable Designs of Programming Models for Exascale SystemsHigh-Performance and Scalable Designs of Programming Models for Exascale Systems
High-Performance and Scalable Designs of Programming Models for Exascale Systems
 
IBM Data Centric Systems & OpenPOWER
IBM Data Centric Systems & OpenPOWERIBM Data Centric Systems & OpenPOWER
IBM Data Centric Systems & OpenPOWER
 
TAU E4S ON OpenPOWER /POWER9 platform
TAU E4S ON OpenPOWER /POWER9 platformTAU E4S ON OpenPOWER /POWER9 platform
TAU E4S ON OpenPOWER /POWER9 platform
 
FD.io Vector Packet Processing (VPP)
FD.io Vector Packet Processing (VPP)FD.io Vector Packet Processing (VPP)
FD.io Vector Packet Processing (VPP)
 
TLDK - FD.io Sept 2016
TLDK - FD.io Sept 2016 TLDK - FD.io Sept 2016
TLDK - FD.io Sept 2016
 
Trends in Systems and How to Get Efficient Performance
Trends in Systems and How to Get Efficient PerformanceTrends in Systems and How to Get Efficient Performance
Trends in Systems and How to Get Efficient Performance
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Cluster
 
Scaling the Container Dataplane
Scaling the Container Dataplane Scaling the Container Dataplane
Scaling the Container Dataplane
 
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand SolutionsMellanox Announces HDR 200 Gb/s InfiniBand Solutions
Mellanox Announces HDR 200 Gb/s InfiniBand Solutions
 
OpenPOWER System Marconi100
OpenPOWER System Marconi100OpenPOWER System Marconi100
OpenPOWER System Marconi100
 
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFV
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFVDPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFV
DPDK Summit - 08 Sept 2014 - Ericsson - A Multi-Socket Ferrari for NFV
 
OpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software StackOpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software Stack
 
RDMA on ARM
RDMA on ARMRDMA on ARM
RDMA on ARM
 
AI is Impacting HPC Everywhere
AI is Impacting HPC EverywhereAI is Impacting HPC Everywhere
AI is Impacting HPC Everywhere
 
Lenovo HPC Strategy Update
Lenovo HPC Strategy UpdateLenovo HPC Strategy Update
Lenovo HPC Strategy Update
 

Semelhante a Overview of HPC Interconnects

Ethernet summit 2011_toe
Ethernet summit 2011_toeEthernet summit 2011_toe
Ethernet summit 2011_toeintilop
 
Heterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of SystemsHeterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of SystemsAnand Haridass
 
PCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA CampPCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA CampFPGA Central
 
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKS
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKSHIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKS
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKSijngnjournal
 
BUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoCBUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoCLinaro
 
Communication Performance Over A Gigabit Ethernet Network
Communication Performance Over A Gigabit Ethernet NetworkCommunication Performance Over A Gigabit Ethernet Network
Communication Performance Over A Gigabit Ethernet NetworkIJERA Editor
 
Network Processing on an SPE Core in Cell Broadband EngineTM
Network Processing on an SPE Core in Cell Broadband EngineTMNetwork Processing on an SPE Core in Cell Broadband EngineTM
Network Processing on an SPE Core in Cell Broadband EngineTMSlide_N
 
Flexible and Scalable Domain-Specific Architectures
Flexible and Scalable Domain-Specific ArchitecturesFlexible and Scalable Domain-Specific Architectures
Flexible and Scalable Domain-Specific ArchitecturesNetronome
 
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011Stephen Foskett
 
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)Ontico
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task ComputingEric Van Hensbergen
 
High Performance Communication for Oracle using InfiniBand
High Performance Communication for Oracle using InfiniBandHigh Performance Communication for Oracle using InfiniBand
High Performance Communication for Oracle using InfiniBandwebhostingguy
 
ETHERNET PACKET PROCESSOR FOR SOC APPLICATION
ETHERNET PACKET PROCESSOR FOR SOC APPLICATIONETHERNET PACKET PROCESSOR FOR SOC APPLICATION
ETHERNET PACKET PROCESSOR FOR SOC APPLICATIONcscpconf
 
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...Netronome
 
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linux
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running LinuxLinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linux
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linuxbrouer
 
Converged Networks: FCoE, iSCSI and the Future of Storage Networking
Converged Networks: FCoE, iSCSI and the Future of Storage NetworkingConverged Networks: FCoE, iSCSI and the Future of Storage Networking
Converged Networks: FCoE, iSCSI and the Future of Storage NetworkingStuart Miniman
 
Mp So C 18 Apr
Mp So C 18 AprMp So C 18 Apr
Mp So C 18 AprFNian
 

Semelhante a Overview of HPC Interconnects (20)

Ethernet summit 2011_toe
Ethernet summit 2011_toeEthernet summit 2011_toe
Ethernet summit 2011_toe
 
Heterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of SystemsHeterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of Systems
 
PCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA CampPCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA Camp
 
The Cell Processor
The Cell ProcessorThe Cell Processor
The Cell Processor
 
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKS
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKSHIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKS
HIGH PERFORMANCE ETHERNET PACKET PROCESSOR CORE FOR NEXT GENERATION NETWORKS
 
BUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoCBUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoC
 
slides
slidesslides
slides
 
Communication Performance Over A Gigabit Ethernet Network
Communication Performance Over A Gigabit Ethernet NetworkCommunication Performance Over A Gigabit Ethernet Network
Communication Performance Over A Gigabit Ethernet Network
 
Network Processing on an SPE Core in Cell Broadband EngineTM
Network Processing on an SPE Core in Cell Broadband EngineTMNetwork Processing on an SPE Core in Cell Broadband EngineTM
Network Processing on an SPE Core in Cell Broadband EngineTM
 
Flexible and Scalable Domain-Specific Architectures
Flexible and Scalable Domain-Specific ArchitecturesFlexible and Scalable Domain-Specific Architectures
Flexible and Scalable Domain-Specific Architectures
 
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011
"FCoE vs. iSCSI - Making the Choice" from Interop Las Vegas 2011
 
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)
Dataplane networking acceleration with OpenDataplane / Максим Уваров (Linaro)
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task Computing
 
High Performance Communication for Oracle using InfiniBand
High Performance Communication for Oracle using InfiniBandHigh Performance Communication for Oracle using InfiniBand
High Performance Communication for Oracle using InfiniBand
 
NWU and HPC
NWU and HPCNWU and HPC
NWU and HPC
 
ETHERNET PACKET PROCESSOR FOR SOC APPLICATION
ETHERNET PACKET PROCESSOR FOR SOC APPLICATIONETHERNET PACKET PROCESSOR FOR SOC APPLICATION
ETHERNET PACKET PROCESSOR FOR SOC APPLICATION
 
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...
Disaggregation a Primer: Optimizing design for Edge Cloud & Bare Metal applic...
 
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linux
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running LinuxLinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linux
LinuxCon2009: 10Gbit/s Bi-Directional Routing on standard hardware running Linux
 
Converged Networks: FCoE, iSCSI and the Future of Storage Networking
Converged Networks: FCoE, iSCSI and the Future of Storage NetworkingConverged Networks: FCoE, iSCSI and the Future of Storage Networking
Converged Networks: FCoE, iSCSI and the Future of Storage Networking
 
Mp So C 18 Apr
Mp So C 18 AprMp So C 18 Apr
Mp So C 18 Apr
 

Mais de inside-BigData.com

Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networksinside-BigData.com
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networksinside-BigData.com
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecastsinside-BigData.com
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Updateinside-BigData.com
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19inside-BigData.com
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODinside-BigData.com
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Erainside-BigData.com
 
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...inside-BigData.com
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolversinside-BigData.com
 
Scientific Applications and Heterogeneous Architectures
Scientific Applications and Heterogeneous ArchitecturesScientific Applications and Heterogeneous Architectures
Scientific Applications and Heterogeneous Architecturesinside-BigData.com
 
SW/HW co-design for near-term quantum computing
SW/HW co-design for near-term quantum computingSW/HW co-design for near-term quantum computing
SW/HW co-design for near-term quantum computinginside-BigData.com
 

Mais de inside-BigData.com (20)

Major Market Shifts in IT
Major Market Shifts in ITMajor Market Shifts in IT
Major Market Shifts in IT
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networks
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networks
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecasts
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Update
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Acceleration
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Era
 
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...
Efficient Model Selection for Deep Neural Networks on Massively Parallel Proc...
 
Data Parallel Deep Learning
Data Parallel Deep LearningData Parallel Deep Learning
Data Parallel Deep Learning
 
Making Supernovae with Jets
Making Supernovae with JetsMaking Supernovae with Jets
Making Supernovae with Jets
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolvers
 
Scientific Applications and Heterogeneous Architectures
Scientific Applications and Heterogeneous ArchitecturesScientific Applications and Heterogeneous Architectures
Scientific Applications and Heterogeneous Architectures
 
SW/HW co-design for near-term quantum computing
SW/HW co-design for near-term quantum computingSW/HW co-design for near-term quantum computing
SW/HW co-design for near-term quantum computing
 
FPGAs and Machine Learning
FPGAs and Machine LearningFPGAs and Machine Learning
FPGAs and Machine Learning
 

Último

How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity PlanDatabarracks
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningLars Bell
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 

Último (20)

How to write a Business Continuity Plan
How to write a Business Continuity PlanHow to write a Business Continuity Plan
How to write a Business Continuity Plan
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine Tuning
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 

Overview of HPC Interconnects

  • 1. Interconnects Ken Raffenetti Software Development Specialist Programming Models and Runtime Systems Group Mathematics and Computer Science Division Argonne National Laboratory
  • 2. U.S. DOE Potential System Architecture Targets ATPESC (07/29/2019) System attributes 2010 2018-2019 2021-2022 System peak 2 Peta 150-200 Petaflop/sec 1 Exaflop/sec Power 6 MW 15 MW 20 MW System memory 0.3 PB 5 PB 32-64 PB Node performance 125 GF 3 TF 30 TF 10 TF 100 TF Node memory BW 25 GB/s 0.1TB/sec 1 TB/sec 0.4TB/sec 4 TB/sec Node concurrency 12 O(100) O(1,000) O(1,000) O(10,000) System size (nodes) 18,700 50,000 5,000 100,000 10,000 Total Node Interconnect BW 1.5 GB/s 20 GB/sec 200GB/sec MTTI days O(1day) O(1 day) Past production Current generation (e.g., CORAL) Exascale Goals [Includes modifications to the DOE Exascale report]
  • 3. General Trends in System Architecture § Number of nodes is increasing, but at a moderate pace § Number of cores/threads on a node is increasing rapidly § Each core is not increasing in speed (clock frequency) § What does this mean for networks? – More cores will drive the network – More sharing of the network infrastructure – The aggregate amount of communication from each node will increase moderately, but will be divided into many smaller messages – A single core may not be able fully saturate the NIC ATPESC (07/29/2019)
  • 4. A Simplified Network Architecture § Hardware components – Processing cores and memory subsystem – I/O bus or links – Network adapters/switches § Software components – Communication stack § Balanced approach required to maximize user-perceived network performance ATPESC (07/29/2019) P0 Core0 Core1 Core2 Core3 P1 Core0 Core1 Core2 Core3 Memory MemoryI / O B u s Network Adapter Network Switch Processing Bottlenecks I/O Interface Bottlenecks Network End-host Bottlenecks Network Fabric Bottlenecks
  • 5. Agenda ATPESC (07/29/2019) Network Adapters Network Topologies Network/Processor/Memory Interactions
  • 6. Bottlenecks on Traditional Network Adapters § Network speeds plateaued at around 1Gbps – Features provided were limited – Commodity networks were not considered scalable enough for very large-scale systems ATPESC (07/29/2019) P0 Core0 Core1 Core2 Core3 P1 Core0 Core1 Core2 Core3 Memory MemoryI / O B u s Network Adapter Network Switch Network Bottlenecks Ethernet (1979 - ) 10 Mbit/sec Fast Ethernet (1993 -) 100 Mbit/sec Gigabit Ethernet (1995 -) 1000 Mbit /sec ATM (1995 -) 155/622/1024 Mbit/sec Myrinet (1993 -) 1 Gbit/sec Fibre Channel (1994 -) 1 Gbit/sec
  • 7. End-host Network Interface Speeds § HPC network technologies provide high bandwidth links – InfiniBand EDR gives 100 Gbps per network link • Will continue to increase (HDR 200 Gbps, etc.) – Multiple network links becoming a common place • ORNL Summit and LLNL Sierra machines, Japanese Post T2K machine • Torus style or other multi-dimensional networks § End-host peak network bandwidth is “mostly” no longer considered a major limitation § Network latency still an issue – That’s a harder problem to solve – limited by physics, not technology • There is some room to improve it in current technology (trimming the fat) • Significant effort in making systems denser so as to reduce network latency § Other important metrics: message rate, congestion, … ATPESC (07/29/2019)
  • 8. Simple Network Architecture (past systems) § Processor, memory, network are all decoupled ATPESC (07/29/2019) P0 Core0 Core1 Core2 Core3 P1 Core0 Core1 Core2 Core3 Memory MemoryI / O B u s Network Adapter Network Switch
  • 9. Integrated Memory Controllers (current systems) § In the past 10 years or so, memory controllers have been integrated on to the processor § Primary purpose was scalable memory bandwidth (NUMA) § Also helps network communication – Data transfer to/from network requires coordination with caches § Several network I/O technologies exist – PCIe, HTX, NVLink – Expected to provide higher bandwidth than what network links will have ATPESC (07/29/2019) P1 Core0 Core1 Core2 Core3 Memory Memory I / O B u s Network Adapter Network Switch Memory Controller P0 Core0 Core1 Core2 Core3 Memory Controller
  • 10. Integrated Networks (future systems?) ATPESC (07/29/2019) Network Switch P0 Core0 Core1 Core2 Core3 Memory Controller Network Interface Off-chip Memory § Several vendors are considering processor-integrated network adapters § May improve network bandwidth – Unclear if the I/O bus would be a bottleneck § Improves network latencies – Control messages between the processor, network, and memory are now on-chip § Improves network functionality – Communication is a first-class citizen and better integrated with processor features – E.g., network atomic operations can be atomic with respect to processor atomics In-package Memory P1 Core0 Core1 Core2 Core3 Memory Controller Network Interface Off-chip Memory In-package Memory
  • 11. Processing Bottlenecks in Traditional Protocols § Ex: TCP/IP, UDP/IP § Generic architecture for all networks § Host processor handles almost all aspects of communication – Data buffering (copies on sender and receiver) – Data integrity (checksum) – Routing aspects (IP routing) § Signaling between different layers – Hardware interrupt on packet arrival or transmission – Software signals between different layers to handle protocol processing in different priority levels ATPESC (07/29/2019) P0 Core0 Core1 Core2 Core3 P1 Core0 Core1 Core2 Core3 Memory MemoryI / O B u s Network Adapter Network Switch Processing Bottlenecks
  • 12. Network Protocol Stacks: The Offload Era § Modern networks are spending more and more network real-estate on offloading various communication features on hardware § Network and transport layers are hardware offloaded for most modern HPC networks – Reliability (retransmissions, CRC checks), packetization – OS-based memory registration, and user-level data transmission ATPESC (07/29/2019)
  • 13. Comparing Offloaded Network Stacks with Traditional Network Stacks ATPESC (07/29/2019) Application Layer MPI, PGAS, File Systems Transport Layer Low-level interface Reliable/unreliable protocols Link Layer Flow-control, Error Detection Physical Layer Hardware Offload Copper or Optical HTTP, FTP, MPI, File Systems Routing Physical Layer Link Layer Network Layer Transport Layer Application Layer Traditional Ethernet Sockets Interface TCP, UDP Flow-control and Error Detection Copper, Optical or Wireless Network Layer Routing Management tools DNS management tools
  • 14. Current State for Network APIs § A large number of network vendor specific APIs – InfiniBand verbs, Intel PSM2, IBM PAMI, Cray Gemini/DMAPP, … § Recent efforts to standardize these low-level communication APIs – Open Fabrics Interface (OFI) • Effort from Intel, CISCO, etc., to provide a unified low-level communication layer that exposes features provided by each network – Unified Communication X (UCX) • Effort from Mellanox, IBM, ORNL, etc., to provide a unified low-level communication layer that allows for efficient MPI and PGAS communication – Portals 4 • Effort from Sandia National Laboratory to provide a network hardware capability centric API ATPESC (07/29/2019)
  • 15. User-level Communication: Memory Registration ATPESC (07/29/2019) 1. Registration Request • Send virtual address and length 2. Kernel handles virtual->physical mapping and pins region into physical memory • Process cannot map memory that it does not own (security !) 3. Network adapter caches the virtual to physical mapping and issues a handle 4. Handle is returned to application Before we do any communication: All memory used for communication must be registered 1 3 4 Process Kernel Network 2
  • 16. User-level Communication: OS Bypass ATPESC (07/29/2019) Process Kernel Network User-level APIs allow direct interaction with network adapters § Contrast with traditional network APIs that trap down to the kernel § Eliminates heavyweight context switch § Memory registration caches allow for fast buffer re-use, further reducing dependence on the kernel
  • 17. Send/Receive Communication ATPESC (07/29/2019) Network Interface Memory Memory Network Interface Send Recv Memory Segment Send entry contains information about the send buffer (multiple non-contiguous segments) Processor Processor Send Recv Memory Segment Receive entry contains information on the receive buffer (multiple non-contiguous segments); Incoming messages have to be matched to a receive entry to know where to place Hardware ACK Memory Segment Memory Segment Memory Segment
  • 18. PUT/GET Communication ATPESC (07/29/2019) Network Interface Memory Memory Network Interface Send Recv Memory Segment Send entry contains information about the send buffer (multiple segments) and the receive buffer (single segment) Processor Processor Send Recv Memory Segment Hardware ACK Memory Segment Memory Segment
  • 19. Atomic Operations ATPESC (07/29/2019) Network Interface Memory Memory Network Interface Send Recv Memory Segment Send entry contains information about the send buffer and the receive buffer Processor Processor Send Recv Source Memory Segment OP Destination Memory Segment
  • 20. Network Protocol Stacks: Specialization § Increasing network specialization is the focus today – The next generation of networks plan to have further support for noncontiguous data movement, and multiple contexts for multithreaded architectures § Some networks, such as the Tofu network, Cray Aries and InfiniBand, are also offloading some MPI and PGAS features on to hardware – E.g., PUT/GET communication has hardware support – Increasing number of atomic operations being offloaded to hardware • Compare-and-swap, fetch-and-add, swap – Collective operations (NIC and switch support) • SHARP collectives – Hardware tag matching for MPI send/recv • Cray Seastar, Bull BXI, Mellanox Infiniband (ConnectX-5 and later) ATPESC (07/29/2019)
  • 21. Agenda ATPESC (07/29/2019) Network Adapters Network Topologies Network/Processor/Memory Interactions
  • 22. Traditional Network Topologies: Crossbar § A network topology describes how different network adapters and switches are interconnected with each other § The ideal network topology (for performance) is a crossbar – Alltoall connection between network adapters – Typically done on a single network ASIC – Current network crossbar ASICs go up to 64 ports; too expensive to scale to higher port counts – All communication is nonblocking ATPESC (07/29/2019)
  • 23. Traditional Network Topologies: Fat-tree § The most common topology for small and medium scale systems is a fat-tree – Nonblocking fat-tree switches available in abundance • Allows for pseudo nonblocking communication • Between all pairs of processes, there exists a completely nonblocking path, but not all paths are nonblocking – More scalable than crossbars, but the number of network links still increases super-linearly with node count • Can get very expensive with scale ATPESC (07/29/2019)
  • 24. Network Topology Trends § Modern topologies are moving towards more “scalability” (with respect to cost, not performance) § Blue Gene, Cray XE/XK, and K supercomputers use a torus-network; Cray XC uses dragonfly – Linear increase in the number of links/routers with system size – Any communication that is more than one hop away has a possibility of interference – congestion is not just possible, but common – Even when there is no congestion, such topologies increase the network diameter causing performance loss § Take-away: topological locality is important and its not going to get better ATPESC (07/29/2019)
  • 25. Network Congestion Behavior: IBM BG/P 0 500 1000 1500 2000 2500 3000 3500 1 2 4 8 16 32 64 128 256 512 1K 2K 4K 8K 16K 32K 64K 128K 256K 512K 1M Bandwidth(Mbps) Message Size (bytes) P2-P5 P3-P4 No overlap ATPESC (07/29/2019) P0 P1 P2 P3 P4 P5 P6 P7
  • 26. 2D Nearest Neighbor: Process Mapping (XYZ) ATPESC (07/29/2019) X-Axis Z-Axis Y-Axis
  • 27. Nearest Neighbor Performance: IBM BG/P ATPESC (07/29/2019) 0 100 200 300 400 500 600 700 800 900 2 4 8 16 32 64 128 256 512 1K ExecutionTime(us) Grid Partition (bytes) System Size : 16K Cores XYZT TXYZ ZYXT TZYX 0 500 1000 1500 2000 2500 2 4 8 16 32 64 128 256 512 1K ExecutionTime(us) Grid Partition (bytes) System Size : 128K Cores XYZT TXYZ ZYXT TZYX 2D Halo Exchange
  • 28. Agenda ATPESC (07/29/2019) Network Adapters Network Topologies Network/Processor/Memory Interactions
  • 29. Network Interactions with Memory/Cache § Most network interfaces understand and work with the cache coherence protocols available on modern systems – Users do not have to ensure that data is flushed from cache before communication – Network and memory controller hardware understand what state the data is in and communicate appropriately ATPESC (07/29/2019)
  • 30. Send-side Network Communication ATPESC (07/29/2019) North BridgeCPU NIC FSB Memory Bus I/OBus Appln. Buffer L3 $ Memory Appln. Buffer
  • 31. Receive-side Network Communication ATPESC (07/29/2019) North BridgeCPU NIC FSB Memory Bus I/OBus Appln. Buffer L3 $ Memory Appln. Buffer
  • 32. Network/Processor Interoperation Trends § Direct cache injection – Most current networks inject data into memory • If data is in cache, they flush cache and then inject to memory – Some networks are investigating direct cache injection • Data can be injected directly into the last-level cache • Can be tricky since it can cause cache pollution if the incoming data is not used immediately § Atomic operations – Current network atomic operations are only atomic with respect to other network operations and not with respect to processor atomics • E.g., network fetch-and-add and processor fetch-and-add might corrupt each other’s data – With network/processor integration, this is expected to be fixed ATPESC (07/29/2019)
  • 33. Network Interactions with Accelerators § PCI Express peer-to-peer capabilities enables network adapters to directly access third-party devices – Coordination between network adapter and accelerator (GPUs, FPGAs, …) – Data does not need to be copied into to/from buffers when going over the network – GPUDirect RDMA for NVIDIA GPUs is one example ATPESC (07/29/2019)
  • 34. Summary § These are interesting times for all components in the overall system architecture: processor, memory, interconnect – And interesting times for computational science on these systems § Interconnect technology is rapidly advancing – More hardware integration is the key to removing bottlenecks and improve functionality • Processor/memory/network integration is already in progress and will continue for the foreseeable future – Offload technologies continue to evolve as we move more functionality to the network hardware – Network topologies are becoming more “shared” (cost saving) ATPESC (07/29/2019)