SlideShare uma empresa Scribd logo
1 de 43
ISC 2014
Dr Paul Calleja
Director Cambridge HPC Service
ISC 2014
• Introduction to Cambridge HPCS
• Overview of the SKA project and the SKA SDP consortium
• Two SDP work pages
• SDP Open Architecture Lab
• SDP operations
Overview
ISC 2014
Cambridge University
• The University of Cambridge is a world leading teaching & research
institution, consistently ranked within the top 3 Universities world wide
• Annual income of £1200M - 40% is research related - one of the largest
R&D budgets within the UK HE sector
• 17000 students, 9,000 staff
• Cambridge is a major technology centre
– 1535 technology companies in surrounding science parks
– £12B annual revenue
– 53000 staff
• The HPCS has a mandate to provide HPC services to both the University
and wider technology company community
ISC 2014
Four domains of activity
Commodity
HPC Centre of
Excellence
Promoting
uptake of HPC
by UK Industry
Driving
Discovery
Advancing
development and
application of HPC
HPC
R& D
ISC 2014
• Over 1000 registered users from 35 departments
• 856 Dell Servers - 450 TF sustained DP performance
• 128 node Westmere (1536 cores) (16 TF)
• 600 node (9600 core) full non blocking Mellanox FDR IB 2,6 GHz sandy bridge
(200 TF) one of the fastest Intel clusters in he UK
• SKA GPU test bed -128 node 256 card NVIDIA K20 GPU
• Fastest GPU system in UK 250 TF
• Designed for maximum I/O throughput and message rate
• Full non blocking Dual rail Mellanox FDR Connect IB
• Full GPUDirect functionality
• Design for maximum energy efficiency
• 2 in Green500 (Nov 2013)
• Most efficient air cooled supercomputer in the world
• 4 PB storage – Lustre parallel file system 50GB/s
Cambridge HPC vital statistics
ISC 2014
CORE – Industrial HPC service & consultancy
ISC 2014
Dell | Cambridge HPC Solution Centre
• The Solution Centre is a Dell Cambridge joint funded HPC centre of
excellence, provide leading edge commodity open source HPC solutions.
ISC 2014
DiRAC national HPC service
ISC 2014
• Cambridge were the first European NVIDIA CUDA COE
• Cambridge has had first large scale GPU cluster in UK for the last four years
• Key technology demonstrator for SKA
• Strong CUDA development skill within HPCS
• New large GPU system– largest GPU system in UK one of the most energy
efficient supercomputers in the world – built to push parallel scalability of
GPU clusters by deploying the best network possible and combining with
GPUDirect
NVIDIA CCOE
ISC 2014
• The HPCS is providing all the data storage and computational recourse for a
major new genomics study
• The study involves the gene sequencing of 20,000 decease patients from
the UK
• This is a major data throughput workload with high data security issues
• It requires building designing an efficient data throughput pipeline in terms of
hardware and software.
20K genome project
ISC 2014
• 5 year research project with JLR
• Drive capability in simulation & data mining
• HPC design, implementation and operation best practice
JLR R&D
ISC 2014
SA CHPC collaboration
• HPCS has a long term strategic
partnership with CHPC
• HPCS has been working closely
with CHPC for last 6 years
• Technology strategy, system design
procurement
• HPC system stack development
• SKA platform development
ISC 2014
• Next generation radio telescope
• 100 x more sensitive
• 1000000 X faster
• 5 square km of dish over 3000 km
• The next big science project
• Currently the worlds most ambitious IT
Project
• First real exascale ready application
• Largest global big-data challenge
Square Kilometre Array - SKA
ISC 2014
SKA location
• Needs a radio-quiet site
• Very low population density
• Large amount of space
• Two sites:
• Western Australia
• Karoo Desert RSA
A Continental sized RadioA Continental sized Radio
TelescopeTelescope
ISC 2014
SKA – Key scientific drivers
Cradle of lifeCosmic Magnetism
Evolution of galaxies
Pulsar survey
gravity waves
Exploring the
dark ages
ISC 2014
But……
Most importantly the SKA will
investigate phenomena
we have not even imagined yet
Most importantly the SKA will
investigate phenomena
we have not even imagined yet
ISC 2014
SKA timeline
2022 Operations SKA1 10% 2025: Operations SKA2 100%
2023-2027 Construction of Full SKA, SKA2 €2 B
2017-2022 10% SKA construction, SKA1 €650M
2012 Site selection
2012 - 2016 Pre-Construction: 1 yr Detailed design €90M
PEP 3 yr Production Readiness
2008 - 2012 System design and refinement of specification
2000 - 2007 Initial concepts stage
1995 - 2000 Preliminary ideas and R&D
ISC 2014
SKA project structure
SKA BoardSKA Board
Director GeneralDirector General
Work Package
Consortium 1
Work Package
Consortium 1
Work Package
Consortium n
Work Package
Consortium n
Advisory Committees
(Science, Engineering,
Finance, Funding …)
Advisory Committees
(Science, Engineering,
Finance, Funding …)
……
Project Office
(OSKAO)
Project Office
(OSKAO)
Locally funded
ISC 2014
Work package breakdown
UK (lead), AU (CSIRO…), NL (ASTRON…)
South Africa SKA, Industry (Intel, IBM…)
UK (lead), AU (CSIRO…), NL (ASTRON…)
South Africa SKA, Industry (Intel, IBM…)
1. System
2. Science
3. Maintenance and support /Operations Plan
4. Site preparation
5. Dishes
6. Aperture arrays
7. Signal transport
8. Data networks
9. Signal processing
10. Science Data Processor
11. Monitor and Control
12. Power
SPO
ISC 2014
SDP = Streaming data processor challenge
• The SDP consortium led by Paul Alexander University of Cambridge
• 3 year design phase has now started (as of November 2013)
• To deliver SKA ICT infrastructure need a strong multi-disciplinary team
• Radio astronomy expertise
• HPC expertise (scalable software implementations; management)
• HPC hardware (heterogeneous processors; interconnects; storage)
• Delivery of data to users (cloud; UI …)
• Building a broad global consortium:
• 11 countries: UK, USA, AUS, NZ, Canada, NL, Germany, China, France,
Spain, South Korea
• Radio astronomy observatories; HPC centres; Multi-national ICT companies;
sub-contractors
ISC 2014
SDP consortium members
Management Groupings Workshare (%)
University of Cambridge (Astrophysics & HPFCS) 9.15
Netherlands Institute for Radio Astronomy 9.25
International Centre for Radio Astronomy Research 8.35
SKA South Africa / CHPC 8.15
STFC Laboratories 4.05
Non-Imaging Processing Team 6.95
University of Manchester
Max-Planck-Institut für Radioastronomie
University of Oxford (Physics)
University of Oxford (OeRC) 4.85
Chinese Universities Collaboration 5.85
New Zealand Universities Collaboration 3.55
Canadian Collaboration 13.65
Forschungszentrum Jülich 2.95
Centre for High Performance Computing South Africa 3.95
iVEC Australia (Pawsey) 1.85
Centro Nacional de Supercomputación 2.25
Fundación Centro de Supercomputación de Castilla y León 1.85
Instituto de Telecomunicações 3.95
University of Southampton 2.35
University College London 2.35
University of Melbourne 1.85
French Universities Collaboration 1.85
Universidad de Chile 1.85
ISC 2014
SDP –strong industrial partnership
• Discussions under way with
• DelI, NVIDIA, Intel, HP IBM, SGI, l, ARM, Microsoft Research
• Xyratex, Mellanox, Cray, DDN
• NAG, Cambridge Consultants, Parallel Scientific
• Amazon, Bull, AMD, Altera, Solar flare, Geomerics, Samsung, CISCO
• Apologies to those I’ve forgotten to list
ISC 2014
SKA data rates
..
Sparse AA
Dense AA
..
Central Processing Facility - CPF
User interface
via Internet
...
To 250 AA Stations
DSP
...
DSP
To 1200 Dishes
...15m Dishes
16 Tb/s
10 Gb/s
Data
Time
Control
70-450 MHz
Wide FoV
0.4-1.4 GHz
Wide FoV
1.2-10 GHz
WB-Single
Pixel feeds
Tile &
Station
Processing
Optical
Data
links
...
AA slice
...
AA slice
...
AA slice
...Dish&AA+DishCorrelation
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
ProcessorBuffer
Dataswitch
......
Data
Archive
Science
Processors
Tb/s Gb/s Gb/s
...
...
Time
Standard
ImagingProcessors
Control Processors & User interface
Pb/s
Correlator UV Processors Image formation Archive
Aperture Array
Station
16 Tb/s 4 Pb/s
24 Tb/s
20 Gb/s
20 Gb/s
1000Tb/s
ISC 2014
SKA conceptual data flow
ISC 2014
Science data processor pipeline
10 Pflop
1 Eflop
100 Pflop
Software
complexity
10 Tb/s 200 Pflop
10 Eflop
…
Incoming
Data from
collectors
Switch
Bufferstore
Switch
Bufferstore
BulkStoreBulkStore
Correlator
Beamformer
UV
Processor
Imaging:
Non-Imaging:
Corner
Turning
Course
Delays
Fine F-step/
Correlation
Visibility
Steering
Observation
Buffer
Gridding
Visibilities Imaging
Image
Storage
Corner
Turning
Course
Delays
Beamforming/
De-dispersion
Beam
Steering
Observation
Buffer
Time-series
Searching
Search
analysis
Object/timing
Storage
HPCscienceHPCscience
processingprocessing
Image
Processor
1000Tb/s 1 Eflop10 EB/ySKA 2
SKA 1 1 EB/y50 PB
2.5 EB
ISC 2014
SDP processing rack – feasibility model
Host processor
Multi-core X86
M-Core->10TFLOP/s
M-Core->10TFLOP/s
To rack
switches
Disk 1
≥1TB
56Gb/s
PCI Bus
Disk 2
≥1TB
Disk 3
≥1TB
Disk 4
≥1TB
Processing blade 1
Processing blade 2
Processing blade 3
Processing blade 4
Processing blade 5
Processing blade 6
Processing blade 7
Processing blade 8
Processing blade 9
Processing blade 10
Processing blade 11
Processing blade 12
Processing blade 13
Processing blade 14
Processing blade 15
Processing blade 16
Processing blade 17
Processing blade 18
Processing blade 19
Processing blade 20
Leaf Switch-1 56Gb/s
Leaf Switch-2 56Gb/s
42U Rack
Processing Blade:
GGPU,
MIC,…?
GGPU,
MIC,…?
Blade SpecificationBlade Specification
ISC 2014
SKA feasibility model
…
…
…
…
…
…
…
…
…
…
……
AA-low
Data 1
1
280
AA-low
Data 2
1
280
Dishes
Data 4
1 2 16
…
1 3 N
…
HPCHPC
BulkBulk
StoreStore
2
SwitchSwitch
Correlator/
UV processor
Further UV
processors
Imaging
Processor
Corner Turner
switches
56Gb/s
each
…
…
AA-low
Data 3
1
280
1
250
……
ISC 2014
SKA conceptual software stack
ISC 2014
• Create a critical mass of HPC & astronomy knowledge combined with HPC equipment and lab staff to
produce a shared resource to drive SKA system development studies and SDP prototyping – Strong
coordination of all activities with COMP
• Provide a coordinated engagement mechanism with Industry to drive SKA platform development studies
Dedicated OAL project manager
• Benchmarking
• Perform standardised benchmarking across range different vendor solutions
• Undertake a consistent analysis of benchmark systems and report into COMP
• Manage a number of industry contracts driving system studies
• Low level software RAID / Lustre performance testing
• Large scale archives
• Software defined networking
• Openstack in HPC environment
• SLUM as a telescope scheduler
Open Architecture Lab function
ISC 2014
• Build prototype systems under direction from COMP
• Undertaken system studies directed from COMP to investigate particular
system aspects Dedicated HPC engineer being hired to drive studies
• Act as managed Lab for COMP and wider SDP work packages., build
systems, perform technical studies, make systems accessible. Dedicated Lab
engineer to service LAB
Open Architecture Lab function
ISC 2014
• Emphasis on testing scalable components in terms of hardware and
software
• Key considerations in architectural studies will be:-
• Energy efficient architectures
• Scalable cost effective storage
• Interconnects
• Scalable system software
• Operations
Open Architecture Lab function
ISC 2014
•Act as interface with industry providing coordinated engagement with SDP
• Benchmark study papers
• Roadmap papers
• Discussion digests
•Industrial system study contracts
• Design papers
• Benchmark papers
•Build target test platform
• Design papers
• benchmark papers
•Managed lab providing a service to SDP consortium
• Service function is output
Open Architecture Lab – Outputs
ISC 2014
• Coordinated by Cambridge jointly run HPCS (Cambridge) and CHPC (SA)
– Dedicated PM
– Dedicated HPC engineer
– Dedicated lab engineer
• Collaborate and coordinate with distributed labs
• Cambridge
• CHPC
• Astron
• Julich
• Oxford
Open Architecture Lab – Organisation
ISC 2014
•Large scale systems
• 600 node (9600 core) full non blocking Mellanox FDR IB 2,6 GHz sandy bridge
(200 TF) one of the fastest Intel clusters in he UK
• SKA GPU test bed -128 node 256 card NVIDIA K20 GPU
• GPU test bed built extensive testing underway – good understanding of GPU –
GPU RDMA functionality - GPU focused engineer in place
• Good understanding of current best practise in energy efficient computing and
data build and design
•4 PB storage – Lustre parallel file system 50GB/s
Cambridge test beds
ISC 2014
• Small scale CPU test beds
• Phi – Installed last week – larger cluster being designed
• Arm – evaluating solutions
• Atom – evaluating solutions
• Agreed Intel Radio Astronomy IPPC 2 head count to be put in place
looking at PHI
• Storage test beds
• Lustre on commodity hardware – H/W RAID - Installed
• Lustre on commodity hardware – SW RAID – under test
• Lustre on proprietary hardware – discussions with vendors
• Lustre on ZFS – under test
• Ceph – under test
• Archive test bed –discussions with vendors
• Distributed file system flash accelerated - in design
• Strong storage test programme underway – head count of 4 in place
driving the programme
Cambridge test beds
ISC 2014
• Networking
• Use current production system as large scale test bed
• Dedicated equipment to be ordered
• IB
• Ethernet
• Software defined networks
• RDMA data transfers co-p to co-p
• Networking SOW being constructed – 1FTE to be hired
• Slurm test bed at Cambridge & CHPC
• Headcount at Cambridge CHPC
• Openstack test bed under construction –
• Openstack development in CHPC/HPCS
• SDP operations – OAL to feed into operation WP
• Data centre issues
Cambridge test beds
ISC 2014
• The SKA SDP compute facility will be at the time of deployment one of the
largest HPC systems in existence
• Operational management of large HPC systems is challenging at the best of
times - When HPC systems are housed in well established research centres
with good IT logistics and experienced Linux HPC staff
• The SKA SDP could be housed in a desert location with little surrounding IT
infrastructure, with poor IT logistics and little prior HPC history at the site
• Potential SKA SDP exascale systems are likely to consist of 100,000 nodes
occupy 800 cabinets and consume 20 MW. This is very large – around 5
times the physical size of Titan Cray at Oakridge national labs.
• The SKA SDP HPC operations will be very challenging but tractable
SKA Exascale computing in the desert
ISC 2014
• We can describe the operational aspects by functional element
Machine room requirements **
SDP data connectivity requirements
SDP workflow requirements
System service level requirements
System management software requirements**
Commissioning & acceptance test procedures
System administration procedure
User access procedures
Security procedure
Maintenance & logistical procedures **
Refresh procedure
System staffing & training procedures **
SKA HPC operations – functional elements
ISC 2014
• Machine room infrastructure for exascale HPC facilities is challenging
• 800 racks, 1600M squared
• 30MW IT load
• ~40 Kw of heat per rack
• Cooling efficiency and heat density management is vital
• Machine infrastructure at this scale is both costly and time consuming
• The power cost alone at todays cost is 10’s of millions (£) per year
• Desert location presents particular problems for data centre
• Hot ambient temperature - difficult for compressor less cooling
• Lack of water - difficult for compressor less cooling
• Very dry air - difficult for humidification
• Remote location - difficult for DC maintenance
Machine room requirements
ISC 2014
• System management software is the vital element in HPC operations
• System management software today does not scale to exascale
• Worldwide coordinated effort to develop system management software for
exascale in HPC community
• We are very interested in leveraging Openstack technologies from non HPC
communities
System management software
ISC 2014
• Current HPC technology MBTF for hardware and system software result in
failure rates of ~ 2 nodes per week on a cluster a ~600 nodes.
• It is expected that SKA exascale systems could contain ~100,000 nodes
• Thus expected failure rates of 300 nodes per week could be realistic
• During system commissioning this will be 3 or 4 X
• Fixing nodes quickly is vital otherwise the system will soon degrade into a
non functional state
• The manual engineering processes for fault detection and diagnosis on 600
will not scale to 100,000 nodes. This needs to be automated by the system
software layer
• Vendor hardware replacement logistics need to cope with high turn around
rates
Maintenance logistics
ISC 2014
• Providing functional staffing levels and experience at remote desert location
will be challenging
• Its hard enough finding good HPC staff to run small scale HPC systems in
Cambridge – finding orders of magnitude more staff to run much more
complicated systems in a remote desert location will be very Challenging
• Operational procedures using a combination of remote system
administration staff and DC smart hands will be needed.
• HPC training programmes need to be implemented to skill up way in
advance
Staffing levels and training
ISC 2014
Early Cambridge SKA solution - EDSAC 1
Maurice Wilkes

Mais conteúdo relacionado

Mais procurados

Mathematical bridges From Old to New
Mathematical bridges From Old to NewMathematical bridges From Old to New
Mathematical bridges From Old to NewMapR Technologies
 
Free Code Friday - Machine Learning with Apache Spark
Free Code Friday - Machine Learning with Apache SparkFree Code Friday - Machine Learning with Apache Spark
Free Code Friday - Machine Learning with Apache SparkMapR Technologies
 
CourboSpark: Decision Tree for Time-series on Spark
CourboSpark: Decision Tree for Time-series on SparkCourboSpark: Decision Tree for Time-series on Spark
CourboSpark: Decision Tree for Time-series on SparkDataWorks Summit
 
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...Codemotion
 
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...zionsaint
 
The Cambridge Research Computing Service
The Cambridge Research Computing ServiceThe Cambridge Research Computing Service
The Cambridge Research Computing Serviceinside-BigData.com
 
Real time-hadoop
Real time-hadoopReal time-hadoop
Real time-hadoopTed Dunning
 
Visualisation of Big Imaging Data
Visualisation of Big Imaging DataVisualisation of Big Imaging Data
Visualisation of Big Imaging DataSlava Kitaeff, PhD
 
Activeeon - Scale Beyond Limits
Activeeon - Scale Beyond LimitsActiveeon - Scale Beyond Limits
Activeeon - Scale Beyond LimitsActiveeon
 
Build a Time Series Application with Apache Spark and Apache HBase
Build a Time Series Application with Apache Spark and Apache  HBaseBuild a Time Series Application with Apache Spark and Apache  HBase
Build a Time Series Application with Apache Spark and Apache HBaseCarol McDonald
 
Dunning time-series-2015
Dunning time-series-2015Dunning time-series-2015
Dunning time-series-2015Ted Dunning
 
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...Carol McDonald
 
What is the past future tense of data?
What is the past future tense of data?What is the past future tense of data?
What is the past future tense of data?Ted Dunning
 
Designing HPC, Deep Learning, and Cloud Middleware for Exascale Systems
Designing HPC, Deep Learning, and Cloud Middleware for Exascale SystemsDesigning HPC, Deep Learning, and Cloud Middleware for Exascale Systems
Designing HPC, Deep Learning, and Cloud Middleware for Exascale Systemsinside-BigData.com
 
Where is Data Going? - RMDC Keynote
Where is Data Going? - RMDC KeynoteWhere is Data Going? - RMDC Keynote
Where is Data Going? - RMDC KeynoteTed Dunning
 

Mais procurados (20)

Mathematical bridges From Old to New
Mathematical bridges From Old to NewMathematical bridges From Old to New
Mathematical bridges From Old to New
 
Free Code Friday - Machine Learning with Apache Spark
Free Code Friday - Machine Learning with Apache SparkFree Code Friday - Machine Learning with Apache Spark
Free Code Friday - Machine Learning with Apache Spark
 
CourboSpark: Decision Tree for Time-series on Spark
CourboSpark: Decision Tree for Time-series on SparkCourboSpark: Decision Tree for Time-series on Spark
CourboSpark: Decision Tree for Time-series on Spark
 
Earth Science Data and Information System (ESDIS) Project Update
Earth Science Data and Information System (ESDIS) Project UpdateEarth Science Data and Information System (ESDIS) Project Update
Earth Science Data and Information System (ESDIS) Project Update
 
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...
Anomaly Detection in Telecom with Spark - Tugdual Grall - Codemotion Amsterda...
 
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
DARPA ERI Summit 2018: The End of Moore’s Law & Faster General Purpose Comput...
 
The Cambridge Research Computing Service
The Cambridge Research Computing ServiceThe Cambridge Research Computing Service
The Cambridge Research Computing Service
 
HDF OPeNDAP Project Update and Demo
HDF OPeNDAP Project Update and DemoHDF OPeNDAP Project Update and Demo
HDF OPeNDAP Project Update and Demo
 
Real time-hadoop
Real time-hadoopReal time-hadoop
Real time-hadoop
 
Visualisation of Big Imaging Data
Visualisation of Big Imaging DataVisualisation of Big Imaging Data
Visualisation of Big Imaging Data
 
Activeeon - Scale Beyond Limits
Activeeon - Scale Beyond LimitsActiveeon - Scale Beyond Limits
Activeeon - Scale Beyond Limits
 
Build a Time Series Application with Apache Spark and Apache HBase
Build a Time Series Application with Apache Spark and Apache  HBaseBuild a Time Series Application with Apache Spark and Apache  HBase
Build a Time Series Application with Apache Spark and Apache HBase
 
HDF & HDF-EOS Data & Support at NSIDC
HDF & HDF-EOS Data & Support at NSIDCHDF & HDF-EOS Data & Support at NSIDC
HDF & HDF-EOS Data & Support at NSIDC
 
Dunning time-series-2015
Dunning time-series-2015Dunning time-series-2015
Dunning time-series-2015
 
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...
Fast, Scalable, Streaming Applications with Spark Streaming, the Kafka API an...
 
What is the past future tense of data?
What is the past future tense of data?What is the past future tense of data?
What is the past future tense of data?
 
Designing HPC, Deep Learning, and Cloud Middleware for Exascale Systems
Designing HPC, Deep Learning, and Cloud Middleware for Exascale SystemsDesigning HPC, Deep Learning, and Cloud Middleware for Exascale Systems
Designing HPC, Deep Learning, and Cloud Middleware for Exascale Systems
 
Apache Spark Overview
Apache Spark OverviewApache Spark Overview
Apache Spark Overview
 
Using IDL with Suomi NPP VIIRS Data
Using IDL with Suomi NPP VIIRS DataUsing IDL with Suomi NPP VIIRS Data
Using IDL with Suomi NPP VIIRS Data
 
Where is Data Going? - RMDC Keynote
Where is Data Going? - RMDC KeynoteWhere is Data Going? - RMDC Keynote
Where is Data Going? - RMDC Keynote
 

Semelhante a The SKA Project - The World's Largest Streaming Data Processor

Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...confluent
 
Archiving data from Durham to RAL using the File Transfer Service (FTS)
Archiving data from Durham to RAL using the File Transfer Service (FTS)Archiving data from Durham to RAL using the File Transfer Service (FTS)
Archiving data from Durham to RAL using the File Transfer Service (FTS)Jisc
 
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)  NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP) Christian Esteve Rothenberg
 
Presentation dr. fanaroff
Presentation dr. fanaroffPresentation dr. fanaroff
Presentation dr. fanaroffPeter Abwao
 
Cloud Standards in the Real World: Cloud Standards Testing for Developers
Cloud Standards in the Real World: Cloud Standards Testing for DevelopersCloud Standards in the Real World: Cloud Standards Testing for Developers
Cloud Standards in the Real World: Cloud Standards Testing for DevelopersAlan Sill
 
Data Infrastructure Development for SKA/Jasper Horrell
Data Infrastructure Development for SKA/Jasper HorrellData Infrastructure Development for SKA/Jasper Horrell
Data Infrastructure Development for SKA/Jasper HorrellAfrican Open Science Platform
 
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...Larry Smarr
 
Horizon 2020 ICT and Advanced Materials & Manufacturing
Horizon 2020 ICT and Advanced Materials & ManufacturingHorizon 2020 ICT and Advanced Materials & Manufacturing
Horizon 2020 ICT and Advanced Materials & ManufacturingInvest Northern Ireland
 
Shared services - the future of HPC and big data facilities for UK research
Shared services - the future of HPC and big data facilities for UK researchShared services - the future of HPC and big data facilities for UK research
Shared services - the future of HPC and big data facilities for UK researchMartin Hamilton
 
Pioneering and Democratizing Scalable HPC+AI at PSC
Pioneering and Democratizing Scalable HPC+AI at PSCPioneering and Democratizing Scalable HPC+AI at PSC
Pioneering and Democratizing Scalable HPC+AI at PSCinside-BigData.com
 
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...OPNFV
 
ClusterVision & Intel: Top500 class Computing at the University of Paderborn
ClusterVision & Intel: Top500 class Computing at the University of PaderbornClusterVision & Intel: Top500 class Computing at the University of Paderborn
ClusterVision & Intel: Top500 class Computing at the University of PaderbornIntel IT Center
 
Pathways for EOSC-hub and MaX collaboration
Pathways for EOSC-hub and MaX collaborationPathways for EOSC-hub and MaX collaboration
Pathways for EOSC-hub and MaX collaborationEOSC-hub project
 
Campus networking
Campus networkingCampus networking
Campus networkingJisc
 
Grid is Dead ? Nimrod on the Cloud
Grid is Dead ? Nimrod on the CloudGrid is Dead ? Nimrod on the Cloud
Grid is Dead ? Nimrod on the CloudAdianto Wibisono
 
Louise McCluskey, Kx Engineer at Kx Systems
Louise McCluskey, Kx Engineer at Kx SystemsLouise McCluskey, Kx Engineer at Kx Systems
Louise McCluskey, Kx Engineer at Kx SystemsDataconomy Media
 

Semelhante a The SKA Project - The World's Largest Streaming Data Processor (20)

EPCC MSc industry projects
EPCC MSc industry projectsEPCC MSc industry projects
EPCC MSc industry projects
 
UKTI Webinar Square Kilometre Array
UKTI Webinar Square Kilometre Array UKTI Webinar Square Kilometre Array
UKTI Webinar Square Kilometre Array
 
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
Enabling Insight to Support World-Class Supercomputing (Stefan Ceballos, Oak ...
 
Archiving data from Durham to RAL using the File Transfer Service (FTS)
Archiving data from Durham to RAL using the File Transfer Service (FTS)Archiving data from Durham to RAL using the File Transfer Service (FTS)
Archiving data from Durham to RAL using the File Transfer Service (FTS)
 
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)  NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)
NECOS Industrial Workshop lightning talk by Marcos Felipe Schwarz (RNP)
 
Presentation dr. fanaroff
Presentation dr. fanaroffPresentation dr. fanaroff
Presentation dr. fanaroff
 
Cloud Standards in the Real World: Cloud Standards Testing for Developers
Cloud Standards in the Real World: Cloud Standards Testing for DevelopersCloud Standards in the Real World: Cloud Standards Testing for Developers
Cloud Standards in the Real World: Cloud Standards Testing for Developers
 
Data Infrastructure Development for SKA/Jasper Horrell
Data Infrastructure Development for SKA/Jasper HorrellData Infrastructure Development for SKA/Jasper Horrell
Data Infrastructure Development for SKA/Jasper Horrell
 
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
 
Horizon 2020 ICT and Advanced Materials & Manufacturing
Horizon 2020 ICT and Advanced Materials & ManufacturingHorizon 2020 ICT and Advanced Materials & Manufacturing
Horizon 2020 ICT and Advanced Materials & Manufacturing
 
Shared services - the future of HPC and big data facilities for UK research
Shared services - the future of HPC and big data facilities for UK researchShared services - the future of HPC and big data facilities for UK research
Shared services - the future of HPC and big data facilities for UK research
 
Pioneering and Democratizing Scalable HPC+AI at PSC
Pioneering and Democratizing Scalable HPC+AI at PSCPioneering and Democratizing Scalable HPC+AI at PSC
Pioneering and Democratizing Scalable HPC+AI at PSC
 
Virtualization for HPC at NCI
Virtualization for HPC at NCIVirtualization for HPC at NCI
Virtualization for HPC at NCI
 
Super computers in_uk
Super computers in_ukSuper computers in_uk
Super computers in_uk
 
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...
Run OPNFV Danube on ODCC Scorpio Multi-node Server - Open Software on Open Ha...
 
ClusterVision & Intel: Top500 class Computing at the University of Paderborn
ClusterVision & Intel: Top500 class Computing at the University of PaderbornClusterVision & Intel: Top500 class Computing at the University of Paderborn
ClusterVision & Intel: Top500 class Computing at the University of Paderborn
 
Pathways for EOSC-hub and MaX collaboration
Pathways for EOSC-hub and MaX collaborationPathways for EOSC-hub and MaX collaboration
Pathways for EOSC-hub and MaX collaboration
 
Campus networking
Campus networkingCampus networking
Campus networking
 
Grid is Dead ? Nimrod on the Cloud
Grid is Dead ? Nimrod on the CloudGrid is Dead ? Nimrod on the Cloud
Grid is Dead ? Nimrod on the Cloud
 
Louise McCluskey, Kx Engineer at Kx Systems
Louise McCluskey, Kx Engineer at Kx SystemsLouise McCluskey, Kx Engineer at Kx Systems
Louise McCluskey, Kx Engineer at Kx Systems
 

Mais de inside-BigData.com

Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...inside-BigData.com
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networksinside-BigData.com
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...inside-BigData.com
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...inside-BigData.com
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networksinside-BigData.com
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoringinside-BigData.com
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecastsinside-BigData.com
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Updateinside-BigData.com
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19inside-BigData.com
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuninginside-BigData.com
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODinside-BigData.com
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Accelerationinside-BigData.com
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficientlyinside-BigData.com
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Erainside-BigData.com
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computinginside-BigData.com
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Clusterinside-BigData.com
 

Mais de inside-BigData.com (20)

Major Market Shifts in IT
Major Market Shifts in ITMajor Market Shifts in IT
Major Market Shifts in IT
 
Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...Preparing to program Aurora at Exascale - Early experiences and future direct...
Preparing to program Aurora at Exascale - Early experiences and future direct...
 
Transforming Private 5G Networks
Transforming Private 5G NetworksTransforming Private 5G Networks
Transforming Private 5G Networks
 
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...
 
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
How to Achieve High-Performance, Scalable and Distributed DNN Training on Mod...
 
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
Evolving Cyberinfrastructure, Democratizing Data, and Scaling AI to Catalyze ...
 
HPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural NetworksHPC Impact: EDA Telemetry Neural Networks
HPC Impact: EDA Telemetry Neural Networks
 
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean MonitoringBiohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
Biohybrid Robotic Jellyfish for Future Applications in Ocean Monitoring
 
Machine Learning for Weather Forecasts
Machine Learning for Weather ForecastsMachine Learning for Weather Forecasts
Machine Learning for Weather Forecasts
 
HPC AI Advisory Council Update
HPC AI Advisory Council UpdateHPC AI Advisory Council Update
HPC AI Advisory Council Update
 
Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19Fugaku Supercomputer joins fight against COVID-19
Fugaku Supercomputer joins fight against COVID-19
 
Energy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic TuningEnergy Efficient Computing using Dynamic Tuning
Energy Efficient Computing using Dynamic Tuning
 
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPODHPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
HPC at Scale Enabled by DDN A3i and NVIDIA SuperPOD
 
State of ARM-based HPC
State of ARM-based HPCState of ARM-based HPC
State of ARM-based HPC
 
Versal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud AccelerationVersal Premium ACAP for Network and Cloud Acceleration
Versal Premium ACAP for Network and Cloud Acceleration
 
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance EfficientlyZettar: Moving Massive Amounts of Data across Any Distance Efficiently
Zettar: Moving Massive Amounts of Data across Any Distance Efficiently
 
Scaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's EraScaling TCO in a Post Moore's Era
Scaling TCO in a Post Moore's Era
 
CUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computingCUDA-Python and RAPIDS for blazing fast scientific computing
CUDA-Python and RAPIDS for blazing fast scientific computing
 
Introducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi ClusterIntroducing HPC with a Raspberry Pi Cluster
Introducing HPC with a Raspberry Pi Cluster
 
Overview of HPC Interconnects
Overview of HPC InterconnectsOverview of HPC Interconnects
Overview of HPC Interconnects
 

Último

Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure servicePooja Nehwal
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...shyamraj55
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 

Último (20)

Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure serviceWhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
WhatsApp 9892124323 ✓Call Girls In Kalyan ( Mumbai ) secure service
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 

The SKA Project - The World's Largest Streaming Data Processor

  • 1. ISC 2014 Dr Paul Calleja Director Cambridge HPC Service
  • 2. ISC 2014 • Introduction to Cambridge HPCS • Overview of the SKA project and the SKA SDP consortium • Two SDP work pages • SDP Open Architecture Lab • SDP operations Overview
  • 3. ISC 2014 Cambridge University • The University of Cambridge is a world leading teaching & research institution, consistently ranked within the top 3 Universities world wide • Annual income of £1200M - 40% is research related - one of the largest R&D budgets within the UK HE sector • 17000 students, 9,000 staff • Cambridge is a major technology centre – 1535 technology companies in surrounding science parks – £12B annual revenue – 53000 staff • The HPCS has a mandate to provide HPC services to both the University and wider technology company community
  • 4. ISC 2014 Four domains of activity Commodity HPC Centre of Excellence Promoting uptake of HPC by UK Industry Driving Discovery Advancing development and application of HPC HPC R& D
  • 5. ISC 2014 • Over 1000 registered users from 35 departments • 856 Dell Servers - 450 TF sustained DP performance • 128 node Westmere (1536 cores) (16 TF) • 600 node (9600 core) full non blocking Mellanox FDR IB 2,6 GHz sandy bridge (200 TF) one of the fastest Intel clusters in he UK • SKA GPU test bed -128 node 256 card NVIDIA K20 GPU • Fastest GPU system in UK 250 TF • Designed for maximum I/O throughput and message rate • Full non blocking Dual rail Mellanox FDR Connect IB • Full GPUDirect functionality • Design for maximum energy efficiency • 2 in Green500 (Nov 2013) • Most efficient air cooled supercomputer in the world • 4 PB storage – Lustre parallel file system 50GB/s Cambridge HPC vital statistics
  • 6. ISC 2014 CORE – Industrial HPC service & consultancy
  • 7. ISC 2014 Dell | Cambridge HPC Solution Centre • The Solution Centre is a Dell Cambridge joint funded HPC centre of excellence, provide leading edge commodity open source HPC solutions.
  • 9. ISC 2014 • Cambridge were the first European NVIDIA CUDA COE • Cambridge has had first large scale GPU cluster in UK for the last four years • Key technology demonstrator for SKA • Strong CUDA development skill within HPCS • New large GPU system– largest GPU system in UK one of the most energy efficient supercomputers in the world – built to push parallel scalability of GPU clusters by deploying the best network possible and combining with GPUDirect NVIDIA CCOE
  • 10. ISC 2014 • The HPCS is providing all the data storage and computational recourse for a major new genomics study • The study involves the gene sequencing of 20,000 decease patients from the UK • This is a major data throughput workload with high data security issues • It requires building designing an efficient data throughput pipeline in terms of hardware and software. 20K genome project
  • 11. ISC 2014 • 5 year research project with JLR • Drive capability in simulation & data mining • HPC design, implementation and operation best practice JLR R&D
  • 12. ISC 2014 SA CHPC collaboration • HPCS has a long term strategic partnership with CHPC • HPCS has been working closely with CHPC for last 6 years • Technology strategy, system design procurement • HPC system stack development • SKA platform development
  • 13. ISC 2014 • Next generation radio telescope • 100 x more sensitive • 1000000 X faster • 5 square km of dish over 3000 km • The next big science project • Currently the worlds most ambitious IT Project • First real exascale ready application • Largest global big-data challenge Square Kilometre Array - SKA
  • 14. ISC 2014 SKA location • Needs a radio-quiet site • Very low population density • Large amount of space • Two sites: • Western Australia • Karoo Desert RSA A Continental sized RadioA Continental sized Radio TelescopeTelescope
  • 15. ISC 2014 SKA – Key scientific drivers Cradle of lifeCosmic Magnetism Evolution of galaxies Pulsar survey gravity waves Exploring the dark ages
  • 16. ISC 2014 But…… Most importantly the SKA will investigate phenomena we have not even imagined yet Most importantly the SKA will investigate phenomena we have not even imagined yet
  • 17. ISC 2014 SKA timeline 2022 Operations SKA1 10% 2025: Operations SKA2 100% 2023-2027 Construction of Full SKA, SKA2 €2 B 2017-2022 10% SKA construction, SKA1 €650M 2012 Site selection 2012 - 2016 Pre-Construction: 1 yr Detailed design €90M PEP 3 yr Production Readiness 2008 - 2012 System design and refinement of specification 2000 - 2007 Initial concepts stage 1995 - 2000 Preliminary ideas and R&D
  • 18. ISC 2014 SKA project structure SKA BoardSKA Board Director GeneralDirector General Work Package Consortium 1 Work Package Consortium 1 Work Package Consortium n Work Package Consortium n Advisory Committees (Science, Engineering, Finance, Funding …) Advisory Committees (Science, Engineering, Finance, Funding …) …… Project Office (OSKAO) Project Office (OSKAO) Locally funded
  • 19. ISC 2014 Work package breakdown UK (lead), AU (CSIRO…), NL (ASTRON…) South Africa SKA, Industry (Intel, IBM…) UK (lead), AU (CSIRO…), NL (ASTRON…) South Africa SKA, Industry (Intel, IBM…) 1. System 2. Science 3. Maintenance and support /Operations Plan 4. Site preparation 5. Dishes 6. Aperture arrays 7. Signal transport 8. Data networks 9. Signal processing 10. Science Data Processor 11. Monitor and Control 12. Power SPO
  • 20. ISC 2014 SDP = Streaming data processor challenge • The SDP consortium led by Paul Alexander University of Cambridge • 3 year design phase has now started (as of November 2013) • To deliver SKA ICT infrastructure need a strong multi-disciplinary team • Radio astronomy expertise • HPC expertise (scalable software implementations; management) • HPC hardware (heterogeneous processors; interconnects; storage) • Delivery of data to users (cloud; UI …) • Building a broad global consortium: • 11 countries: UK, USA, AUS, NZ, Canada, NL, Germany, China, France, Spain, South Korea • Radio astronomy observatories; HPC centres; Multi-national ICT companies; sub-contractors
  • 21. ISC 2014 SDP consortium members Management Groupings Workshare (%) University of Cambridge (Astrophysics & HPFCS) 9.15 Netherlands Institute for Radio Astronomy 9.25 International Centre for Radio Astronomy Research 8.35 SKA South Africa / CHPC 8.15 STFC Laboratories 4.05 Non-Imaging Processing Team 6.95 University of Manchester Max-Planck-Institut für Radioastronomie University of Oxford (Physics) University of Oxford (OeRC) 4.85 Chinese Universities Collaboration 5.85 New Zealand Universities Collaboration 3.55 Canadian Collaboration 13.65 Forschungszentrum Jülich 2.95 Centre for High Performance Computing South Africa 3.95 iVEC Australia (Pawsey) 1.85 Centro Nacional de Supercomputación 2.25 Fundación Centro de Supercomputación de Castilla y León 1.85 Instituto de Telecomunicações 3.95 University of Southampton 2.35 University College London 2.35 University of Melbourne 1.85 French Universities Collaboration 1.85 Universidad de Chile 1.85
  • 22. ISC 2014 SDP –strong industrial partnership • Discussions under way with • DelI, NVIDIA, Intel, HP IBM, SGI, l, ARM, Microsoft Research • Xyratex, Mellanox, Cray, DDN • NAG, Cambridge Consultants, Parallel Scientific • Amazon, Bull, AMD, Altera, Solar flare, Geomerics, Samsung, CISCO • Apologies to those I’ve forgotten to list
  • 23. ISC 2014 SKA data rates .. Sparse AA Dense AA .. Central Processing Facility - CPF User interface via Internet ... To 250 AA Stations DSP ... DSP To 1200 Dishes ...15m Dishes 16 Tb/s 10 Gb/s Data Time Control 70-450 MHz Wide FoV 0.4-1.4 GHz Wide FoV 1.2-10 GHz WB-Single Pixel feeds Tile & Station Processing Optical Data links ... AA slice ... AA slice ... AA slice ...Dish&AA+DishCorrelation ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer ProcessorBuffer Dataswitch ...... Data Archive Science Processors Tb/s Gb/s Gb/s ... ... Time Standard ImagingProcessors Control Processors & User interface Pb/s Correlator UV Processors Image formation Archive Aperture Array Station 16 Tb/s 4 Pb/s 24 Tb/s 20 Gb/s 20 Gb/s 1000Tb/s
  • 25. ISC 2014 Science data processor pipeline 10 Pflop 1 Eflop 100 Pflop Software complexity 10 Tb/s 200 Pflop 10 Eflop … Incoming Data from collectors Switch Bufferstore Switch Bufferstore BulkStoreBulkStore Correlator Beamformer UV Processor Imaging: Non-Imaging: Corner Turning Course Delays Fine F-step/ Correlation Visibility Steering Observation Buffer Gridding Visibilities Imaging Image Storage Corner Turning Course Delays Beamforming/ De-dispersion Beam Steering Observation Buffer Time-series Searching Search analysis Object/timing Storage HPCscienceHPCscience processingprocessing Image Processor 1000Tb/s 1 Eflop10 EB/ySKA 2 SKA 1 1 EB/y50 PB 2.5 EB
  • 26. ISC 2014 SDP processing rack – feasibility model Host processor Multi-core X86 M-Core->10TFLOP/s M-Core->10TFLOP/s To rack switches Disk 1 ≥1TB 56Gb/s PCI Bus Disk 2 ≥1TB Disk 3 ≥1TB Disk 4 ≥1TB Processing blade 1 Processing blade 2 Processing blade 3 Processing blade 4 Processing blade 5 Processing blade 6 Processing blade 7 Processing blade 8 Processing blade 9 Processing blade 10 Processing blade 11 Processing blade 12 Processing blade 13 Processing blade 14 Processing blade 15 Processing blade 16 Processing blade 17 Processing blade 18 Processing blade 19 Processing blade 20 Leaf Switch-1 56Gb/s Leaf Switch-2 56Gb/s 42U Rack Processing Blade: GGPU, MIC,…? GGPU, MIC,…? Blade SpecificationBlade Specification
  • 27. ISC 2014 SKA feasibility model … … … … … … … … … … …… AA-low Data 1 1 280 AA-low Data 2 1 280 Dishes Data 4 1 2 16 … 1 3 N … HPCHPC BulkBulk StoreStore 2 SwitchSwitch Correlator/ UV processor Further UV processors Imaging Processor Corner Turner switches 56Gb/s each … … AA-low Data 3 1 280 1 250 ……
  • 28. ISC 2014 SKA conceptual software stack
  • 29. ISC 2014 • Create a critical mass of HPC & astronomy knowledge combined with HPC equipment and lab staff to produce a shared resource to drive SKA system development studies and SDP prototyping – Strong coordination of all activities with COMP • Provide a coordinated engagement mechanism with Industry to drive SKA platform development studies Dedicated OAL project manager • Benchmarking • Perform standardised benchmarking across range different vendor solutions • Undertake a consistent analysis of benchmark systems and report into COMP • Manage a number of industry contracts driving system studies • Low level software RAID / Lustre performance testing • Large scale archives • Software defined networking • Openstack in HPC environment • SLUM as a telescope scheduler Open Architecture Lab function
  • 30. ISC 2014 • Build prototype systems under direction from COMP • Undertaken system studies directed from COMP to investigate particular system aspects Dedicated HPC engineer being hired to drive studies • Act as managed Lab for COMP and wider SDP work packages., build systems, perform technical studies, make systems accessible. Dedicated Lab engineer to service LAB Open Architecture Lab function
  • 31. ISC 2014 • Emphasis on testing scalable components in terms of hardware and software • Key considerations in architectural studies will be:- • Energy efficient architectures • Scalable cost effective storage • Interconnects • Scalable system software • Operations Open Architecture Lab function
  • 32. ISC 2014 •Act as interface with industry providing coordinated engagement with SDP • Benchmark study papers • Roadmap papers • Discussion digests •Industrial system study contracts • Design papers • Benchmark papers •Build target test platform • Design papers • benchmark papers •Managed lab providing a service to SDP consortium • Service function is output Open Architecture Lab – Outputs
  • 33. ISC 2014 • Coordinated by Cambridge jointly run HPCS (Cambridge) and CHPC (SA) – Dedicated PM – Dedicated HPC engineer – Dedicated lab engineer • Collaborate and coordinate with distributed labs • Cambridge • CHPC • Astron • Julich • Oxford Open Architecture Lab – Organisation
  • 34. ISC 2014 •Large scale systems • 600 node (9600 core) full non blocking Mellanox FDR IB 2,6 GHz sandy bridge (200 TF) one of the fastest Intel clusters in he UK • SKA GPU test bed -128 node 256 card NVIDIA K20 GPU • GPU test bed built extensive testing underway – good understanding of GPU – GPU RDMA functionality - GPU focused engineer in place • Good understanding of current best practise in energy efficient computing and data build and design •4 PB storage – Lustre parallel file system 50GB/s Cambridge test beds
  • 35. ISC 2014 • Small scale CPU test beds • Phi – Installed last week – larger cluster being designed • Arm – evaluating solutions • Atom – evaluating solutions • Agreed Intel Radio Astronomy IPPC 2 head count to be put in place looking at PHI • Storage test beds • Lustre on commodity hardware – H/W RAID - Installed • Lustre on commodity hardware – SW RAID – under test • Lustre on proprietary hardware – discussions with vendors • Lustre on ZFS – under test • Ceph – under test • Archive test bed –discussions with vendors • Distributed file system flash accelerated - in design • Strong storage test programme underway – head count of 4 in place driving the programme Cambridge test beds
  • 36. ISC 2014 • Networking • Use current production system as large scale test bed • Dedicated equipment to be ordered • IB • Ethernet • Software defined networks • RDMA data transfers co-p to co-p • Networking SOW being constructed – 1FTE to be hired • Slurm test bed at Cambridge & CHPC • Headcount at Cambridge CHPC • Openstack test bed under construction – • Openstack development in CHPC/HPCS • SDP operations – OAL to feed into operation WP • Data centre issues Cambridge test beds
  • 37. ISC 2014 • The SKA SDP compute facility will be at the time of deployment one of the largest HPC systems in existence • Operational management of large HPC systems is challenging at the best of times - When HPC systems are housed in well established research centres with good IT logistics and experienced Linux HPC staff • The SKA SDP could be housed in a desert location with little surrounding IT infrastructure, with poor IT logistics and little prior HPC history at the site • Potential SKA SDP exascale systems are likely to consist of 100,000 nodes occupy 800 cabinets and consume 20 MW. This is very large – around 5 times the physical size of Titan Cray at Oakridge national labs. • The SKA SDP HPC operations will be very challenging but tractable SKA Exascale computing in the desert
  • 38. ISC 2014 • We can describe the operational aspects by functional element Machine room requirements ** SDP data connectivity requirements SDP workflow requirements System service level requirements System management software requirements** Commissioning & acceptance test procedures System administration procedure User access procedures Security procedure Maintenance & logistical procedures ** Refresh procedure System staffing & training procedures ** SKA HPC operations – functional elements
  • 39. ISC 2014 • Machine room infrastructure for exascale HPC facilities is challenging • 800 racks, 1600M squared • 30MW IT load • ~40 Kw of heat per rack • Cooling efficiency and heat density management is vital • Machine infrastructure at this scale is both costly and time consuming • The power cost alone at todays cost is 10’s of millions (£) per year • Desert location presents particular problems for data centre • Hot ambient temperature - difficult for compressor less cooling • Lack of water - difficult for compressor less cooling • Very dry air - difficult for humidification • Remote location - difficult for DC maintenance Machine room requirements
  • 40. ISC 2014 • System management software is the vital element in HPC operations • System management software today does not scale to exascale • Worldwide coordinated effort to develop system management software for exascale in HPC community • We are very interested in leveraging Openstack technologies from non HPC communities System management software
  • 41. ISC 2014 • Current HPC technology MBTF for hardware and system software result in failure rates of ~ 2 nodes per week on a cluster a ~600 nodes. • It is expected that SKA exascale systems could contain ~100,000 nodes • Thus expected failure rates of 300 nodes per week could be realistic • During system commissioning this will be 3 or 4 X • Fixing nodes quickly is vital otherwise the system will soon degrade into a non functional state • The manual engineering processes for fault detection and diagnosis on 600 will not scale to 100,000 nodes. This needs to be automated by the system software layer • Vendor hardware replacement logistics need to cope with high turn around rates Maintenance logistics
  • 42. ISC 2014 • Providing functional staffing levels and experience at remote desert location will be challenging • Its hard enough finding good HPC staff to run small scale HPC systems in Cambridge – finding orders of magnitude more staff to run much more complicated systems in a remote desert location will be very Challenging • Operational procedures using a combination of remote system administration staff and DC smart hands will be needed. • HPC training programmes need to be implemented to skill up way in advance Staffing levels and training
  • 43. ISC 2014 Early Cambridge SKA solution - EDSAC 1 Maurice Wilkes