SlideShare uma empresa Scribd logo
1 de 31
Baixar para ler offline
Gary Robertson, LCA14, Macau
LCA14-506: Comparative analysis of
preemption vs preempt-rt
In this presentation we will try to illustrate the pro’s, con’s,
and latency characteristics of several Linux kernel
preemption models, and provide some guidance in
selecting an appropriate preemption model for a given
category of application.
Overview of Topics Presented
Questions we will address include:
• Which preemption model provides the best throughput?
• Which model offers the lowest average latencies?
• Which model offers the lowest maximum latencies?
• Which model offers the most predictable latencies?
• How do load conditions impact the respective latency
performance of the various models?
• What impact does CPU Frequency Scaling or CPU
sleep states have on latency performance?
• What is the best model for a given application type?
Overview of Topics - continued
Our intent is to show relative trends between the
preemption models under the same conditions… so the
data presented were gathered thusly:
• Each preemption model configuration was tested using
identical tests running on the same InSignal Arndale.
• Cyclictest was used for a run duration of two hours with
a single thread executing at a SCHED_FIFO priority of
80 to realistically represent scheduling latency for a real-
time process.
• A cyclictest run was done with no system load, then
another with an externally-applied ping flood, and
another with back-to-back executions of hackbench
running to represent maximum system loading.
Test Rationale and Methodology
Latency Impact of CPU Frequency Scaling
Only three Linux preemption models are really interesting
for anything other than desktop use:
• the Server preemption model provides optimal
throughput for applications where latencies are not an
issue
• the Low Latency Desktop preemption model provides
low average latencies for interactive and ‘soft real-time’
applications
• the Full RT preemption model provides the highest level
of latency determinism for ‘hard real-time’ applications
Tested Linux Preemption Models
Server Preemption Model Latencies
Cyclictest with no system load
CPU frequency scaling disabled
Minimum Latency: 16 usec
Average Latency: 24 usec
Most Frequent Latency: 24 usec
Maximum Latency: 572 usec
Standard Deviation: 1.211041
Almost all latencies between 20 usec and
28 usec
However, even at light loads, latencies
out to 572 usec were observed. This is a
consequence of all code paths through
the kernel being non-preemptible.
Server Preemption Model Latencies
Cyclictest with ping flood load
CPU frequency scaling disabled
Minimum Latency: 15 usec
Average Latency: 23 usec
Most Frequent Latency: 24 usec
Maximum Latency: 592 usec
Standard Deviation: 1.580778
Almost all latencies between 20 usec and
28 usec
Note, however, that much longer
latencies continue to be observed due to
lack of any design efforts to avoid them.
Also note that maximum latency is
already beginning to creep upwards.
Server Preemption Model Latencies
Cyclictest with hackbench load
CPU frequency scaling disabled
Minimum Latency: 17 usec
Average Latency: 150655 usec
Most Frequent Latency: 22 usec
Maximum Latency: 2587753 usec
Standard Deviation: 493977.9
The majority of latencies were between
21 usec and 25 usec, gradually tapering
off to single digit frequencies at 204
usec. Note the duration of the max
latency is 4000 times longer than under
no load!
Note also the much lower frequency
percentage for the peak occurrence.
This means a larger percentage of the
higher latencies were observed, and
illustrates the serious degradation of
latency determinism under load in a non-
preemptible kernel where latency was
not a primary design consideration.
Low Latency Desktop Model Latencies
Cyclictest with no system load
CPU frequency scaling disabled
Minimum Latency: 19 usec
Average Latency: 28 usec
Most Frequent Latency: 29 usec
Maximum Latency: 57 usec
Standard Deviation: 0.8698308
The majority of latencies were between
28 usec and 31 usec, quickly tapering off
to single digit frequencies at 42 usec.
Maximum latency was reduced tenfold
under light loads vs. the Server model.
This illustrates the significant
improvements in latency performance
under light loads with kernel preemption
enabled.
Low Latency Desktop Model Latencies
Cyclictest with ping flood
CPU frequency scaling disabled
Minimum Latency: 18 usec
Average Latency: 29 usec
Most Frequent Latency: 29 usec
Maximum Latency: 131 usec
Standard Deviation: 1.79573
The majority of latencies were between
28 usec and 32 usec, quickly tapering off
to single digit frequencies at 80 usec.
The reduced range of observed latencies
indicates improved latency performance
and predictability at moderate loads
versus the Server model. However, as
the next slide will show, latency
performance in this model degrades
seriously under heavy load, making Full
RT a better choice for latency
performance under heavy load
conditions.
Low Latency Desktop Model Latencies
Cyclictest with hackbench
CPU frequency scaling disabled
Minimum Latency: 19 usec
Average Latency: 370606 usec
Most Frequent Latency: 25 usec
Maximum Latency: 4122148 usec
Standard Deviation: 826092
The majority of latencies were between
24 usec and 26 usec, gradually tapering
off to single digit frequencies at 105
usec. Note that the max latency was
70,000 times longer than with this model
under no load!
Max latencies were nearly double that for
a Server model under heavy load, and
latency predictability is low. This
illustrates the combined impacts under
heavy load of increased context switches
without addressing priority inversion or
FIFO queueing disciplines.
Full RT Preemption Model Latencies
Cyclictest with no system load
CPU frequency scaling disabled
Minimum Latency: 19 usec
Average Latency: 29 usec
Most Frequent Latency: 29 usec
Maximum Latency: 53 usec
Standard Deviation: 1.031893
The majority of latencies were between
29 usec and 31 usec, quickly tapering off
to single digit frequencies at 50 usec.
Maximum latency was reduced tenfold
under light loads vs. the Server model.
This illustrates the significant
improvements in latency performance
under light loads with kernel preemption
enabled.
Under light load performance is very
similar to that of the Low Latency
Desktop model.
Full RT Preemption Model Latencies
Cyclictest with ping flood
CPU frequency scaling disabled
Minimum Latency: 19 usec
Average Latency: 29 usec
Most Frequent Latency: 30 usec
Maximum Latency: 59 usec
Standard Deviation: 2.698587
The majority of latencies were between
29 usec and 31 usec, quickly tapering off
to single digit frequencies at 53 usec.
The reduced range of observed latencies
indicates improved latency performance
and predictability at moderate loads
versus the Server model.
Note that even at moderate loads the
maximum latencies are less than half the
duration of those seen in the Low
Latency Desktop model.
Full RT Preemption Model Latencies
Cyclictest with hackbench load
CPU frequency scaling disabled
Minimum Latency: 21 usec
Average Latency: 29 usec
Most Frequent Latency: 25 usec
Maximum Latency: 156 usec
Standard Deviation: 7.69571
The majority of latencies were between
24 usec and 26 usec, with a second
group peaking between 43 and 44 usec,
and quickly tapering off to single digit
frequencies at 134 usec.
Latency performance under heavy load is
much better than in any of the other
preemption models. With threaded
interrupt handlers, priority inheritance
and priority-based queuing disciplines,
the real-time process is still able to meet
much tighter scheduling deadlines
despite heavy activity of other lower-
priority threads.
Comparative Latency Performance
Comparative Latency Performance
Comparative Latency Performance
• For applications in which throughput and not latencies
are the primary consideration, opt for the Server model
• If quality of service is important but missed latency
deadlines will not result in catastrophic failures, opt for
the Low Latency Desktop model and size the hardware
capacity to keep loading moderate
• For host environments for ‘zero overhead Linux’ (ODP
for example), Low Latency Desktop is a good choice
• If latencies must be consistent even under high load
conditions Full RT may be required
• For applications based on POSIX real-time scheduling
and priority-based preemption, use Full RT for best
results
What Preemption Model Is Best for Me?
The test scripts, data files, and graphs used to provide
reference data for this presentation may be accessed
online at the following URL:
http://people.linaro.org/~gary.robertson/LCA14
Data References
More about Linaro Connect: http://connect.linaro.org
More about Linaro: http://www.linaro.org/about/
More about Linaro engineering: http://www.linaro.org/engineering/
Linaro members: www.linaro.org/members
Preemption Model Characteristics
Appendix A
The Server preemption model lies at one extreme of the
latency vs. throughput continuum.
Pro’s include:
• Simplicity, maturity and robustness make this a very
reliable platform
• With no preemption the reduced number of context
switches minimizes system overhead and maximizes
overall throughput
Server Model Characteristics
Con’s include:
• The lack of preemption results in low average
latencies under low loads but much higher latencies
when the system is heavily loaded
• The latencies imposed by different execution paths
through the kernel result in a wide range of latency
durations and low latency determinism
Server Model Characteristics
The Low Latency Desktop preemption model holds the
middle ground in the latency vs. throughput continuum.
Pro’s include:
• Under low to moderate load, latency range and
predictability are significantly improved vs. the Server
model
• This preemption model is supported as part of the
mainstream kernel and tends to be less trouble-prone
than Full RT preemption
Low Latency Desktop Model Characteristics
Con’s include:
• The preemption of kernel operations and increased
number of context switches create increased
overhead and reduced performance relative to the
Server model
• The preemption of kernel operations results in higher
average latencies vs. the Server preemption model
• This preemption model does not perform as well
under heavy system loads as other models
Low Latency Desktop Model Characteristics
The following software-induced latency sources remain
problematic in the Low Latency Desktop preemption
model:
• Exceptions, software interrupts, and device service
request interrupts execute outside of scheduler control
• Most mutual exclusion locking primitives are subject to
priority inversion
• Shared resources use FIFO-based queueing disciplines,
meaning high-priority threads may have to wait behind
lower priority threads for access to the resources
These factors result in lower levels of latency determinism.
Low Latency Desktop Model - continued
The Full RT preemption model represents the latency-
centric end of the latency vs. throughput continuum. It
attempts to mitigate all the remaining software-induced
sources of latency.
• Handlers for exceptions, software interrupts, and
device service request interrupts are encapsulated
inside threads which are under scheduler control
• Priority inheritance is added for most mutual
exclusion locking primitives to prevent priority
inversions
• Shared resources use priority-based queueing
disciplines so that the highest-priority thread always
gets first access to the resources
Full RT Model Characteristics
The Full RT preemption model inevitably suffers from
reduced overall throughput as a consequence of its
efforts to maximize latency determinism:
• Schedulable ‘threaded’ ISRs result in the highest
levels of preemption and context switch overhead
• Priority inheritance involves iterative logic to
temporarily boost the priorities of all lock holders to
equal that of the highest-priority lock waiters. This
adds significant overhead to locking primitive code.
• Priority-based queueing requires sorting the queue
each time a new waiting thread is added
Full RT Model Characteristics - continued
Pro’s include:
• The most consistent and predictable latency
performance available in any preemption model
• The best support environment for creating priority-
based multi-layered applications
• The best hard real-time support available in a Linux
environment
Full RT Model Characteristics - continued
Con’s include:
• Full RT preemption is supported only with a
separately maintained kernel patch set
• The latest supported RT kernel version always lags
behind mainstream development
• Mainstream drivers, libraries, and applications may
not always function properly in the Full RT
environment
• Poorly designed or written real-time threads may
starve out threaded interrupt handlers
Full RT Model Characteristics - continued

Mais conteúdo relacionado

Mais procurados

DevOops - Lessons Learned from an OpenStack Network Architect
DevOops - Lessons Learned from an OpenStack Network ArchitectDevOops - Lessons Learned from an OpenStack Network Architect
DevOops - Lessons Learned from an OpenStack Network ArchitectJames Denton
 
Troubleshooting Common Network Related Issues with NetScaler
Troubleshooting Common Network Related Issues with NetScalerTroubleshooting Common Network Related Issues with NetScaler
Troubleshooting Common Network Related Issues with NetScalerDavid McGeough
 
Ensuring performance for real time packet processing in open stack white paper
Ensuring performance for real time packet processing in open stack white paperEnsuring performance for real time packet processing in open stack white paper
Ensuring performance for real time packet processing in open stack white paperhptoga
 
NetScaler TCP Performance Tuning
NetScaler TCP Performance TuningNetScaler TCP Performance Tuning
NetScaler TCP Performance TuningKevin Mason
 
Alibaba cloud benchmarking report ecs rds limton xavier
Alibaba cloud benchmarking report ecs  rds limton xavierAlibaba cloud benchmarking report ecs  rds limton xavier
Alibaba cloud benchmarking report ecs rds limton xavierLimton Xavier
 
Stevens Hfc2010
Stevens Hfc2010Stevens Hfc2010
Stevens Hfc2010jzw200
 
Training Slides: Basics 102: Introduction to Tungsten Clustering
Training Slides: Basics 102: Introduction to Tungsten ClusteringTraining Slides: Basics 102: Introduction to Tungsten Clustering
Training Slides: Basics 102: Introduction to Tungsten ClusteringContinuent
 
Network performance overview
Network  performance overviewNetwork  performance overview
Network performance overviewMy cp
 
Load Balancing In Distributed Computing
Load Balancing In Distributed ComputingLoad Balancing In Distributed Computing
Load Balancing In Distributed ComputingRicha Singh
 
Q2.12: Research Update on big.LITTLE MP Scheduling
Q2.12: Research Update on big.LITTLE MP SchedulingQ2.12: Research Update on big.LITTLE MP Scheduling
Q2.12: Research Update on big.LITTLE MP SchedulingLinaro
 
Non-Kafkaesque Apache Kafka - Yottabyte 2018
Non-Kafkaesque Apache Kafka - Yottabyte 2018Non-Kafkaesque Apache Kafka - Yottabyte 2018
Non-Kafkaesque Apache Kafka - Yottabyte 2018Otávio Carvalho
 
Netflix SRE perf meetup_slides
Netflix SRE perf meetup_slidesNetflix SRE perf meetup_slides
Netflix SRE perf meetup_slidesEd Hunter
 
OCP Server Memory Channel Testing DRAFT
OCP Server Memory Channel Testing DRAFTOCP Server Memory Channel Testing DRAFT
OCP Server Memory Channel Testing DRAFTBarbara Aichinger
 
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...The Linux Foundation
 
246174 LDAV 2016_Poster_FNL_hi-res(1)
246174 LDAV 2016_Poster_FNL_hi-res(1)246174 LDAV 2016_Poster_FNL_hi-res(1)
246174 LDAV 2016_Poster_FNL_hi-res(1)Jie Jiang
 
Improving Passive Packet Capture : Beyond Device Polling
Improving Passive Packet Capture : Beyond Device PollingImproving Passive Packet Capture : Beyond Device Polling
Improving Passive Packet Capture : Beyond Device PollingHargyo T. Nugroho
 
Inside the Volta GPU Architecture and CUDA 9
Inside the Volta GPU Architecture and CUDA 9Inside the Volta GPU Architecture and CUDA 9
Inside the Volta GPU Architecture and CUDA 9inside-BigData.com
 
Steps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesSteps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesAshwin Pawar
 

Mais procurados (20)

DevOops - Lessons Learned from an OpenStack Network Architect
DevOops - Lessons Learned from an OpenStack Network ArchitectDevOops - Lessons Learned from an OpenStack Network Architect
DevOops - Lessons Learned from an OpenStack Network Architect
 
Troubleshooting Common Network Related Issues with NetScaler
Troubleshooting Common Network Related Issues with NetScalerTroubleshooting Common Network Related Issues with NetScaler
Troubleshooting Common Network Related Issues with NetScaler
 
Ensuring performance for real time packet processing in open stack white paper
Ensuring performance for real time packet processing in open stack white paperEnsuring performance for real time packet processing in open stack white paper
Ensuring performance for real time packet processing in open stack white paper
 
NetScaler TCP Performance Tuning
NetScaler TCP Performance TuningNetScaler TCP Performance Tuning
NetScaler TCP Performance Tuning
 
Demystifying openvswitch
Demystifying openvswitchDemystifying openvswitch
Demystifying openvswitch
 
Alibaba cloud benchmarking report ecs rds limton xavier
Alibaba cloud benchmarking report ecs  rds limton xavierAlibaba cloud benchmarking report ecs  rds limton xavier
Alibaba cloud benchmarking report ecs rds limton xavier
 
Stevens Hfc2010
Stevens Hfc2010Stevens Hfc2010
Stevens Hfc2010
 
Training Slides: Basics 102: Introduction to Tungsten Clustering
Training Slides: Basics 102: Introduction to Tungsten ClusteringTraining Slides: Basics 102: Introduction to Tungsten Clustering
Training Slides: Basics 102: Introduction to Tungsten Clustering
 
Network performance overview
Network  performance overviewNetwork  performance overview
Network performance overview
 
Решения WANDL и NorthStar для операторов
Решения WANDL и NorthStar для операторовРешения WANDL и NorthStar для операторов
Решения WANDL и NorthStar для операторов
 
Load Balancing In Distributed Computing
Load Balancing In Distributed ComputingLoad Balancing In Distributed Computing
Load Balancing In Distributed Computing
 
Q2.12: Research Update on big.LITTLE MP Scheduling
Q2.12: Research Update on big.LITTLE MP SchedulingQ2.12: Research Update on big.LITTLE MP Scheduling
Q2.12: Research Update on big.LITTLE MP Scheduling
 
Non-Kafkaesque Apache Kafka - Yottabyte 2018
Non-Kafkaesque Apache Kafka - Yottabyte 2018Non-Kafkaesque Apache Kafka - Yottabyte 2018
Non-Kafkaesque Apache Kafka - Yottabyte 2018
 
Netflix SRE perf meetup_slides
Netflix SRE perf meetup_slidesNetflix SRE perf meetup_slides
Netflix SRE perf meetup_slides
 
OCP Server Memory Channel Testing DRAFT
OCP Server Memory Channel Testing DRAFTOCP Server Memory Channel Testing DRAFT
OCP Server Memory Channel Testing DRAFT
 
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
XPDS13: On Paravirualizing TCP - Congestion Control on Xen VMs - Luwei Cheng,...
 
246174 LDAV 2016_Poster_FNL_hi-res(1)
246174 LDAV 2016_Poster_FNL_hi-res(1)246174 LDAV 2016_Poster_FNL_hi-res(1)
246174 LDAV 2016_Poster_FNL_hi-res(1)
 
Improving Passive Packet Capture : Beyond Device Polling
Improving Passive Packet Capture : Beyond Device PollingImproving Passive Packet Capture : Beyond Device Polling
Improving Passive Packet Capture : Beyond Device Polling
 
Inside the Volta GPU Architecture and CUDA 9
Inside the Volta GPU Architecture and CUDA 9Inside the Volta GPU Architecture and CUDA 9
Inside the Volta GPU Architecture and CUDA 9
 
Steps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issuesSteps to identify ONTAP latency related issues
Steps to identify ONTAP latency related issues
 

Semelhante a LCA14: LCA14-506: Comparative analysis of preemption vs preempt-rt

Four Ways to Improve Linux Performance IEEE Webinar, R2.0
Four Ways to Improve Linux Performance IEEE Webinar, R2.0Four Ways to Improve Linux Performance IEEE Webinar, R2.0
Four Ways to Improve Linux Performance IEEE Webinar, R2.0Michael Christofferson
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCoburn Watson
 
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Amazon Web Services
 
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...Continuent
 
Configuration Optimization for Big Data Software
Configuration Optimization for Big Data SoftwareConfiguration Optimization for Big Data Software
Configuration Optimization for Big Data SoftwarePooyan Jamshidi
 
참여기관_발표자료-국민대학교 201301 정기회의
참여기관_발표자료-국민대학교 201301 정기회의참여기관_발표자료-국민대학교 201301 정기회의
참여기관_발표자료-국민대학교 201301 정기회의DzH QWuynh
 
Parallel Algorithms Advantages and Disadvantages
Parallel Algorithms Advantages and DisadvantagesParallel Algorithms Advantages and Disadvantages
Parallel Algorithms Advantages and DisadvantagesMurtadha Alsabbagh
 
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...Joseph Luchette
 
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5Ginsbourg.com - Performance and Load Test Report Template LTR 1.5
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5Shay Ginsbourg
 
Continuous Performance Testing
Continuous Performance TestingContinuous Performance Testing
Continuous Performance TestingC4Media
 
Cloud Performance Benchmarking
Cloud Performance BenchmarkingCloud Performance Benchmarking
Cloud Performance BenchmarkingSantanu Dey
 
week_2Lec02_CS422.pptx
week_2Lec02_CS422.pptxweek_2Lec02_CS422.pptx
week_2Lec02_CS422.pptxmivomi1
 
VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld
 
Deterministic and high throughput data processing for CubeSats
Deterministic and high throughput data processing for CubeSatsDeterministic and high throughput data processing for CubeSats
Deterministic and high throughput data processing for CubeSatsPablo Ghiglino
 

Semelhante a LCA14: LCA14-506: Comparative analysis of preemption vs preempt-rt (20)

Chap2 slides
Chap2 slidesChap2 slides
Chap2 slides
 
Dasia 2022
Dasia 2022Dasia 2022
Dasia 2022
 
Four Ways to Improve Linux Performance IEEE Webinar, R2.0
Four Ways to Improve Linux Performance IEEE Webinar, R2.0Four Ways to Improve Linux Performance IEEE Webinar, R2.0
Four Ways to Improve Linux Performance IEEE Webinar, R2.0
 
CPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performanceCPN302 your-linux-ami-optimization-and-performance
CPN302 your-linux-ami-optimization-and-performance
 
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
Your Linux AMI: Optimization and Performance (CPN302) | AWS re:Invent 2013
 
Presentation
PresentationPresentation
Presentation
 
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...
Webinar Slides: Tungsten Connector / Proxy – The Secret Sauce Behind Zero-Dow...
 
Chap2 slides
Chap2 slidesChap2 slides
Chap2 slides
 
Configuration Optimization for Big Data Software
Configuration Optimization for Big Data SoftwareConfiguration Optimization for Big Data Software
Configuration Optimization for Big Data Software
 
Super Computer
Super ComputerSuper Computer
Super Computer
 
참여기관_발표자료-국민대학교 201301 정기회의
참여기관_발표자료-국민대학교 201301 정기회의참여기관_발표자료-국민대학교 201301 정기회의
참여기관_발표자료-국민대학교 201301 정기회의
 
Parallel Algorithms Advantages and Disadvantages
Parallel Algorithms Advantages and DisadvantagesParallel Algorithms Advantages and Disadvantages
Parallel Algorithms Advantages and Disadvantages
 
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...
Unlimited Virtual Computing Capacity using the Cloud for Automated Parameter ...
 
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5Ginsbourg.com - Performance and Load Test Report Template LTR 1.5
Ginsbourg.com - Performance and Load Test Report Template LTR 1.5
 
Continuous Performance Testing
Continuous Performance TestingContinuous Performance Testing
Continuous Performance Testing
 
ECI OpenFlow 2.0 the Future of SDN
ECI OpenFlow 2.0 the Future of SDN ECI OpenFlow 2.0 the Future of SDN
ECI OpenFlow 2.0 the Future of SDN
 
Cloud Performance Benchmarking
Cloud Performance BenchmarkingCloud Performance Benchmarking
Cloud Performance Benchmarking
 
week_2Lec02_CS422.pptx
week_2Lec02_CS422.pptxweek_2Lec02_CS422.pptx
week_2Lec02_CS422.pptx
 
VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series VMworld 2014: Extreme Performance Series
VMworld 2014: Extreme Performance Series
 
Deterministic and high throughput data processing for CubeSats
Deterministic and high throughput data processing for CubeSatsDeterministic and high throughput data processing for CubeSats
Deterministic and high throughput data processing for CubeSats
 

Mais de Linaro

Deep Learning Neural Network Acceleration at the Edge - Andrea Gallo
Deep Learning Neural Network Acceleration at the Edge - Andrea GalloDeep Learning Neural Network Acceleration at the Edge - Andrea Gallo
Deep Learning Neural Network Acceleration at the Edge - Andrea GalloLinaro
 
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta Vekaria
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta VekariaArm Architecture HPC Workshop Santa Clara 2018 - Kanta Vekaria
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta VekariaLinaro
 
Huawei’s requirements for the ARM based HPC solution readiness - Joshua Mora
Huawei’s requirements for the ARM based HPC solution readiness - Joshua MoraHuawei’s requirements for the ARM based HPC solution readiness - Joshua Mora
Huawei’s requirements for the ARM based HPC solution readiness - Joshua MoraLinaro
 
Bud17 113: distribution ci using qemu and open qa
Bud17 113: distribution ci using qemu and open qaBud17 113: distribution ci using qemu and open qa
Bud17 113: distribution ci using qemu and open qaLinaro
 
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018Linaro
 
HPC network stack on ARM - Linaro HPC Workshop 2018
HPC network stack on ARM - Linaro HPC Workshop 2018HPC network stack on ARM - Linaro HPC Workshop 2018
HPC network stack on ARM - Linaro HPC Workshop 2018Linaro
 
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...Linaro
 
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...Linaro
 
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...Linaro
 
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...Linaro
 
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineHKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineLinaro
 
HKG18-100K1 - George Grey: Opening Keynote
HKG18-100K1 - George Grey: Opening KeynoteHKG18-100K1 - George Grey: Opening Keynote
HKG18-100K1 - George Grey: Opening KeynoteLinaro
 
HKG18-318 - OpenAMP Workshop
HKG18-318 - OpenAMP WorkshopHKG18-318 - OpenAMP Workshop
HKG18-318 - OpenAMP WorkshopLinaro
 
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineHKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineLinaro
 
HKG18-315 - Why the ecosystem is a wonderful thing, warts and all
HKG18-315 - Why the ecosystem is a wonderful thing, warts and allHKG18-315 - Why the ecosystem is a wonderful thing, warts and all
HKG18-315 - Why the ecosystem is a wonderful thing, warts and allLinaro
 
HKG18- 115 - Partitioning ARM Systems with the Jailhouse Hypervisor
HKG18- 115 - Partitioning ARM Systems with the Jailhouse HypervisorHKG18- 115 - Partitioning ARM Systems with the Jailhouse Hypervisor
HKG18- 115 - Partitioning ARM Systems with the Jailhouse HypervisorLinaro
 
HKG18-TR08 - Upstreaming SVE in QEMU
HKG18-TR08 - Upstreaming SVE in QEMUHKG18-TR08 - Upstreaming SVE in QEMU
HKG18-TR08 - Upstreaming SVE in QEMULinaro
 
HKG18-113- Secure Data Path work with i.MX8M
HKG18-113- Secure Data Path work with i.MX8MHKG18-113- Secure Data Path work with i.MX8M
HKG18-113- Secure Data Path work with i.MX8MLinaro
 
HKG18-120 - Devicetree Schema Documentation and Validation
HKG18-120 - Devicetree Schema Documentation and Validation HKG18-120 - Devicetree Schema Documentation and Validation
HKG18-120 - Devicetree Schema Documentation and Validation Linaro
 
HKG18-223 - Trusted FirmwareM: Trusted boot
HKG18-223 - Trusted FirmwareM: Trusted bootHKG18-223 - Trusted FirmwareM: Trusted boot
HKG18-223 - Trusted FirmwareM: Trusted bootLinaro
 

Mais de Linaro (20)

Deep Learning Neural Network Acceleration at the Edge - Andrea Gallo
Deep Learning Neural Network Acceleration at the Edge - Andrea GalloDeep Learning Neural Network Acceleration at the Edge - Andrea Gallo
Deep Learning Neural Network Acceleration at the Edge - Andrea Gallo
 
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta Vekaria
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta VekariaArm Architecture HPC Workshop Santa Clara 2018 - Kanta Vekaria
Arm Architecture HPC Workshop Santa Clara 2018 - Kanta Vekaria
 
Huawei’s requirements for the ARM based HPC solution readiness - Joshua Mora
Huawei’s requirements for the ARM based HPC solution readiness - Joshua MoraHuawei’s requirements for the ARM based HPC solution readiness - Joshua Mora
Huawei’s requirements for the ARM based HPC solution readiness - Joshua Mora
 
Bud17 113: distribution ci using qemu and open qa
Bud17 113: distribution ci using qemu and open qaBud17 113: distribution ci using qemu and open qa
Bud17 113: distribution ci using qemu and open qa
 
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018
OpenHPC Automation with Ansible - Renato Golin - Linaro Arm HPC Workshop 2018
 
HPC network stack on ARM - Linaro HPC Workshop 2018
HPC network stack on ARM - Linaro HPC Workshop 2018HPC network stack on ARM - Linaro HPC Workshop 2018
HPC network stack on ARM - Linaro HPC Workshop 2018
 
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...
It just keeps getting better - SUSE enablement for Arm - Linaro HPC Workshop ...
 
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...
Intelligent Interconnect Architecture to Enable Next Generation HPC - Linaro ...
 
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...
Yutaka Ishikawa - Post-K and Arm HPC Ecosystem - Linaro Arm HPC Workshop Sant...
 
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...
Andrew J Younge - Vanguard Astra - Petascale Arm Platform for U.S. DOE/ASC Su...
 
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineHKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
 
HKG18-100K1 - George Grey: Opening Keynote
HKG18-100K1 - George Grey: Opening KeynoteHKG18-100K1 - George Grey: Opening Keynote
HKG18-100K1 - George Grey: Opening Keynote
 
HKG18-318 - OpenAMP Workshop
HKG18-318 - OpenAMP WorkshopHKG18-318 - OpenAMP Workshop
HKG18-318 - OpenAMP Workshop
 
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainlineHKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
HKG18-501 - EAS on Common Kernel 4.14 and getting (much) closer to mainline
 
HKG18-315 - Why the ecosystem is a wonderful thing, warts and all
HKG18-315 - Why the ecosystem is a wonderful thing, warts and allHKG18-315 - Why the ecosystem is a wonderful thing, warts and all
HKG18-315 - Why the ecosystem is a wonderful thing, warts and all
 
HKG18- 115 - Partitioning ARM Systems with the Jailhouse Hypervisor
HKG18- 115 - Partitioning ARM Systems with the Jailhouse HypervisorHKG18- 115 - Partitioning ARM Systems with the Jailhouse Hypervisor
HKG18- 115 - Partitioning ARM Systems with the Jailhouse Hypervisor
 
HKG18-TR08 - Upstreaming SVE in QEMU
HKG18-TR08 - Upstreaming SVE in QEMUHKG18-TR08 - Upstreaming SVE in QEMU
HKG18-TR08 - Upstreaming SVE in QEMU
 
HKG18-113- Secure Data Path work with i.MX8M
HKG18-113- Secure Data Path work with i.MX8MHKG18-113- Secure Data Path work with i.MX8M
HKG18-113- Secure Data Path work with i.MX8M
 
HKG18-120 - Devicetree Schema Documentation and Validation
HKG18-120 - Devicetree Schema Documentation and Validation HKG18-120 - Devicetree Schema Documentation and Validation
HKG18-120 - Devicetree Schema Documentation and Validation
 
HKG18-223 - Trusted FirmwareM: Trusted boot
HKG18-223 - Trusted FirmwareM: Trusted bootHKG18-223 - Trusted FirmwareM: Trusted boot
HKG18-223 - Trusted FirmwareM: Trusted boot
 

Último

Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsEnterprise Knowledge
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...shyamraj55
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...gurkirankumar98700
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Paola De la Torre
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 

Último (20)

Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
IAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI SolutionsIAC 2024 - IA Fast Track to Search Focused AI Solutions
IAC 2024 - IA Fast Track to Search Focused AI Solutions
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 

LCA14: LCA14-506: Comparative analysis of preemption vs preempt-rt

  • 1. Gary Robertson, LCA14, Macau LCA14-506: Comparative analysis of preemption vs preempt-rt
  • 2. In this presentation we will try to illustrate the pro’s, con’s, and latency characteristics of several Linux kernel preemption models, and provide some guidance in selecting an appropriate preemption model for a given category of application. Overview of Topics Presented
  • 3. Questions we will address include: • Which preemption model provides the best throughput? • Which model offers the lowest average latencies? • Which model offers the lowest maximum latencies? • Which model offers the most predictable latencies? • How do load conditions impact the respective latency performance of the various models? • What impact does CPU Frequency Scaling or CPU sleep states have on latency performance? • What is the best model for a given application type? Overview of Topics - continued
  • 4. Our intent is to show relative trends between the preemption models under the same conditions… so the data presented were gathered thusly: • Each preemption model configuration was tested using identical tests running on the same InSignal Arndale. • Cyclictest was used for a run duration of two hours with a single thread executing at a SCHED_FIFO priority of 80 to realistically represent scheduling latency for a real- time process. • A cyclictest run was done with no system load, then another with an externally-applied ping flood, and another with back-to-back executions of hackbench running to represent maximum system loading. Test Rationale and Methodology
  • 5. Latency Impact of CPU Frequency Scaling
  • 6. Only three Linux preemption models are really interesting for anything other than desktop use: • the Server preemption model provides optimal throughput for applications where latencies are not an issue • the Low Latency Desktop preemption model provides low average latencies for interactive and ‘soft real-time’ applications • the Full RT preemption model provides the highest level of latency determinism for ‘hard real-time’ applications Tested Linux Preemption Models
  • 7. Server Preemption Model Latencies Cyclictest with no system load CPU frequency scaling disabled Minimum Latency: 16 usec Average Latency: 24 usec Most Frequent Latency: 24 usec Maximum Latency: 572 usec Standard Deviation: 1.211041 Almost all latencies between 20 usec and 28 usec However, even at light loads, latencies out to 572 usec were observed. This is a consequence of all code paths through the kernel being non-preemptible.
  • 8. Server Preemption Model Latencies Cyclictest with ping flood load CPU frequency scaling disabled Minimum Latency: 15 usec Average Latency: 23 usec Most Frequent Latency: 24 usec Maximum Latency: 592 usec Standard Deviation: 1.580778 Almost all latencies between 20 usec and 28 usec Note, however, that much longer latencies continue to be observed due to lack of any design efforts to avoid them. Also note that maximum latency is already beginning to creep upwards.
  • 9. Server Preemption Model Latencies Cyclictest with hackbench load CPU frequency scaling disabled Minimum Latency: 17 usec Average Latency: 150655 usec Most Frequent Latency: 22 usec Maximum Latency: 2587753 usec Standard Deviation: 493977.9 The majority of latencies were between 21 usec and 25 usec, gradually tapering off to single digit frequencies at 204 usec. Note the duration of the max latency is 4000 times longer than under no load! Note also the much lower frequency percentage for the peak occurrence. This means a larger percentage of the higher latencies were observed, and illustrates the serious degradation of latency determinism under load in a non- preemptible kernel where latency was not a primary design consideration.
  • 10. Low Latency Desktop Model Latencies Cyclictest with no system load CPU frequency scaling disabled Minimum Latency: 19 usec Average Latency: 28 usec Most Frequent Latency: 29 usec Maximum Latency: 57 usec Standard Deviation: 0.8698308 The majority of latencies were between 28 usec and 31 usec, quickly tapering off to single digit frequencies at 42 usec. Maximum latency was reduced tenfold under light loads vs. the Server model. This illustrates the significant improvements in latency performance under light loads with kernel preemption enabled.
  • 11. Low Latency Desktop Model Latencies Cyclictest with ping flood CPU frequency scaling disabled Minimum Latency: 18 usec Average Latency: 29 usec Most Frequent Latency: 29 usec Maximum Latency: 131 usec Standard Deviation: 1.79573 The majority of latencies were between 28 usec and 32 usec, quickly tapering off to single digit frequencies at 80 usec. The reduced range of observed latencies indicates improved latency performance and predictability at moderate loads versus the Server model. However, as the next slide will show, latency performance in this model degrades seriously under heavy load, making Full RT a better choice for latency performance under heavy load conditions.
  • 12. Low Latency Desktop Model Latencies Cyclictest with hackbench CPU frequency scaling disabled Minimum Latency: 19 usec Average Latency: 370606 usec Most Frequent Latency: 25 usec Maximum Latency: 4122148 usec Standard Deviation: 826092 The majority of latencies were between 24 usec and 26 usec, gradually tapering off to single digit frequencies at 105 usec. Note that the max latency was 70,000 times longer than with this model under no load! Max latencies were nearly double that for a Server model under heavy load, and latency predictability is low. This illustrates the combined impacts under heavy load of increased context switches without addressing priority inversion or FIFO queueing disciplines.
  • 13. Full RT Preemption Model Latencies Cyclictest with no system load CPU frequency scaling disabled Minimum Latency: 19 usec Average Latency: 29 usec Most Frequent Latency: 29 usec Maximum Latency: 53 usec Standard Deviation: 1.031893 The majority of latencies were between 29 usec and 31 usec, quickly tapering off to single digit frequencies at 50 usec. Maximum latency was reduced tenfold under light loads vs. the Server model. This illustrates the significant improvements in latency performance under light loads with kernel preemption enabled. Under light load performance is very similar to that of the Low Latency Desktop model.
  • 14. Full RT Preemption Model Latencies Cyclictest with ping flood CPU frequency scaling disabled Minimum Latency: 19 usec Average Latency: 29 usec Most Frequent Latency: 30 usec Maximum Latency: 59 usec Standard Deviation: 2.698587 The majority of latencies were between 29 usec and 31 usec, quickly tapering off to single digit frequencies at 53 usec. The reduced range of observed latencies indicates improved latency performance and predictability at moderate loads versus the Server model. Note that even at moderate loads the maximum latencies are less than half the duration of those seen in the Low Latency Desktop model.
  • 15. Full RT Preemption Model Latencies Cyclictest with hackbench load CPU frequency scaling disabled Minimum Latency: 21 usec Average Latency: 29 usec Most Frequent Latency: 25 usec Maximum Latency: 156 usec Standard Deviation: 7.69571 The majority of latencies were between 24 usec and 26 usec, with a second group peaking between 43 and 44 usec, and quickly tapering off to single digit frequencies at 134 usec. Latency performance under heavy load is much better than in any of the other preemption models. With threaded interrupt handlers, priority inheritance and priority-based queuing disciplines, the real-time process is still able to meet much tighter scheduling deadlines despite heavy activity of other lower- priority threads.
  • 19. • For applications in which throughput and not latencies are the primary consideration, opt for the Server model • If quality of service is important but missed latency deadlines will not result in catastrophic failures, opt for the Low Latency Desktop model and size the hardware capacity to keep loading moderate • For host environments for ‘zero overhead Linux’ (ODP for example), Low Latency Desktop is a good choice • If latencies must be consistent even under high load conditions Full RT may be required • For applications based on POSIX real-time scheduling and priority-based preemption, use Full RT for best results What Preemption Model Is Best for Me?
  • 20. The test scripts, data files, and graphs used to provide reference data for this presentation may be accessed online at the following URL: http://people.linaro.org/~gary.robertson/LCA14 Data References
  • 21. More about Linaro Connect: http://connect.linaro.org More about Linaro: http://www.linaro.org/about/ More about Linaro engineering: http://www.linaro.org/engineering/ Linaro members: www.linaro.org/members
  • 23. The Server preemption model lies at one extreme of the latency vs. throughput continuum. Pro’s include: • Simplicity, maturity and robustness make this a very reliable platform • With no preemption the reduced number of context switches minimizes system overhead and maximizes overall throughput Server Model Characteristics
  • 24. Con’s include: • The lack of preemption results in low average latencies under low loads but much higher latencies when the system is heavily loaded • The latencies imposed by different execution paths through the kernel result in a wide range of latency durations and low latency determinism Server Model Characteristics
  • 25. The Low Latency Desktop preemption model holds the middle ground in the latency vs. throughput continuum. Pro’s include: • Under low to moderate load, latency range and predictability are significantly improved vs. the Server model • This preemption model is supported as part of the mainstream kernel and tends to be less trouble-prone than Full RT preemption Low Latency Desktop Model Characteristics
  • 26. Con’s include: • The preemption of kernel operations and increased number of context switches create increased overhead and reduced performance relative to the Server model • The preemption of kernel operations results in higher average latencies vs. the Server preemption model • This preemption model does not perform as well under heavy system loads as other models Low Latency Desktop Model Characteristics
  • 27. The following software-induced latency sources remain problematic in the Low Latency Desktop preemption model: • Exceptions, software interrupts, and device service request interrupts execute outside of scheduler control • Most mutual exclusion locking primitives are subject to priority inversion • Shared resources use FIFO-based queueing disciplines, meaning high-priority threads may have to wait behind lower priority threads for access to the resources These factors result in lower levels of latency determinism. Low Latency Desktop Model - continued
  • 28. The Full RT preemption model represents the latency- centric end of the latency vs. throughput continuum. It attempts to mitigate all the remaining software-induced sources of latency. • Handlers for exceptions, software interrupts, and device service request interrupts are encapsulated inside threads which are under scheduler control • Priority inheritance is added for most mutual exclusion locking primitives to prevent priority inversions • Shared resources use priority-based queueing disciplines so that the highest-priority thread always gets first access to the resources Full RT Model Characteristics
  • 29. The Full RT preemption model inevitably suffers from reduced overall throughput as a consequence of its efforts to maximize latency determinism: • Schedulable ‘threaded’ ISRs result in the highest levels of preemption and context switch overhead • Priority inheritance involves iterative logic to temporarily boost the priorities of all lock holders to equal that of the highest-priority lock waiters. This adds significant overhead to locking primitive code. • Priority-based queueing requires sorting the queue each time a new waiting thread is added Full RT Model Characteristics - continued
  • 30. Pro’s include: • The most consistent and predictable latency performance available in any preemption model • The best support environment for creating priority- based multi-layered applications • The best hard real-time support available in a Linux environment Full RT Model Characteristics - continued
  • 31. Con’s include: • Full RT preemption is supported only with a separately maintained kernel patch set • The latest supported RT kernel version always lags behind mainstream development • Mainstream drivers, libraries, and applications may not always function properly in the Full RT environment • Poorly designed or written real-time threads may starve out threaded interrupt handlers Full RT Model Characteristics - continued