SlideShare a Scribd company logo
1 of 28
Parallel Analytics as a Service
Petrie Wong, Andy He, Eric Lo
Department of Computing
The Hong Kong Polytechnic University
{cskfwong, cszahe, ericlo}@comp.polyu.edu.hk
Motivation
Massively Parallel Processing Relational
Database Systems
o SQL interface
o easy to scale-out
o fast query time
o expensive
 hardware cost, operation cost and
 software license
Vertica, Microsoft PDW, Greenplum, ...
about USD 15K per core or
USD 50K per TB of data
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 2
Motivation
● MPPDB-as-a-Service
o service provider
 consolidate tenants to fewer machines
● achieves a lower operation cost
o tenants
 enjoy parallel analytics at a lower price
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 3
MPPDB as a service: Example
● Shivnath: requests a 64-node PDB
● Reynold Xin: requests a 128-node PDB
● Eric: requests a 32-node PDB
● Donald Kossmann: ...
● We: service provider
o have 1,000 nodes
● Problems to Solve:
o Tenant Consolidation
 Performance (SLA Guarantee P)
 High Availability (Replication Factor R)
● P% (e.g., 99.9%) of time a tenant finishes its queries
without feeling any performance delay
o Shivnath's Q1 finishes in 100 seconds (isolation)
o After consolidation: Q1 finishes in 100 seconds
o After Consolidation
 Query Routing
 Monitoring, Elastic Scaling, and Re-consolidation
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 4
Virtual Machine vs. Shared-process
Virtual Machine*
● offers hard
isolation
● incurs significant
redundancy on
resources
o multiple copies of
DBMS
Shared-process
● obtains higher
consolidation
effectiveness
● previous research
o single DB
o OLTP workloads
● This paper
o MPPDB
o OLAP workloads
* We believe Amazon's RedShift is based on VM
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 5
DaaS
● OLTP workload
o queries touch small data
● SLA
o throughput (e.g., 100tps)
b4-&-after consolidation
is the same
Previous Works (DaaS)
vs. MPPDBaaS
MPPDBaaS
● OLAP workload
o I/O-intensive
● SLA
o query latency b4-&-after
consolidation is the same
after consolidation,
many tenants active together?
SLA OK!
after consolidation,
many tenants active together?
Easily Violate SLA!
● many tenants
o small data, single node
● many tenants
o "big" data, multiple nodes
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 6
Requirements of a MPPDBaaS
R1. Support 10000+ of tenants and machine nodes
R2. Offer high availability services
o replicate tenants' data
R3. Offer query-latency based performance SLA
guarantees
o need performance prediction techniques
R4. Support general tenants:
o tenant's query templates may not be known in
advance
o voids R3
 because most performance prediction techniques
require query templates be given
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 7
Our solution:
Tenant-Driven Design
● active tenant ratio in real multi-tenant
environments is low
o in IBM DaaS: ~10% active at same time
● and come up with a deployment plan that
o let each active tenant exclusively use an MPPDB
B. Reinwald. Multitenancy. University of
Washington and Microsoft Research Summer
Institute, 2010.
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 8
● R = 3, TDD creates 3 MPPDBs
● # of nodes of each MPPDB
= the tenant that requests the max # of
nodes
Tenant Driven Design:
A Toy Example
● 10 tenants: T1, T2, ... T10
Tenant T1 T2 T3 T4 T5 ... T10 Total
Requested
Nodes
6 6 5 5 2 ... 2 34
Group MPPDB0 MPPDB1 MPPDB2
Nodes 6 6 6
Hosting T1 - T10 T1 - T10 T1 - T10
TDD property:
replication factor
= A (# of active tenant in peak hour)
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 9
T2
T2
Tenant Driven Design
Query Routing - Example
T2
T1
T5
00:00 GMT 24:00 GMT
T1
Q1
Q3
Q2
Q4
T2
Q7
Q5
Q6
MPPDB1
MPPDB0
MPPDB2
Routing Principle:
Let a tenant exclusively use a MPPDB!
non-linear scale-out
A=3
TDD's SLA Guarantee: NO matter what query,
ad-hoc query (a new query
that is seen the first time)
linear scale-out query with template given
submitted sequentially
(interactive queries)
submitted in a batch
at any multi-programming-level
the SLAs of a maximum of A active tenants are met.
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 10
Scaling up TDD
Realistic Example
● 5,000 tenants
● Average 10% of tenants active
● 500 active tenants
500+ Replicas !!!
TDD guarantee:
the SLAs of a maximum of A active tenants are met.
TDD property:
replication factor = A (# of active tenant in peak hour)
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 11
TDD guarantee:
the SLAs of a maximum of 500 active tenants are met.
TDD property:
replication factor = 500 (# of active tenant in peak hour)
Scaling up TDD: solution
● Further divide the tenants into groups
● Each group uses one TDD
● Considerations
o minimize the total # of nodes used
o reasonable replication factor
 # active tenant in each tenant-group is low
e.g. only 3 tenants active together
e.g. divide 5000 tenant into 300 groups
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 12
Scaling up TDD
● Formulate as
o Largest Item
Vector Bin
Packing Problem
with Fuzzy
Capacity
● Devise effective
heuristic solution
called 2-step
Grouping
algorithm
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 13
Thrifty - System Architecture
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 14
Experimental Result -- Overview
● Setup: MulTe-like
● 5,000 tenants
● 2 - 32 nodes
● 200GB to 3.2TB
~89,300 nodes requested
(without replica)
saves 83.3% of nodes !!!
~15,000 nodes only
including high availability
already (3 replicas!)
after
consolidation
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 15
Consolidation Effectiveness
86.5%
79.3%
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 16
Thrifty - Prototype of MPPDB as a Service
● Shared-Process Consolidation
● provide High Availability
● guarantee Query Performance SLA
● support General Tenants
Conclusions
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 17
Thank You
any questions?
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University END
Tenant-Driven Design (TDD)
● Cluster Design
o divide the machine nodes in the cluster into A
groups
 each group of nodes deploys a separate
MPPDB
● Tenant Placement
o each MPPDB hosts all tenants
● Query Routing
o route a tenant's query to a MPPDB that contains its
dataTDD's property :
enforces a replication factor of A for each tenant
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-1
Consolidation Effectiveness -
under Higher Active Tenant Ratio
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-2
Heuristic Algorithm -
2-Step Grouping
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-3
2-Step Grouping
vs. Standard Heuristics
● get higher
consolidation ratio in
all dimensions
● get result in few
hours only for 10k
tenants
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-4
Lightweight Elastic Scaling
● Reactive approach
o If the SLA is lower than 99.9% based in past 24hrs
statistic, the system will trigger elastic scaling.
o a new MPPDB will be online and the system will
o deploy the tenant Tx to that MPPDB
o which is most active in past 24hrs to.
o all queries of the Tx will be route to the new MPPDB
after its online.
● Example
o three 10-node MPPDBs serves 20 tenants
o SLA guarantee drops from 99.95% to 99.0%
 solution: starts a new 10-node MPPDB
● cost 30 minutes for 1TB data (with parallel loading)Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-5
Lightweight Elastic Scaling
trigger finishtrigger finish
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-6
Manual Tuning
System administrator may instead use
the parameter U to manually
Use case
● tiny time percentage of possibly SLA
violations
o three 10-node MPPDBs
o serves 20 tenants
o SLA guarantee drops from 99.9% to 99.8%
 solution 1: starts a new 10-node MPPDB (+10)
 solution 2: increase 10-nodes to 11-nodes (+3)
● shorter query latency
saves 7 nodes (17.5%)
number of nodes in a MPPDB
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-7
Linear and Non-linear Scale-out
Queries - Concurrent Execution
consolidation opportunity by merging clusters
● query template required
SLA
SLA
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-8
Contribution
● Thrifty
o a prototype implementation of MPPDBaaS
o Tenant-driven Design (TDD)
 Smartly organize the tenants into groups
 Each group has R MPPDBs
 Each MPPDB
● host a group of tenants
 Prove that TDD solves R2 to R4 in a row!
o Scale TDD to 10000+ of tenants and machine nodes
(R1)?
 Cast the problem as a new variant of
vector bin packing problem
 Provide an efficient heuristics algorithm
[R = replication factor]
R2. high availability
R3. query latency SLA
R4. generality
Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-9
Replications
● Vary R

More Related Content

What's hot

FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...
FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...
FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...Flink Forward
 
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNet
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNetFrom Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNet
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNetEric Haibin Lin
 
MLConf 2016 SigOpt Talk by Scott Clark
MLConf 2016 SigOpt Talk by Scott ClarkMLConf 2016 SigOpt Talk by Scott Clark
MLConf 2016 SigOpt Talk by Scott ClarkSigOpt
 
PyConUK 2018 - Journey from HTTP to gRPC
PyConUK 2018 - Journey from HTTP to gRPCPyConUK 2018 - Journey from HTTP to gRPC
PyConUK 2018 - Journey from HTTP to gRPCTatiana Al-Chueyr
 
Serving BERT Models in Production with TorchServe
Serving BERT Models in Production with TorchServeServing BERT Models in Production with TorchServe
Serving BERT Models in Production with TorchServeNidhin Pattaniyil
 
And Then There Are Algorithms
And Then There Are AlgorithmsAnd Then There Are Algorithms
And Then There Are AlgorithmsInfluxData
 
Understand and Harness the Capabilities of Intel® Xeon Phi™ Processors
Understand and Harness the Capabilities of Intel® Xeon Phi™ ProcessorsUnderstand and Harness the Capabilities of Intel® Xeon Phi™ Processors
Understand and Harness the Capabilities of Intel® Xeon Phi™ ProcessorsIntel® Software
 
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...IOSR Journals
 
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache Zeppelin
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache ZeppelinMoon soo Lee – Data Science Lifecycle with Apache Flink and Apache Zeppelin
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache ZeppelinFlink Forward
 
Shuffle phase as the bottleneck in Hadoop Terasort
Shuffle phase as the bottleneck in Hadoop TerasortShuffle phase as the bottleneck in Hadoop Terasort
Shuffle phase as the bottleneck in Hadoop Terasortpramodbiligiri
 
Maximilian Michels – Google Cloud Dataflow on Top of Apache Flink
Maximilian Michels – Google Cloud Dataflow on Top of Apache FlinkMaximilian Michels – Google Cloud Dataflow on Top of Apache Flink
Maximilian Michels – Google Cloud Dataflow on Top of Apache FlinkFlink Forward
 
Project Planning using Little’s Law
Project Planning using Little’s LawProject Planning using Little’s Law
Project Planning using Little’s LawDimitar Bakardzhiev
 
Hadoop Network Performance profile
Hadoop Network Performance profileHadoop Network Performance profile
Hadoop Network Performance profilepramodbiligiri
 
Task Parallel Library Data Flows
Task Parallel Library Data FlowsTask Parallel Library Data Flows
Task Parallel Library Data FlowsSANKARSAN BOSE
 
Introducing the TPCx-HS Benchmark for Big Data
Introducing the TPCx-HS Benchmark for Big DataIntroducing the TPCx-HS Benchmark for Big Data
Introducing the TPCx-HS Benchmark for Big Datainside-BigData.com
 

What's hot (19)

FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...
FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...
FlinkDTW: Time-series Pattern Search at Scale Using Dynamic Time Warping - Ch...
 
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNet
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNetFrom Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNet
From Hours to Minutes: The Journey of Optimizing Mask-RCNN and BERT Using MXNet
 
Manycores for the Masses
Manycores for the MassesManycores for the Masses
Manycores for the Masses
 
MLConf 2016 SigOpt Talk by Scott Clark
MLConf 2016 SigOpt Talk by Scott ClarkMLConf 2016 SigOpt Talk by Scott Clark
MLConf 2016 SigOpt Talk by Scott Clark
 
PyConUK 2018 - Journey from HTTP to gRPC
PyConUK 2018 - Journey from HTTP to gRPCPyConUK 2018 - Journey from HTTP to gRPC
PyConUK 2018 - Journey from HTTP to gRPC
 
Serving BERT Models in Production with TorchServe
Serving BERT Models in Production with TorchServeServing BERT Models in Production with TorchServe
Serving BERT Models in Production with TorchServe
 
And Then There Are Algorithms
And Then There Are AlgorithmsAnd Then There Are Algorithms
And Then There Are Algorithms
 
RL-Cache: Learning-Based Cache Admission for Content Delivery
RL-Cache: Learning-Based Cache Admission for Content DeliveryRL-Cache: Learning-Based Cache Admission for Content Delivery
RL-Cache: Learning-Based Cache Admission for Content Delivery
 
Understand and Harness the Capabilities of Intel® Xeon Phi™ Processors
Understand and Harness the Capabilities of Intel® Xeon Phi™ ProcessorsUnderstand and Harness the Capabilities of Intel® Xeon Phi™ Processors
Understand and Harness the Capabilities of Intel® Xeon Phi™ Processors
 
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...
Performance Evaluation of IPv4 Vs Ipv6 and Tunnelling Techniques Using Optimi...
 
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache Zeppelin
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache ZeppelinMoon soo Lee – Data Science Lifecycle with Apache Flink and Apache Zeppelin
Moon soo Lee – Data Science Lifecycle with Apache Flink and Apache Zeppelin
 
HTTP Adaptive Streaming – Quo Vadis?
HTTP Adaptive Streaming – Quo Vadis?HTTP Adaptive Streaming – Quo Vadis?
HTTP Adaptive Streaming – Quo Vadis?
 
NTCIR15WWW3overview
NTCIR15WWW3overviewNTCIR15WWW3overview
NTCIR15WWW3overview
 
Shuffle phase as the bottleneck in Hadoop Terasort
Shuffle phase as the bottleneck in Hadoop TerasortShuffle phase as the bottleneck in Hadoop Terasort
Shuffle phase as the bottleneck in Hadoop Terasort
 
Maximilian Michels – Google Cloud Dataflow on Top of Apache Flink
Maximilian Michels – Google Cloud Dataflow on Top of Apache FlinkMaximilian Michels – Google Cloud Dataflow on Top of Apache Flink
Maximilian Michels – Google Cloud Dataflow on Top of Apache Flink
 
Project Planning using Little’s Law
Project Planning using Little’s LawProject Planning using Little’s Law
Project Planning using Little’s Law
 
Hadoop Network Performance profile
Hadoop Network Performance profileHadoop Network Performance profile
Hadoop Network Performance profile
 
Task Parallel Library Data Flows
Task Parallel Library Data FlowsTask Parallel Library Data Flows
Task Parallel Library Data Flows
 
Introducing the TPCx-HS Benchmark for Big Data
Introducing the TPCx-HS Benchmark for Big DataIntroducing the TPCx-HS Benchmark for Big Data
Introducing the TPCx-HS Benchmark for Big Data
 

Similar to Parallel analytics as a service

SDN in the Management Plane: OpenConfig and Streaming Telemetry
SDN in the Management Plane: OpenConfig and Streaming TelemetrySDN in the Management Plane: OpenConfig and Streaming Telemetry
SDN in the Management Plane: OpenConfig and Streaming TelemetryAnees Shaikh
 
IPLC Analytic Dashboard - Mohd Rizal bin Mohd Ramly
IPLC Analytic Dashboard - Mohd Rizal bin Mohd RamlyIPLC Analytic Dashboard - Mohd Rizal bin Mohd Ramly
IPLC Analytic Dashboard - Mohd Rizal bin Mohd RamlyMyNOG
 
Stream processing with Apache Flink @ OfferUp
Stream processing with Apache Flink @ OfferUpStream processing with Apache Flink @ OfferUp
Stream processing with Apache Flink @ OfferUpBowen Li
 
Optimization of Continuous Queries in Federated Database and Stream Processin...
Optimization of Continuous Queries in Federated Database and Stream Processin...Optimization of Continuous Queries in Federated Database and Stream Processin...
Optimization of Continuous Queries in Federated Database and Stream Processin...Zbigniew Jerzak
 
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko Stream Data Processing at Big Data Landscape by Oleksandr Fedirko
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko GlobalLogic Ukraine
 
ICLR 2020 Recap
ICLR 2020 RecapICLR 2020 Recap
ICLR 2020 RecapSri Ambati
 
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...Automating Speed: A Proven Approach to Preventing Performance Regressions in ...
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...HostedbyConfluent
 
Triantafyllia Voulibasi
Triantafyllia VoulibasiTriantafyllia Voulibasi
Triantafyllia VoulibasiISSEL
 
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe Engine
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe EngineElastic Scaling of a High-Throughput Content-Based Publish/Subscribe Engine
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe EngineZbigniew Jerzak
 
Enabling Presto Caching at Uber with Alluxio
Enabling Presto Caching at Uber with AlluxioEnabling Presto Caching at Uber with Alluxio
Enabling Presto Caching at Uber with AlluxioAlluxio, Inc.
 
Cloud-Scale BGP and NetFlow Analysis
Cloud-Scale BGP and NetFlow AnalysisCloud-Scale BGP and NetFlow Analysis
Cloud-Scale BGP and NetFlow AnalysisAlex Henthorn-Iwane
 
CloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning
 
High-Speed Reactive Microservices
High-Speed Reactive MicroservicesHigh-Speed Reactive Microservices
High-Speed Reactive MicroservicesRick Hightower
 
Dubbo and Weidian's practice on micro-service architecture
Dubbo and Weidian's practice on micro-service architectureDubbo and Weidian's practice on micro-service architecture
Dubbo and Weidian's practice on micro-service architectureHuxing Zhang
 
IEEEGlobecom'22-OL-RICHTER.pdf
IEEEGlobecom'22-OL-RICHTER.pdfIEEEGlobecom'22-OL-RICHTER.pdf
IEEEGlobecom'22-OL-RICHTER.pdfReza Farahani
 
Model based transaction-aware cloud resources management case study and met...
Model based transaction-aware cloud resources management   case study and met...Model based transaction-aware cloud resources management   case study and met...
Model based transaction-aware cloud resources management case study and met...Leonid Grinshpan, Ph.D.
 
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...IEEEFINALYEARSTUDENTPROJECT
 

Similar to Parallel analytics as a service (20)

SDN in the Management Plane: OpenConfig and Streaming Telemetry
SDN in the Management Plane: OpenConfig and Streaming TelemetrySDN in the Management Plane: OpenConfig and Streaming Telemetry
SDN in the Management Plane: OpenConfig and Streaming Telemetry
 
rerngvit_phd_seminar
rerngvit_phd_seminarrerngvit_phd_seminar
rerngvit_phd_seminar
 
IPLC Analytic Dashboard - Mohd Rizal bin Mohd Ramly
IPLC Analytic Dashboard - Mohd Rizal bin Mohd RamlyIPLC Analytic Dashboard - Mohd Rizal bin Mohd Ramly
IPLC Analytic Dashboard - Mohd Rizal bin Mohd Ramly
 
Stream processing with Apache Flink @ OfferUp
Stream processing with Apache Flink @ OfferUpStream processing with Apache Flink @ OfferUp
Stream processing with Apache Flink @ OfferUp
 
Optimization of Continuous Queries in Federated Database and Stream Processin...
Optimization of Continuous Queries in Federated Database and Stream Processin...Optimization of Continuous Queries in Federated Database and Stream Processin...
Optimization of Continuous Queries in Federated Database and Stream Processin...
 
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko Stream Data Processing at Big Data Landscape by Oleksandr Fedirko
Stream Data Processing at Big Data Landscape by Oleksandr Fedirko
 
ICLR 2020 Recap
ICLR 2020 RecapICLR 2020 Recap
ICLR 2020 Recap
 
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...Automating Speed: A Proven Approach to Preventing Performance Regressions in ...
Automating Speed: A Proven Approach to Preventing Performance Regressions in ...
 
Triantafyllia Voulibasi
Triantafyllia VoulibasiTriantafyllia Voulibasi
Triantafyllia Voulibasi
 
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe Engine
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe EngineElastic Scaling of a High-Throughput Content-Based Publish/Subscribe Engine
Elastic Scaling of a High-Throughput Content-Based Publish/Subscribe Engine
 
Enabling Presto Caching at Uber with Alluxio
Enabling Presto Caching at Uber with AlluxioEnabling Presto Caching at Uber with Alluxio
Enabling Presto Caching at Uber with Alluxio
 
BAXTER phase 1b
BAXTER phase 1bBAXTER phase 1b
BAXTER phase 1b
 
Cloud-Scale BGP and NetFlow Analysis
Cloud-Scale BGP and NetFlow AnalysisCloud-Scale BGP and NetFlow Analysis
Cloud-Scale BGP and NetFlow Analysis
 
CloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use CaseCloudLightning and the OPM-based Use Case
CloudLightning and the OPM-based Use Case
 
High-Speed Reactive Microservices
High-Speed Reactive MicroservicesHigh-Speed Reactive Microservices
High-Speed Reactive Microservices
 
cikm14
cikm14cikm14
cikm14
 
Dubbo and Weidian's practice on micro-service architecture
Dubbo and Weidian's practice on micro-service architectureDubbo and Weidian's practice on micro-service architecture
Dubbo and Weidian's practice on micro-service architecture
 
IEEEGlobecom'22-OL-RICHTER.pdf
IEEEGlobecom'22-OL-RICHTER.pdfIEEEGlobecom'22-OL-RICHTER.pdf
IEEEGlobecom'22-OL-RICHTER.pdf
 
Model based transaction-aware cloud resources management case study and met...
Model based transaction-aware cloud resources management   case study and met...Model based transaction-aware cloud resources management   case study and met...
Model based transaction-aware cloud resources management case study and met...
 
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...
2014 IEEE JAVA NETWORKING COMPUTING PROJECT Cost effective resource allocatio...
 

Recently uploaded

EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptxthyngster
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998YohFuh
 
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /WhatsappsBeautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsappssapnasaifi408
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingNeil Barnes
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxolyaivanovalion
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxolyaivanovalion
 
Call Girls In Mahipalpur O9654467111 Escorts Service
Call Girls In Mahipalpur O9654467111  Escorts ServiceCall Girls In Mahipalpur O9654467111  Escorts Service
Call Girls In Mahipalpur O9654467111 Escorts ServiceSapana Sha
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Callshivangimorya083
 
Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystSamantha Rae Coolbeth
 
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfSocial Samosa
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130Suhani Kapoor
 
定制英国白金汉大学毕业证(UCB毕业证书) 成绩单原版一比一
定制英国白金汉大学毕业证(UCB毕业证书)																			成绩单原版一比一定制英国白金汉大学毕业证(UCB毕业证书)																			成绩单原版一比一
定制英国白金汉大学毕业证(UCB毕业证书) 成绩单原版一比一ffjhghh
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...Suhani Kapoor
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionfulawalesam
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptSonatrach
 
B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxStephen266013
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxolyaivanovalion
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...Suhani Kapoor
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysismanisha194592
 

Recently uploaded (20)

EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998
 
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /WhatsappsBeautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data Storytelling
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptx
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptx
 
Call Girls In Mahipalpur O9654467111 Escorts Service
Call Girls In Mahipalpur O9654467111  Escorts ServiceCall Girls In Mahipalpur O9654467111  Escorts Service
Call Girls In Mahipalpur O9654467111 Escorts Service
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
 
Unveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data AnalystUnveiling Insights: The Role of a Data Analyst
Unveiling Insights: The Role of a Data Analyst
 
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdfKantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
Kantar AI Summit- Under Embargo till Wednesday, 24th April 2024, 4 PM, IST.pdf
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
 
定制英国白金汉大学毕业证(UCB毕业证书) 成绩单原版一比一
定制英国白金汉大学毕业证(UCB毕业证书)																			成绩单原版一比一定制英国白金汉大学毕业证(UCB毕业证书)																			成绩单原版一比一
定制英国白金汉大学毕业证(UCB毕业证书) 成绩单原版一比一
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interaction
 
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in  KishangarhDelhi 99530 vip 56974 Genuine Escort Service Call Girls in  Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
 
B2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docxB2 Creative Industry Response Evaluation.docx
B2 Creative Industry Response Evaluation.docx
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptx
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
 
April 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's AnalysisApril 2024 - Crypto Market Report's Analysis
April 2024 - Crypto Market Report's Analysis
 

Parallel analytics as a service

  • 1. Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo Department of Computing The Hong Kong Polytechnic University {cskfwong, cszahe, ericlo}@comp.polyu.edu.hk
  • 2. Motivation Massively Parallel Processing Relational Database Systems o SQL interface o easy to scale-out o fast query time o expensive  hardware cost, operation cost and  software license Vertica, Microsoft PDW, Greenplum, ... about USD 15K per core or USD 50K per TB of data Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 2
  • 3. Motivation ● MPPDB-as-a-Service o service provider  consolidate tenants to fewer machines ● achieves a lower operation cost o tenants  enjoy parallel analytics at a lower price Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 3
  • 4. MPPDB as a service: Example ● Shivnath: requests a 64-node PDB ● Reynold Xin: requests a 128-node PDB ● Eric: requests a 32-node PDB ● Donald Kossmann: ... ● We: service provider o have 1,000 nodes ● Problems to Solve: o Tenant Consolidation  Performance (SLA Guarantee P)  High Availability (Replication Factor R) ● P% (e.g., 99.9%) of time a tenant finishes its queries without feeling any performance delay o Shivnath's Q1 finishes in 100 seconds (isolation) o After consolidation: Q1 finishes in 100 seconds o After Consolidation  Query Routing  Monitoring, Elastic Scaling, and Re-consolidation Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 4
  • 5. Virtual Machine vs. Shared-process Virtual Machine* ● offers hard isolation ● incurs significant redundancy on resources o multiple copies of DBMS Shared-process ● obtains higher consolidation effectiveness ● previous research o single DB o OLTP workloads ● This paper o MPPDB o OLAP workloads * We believe Amazon's RedShift is based on VM Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 5
  • 6. DaaS ● OLTP workload o queries touch small data ● SLA o throughput (e.g., 100tps) b4-&-after consolidation is the same Previous Works (DaaS) vs. MPPDBaaS MPPDBaaS ● OLAP workload o I/O-intensive ● SLA o query latency b4-&-after consolidation is the same after consolidation, many tenants active together? SLA OK! after consolidation, many tenants active together? Easily Violate SLA! ● many tenants o small data, single node ● many tenants o "big" data, multiple nodes Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 6
  • 7. Requirements of a MPPDBaaS R1. Support 10000+ of tenants and machine nodes R2. Offer high availability services o replicate tenants' data R3. Offer query-latency based performance SLA guarantees o need performance prediction techniques R4. Support general tenants: o tenant's query templates may not be known in advance o voids R3  because most performance prediction techniques require query templates be given Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 7
  • 8. Our solution: Tenant-Driven Design ● active tenant ratio in real multi-tenant environments is low o in IBM DaaS: ~10% active at same time ● and come up with a deployment plan that o let each active tenant exclusively use an MPPDB B. Reinwald. Multitenancy. University of Washington and Microsoft Research Summer Institute, 2010. Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 8
  • 9. ● R = 3, TDD creates 3 MPPDBs ● # of nodes of each MPPDB = the tenant that requests the max # of nodes Tenant Driven Design: A Toy Example ● 10 tenants: T1, T2, ... T10 Tenant T1 T2 T3 T4 T5 ... T10 Total Requested Nodes 6 6 5 5 2 ... 2 34 Group MPPDB0 MPPDB1 MPPDB2 Nodes 6 6 6 Hosting T1 - T10 T1 - T10 T1 - T10 TDD property: replication factor = A (# of active tenant in peak hour) Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 9
  • 10. T2 T2 Tenant Driven Design Query Routing - Example T2 T1 T5 00:00 GMT 24:00 GMT T1 Q1 Q3 Q2 Q4 T2 Q7 Q5 Q6 MPPDB1 MPPDB0 MPPDB2 Routing Principle: Let a tenant exclusively use a MPPDB! non-linear scale-out A=3 TDD's SLA Guarantee: NO matter what query, ad-hoc query (a new query that is seen the first time) linear scale-out query with template given submitted sequentially (interactive queries) submitted in a batch at any multi-programming-level the SLAs of a maximum of A active tenants are met. Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 10
  • 11. Scaling up TDD Realistic Example ● 5,000 tenants ● Average 10% of tenants active ● 500 active tenants 500+ Replicas !!! TDD guarantee: the SLAs of a maximum of A active tenants are met. TDD property: replication factor = A (# of active tenant in peak hour) Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 11 TDD guarantee: the SLAs of a maximum of 500 active tenants are met. TDD property: replication factor = 500 (# of active tenant in peak hour)
  • 12. Scaling up TDD: solution ● Further divide the tenants into groups ● Each group uses one TDD ● Considerations o minimize the total # of nodes used o reasonable replication factor  # active tenant in each tenant-group is low e.g. only 3 tenants active together e.g. divide 5000 tenant into 300 groups Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 12
  • 13. Scaling up TDD ● Formulate as o Largest Item Vector Bin Packing Problem with Fuzzy Capacity ● Devise effective heuristic solution called 2-step Grouping algorithm Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 13
  • 14. Thrifty - System Architecture Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 14
  • 15. Experimental Result -- Overview ● Setup: MulTe-like ● 5,000 tenants ● 2 - 32 nodes ● 200GB to 3.2TB ~89,300 nodes requested (without replica) saves 83.3% of nodes !!! ~15,000 nodes only including high availability already (3 replicas!) after consolidation Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 15
  • 16. Consolidation Effectiveness 86.5% 79.3% Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 16
  • 17. Thrifty - Prototype of MPPDB as a Service ● Shared-Process Consolidation ● provide High Availability ● guarantee Query Performance SLA ● support General Tenants Conclusions Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University 17
  • 18. Thank You any questions? Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University END
  • 19. Tenant-Driven Design (TDD) ● Cluster Design o divide the machine nodes in the cluster into A groups  each group of nodes deploys a separate MPPDB ● Tenant Placement o each MPPDB hosts all tenants ● Query Routing o route a tenant's query to a MPPDB that contains its dataTDD's property : enforces a replication factor of A for each tenant Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-1
  • 20. Consolidation Effectiveness - under Higher Active Tenant Ratio Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-2
  • 21. Heuristic Algorithm - 2-Step Grouping Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-3
  • 22. 2-Step Grouping vs. Standard Heuristics ● get higher consolidation ratio in all dimensions ● get result in few hours only for 10k tenants Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-4
  • 23. Lightweight Elastic Scaling ● Reactive approach o If the SLA is lower than 99.9% based in past 24hrs statistic, the system will trigger elastic scaling. o a new MPPDB will be online and the system will o deploy the tenant Tx to that MPPDB o which is most active in past 24hrs to. o all queries of the Tx will be route to the new MPPDB after its online. ● Example o three 10-node MPPDBs serves 20 tenants o SLA guarantee drops from 99.95% to 99.0%  solution: starts a new 10-node MPPDB ● cost 30 minutes for 1TB data (with parallel loading)Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-5
  • 24. Lightweight Elastic Scaling trigger finishtrigger finish Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-6
  • 25. Manual Tuning System administrator may instead use the parameter U to manually Use case ● tiny time percentage of possibly SLA violations o three 10-node MPPDBs o serves 20 tenants o SLA guarantee drops from 99.9% to 99.8%  solution 1: starts a new 10-node MPPDB (+10)  solution 2: increase 10-nodes to 11-nodes (+3) ● shorter query latency saves 7 nodes (17.5%) number of nodes in a MPPDB Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-7
  • 26. Linear and Non-linear Scale-out Queries - Concurrent Execution consolidation opportunity by merging clusters ● query template required SLA SLA Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-8
  • 27. Contribution ● Thrifty o a prototype implementation of MPPDBaaS o Tenant-driven Design (TDD)  Smartly organize the tenants into groups  Each group has R MPPDBs  Each MPPDB ● host a group of tenants  Prove that TDD solves R2 to R4 in a row! o Scale TDD to 10000+ of tenants and machine nodes (R1)?  Cast the problem as a new variant of vector bin packing problem  Provide an efficient heuristics algorithm [R = replication factor] R2. high availability R3. query latency SLA R4. generality Parallel Analytics as a Service Petrie Wong, Andy He, Eric Lo ,The Hong Kong Polytechnic University b-9

Editor's Notes

  1. Recently, massively parallel processing relational database systems (MPPDBs) like Vertica, Microsoft SQL Server Parallel Database Warehouse, and Greenplum have gained much momentum in the big data analytic market. These MPPDBs feature SQL interface, fast data loading, easy scale-out, fast recovery, and short analytical query processing time. become attractive for companies having analytical tasks on hundreds gigabytes to some ten terabytes of data . From the viewpoint of tenants who rent the service, they can enjoy par- allel analytics on their “big” data, but the (e.g., administration), and especially the software li- cense (which costs about USD 15K per core or USD 50K per TB of data for a commercial MPPDB we know) they pay for a share of the service are likely to be much lower than running everything themselves. From the viewpoint of the service providers, they can achieve a lower total cost of ownership by consolidating the ten- ants on to a shared hardware infrastructure, so that the service can be run using far fewer machine nodes than the sum of the number of the machine nodes requested by the tenants. When sharing re- sources among tenants, it is challenging to ensure the service level agreements (SLA) for the tenants are met after consolidation. Ide- ally, a tenant that rents a 4-node MPPDB should have the feeling that the queries are really executed by a MPPDB that runs on four dedicated machine nodes. Therefore, a service provider that offers MPPDB-as-a-service should regard the query latency before con- solidation as the performance SLA.
  2. One approach to offer database-as-a-service (DaaS) is based on the use of virtual machines (VM) [24]. The VM approach offers hard isolation among tenants and thus performance isolation among tenants could be easily achieved. However, the VM approach in- curs significant redundancy on the resources (e.g., multiple copies of the OS and DBMS, etc.) [8, 7]. Therefore, the recent trend for DaaS is to use the shared-process multi-tenancy [13] model (i.e., sharing the same database process among tenants). Relational Cloud [8, 7] and SQL Azure [4] are using this model so that a much higher consolidation effectiveness could be obtained. But so far they focus on tenants with transactional workloads.
  3. Transactional (OLTP) workloads generally touch a small amount of data per transaction. So, OLTP-database-as-a-service using the shared-process model could still support a large number of concur- rent query executions after consolidation [8, 7]. In contrast, analytical workloads are I/O-intensive, so concurrent query executions on the same database instance easily increase the query latency and violate the SLA. In Figure 1a, the lines 2T-CON and 4T-CON show the speedup of Q1 (with respect to 1-node) when there are re- spectively two tenants and four tenants sharing the same database instance, and the tenants submit Q1 together. We can see that Q1’s performance are 2× (in 2-tenant setting) and 4× slower (in 4-tenant setting). That is, once the tenants submit queries together, none of their SLAs could be met. This is a challenge of using the shared- process multi-tenancy model to serve analytical workloads.
  4. R1 and R2 are two general cloud computing requirements. R3 is a DaaS requirement specific to analytical workloads.1 R4 is spe- cific to MPPDBaaS and R5 is necessary if we want to offer a real MPPDBaaS. These requirements together make the design and im- plementation of Thrifty very challenging. For example, sharing the database process among tenants (to enjoy higher consolidation ra- tio) makes R3 challenging because there is no more hard isolation. While performance prediction techniques for concurrent analytical workload [9, 1, 21, 16, 2] may help there on the surface, these techniques require the knowledge of the query sets as input, which contradicts with R5 above.
  5. (a) Tenant Activity Monitor: The Tenant Activity Monitor auto- matically collects the query logs of the deployed MPPDBs, derives the tenant activities, and summarizes the query char- acteristics of individual tenants. For example, it continuously monitors the active tenant ratio of all tenants in the past 30 days. These information is passed to the Deployment Advisor for further processing, and is available to the system adminis- trator for advanced system tuning. (b) Deployment Advisor: The Deployment Advisor takes as inputs the tenant activity statistics, the individual tenant information (e.g., number of nodes requested), a replication factor R (spec- ified by a system administrator for high availability), and a per- formance SLA guarantee P (also specified by the system ad- ministrator which ensures that, after consolidation, the tenants can still meet their SLA for a P %, e.g., 99.9%, of time). It re- turns as output a deployment plan. The deployment plan con- sists of two parts: cluster design and tenant placement. The cluster design details how the machine nodes in the cluster are arranged into groups. Each group of nodes then runs a single MPPDB instance. The tenant placement specifies which MP- PDB instances a tenant should be deployed on; and a tenant is deployed on R MPPDB instances. At run-time, the Deploy- ment Advisor may fine-tune the deployment plan according to the live-updated tenant activities supplied by the Tenant Activ- ity Monitor. Currently, Thrifty assumes all nodes in the cluster are identical in configurations (e.g., CPU, RAM). (c) Deployment Master: The Deployment Master follows the de- ployment plan devised by the Deployment Advisor to start the MPPDB instances and deploy the tenants onto them. It also switches off/hibernates nodes that are not listed in the deploy- ment plan. The deployment is supposed to be static for days. A (re)-consolidation process is expected to be executed period- ically, because it is expected that there are new tenants register with and existing tenants de-register with the service. (d) Query Router: The Query Router accepts queries from tenants and routes the queries to the proper MPPDB instance according to the tenant placement. The crux of Thrifty is to devise a good deployment plan to serve thousands of tenants using fewer machine nodes. Thrifty targets on tenants with temporal skew [19] analytical workloads, e.g, tenants from different time zones and mostly submit queries within a cer- tain period (e.g., 9am to 5pm) but not in a 24 × 7 fashion. That type of temporal skew workload is not uncommon in today’s data analytical market [5]. The target tenants are supposed to have up to only some tens terabytes of data but hope to enjoy the paral- lel speedup offered by the expensive MPPDB products. These are what enable consolidation. Tenants that are always active and/or with more than terabytes of data could be detected by Thrifty and they will be excluded from consolidation.2
  6. At run-time, tenants’ activities may deviate from the history. Consider a tenant-group having a TTP of 99.95% during tenant- grouping. At run-time, some tenants in that group may become more active than what the past tenant activities have indicated, and that will result in a lower run-time TTP (RT-TTP). When the RT- TTP drops to, say 98%, that implies there is 2% of time the tenants are not exclusively served by dedicated MPPDBs and some of their queries may not meet the SLA. Thrifty handles this issue using a standard elastic scaling approach as in other cloud systems. The Tenant Activity Monitor monitors the RT-TTP of each tenant-group using a time window (e.g., 24-hour). When it detects the RT-TTP of a tenant-group of the past 24 hours has just dropped below the performance SLA guarantee P , Thrifty will call the Deployment Advisor to take action. One possible action is to elastically scale up A by one, i.e., adding one more MPPDB. In MPPDBaaS, however, scaling up A is not as simple and efficient as scaling resources (e.g., RAM) in VM-based cloud environment. Specifically, it takes hours to start a new MPPDB and bulk load all tenants’ data before it is ready to use. Table 1 shows the time of starting up a MPPDB and bulk load- ing tenant data into a MPPDB in our environment. We see that the data loading time dominates the times of starting the machines and creating the MPPDB instances, although it has already achieved a reasonable loading rate (about 1.2GB/min). Take starting a 10- node MPPDB that contains 1TB of tenant data as an example. Af- ter detecting the RT-TTP of a tenant-group of the past 24 hours has dropped below the, say, 99.9% performance SLA guarantee thresh- old, Thrifty needs about 14.5 hours (50446s+1779s) to prepare the new MPPDB. Assume that the RT-TTP of the tenant group is con- sistently 98% during those 14.5 hours, that would result in about 17.4 minutes (14.5 hours × 2%) having concurrent query process- ing at MPPDB0 . Consider a reasonable time frame of, say, one month, Thrifty has about 43 minutes (1 month × 0.1%) of “grace period” for not meeting the SLA guarantee of a tenant group.3 So, that grace period is barely enough to tolerate the scaling of A twice. Therefore, we look for a more lightweight approach to implement the elastic scale-up operation such that we can tolerate more scale- up operations or possibly other performance-influencing operations (e.g., node failure) in the grace period. Our lightweight approach is to identify the “over-active” ten- ant(s) in that tenant group and add a new MPPDB to serve only those tenant(s), instead of adding a MPPDB that serves the whole tenant-group. We see from Table 1 the data loading time scales linearly with the data size. As the data size of a tenant in a tenant- group must always be less than the data size of all tenants in that group, the elastic scale-up operation that starts a new MPPDB and loads data for only the over-active tenants would be more effi- cient. Specifically, on detecting the RT-TTP of a tenant-group of the past 24 hours has dropped below P , Thrifty will run an over- active-tenant-identification algorithm to identify the tenant(s) that are more active than the history indicated, and create a new MP- PDB to serve them. The over-active-tenant-identification algorithm is similar to the tenant-grouping algorithm presented in Algorithm 2, with the only change that takes only the tenants of a particular tenant-group as inputs. It is expected some tenant(s) cannot join the same tenant group anymore, and they are identified as over-active and their data are bulk loaded4 to a new MPPDB. When the new MPPDB is ready, the Deployment Advisor will notify the Query Router to route queries from the over-active tenant(s) to the new MPPDB. For any tenant-group that has ever gone through the above elas- tic scaling process, Thrifty will add tenants in those tenant-groups to a re-consolidation list. Those tenants, together with new tenants, over-active tenants, and tenants in tenant-groups with de-registered tenants, will get re-consolidated in the next (re)-consolidation cy- cle. Finally, tenants with regular bursts in tenant activity (e.g., there are usually bursts near the end of a fiscal year) could be identified by Thrifty’s regular activity monitoring and they would be excluded from consolidation before the bursts arrive. We regard our approach as a reactive approach to maintain the SLA performance guarantee. A pessimistic approach is to use A + 1 MPPDBs upfront to serve a tenant group, which is not cost- effective. A proactive approach is to predict at run-time whether the RT-TTP will soon drop below P and proactively trigger lightweight elastic scaling if so. That approach, however, is subjected to pre- diction error and spikes (e.g, sharp drop of RT-TTP followed by sharp rise) in tenant activities.
  7. The elastic scaling operation of Thrifty automatically adds a MP- PDB to serve the over-active tenants when the RT-TTP of a tenant- group of the past 24 hours stays below the SLA performance guar- antee P . That MPPDB will be there until the next re-consolidation cycle (to minimize the overhead of scaling up and down). A system administrator may override this action. Consider a performance SLA guarantee of 99.9%, and a tenant-group of tenants served by three 10-node MPPDBs (A = R = 3). If there is a minor in- crease of active tenant ratio of that group, resulting in a RT-TTP of, say, 99.8%, in the past 24 hours, and if a system administra- tor realizes from the Tenant Activity Monitor that the RT-TTP of that group is not continuously dropping but stays flat, she may think that Thrifty’s elastic scaling that starts a new 10-node MP- PDB for the over-active tenants is overkilling (because there is only 99.9%−99.8%=0.1% of tiny time percentage of possibly SLA vio- lations). In situations like this, the system administrator may instead use the parameter U to manually tune the performance SLA guarantee of that tenant-group. Recall that the tenant-driven design reserves a parameter U to control the number of nodes used by the MPPDB0 (Section 3). In our example above, the system administrator may override Thrifty’s elastic scaling but increase U from 10 to, say, 12. Recall that the query routing algorithm of Thrifty routes a query to MPPDB0 if all MPPDBs are busy (Algorithm 1 Line 10). When- ever the 0.2% of time the fourth (or the fifth...) active tenant sub- mits a query, that is routed to MPPDB0 for concurrent processing. So, it is possible that the 99.9% SLAs can be met empirically by having a higher degree of parallelism and more computation power (U +2) in MPPDB0 (like Point C in Figure 1b).
  8. The second consolidation opportunity can be observed by look- ing at Figure 1b, which shows the query latency of Q1. First, as- sume there are four tenants, each rents a 2-node MPPDB to process a TPC-H scale factor 100 dataset. The service provider basically re- quires a total of 4×2 = 8 separate nodes to host them. The SLA of TPC-H Q1 for each tenant is then A seconds, as illustrated in Fig- ure 1b. What if the service provider uses a 6-node MPPDB to host all four tenants (i.e., every tenant partitions their data on to those six nodes)? Let us assume that only one out of the four tenant is active (e.g., submit a Q1 instance) and the rest of them are inactive, i.e., not submitting any query. That active tenant then can obtain the result of Q1 in B seconds — the SLA is met. When two out of the four tenants are active together and they concurrently submit an instance of Q1 to the 6-node MPPDB, each tenant can obtain the result of Q1 in C seconds — their SLAs are also met. The above shows that it is a consolidation opportunity that is more specific to parallel analytic workload.