SlideShare uma empresa Scribd logo
1 de 38
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Murtaza Doctor Principal Architect
Giang Nguyen Sr. Software Engineer
How we solved
Real-Time User
Segmentation
using HBase ?
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Outline
• {rr} Story
• {rr} Personalization Platform
• User Segmentation: Problem Statement
• Design & Architecture
• Performance Metrics
• Q&A
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Multiple Personalization Placements
Personalized Product
Recommendations Targeted Promotions Relevant Brand Advertising
Inventory and
Margin Data
Catalog
Attribute Data
Real time user
behavior
Multi-Channel
purchase
history
3rd Party Data
Sources
Future Data
Sources
Input
Data
100+ algorithms dynamically targeted by
page, context, and user behavior, across
all retail channels.
Targeted content optimization to
customer segments
Monetization through relevant
brand advertising
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
 One place where all data
about your customer lives
(e.g. loyalty, customer
service, offline, catalogue)
 Direct access to data for
your entire enterprise via API
 Real-time actionable data
and business intelligence
 Link customer activity across
any channel
 Leverage 3rd party data (e.g.
weather, geography, census)
Delivering a
Customer-centric
Omni-channel
Data model
RichRelevanceDataMesh
Real-time
Segmentation
DMP
Integration
RichRelevance DataMesh Cloud Platform
Delivering a Single View of your Customer
Event-based
Triggers
Ad Hoc
Reporting
Omni-channel
Personalization
Loyalty
Integration
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Our cloud-based platform supports both real-time processes
and analytical use cases, utilizing technologies to name a
few: Crunch, Hive, HBase, Avro, Azkaban, Voldemort, Kafka
Someone clicks on a {rr} recommendation
every 21 milliseconds
Did You Know?
Our data capacity includes a 1.5 PB Hadoop infrastructure,
which enables us to employ 100+ algorithms in real-time
In the US, we serve 7000 requests per second with an average
response time of 50 ms
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Real-Time User Segmentation
Finding and Targeting the Right Audience
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Meet Amanda and Jessica
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Because We Know What They Like!
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
• Utilizes valuable targeting
data such as event logs,
off-line data, DMA,
demographics, and etc.
• Finds highly qualified
consumers and buckets
them into segments
What is the {rr} Segment Builder?
Example: for Retail, Segment Builder supports view,
click, purchase on products, categories, and brands
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Segment Builder
Segment’s List page
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Segment Builder
Add/Edit Segment page
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
• Create segments to capture the audience via UI
• Each behavior is captured by a rule
• Each rule corresponds to a row key in HBase
• Each rule returns the users
• Rules are joined using set union or intersection
• Segment definition consists of one or more rules
Design: Segment Evaluation Engine
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Our choice Avro and Kafka
Real-Time Data Ingestion
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
• User interacts with a retail site
• Events are triggered in real-
time, each event is converted
into an Avro record
• Events are ingested using our
real-time framework built using
Apache Kafka
Real-time data ingestion
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Design Principles: Real-Time Solution
• Streaming: Support for streaming versus batch
• Reliability: no loss of events and at least once
delivery of messages
• Scalable: add more brokers, add more consumers,
partition data
• Distributed system with central coordinator like
zookeeper
• Persistence of messages
• Push & pull mechanism
• Support for compression
• Low operational complexity & easy to monitor
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Our Decision – Apache Kafka
• Distributed pub-sub messaging system
• More generic concept (Ingest & Publish)
• Can support both offline & online use-cases
• Designed for persistent messages as the common
case
• Guarantee ordering of events
• Supports gzip & snappy (0.8) compression protocols
• Supports rewind offset & re-consumption of data
• No concept of master node, brokers are peers
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Kafka Architecture
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Common Consumer Framework
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Volume Facts?
Daily clickstream event data 150 GB
Average size of message 2 KB
Batch Size 5000 messages
Producer throughput 5000 messages/sec
Real time HBase Consumer 7000 messages/sec
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
End to End real-time story…
• User exhibits a behavior
• Events are generated at front-end data-center
• Events are streamed to backend data center
via Kafka
• Events are consumed and indexed in HBase
tables
• Takes seconds from event origination to
indexing in HBase
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
End to End real-time story…
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
User Segmentation Engines
Features {rr} Engine Other Engines
User’s behavior ingestion Real-time Not Real-time
Batch style processing is done Immediately At end of a day
When segment membership is changed
notifications will be
Event driven N/A
Technologies used Scalable and
open source
Unscalable and
proprietary
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Use Cases
Use Case #1
Users exhibit behaviors
Behaviors ingested and indexed in
real time
Users are now in corresponding
segments
Retrieving users takes seconds
Use Case #2
Users exhibit behaviors
Segment membership calculated in
real time
Notifications are sent on segment
membership change
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Our choice HBase?
• Real time segment evaluation
• Optimized for read and scan
• Column cell supports frequency use case
• Eventual consistency does not work
• Seamless integration with Hadoop
• Possible with good row key design
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
HBase Row Key Design
• Took a few attempts
• Design considerations
– Timestamp in row or columns
– Partition behavior by date
– Optimized for read or write
– Hot spotting issues
– Uniform key distribution
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Design: First Attempt
• Row key represents behavior
• Columns store the user id
• Cell stores behavior time and capture frequency
• One column family U
RowKey Columns
338VBChanel 23b93laddf82:1370377143973
Hd92jslahd0a:1313323414192
338CCElectronic z3be3la2dfa2:1370477142970
kd9zjsla3d01:1313323414192
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Design: First Attempt Issues
• Row too wide
• May exceed HFile size
• Terrible write/read performance
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
• Partition behavior by date
• Reduce row size
• Gained ability to scan across dates
Design: Second Attempt
Rowkey Columns
338VBChanel1370377143 23b93laddf82:1370377143973
Hd92jslahd0a:1313323414192
338CCElectronic1370377143 z3be3la2dfa2:1370477142970
kd9zjsla3d01:1313323414192
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Design: Second Attempt Issues
• Hot spotting
• Popular products or high level categories can
have millions of users, each day
• One region serving same dimension type
• Terrible write/read performance
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
OK…I need a BREAK!!!
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Design: Final
• Shard row to prevent hot-spotting
• Shard into N number of regions
• Significant improvement in read/write
• Prepend a salt to each key
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
[salt]_len_[siteId]_len_[metric]_len_[dimension]_len_[value][timestamp]
Design: Final
_len_ is the integer length of the field following it
timestamp is stored in day granularity
[salt] is computed by first creating a hash from the
siteId, metric, and dimension, then combining this
with a random number between 0 and N (number
of shards) for sharding
Row key contains
attribute value siteId metric attribute timestamp userGUID
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
• Complex segments contain many rules
• Each rule = one behavior = one row key
• Each row key returns a set of users
• OR = Full outer join
• AND = Inner join
• Done in memory for small rules
• Merged on disk for large rules
Design: Behavior Joins
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
HBase Consumer Sync versus Async API
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Segmentation Performance
• Seconds latency
• 40K puts/sec over 2 tables, 8 regions per table
• Scaling achieved through addition of regions
• Small segments calculated in msecs
• Mid-size segments in seconds
• Large segments calculated in 10s of seconds
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
© 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
Thank You

Mais conteúdo relacionado

Mais procurados

Mais procurados (20)

The columnar roadmap: Apache Parquet and Apache Arrow
The columnar roadmap: Apache Parquet and Apache ArrowThe columnar roadmap: Apache Parquet and Apache Arrow
The columnar roadmap: Apache Parquet and Apache Arrow
 
Big Query Basics
Big Query BasicsBig Query Basics
Big Query Basics
 
Greenplum and Kafka: Real-time Streaming to Greenplum - Greenplum Summit 2019
Greenplum and Kafka: Real-time Streaming to Greenplum - Greenplum Summit 2019Greenplum and Kafka: Real-time Streaming to Greenplum - Greenplum Summit 2019
Greenplum and Kafka: Real-time Streaming to Greenplum - Greenplum Summit 2019
 
Apache Kafka Architecture & Fundamentals Explained
Apache Kafka Architecture & Fundamentals ExplainedApache Kafka Architecture & Fundamentals Explained
Apache Kafka Architecture & Fundamentals Explained
 
Apache Airflow
Apache AirflowApache Airflow
Apache Airflow
 
Apache Airflow
Apache AirflowApache Airflow
Apache Airflow
 
Batch Processing at Scale with Flink & Iceberg
Batch Processing at Scale with Flink & IcebergBatch Processing at Scale with Flink & Iceberg
Batch Processing at Scale with Flink & Iceberg
 
Getting Started with Apache Spark on Kubernetes
Getting Started with Apache Spark on KubernetesGetting Started with Apache Spark on Kubernetes
Getting Started with Apache Spark on Kubernetes
 
Microservices Integration Patterns with Kafka
Microservices Integration Patterns with KafkaMicroservices Integration Patterns with Kafka
Microservices Integration Patterns with Kafka
 
Practical learnings from running thousands of Flink jobs
Practical learnings from running thousands of Flink jobsPractical learnings from running thousands of Flink jobs
Practical learnings from running thousands of Flink jobs
 
Apache Iceberg: An Architectural Look Under the Covers
Apache Iceberg: An Architectural Look Under the CoversApache Iceberg: An Architectural Look Under the Covers
Apache Iceberg: An Architectural Look Under the Covers
 
InnoDB MVCC Architecture (by 권건우)
InnoDB MVCC Architecture (by 권건우)InnoDB MVCC Architecture (by 권건우)
InnoDB MVCC Architecture (by 권건우)
 
Spark + S3 + R3를 이용한 데이터 분석 시스템 만들기
Spark + S3 + R3를 이용한 데이터 분석 시스템 만들기Spark + S3 + R3를 이용한 데이터 분석 시스템 만들기
Spark + S3 + R3를 이용한 데이터 분석 시스템 만들기
 
Presto overview
Presto overviewPresto overview
Presto overview
 
Dive into PySpark
Dive into PySparkDive into PySpark
Dive into PySpark
 
kafka
kafkakafka
kafka
 
Kafka 101
Kafka 101Kafka 101
Kafka 101
 
Building Better Data Pipelines using Apache Airflow
Building Better Data Pipelines using Apache AirflowBuilding Better Data Pipelines using Apache Airflow
Building Better Data Pipelines using Apache Airflow
 
A Deep Dive into Kafka Controller
A Deep Dive into Kafka ControllerA Deep Dive into Kafka Controller
A Deep Dive into Kafka Controller
 
Dynamically Scaling Data Streams across Multiple Kafka Clusters with Zero Fli...
Dynamically Scaling Data Streams across Multiple Kafka Clusters with Zero Fli...Dynamically Scaling Data Streams across Multiple Kafka Clusters with Zero Fli...
Dynamically Scaling Data Streams across Multiple Kafka Clusters with Zero Fli...
 

Semelhante a How we solved Real-time User Segmentation using HBase

Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
Greg Makowski
 
Hello Streams Overview
Hello Streams OverviewHello Streams Overview
Hello Streams Overview
psanet
 

Semelhante a How we solved Real-time User Segmentation using HBase (20)

HBaseCon 2013: Realtime User Segmentation using Apache HBase -- Architectural...
HBaseCon 2013: Realtime User Segmentation using Apache HBase -- Architectural...HBaseCon 2013: Realtime User Segmentation using Apache HBase -- Architectural...
HBaseCon 2013: Realtime User Segmentation using Apache HBase -- Architectural...
 
OC Big Data Monthly Meetup #6 - Session 1 - IBM
OC Big Data Monthly Meetup #6 - Session 1 - IBMOC Big Data Monthly Meetup #6 - Session 1 - IBM
OC Big Data Monthly Meetup #6 - Session 1 - IBM
 
SD Big Data Monthly Meetup #4 - Session 1 - IBM
SD Big Data Monthly Meetup #4 - Session 1 - IBMSD Big Data Monthly Meetup #4 - Session 1 - IBM
SD Big Data Monthly Meetup #4 - Session 1 - IBM
 
HBaseCon 2013: Evolving a First-Generation Apache HBase Deployment to Second...
HBaseCon 2013:  Evolving a First-Generation Apache HBase Deployment to Second...HBaseCon 2013:  Evolving a First-Generation Apache HBase Deployment to Second...
HBaseCon 2013: Evolving a First-Generation Apache HBase Deployment to Second...
 
HBaseCon 2013: ETL for Apache HBase
HBaseCon 2013: ETL for Apache HBaseHBaseCon 2013: ETL for Apache HBase
HBaseCon 2013: ETL for Apache HBase
 
Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
Powering Real­time Decision Engines in Finance and Healthcare using Open Sour...
 
Gcp intro-20160721
Gcp intro-20160721Gcp intro-20160721
Gcp intro-20160721
 
Hello Streams Overview
Hello Streams OverviewHello Streams Overview
Hello Streams Overview
 
GPSBUS214-Key Considerations for Cloud Procurement in the Public Sector
GPSBUS214-Key Considerations for Cloud Procurement in the Public SectorGPSBUS214-Key Considerations for Cloud Procurement in the Public Sector
GPSBUS214-Key Considerations for Cloud Procurement in the Public Sector
 
Choosing the Right Database for My Workload: Purpose-Built Databases
Choosing the Right Database for My Workload: Purpose-Built Databases Choosing the Right Database for My Workload: Purpose-Built Databases
Choosing the Right Database for My Workload: Purpose-Built Databases
 
Non-Relational Revolution: Database Week SF
Non-Relational Revolution: Database Week SFNon-Relational Revolution: Database Week SF
Non-Relational Revolution: Database Week SF
 
Non-Relational Revolution - Joseph Idziorek
Non-Relational Revolution - Joseph IdziorekNon-Relational Revolution - Joseph Idziorek
Non-Relational Revolution - Joseph Idziorek
 
Breaking Down the 'Monowhat'
Breaking Down the 'Monowhat'Breaking Down the 'Monowhat'
Breaking Down the 'Monowhat'
 
MBL204_Architecting Cost-Effective Mobile Backends for Scale, Security, and P...
MBL204_Architecting Cost-Effective Mobile Backends for Scale, Security, and P...MBL204_Architecting Cost-Effective Mobile Backends for Scale, Security, and P...
MBL204_Architecting Cost-Effective Mobile Backends for Scale, Security, and P...
 
Non-Relational Revolution
Non-Relational RevolutionNon-Relational Revolution
Non-Relational Revolution
 
Implementing Advanced Analytics Platform
Implementing Advanced Analytics PlatformImplementing Advanced Analytics Platform
Implementing Advanced Analytics Platform
 
Best Practices for Monitoring Cloud Networks
Best Practices for Monitoring Cloud NetworksBest Practices for Monitoring Cloud Networks
Best Practices for Monitoring Cloud Networks
 
Datastax - The Architect's guide to customer experience (CX)
Datastax - The Architect's guide to customer experience (CX)Datastax - The Architect's guide to customer experience (CX)
Datastax - The Architect's guide to customer experience (CX)
 
Microservices: Data & Design - Miguel Cervantes
Microservices: Data & Design - Miguel CervantesMicroservices: Data & Design - Miguel Cervantes
Microservices: Data & Design - Miguel Cervantes
 
Vistara 3.1 - Delivering Unified IT Operations
Vistara 3.1 - Delivering Unified IT OperationsVistara 3.1 - Delivering Unified IT Operations
Vistara 3.1 - Delivering Unified IT Operations
 

Mais de DataWorks Summit

HBase Global Indexing to support large-scale data ingestion at Uber
HBase Global Indexing to support large-scale data ingestion at UberHBase Global Indexing to support large-scale data ingestion at Uber
HBase Global Indexing to support large-scale data ingestion at Uber
DataWorks Summit
 
Security Framework for Multitenant Architecture
Security Framework for Multitenant ArchitectureSecurity Framework for Multitenant Architecture
Security Framework for Multitenant Architecture
DataWorks Summit
 
Computer Vision: Coming to a Store Near You
Computer Vision: Coming to a Store Near YouComputer Vision: Coming to a Store Near You
Computer Vision: Coming to a Store Near You
DataWorks Summit
 

Mais de DataWorks Summit (20)

Data Science Crash Course
Data Science Crash CourseData Science Crash Course
Data Science Crash Course
 
Floating on a RAFT: HBase Durability with Apache Ratis
Floating on a RAFT: HBase Durability with Apache RatisFloating on a RAFT: HBase Durability with Apache Ratis
Floating on a RAFT: HBase Durability with Apache Ratis
 
Tracking Crime as It Occurs with Apache Phoenix, Apache HBase and Apache NiFi
Tracking Crime as It Occurs with Apache Phoenix, Apache HBase and Apache NiFiTracking Crime as It Occurs with Apache Phoenix, Apache HBase and Apache NiFi
Tracking Crime as It Occurs with Apache Phoenix, Apache HBase and Apache NiFi
 
HBase Tales From the Trenches - Short stories about most common HBase operati...
HBase Tales From the Trenches - Short stories about most common HBase operati...HBase Tales From the Trenches - Short stories about most common HBase operati...
HBase Tales From the Trenches - Short stories about most common HBase operati...
 
Optimizing Geospatial Operations with Server-side Programming in HBase and Ac...
Optimizing Geospatial Operations with Server-side Programming in HBase and Ac...Optimizing Geospatial Operations with Server-side Programming in HBase and Ac...
Optimizing Geospatial Operations with Server-side Programming in HBase and Ac...
 
Managing the Dewey Decimal System
Managing the Dewey Decimal SystemManaging the Dewey Decimal System
Managing the Dewey Decimal System
 
Practical NoSQL: Accumulo's dirlist Example
Practical NoSQL: Accumulo's dirlist ExamplePractical NoSQL: Accumulo's dirlist Example
Practical NoSQL: Accumulo's dirlist Example
 
HBase Global Indexing to support large-scale data ingestion at Uber
HBase Global Indexing to support large-scale data ingestion at UberHBase Global Indexing to support large-scale data ingestion at Uber
HBase Global Indexing to support large-scale data ingestion at Uber
 
Scaling Cloud-Scale Translytics Workloads with Omid and Phoenix
Scaling Cloud-Scale Translytics Workloads with Omid and PhoenixScaling Cloud-Scale Translytics Workloads with Omid and Phoenix
Scaling Cloud-Scale Translytics Workloads with Omid and Phoenix
 
Building the High Speed Cybersecurity Data Pipeline Using Apache NiFi
Building the High Speed Cybersecurity Data Pipeline Using Apache NiFiBuilding the High Speed Cybersecurity Data Pipeline Using Apache NiFi
Building the High Speed Cybersecurity Data Pipeline Using Apache NiFi
 
Supporting Apache HBase : Troubleshooting and Supportability Improvements
Supporting Apache HBase : Troubleshooting and Supportability ImprovementsSupporting Apache HBase : Troubleshooting and Supportability Improvements
Supporting Apache HBase : Troubleshooting and Supportability Improvements
 
Security Framework for Multitenant Architecture
Security Framework for Multitenant ArchitectureSecurity Framework for Multitenant Architecture
Security Framework for Multitenant Architecture
 
Presto: Optimizing Performance of SQL-on-Anything Engine
Presto: Optimizing Performance of SQL-on-Anything EnginePresto: Optimizing Performance of SQL-on-Anything Engine
Presto: Optimizing Performance of SQL-on-Anything Engine
 
Introducing MlFlow: An Open Source Platform for the Machine Learning Lifecycl...
Introducing MlFlow: An Open Source Platform for the Machine Learning Lifecycl...Introducing MlFlow: An Open Source Platform for the Machine Learning Lifecycl...
Introducing MlFlow: An Open Source Platform for the Machine Learning Lifecycl...
 
Extending Twitter's Data Platform to Google Cloud
Extending Twitter's Data Platform to Google CloudExtending Twitter's Data Platform to Google Cloud
Extending Twitter's Data Platform to Google Cloud
 
Event-Driven Messaging and Actions using Apache Flink and Apache NiFi
Event-Driven Messaging and Actions using Apache Flink and Apache NiFiEvent-Driven Messaging and Actions using Apache Flink and Apache NiFi
Event-Driven Messaging and Actions using Apache Flink and Apache NiFi
 
Securing Data in Hybrid on-premise and Cloud Environments using Apache Ranger
Securing Data in Hybrid on-premise and Cloud Environments using Apache RangerSecuring Data in Hybrid on-premise and Cloud Environments using Apache Ranger
Securing Data in Hybrid on-premise and Cloud Environments using Apache Ranger
 
Big Data Meets NVM: Accelerating Big Data Processing with Non-Volatile Memory...
Big Data Meets NVM: Accelerating Big Data Processing with Non-Volatile Memory...Big Data Meets NVM: Accelerating Big Data Processing with Non-Volatile Memory...
Big Data Meets NVM: Accelerating Big Data Processing with Non-Volatile Memory...
 
Computer Vision: Coming to a Store Near You
Computer Vision: Coming to a Store Near YouComputer Vision: Coming to a Store Near You
Computer Vision: Coming to a Store Near You
 
Big Data Genomics: Clustering Billions of DNA Sequences with Apache Spark
Big Data Genomics: Clustering Billions of DNA Sequences with Apache SparkBig Data Genomics: Clustering Billions of DNA Sequences with Apache Spark
Big Data Genomics: Clustering Billions of DNA Sequences with Apache Spark
 

Último

Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
WSO2
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Victor Rentea
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
panagenda
 

Último (20)

EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
Architecting Cloud Native Applications
Architecting Cloud Native ApplicationsArchitecting Cloud Native Applications
Architecting Cloud Native Applications
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
Navigating the Deluge_ Dubai Floods and the Resilience of Dubai International...
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontology
 
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
Biography Of Angeliki Cooney | Senior Vice President Life Sciences | Albany, ...
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
Platformless Horizons for Digital Adaptability
Platformless Horizons for Digital AdaptabilityPlatformless Horizons for Digital Adaptability
Platformless Horizons for Digital Adaptability
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 

How we solved Real-time User Segmentation using HBase

  • 1. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Murtaza Doctor Principal Architect Giang Nguyen Sr. Software Engineer How we solved Real-Time User Segmentation using HBase ?
  • 2. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Outline • {rr} Story • {rr} Personalization Platform • User Segmentation: Problem Statement • Design & Architecture • Performance Metrics • Q&A
  • 3. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
  • 4. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Multiple Personalization Placements Personalized Product Recommendations Targeted Promotions Relevant Brand Advertising Inventory and Margin Data Catalog Attribute Data Real time user behavior Multi-Channel purchase history 3rd Party Data Sources Future Data Sources Input Data 100+ algorithms dynamically targeted by page, context, and user behavior, across all retail channels. Targeted content optimization to customer segments Monetization through relevant brand advertising
  • 5. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential.  One place where all data about your customer lives (e.g. loyalty, customer service, offline, catalogue)  Direct access to data for your entire enterprise via API  Real-time actionable data and business intelligence  Link customer activity across any channel  Leverage 3rd party data (e.g. weather, geography, census) Delivering a Customer-centric Omni-channel Data model RichRelevanceDataMesh Real-time Segmentation DMP Integration RichRelevance DataMesh Cloud Platform Delivering a Single View of your Customer Event-based Triggers Ad Hoc Reporting Omni-channel Personalization Loyalty Integration
  • 6. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Our cloud-based platform supports both real-time processes and analytical use cases, utilizing technologies to name a few: Crunch, Hive, HBase, Avro, Azkaban, Voldemort, Kafka Someone clicks on a {rr} recommendation every 21 milliseconds Did You Know? Our data capacity includes a 1.5 PB Hadoop infrastructure, which enables us to employ 100+ algorithms in real-time In the US, we serve 7000 requests per second with an average response time of 50 ms
  • 7. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Real-Time User Segmentation Finding and Targeting the Right Audience
  • 8. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Meet Amanda and Jessica
  • 9. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Because We Know What They Like!
  • 10. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. • Utilizes valuable targeting data such as event logs, off-line data, DMA, demographics, and etc. • Finds highly qualified consumers and buckets them into segments What is the {rr} Segment Builder? Example: for Retail, Segment Builder supports view, click, purchase on products, categories, and brands
  • 11. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Segment Builder Segment’s List page
  • 12. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Segment Builder Add/Edit Segment page
  • 13. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. • Create segments to capture the audience via UI • Each behavior is captured by a rule • Each rule corresponds to a row key in HBase • Each rule returns the users • Rules are joined using set union or intersection • Segment definition consists of one or more rules Design: Segment Evaluation Engine
  • 14. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Our choice Avro and Kafka Real-Time Data Ingestion
  • 15. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. • User interacts with a retail site • Events are triggered in real- time, each event is converted into an Avro record • Events are ingested using our real-time framework built using Apache Kafka Real-time data ingestion
  • 16. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Design Principles: Real-Time Solution • Streaming: Support for streaming versus batch • Reliability: no loss of events and at least once delivery of messages • Scalable: add more brokers, add more consumers, partition data • Distributed system with central coordinator like zookeeper • Persistence of messages • Push & pull mechanism • Support for compression • Low operational complexity & easy to monitor
  • 17. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Our Decision – Apache Kafka • Distributed pub-sub messaging system • More generic concept (Ingest & Publish) • Can support both offline & online use-cases • Designed for persistent messages as the common case • Guarantee ordering of events • Supports gzip & snappy (0.8) compression protocols • Supports rewind offset & re-consumption of data • No concept of master node, brokers are peers
  • 18. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Kafka Architecture
  • 19. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Common Consumer Framework
  • 20. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Volume Facts? Daily clickstream event data 150 GB Average size of message 2 KB Batch Size 5000 messages Producer throughput 5000 messages/sec Real time HBase Consumer 7000 messages/sec
  • 21. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. End to End real-time story… • User exhibits a behavior • Events are generated at front-end data-center • Events are streamed to backend data center via Kafka • Events are consumed and indexed in HBase tables • Takes seconds from event origination to indexing in HBase
  • 22. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. End to End real-time story…
  • 23. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. User Segmentation Engines Features {rr} Engine Other Engines User’s behavior ingestion Real-time Not Real-time Batch style processing is done Immediately At end of a day When segment membership is changed notifications will be Event driven N/A Technologies used Scalable and open source Unscalable and proprietary
  • 24. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Use Cases Use Case #1 Users exhibit behaviors Behaviors ingested and indexed in real time Users are now in corresponding segments Retrieving users takes seconds Use Case #2 Users exhibit behaviors Segment membership calculated in real time Notifications are sent on segment membership change
  • 25. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Our choice HBase? • Real time segment evaluation • Optimized for read and scan • Column cell supports frequency use case • Eventual consistency does not work • Seamless integration with Hadoop • Possible with good row key design
  • 26. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. HBase Row Key Design • Took a few attempts • Design considerations – Timestamp in row or columns – Partition behavior by date – Optimized for read or write – Hot spotting issues – Uniform key distribution
  • 27. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Design: First Attempt • Row key represents behavior • Columns store the user id • Cell stores behavior time and capture frequency • One column family U RowKey Columns 338VBChanel 23b93laddf82:1370377143973 Hd92jslahd0a:1313323414192 338CCElectronic z3be3la2dfa2:1370477142970 kd9zjsla3d01:1313323414192
  • 28. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Design: First Attempt Issues • Row too wide • May exceed HFile size • Terrible write/read performance
  • 29. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. • Partition behavior by date • Reduce row size • Gained ability to scan across dates Design: Second Attempt Rowkey Columns 338VBChanel1370377143 23b93laddf82:1370377143973 Hd92jslahd0a:1313323414192 338CCElectronic1370377143 z3be3la2dfa2:1370477142970 kd9zjsla3d01:1313323414192
  • 30. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Design: Second Attempt Issues • Hot spotting • Popular products or high level categories can have millions of users, each day • One region serving same dimension type • Terrible write/read performance
  • 31. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. OK…I need a BREAK!!!
  • 32. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Design: Final • Shard row to prevent hot-spotting • Shard into N number of regions • Significant improvement in read/write • Prepend a salt to each key
  • 33. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. [salt]_len_[siteId]_len_[metric]_len_[dimension]_len_[value][timestamp] Design: Final _len_ is the integer length of the field following it timestamp is stored in day granularity [salt] is computed by first creating a hash from the siteId, metric, and dimension, then combining this with a random number between 0 and N (number of shards) for sharding Row key contains attribute value siteId metric attribute timestamp userGUID
  • 34. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. • Complex segments contain many rules • Each rule = one behavior = one row key • Each row key returns a set of users • OR = Full outer join • AND = Inner join • Done in memory for small rules • Merged on disk for large rules Design: Behavior Joins
  • 35. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. HBase Consumer Sync versus Async API
  • 36. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Segmentation Performance • Seconds latency • 40K puts/sec over 2 tables, 8 regions per table • Scaling achieved through addition of regions • Small segments calculated in msecs • Mid-size segments in seconds • Large segments calculated in 10s of seconds
  • 37. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential.
  • 38. © 2013 RichRelevance, Inc. All Rights Reserved. Confidential. Thank You

Notas do Editor

  1. How we plan to go over stuff
  2. The RichRelevance DataMesh Cloud Platform delivers a single view of your customer by:Giving you one single place to house unlimited sets of dataExample use cases:Create your own run-time strategies (predictive models)Create and manage segments via toolAutomatic & real-time segment creationView performance of strategies against KPIs Run adhoc queries using SQL-like toolImport into offline toolsOLAP capabilitiesMarket Basket AnalysisCustomer Lifetime ValueSequential Pattern miningManage APIs, build products & applications
  3. Nuggets or Data Points1.5PB not as big as yahoo or facebook – huge from a retail industry perspective
  4. Distributed System:: i.e. producers, brokers and consumer entities can all be deployed to different hosts in different colos in a truly distributed fashion and coordination controlled through zookeeperPersistence of Messages: messages need to be persisted on the broker for reliability, replay and temporary storagePush & Pull Mechanism:: i.e. push data to Kafka server and pull data from it using a consumer. This allows for two different rates: rate at which messages are transferred to the kafka server and the rate at which the messages are consumed.: Kafka supports GZIP and version 0.8 will additionally support Snappy compression.