ScyllaDB Migration: How to design high throughput and low latency NoSQL deployments
Yuan Ren, Head of Data Science at mParticle, discusses our ScyllaDB migration and how to process 50 billion monthly messages via NoSQL deployments.
mParticle processes and relays incoming messages in near real time, and thus requires a data store that supports querying historical data as well as storing incoming data streams with low latencies. Read on to learn about our ScyllaDB migration and tips for designing high throughput and low latency NoSQL deployments.
mParticle’s ScyllaDB migration:
We initially built the system on top of Apache Cassandra, but as the business grew, we struggled to tune the cluster to keep up with our growing data ingestion rate. Scylla, the next generation of Cassandra, helped us meet our SLA requirements without having to forego the benefits we love from Cassandra. With Scylla’s auto-tuning capability, we can focus our engineering efforts on building products instead of cluster performance.
Check out the corresponding SQL NYC, The NoSQL & NewSQL Database Meetup materials below to learn:
- How to stop worrying about mix workloads, read-modify-write, compaction and tuning
- Why mParticle decided on the Apache Cassandra to ScyllaDB migration, an open source database that is fully compatible with Apache Cassandra
- How to more effectively utilize high capacity Amazon Web Services (AWS) hardware, including the 64 vcpu, 15TB NVMe i3.16xl instances to run a 100TB dataset with just 10 nodes
The mParticle platform captures more than $5 billion in e-commerce transactions over 1 billion monthly active users. Despite its challenging workload, with read-modify-write patterns on huge payload and strict read latency SLAs, we are able to achieve breakneck throughput and low latency without compromising on global database availability….With a little help from our ScyllaDB migration, of course.
Want to solve for the mobile ecosystem’s biggest engineering problems? Learn why you should join the mParticle team and view open roles on our careers page.
View The Deck:
Watch The Presentation:
Latest from mParticle
Get your flywheel in motion with Data Master
Learn how mParticle's Data Master enables you to increase data quality throughout the customer data pipeline, allowing insights to compound, and making every campaign and product launch better than the last.
GOAT: Lifecycle marketing for scalable growth
Learn how GOAT uses mParticle to streamline their data pipeline and increase Customer Lifetime Value.
mParticle launches new features to help brands create ‘data flywheel’
New features for seamless data quality management, and transformation to serve as a foundation for improved customer experience and better insights.
Better data, better insights, better results: Helping brands create a data flywheel
Introduce total quality management and enforcement into your customer data pipeline with new Data Master features and Calculated Attributes. With Data Master and Calculated Attributes, establishing a source of reliable customer data that will create a customer data flywheel, where the data quality and data’s impact on the product cycle will continuously improve over time.