Introduction to Scaling Memcache
Learn what is required to scale a key-value store like Memcache.
Key-value stores and big data
When designing large-scale social media websites, we need to understand that there are huge volumes of data and that a number of requests are being made. The large scale of data on such platforms entails very intense load and spontaneous I/O requests, which might not be optimized using
This is where caching in memory can be used to allow for linear scaling. A key-value store is one way to cache data. Key-value stores can run on network-enabled systems that share information with other machines without requiring specialized hardware. This allows architects to scale their key-value stores by adding more front-end servers. One such key-value store is Memcached.
Note: We'll call the distributed key-value store as
and the source code or the running binary as Memcached. Memcache Memcache without the 'd'.
Why scale Memcache?
Memcache allows developers to add caching to their applications while having simple operations like set
, get
, and delete
. Although it comes with a lot of features for single-server use, there are major hurdles in scaling it to a global level since it doesn't have built-in features (like replication, or clustering) that allow it to scale out for large-scale applications.
Scaling Memcache requires us to solve novel problems like:
Managing hot spots that arise due to highly popular content.
Batching requests to reduce network load.
Managing stale sets and thundering herds to reduce load.
Requirements
To keep up with the cache demands of a large-scale service, we need to consider the following requirements.
Functional requirements
In key-value stores, we have the usual requirements for the set
, get
and delete
functions. Additionally, we have the following needs:
Accessing and updating highly viewed content: Sometimes, many clients start to access the same key-value pairs in Memcached servers. This can bring an enormous load on the Memcached servers themselves, and can also cause too many cache misses, which can in turn may bring the load to the storage layer. Our system should be able to manage such workloads efficiently.
Handling different workloads: Different key-value pairs have different characteristics. Some have very large keys while others are very popular for just a few hours. We need efficient methods to manage these differences.
Maintaining hit rate: We want to maintain the hit rate while the service is live.
Reducing network load: At the cluster level, one web request can make hundreds of Memcache calls within a cluster, which could lead to an extensive load on the network. Our system should have an intelligent placement of Memcache components and good request routing to appropriate components.
Insulating storage layer: Too many cache misses can overload the storage layer, so we want to reduce the frequency of the key-value store fetches to the database.
Non-functional requirements
Near real-time communication: Latency should be low throughout the system, but we'll have to make a trade-off to maintain consistency.
Real-time feed generation from multiple sources: Front-end servers should be able to communicate with hundreds of Memcached servers quickly to respond to a request.
Memory efficient: The caching system from the cluster level to the cross-regional level should be memory efficient.
Replication: To cater to high-demand popular content, replication becomes necessary to process millions of user requests per second. Replication occurs on the regional and cross-regional levels.
Eventual consistency: All replica regions across the globe should eventually become consistent.
Read-after-write consistency: In a local context, clients ought to be able to update an item and then view the changes made to an item.
Handling failures: We need to safely handle a cluster failure to avoid a cascading failure of clusters.
Problem statement
When we have multiple key-value (Memcached) servers throughout the world, how can we replicate data and maintain its consistency while balancing the tradeoffs (for example, between consistency and latency)?
Level up your interview prep. Join Educative to access 80+ hands-on prep courses.