InterviewBit Problems   Design Cache

About the Design Cache category (1)
I see a problem in this design. Here we are assuming that 10M Request per second (4)
What is the kind of QPS we expect for the system? (3)
Will our machines be able to handle qps of 23000? ( 2 ) (21)
How would you implement HashMap? (12)
What if we shard among machines with 16GB of RAM? (10)
I was asked this problem in Google's interview for L5 position (4)
Consistency vs Availability? (11)
Please explain Number of shards = 30 * 1000 / 16 = 1875 (1)
10M/420 machines gives 2381 QPS per machine (2)
What is 4 in below statement? (3)
A very good resource for designing distributed cache (1)
What if we never had to remove entries from the LRU cache because we had enough space, what would you use to support and get and set? (17)
What is the meaning of "1M QPS" here? (1)
Confused about the statement that QPS of 23K is not easily feasible (2)
How the QPS is calculated here? I am not getting the exact calculation. Can someb (10)
Diagram is the first problem (1)
I dont understand how are you updating the linked list in the LRU (1)
What is the number of machines required to cache? (11)
Benefits of Write Through Cache (1)
LRU cache on a single machine which is multi threaded - how does the LRU part work? (2)
Shard resources (1)
Where to submit solution (1)
Submit Design Problem (1)
How would a LRU cache work on a single machine which is multi threaded? (8)
LRU for a distributed cache? (1)
How would you prioritize above operations to keep latency to a minimum for our system? (7)
What happens when a machine handling a shard goes down? (7)
How would a LRU cache work on a single machine which is single threaded? (8)
What about sharding algorithms ? How does the caller know which server to go for (4)