Finding Kafka throughput limit in infrastructure at Dropbox

This commit is contained in:
Binh Nguyen 2019-02-17 14:16:42 +08:00
parent 68fb31cb90
commit f69a413ee4
1 changed files with 1 additions and 0 deletions

View File

@ -267,6 +267,7 @@ An organized reading list for illustrating the patterns behind scalable, reliabl
* [Kafka at The New York Times](https://open.nytimes.com/publishing-with-apache-kafka-at-the-new-york-times-7f0e3b7d2077)
* [Reprocessing and Dead Letter Queues with Kafka at Uber](https://eng.uber.com/reliable-reprocessing/)
* [Chaperone: Audit Kafka End-to-End at Uber](https://eng.uber.com/chaperone/)
* [Finding Kafka throughput limit in infrastructure at Dropbox](https://blogs.dropbox.com/tech/2019/01/finding-kafkas-throughput-limit-in-dropbox-infrastructure/)
* [Stream Data Deduplication](https://en.wikipedia.org/wiki/Data_deduplication)
* [Exactly-once Semantics with Kafka](https://www.confluent.io/blog/exactly-once-semantics-are-possible-heres-how-apache-kafka-does-it/)
* [Real-time Deduping at Tapjoy](http://eng.tapjoy.com/blog-list/real-time-deduping-at-scale)