open-menu closeme
Home
RetroComp icon
Retro Computing Articles Composition Notebooks 8bit
DevOps icon
Automation K8s Done Right
GenAI icon
Fatlama Newsletter Generative AI
Projects icon
Software Projects Arduino Misterio Esp8266 SQLite
Status
Links icon
GitHub LinkedIn
About
github linkedin rss
  • Avoid Kafka if unsure (think twice series)

    calendar Dec 2, 2019 · 2 min read
     en featured knowledgebase sql  · java nosql kafka
     ·
    Share on: twitter facebook linkedin copy
    Avoid Kafka if unsure (think twice series)

    Some co-workers started using Apache Kafka con a bunch of our Customers.

    Apache Kafka is a community distributed event streaming platform capable of handling trillions of events a day. Initially conceived as a messaging queue, Kafka is based on an abstraction of a distributed commit log[*].
    To get this goal, Apache Kafka needs a complex servers setup, even more complex if you want the certification for the producing company (Confluent).  Now, if you are planning to use Kafka like a simple JavaMessaeSystem (JMS) implementation, think twice before going on this route.

    PostgreSQL 12 offers a fair (and open source) partition implementation, whereas if money are not a problem, Oracle 12c can happy scale on billions of record before running into troubles (and ExaData can scale even more).

    PostgreSQL and Oracle offer optimizations for partitioned data, called “Partition Pruning” in PostreSQL teminology:

    With partition pruning enabled, the planner will examine the definition of each partition and prove that the partition need not be scanned because it could not contain any rows meeting the query's WHERE clause. When the planner can prove this, it excludes (prunes) the partition from the query plan.
    This feature is quite brand new (popped in PostreSQL 11) but it is essential to a successful partition strategy. Before these feature, partitioning was a black magic art. Now it is simpler to manage.
    Read More
  • Finding Kafka’s throughput limit in Dropbox infrastructure | Dropbox Tech Blog

    calendar Feb 14, 2019 · 1 min read
     en  · kafka
     ·
    Share on: twitter facebook linkedin copy
    At Dropbox, Kafka clusters are managed by the Jetstream team, whose primary responsibility is to provide high quality Kafka services. Understanding Kafka’s throughput limit in Dropbox infrastructure is crucial in making proper provisioning decision for different use cases, and this has been an important goal for the team. Recently, we created an automated testing platform to achieve this objective. In this post, we would like to share our method and findings.
    Read Finding Kafka’s throughput limit in Dropbox infrastructure | Dropbox Tech Blog
  • ?FileSystem is faster than RAM [under your Operating System]

    calendar Jan 18, 2019 · 2 min read
     en featured  · development great-ideas java kafka
     ·
    Share on: twitter facebook linkedin copy

    I am studying Apache Kafka (a "distributed streaming platform") and I stumbled upon this conclusion: the "disk read fear" a lot of projects have in the past, is a fake.


    Read More

Recent Posts

  • La ruota del Destino
  • Darth Android
  • Vps Provider
  • Chat
  • RSS Readers
  • Vibe Coding: easy to say, difficult to survive
  • Basic V2 rocks

Latest comments

    Giovanni Giorgi

    Copyright 1999-  GIOVANNI GIORGI. All Rights Reserved

    to-top