DEV Community

Cover image for Top 8 to cover before taking your messaging queue to production
avital trifsik for Memphis.dev

Posted on • Edited on • Originally published at memphis.dev

Top 8 to cover before taking your messaging queue to production

Introduction
Message brokers are a very powerful components in the architecture of distributed applications, which should absorb an indefinite, unknown, or unforeseen amount of workloads and data movement.

When using a message broker or a messaging queue in production, it is often the crucial component that has to be alive for the rest of the app or platform to be fully functional. Therefore, plenty of handlers need to be considered when you’re using a typical message broker before reaching production.


The top 8

  1. Centralized Monitoring from clients to broker, all the way to the infrastructure. Many different events can cause infrastructure and client breaks, and often it’s hard to get to root-cause. Having a unified tool that can correlate the events can reduce the fixed time significantly.

  2. Auto-scaling. Unpredictable scales and workloads are commonplace in data streaming. You should be able to handle it using code and predefined policy.

  3. Self-healing. In streaming, there are multiple crash scenarios like network lags, unbalanced brokers, consumer crashes, and schema transformation. Having an auto-scaling policy, real-time notifications, and a self-healing mechanism installed and configured can save you a lot of effort and time in both keeping your broker alive during peak events and unexpected workloads.

  4. A Dead-letter queue. For auditing and root-cause fixes. Depending on the use case, having a “recycle bin” with a “restore” option is a much easier way to debug an async issue, as debugging data usually requires multiple cycles of reproducing the “bad” event to the client.

  5. Retransmit mechanism. Consumers have three states –
    Listening, processing, acknowledging. If a consumer crashes during the processing stage before acknowledging the message, it should be known to both the consumer and the broker to retransmit only the “unfinished” message, not send the next one.

  6. Storing offsets. To achieve the above section, it is the client’s responsibility to preserve the already read and acked offsets (or message ids), to understand where the needle is, and what needs to be retransmitted/reasked.

  7. Enforce schemas. Messages are being pushed from everywhere. The ability to enforce a single standard across the different producers is a must to enable a healthy scale and proper governance.

  8. Real-time notifications. Make sure you are aware of every bit and component. In the best case, if something went wrong on your streaming layer – you have failed to submit a BI report. In the worst case – you have failed to process payment transactions, and the quicker you will respond, the lower the impact will be.


In case you prefer not to handle it yourself -
Memphis.dev offers the above out-of-the-box.


Follow Us to get the latest updates!
GithubDocsDiscord


Join 4500+ others and sign up for our data engineering newsletter.


Originally published at Memphis.dev By Yaniv Ben Hemo, Co-Founder & CEO at Memphis.dev.

Top comments (0)