Redis Vs RabbitMQ as a data broker/messaging system in between Logstash and elasticsearch
After evaluating both Redis and RabbitMQ I chose RabbitMQ as our broker for the following reasons:
- RabbitMQ allows you to use a built in layer of security by using SSL certificates to encrypt the data that you are sending to the broker and it means that no one will sniff your data and have access to your vital organizational data.
- RabbitMQ is a very stable product that can handle large amounts of events per seconds and many connections without being the bottle neck.
- In our organization we already used RabbitMQ and had good internal knowledge about using it and an already prepared integration with chef.
Regarding scaling, RabbitMQ has a built in cluster implementation that you can use in addition to a load balancer in order to implement a redundant broker environment.
Is my RabbitMQ cluster Active Active or Active Passive?
Now to the weaker point of using RabbitMQ:
- most Logstash shippers do not support RabbitMQ but on the other hand, the best one, named Beaver, has an implementation that will send data to RabbitMQ without a problem.
- The implementation that Beaver has with RabbitMQ in its current version is a little slow on performance (for my purposes) and was not able to handle the rate of 3000 events/sec from one server and from time to time the service crashed.
- Right now I am working on a fix that will solve the performance problem for RabbitMQ and make the Beaver shipper more stable. The first solution is to add more processes that can run simultaneously and will give the shipper more power. The second solution is to change Beaver to send data to RabbitMQ asynchronously which theoretically should be much faster. I hope that I’ll finish implementing both solutions by the end of this week.
You can follow the issue here: https://github.com/josegonzalez/python-beaver/issues/323
And check the pull request here: https://github.com/josegonzalez/python-beaver/pull/324
If you have more questions feel free to leave a comment.
Redis is created as a key value data store despite having some basic message broker capabilities.
RabbitMQ is created as a message broker. It has lots of message broker capabilities naturally.
I have been doing some research on this topic. If performance is important and persistence is not, RabbitMQ is a perfect choice. Redis is a technology developed with a different intent.
Following is a list of pros for using RabbitMQ over Redis:
- RabbitMQ uses Advanced Message Queuing Protocol (AMQP) which can be configured to use SSL, additional layer of security.
- RabbitMQ takes approximately 75% of the time Redis takes in accepting messages.
- RabbitMQ supports priorities for messages, which can be used by workers to consume high priority messages first.
- There is no chance of loosing the message if any worker crashes after consuming the message, which is not the case with Redis.
- RabbitMQ has a good routing system to direct messages to different queues.
A few cons for using RabbitMQ:
- RabbitMQ might be a little hard to maintain, hard to debug crashes.
- node-name or node-ip fluctuations can cause data loss, but if managed well, durable messages can solve the problem.