We use this product for general purpose messaging in cloud-based environments and as an implementation to MTP spec. We are customers of VMware and I'm a senior technical consultant.
Sr Technical Consultant at a tech services company with 1,001-5,000 employees
Can be a very fast message broker. Great stability, built-in admin tools and plugin architecture
Pros and Cons
- "It can be configured to be a very fast message broker. I like the stability, the built-in admin tools and plugin architecture."
- "If you're outside IP address range, the clustering no longer has all the features which is problematic."
What is our primary use case?
How has it helped my organization?
One of the key benefits for us has been the ability to use this solution for microservice architecture communications because it provides great flexibility. One of our clients was able to push messages from the source which were replicated and forwarded to all the other brokers nationwide. Everyone who needed it, received it, and it's very cost-effective.
What is most valuable?
The high availability and not having to replicate is valuable as is the message consumer. It can be configured depending on the use case to be a very fast message broker. I like the stability, the built-in admin tools and the plugin architecture. One of the things that makes it unique is that all of the components for messaging can be created programmatically, meaning you can have services or applications that get spun up or have auto incrementing instances. If you're in an elastic environment, you don't have to pre-configure the messaging system and the keys don't have to be known ahead of time.
What needs improvement?
One of the issues is that as soon as you go outside of a switch or not in IP address range, the clustering no longer has all the wonderful features so clustering outside of network boundaries is a problem. I'd like to see stream processing as an additional feature. Kafka has a streaming API and I'd like Rabbit to have that too.
Buyer's Guide
VMware Tanzu Data Solutions
March 2025

Learn what your peers think about VMware Tanzu Data Solutions. Get advice and tips from experienced pros sharing their opinions. Updated: March 2025.
848,716 professionals have used our research since 2012.
For how long have I used the solution?
I've been using this solution for nine years.
What do I think about the stability of the solution?
This solution is very stable, no problems there.
What do I think about the scalability of the solution?
This solution is very scalable and the number of users really depends on the organization.
How are customer service and support?
Technical support is good, very good. But because the underlying implementation technology is Erlang, sometimes the technical problems are at that level, in which case there's one major technical solution provider called Erlang Solutions. They're okay but if the problem goes past the product level and into the technology level, then there can be a delay in getting support because you're dealing with two companies and two technical support services.
How was the initial setup?
These days the initial setup is moderately complex because it uses a technology that is worldwide, Erlang, which is obscure. You have to install Erlang first and that is moderately difficult. Deployment takes about a day.
What's my experience with pricing, setup cost, and licensing?
They use a credit based system for licensingwhere you purchase credits. People don't like it.
What other advice do I have?
My advice would be to have the messaging topology mapped out before you deploy to make the process from installation to a functioning solution more efficient. If you start looking at the topology from the revenue perspective, it usually ends up with more iterations to implement the correct topology, whereas if you start off mapping and then install, it's a more efficient way to go about it.
I rate this solution an eight out of 10.
Which deployment model are you using for this solution?
On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Co-Founder, Chief of Operations with 10,001+ employees
A scalable and future-proof solution for data warehousing
Pros and Cons
- "We chose Greenplum because of the architecture in terms of clustering databases and being able to have, or at least utilize the resources that are sitting on a database."
- "The installation is difficult and should be made easier."
What is our primary use case?
We install this solution for our clients. At the moment we are in the middle of an installation for a data warehouse that will be used by a telecommunications company that is based in Lesotho. We have not gone into production yet, but we have used it in a test environment and it works very well.
We are a technology company, so we handle software development, software implementation, data warehousing, and business intelligence.
We are using the on-premise deployment model. In Africa, there isn't much adoption of cloud services, so most of our clients are expecting on-premise implementation.
What is most valuable?
We chose Greenplum because of the architecture in terms of clustering databases and being able to have, or at least utilize the resources that are sitting on a database.
What needs improvement?
The installation is difficult and should be made easier. Maybe if the process was simpler it would have a quicker adoption by other developers. This could also be accomplished by providing training aids, such as videos to help with installation or using certain features. There are resources currently available on their website, but you have to search through a lot of documentation.
For how long have I used the solution?
We are currently implementing this solution.
What do I think about the scalability of the solution?
Our expectation is that the scalability will be good, as it is one of the main reasons that we have invested in this solution.
How are customer service and technical support?
To this point, I have referenced the material on the website but have not really interacted with technical support.
How was the initial setup?
The initial setup of this solution is not very simple. You need to properly follow the steps in terms of getting the whole architecture put together. We have a team of five people who are working on different aspects of the implementation.
Currently, we are focusing on the data layer. Next will be the ETL layer.
What about the implementation team?
We are using our in-house team to implement this solution for our client.
Which other solutions did I evaluate?
We have used Oracle and Microsoft SQL, but we haven't had much success. We found that Oracle was not as scalable and we were having some performance bottlenecks. Also, from a licensing perspective, Greenplum was a better choice. For all of these reasons, we have chosen to invest heavily in Greenplum.
What other advice do I have?
I would recommend this solution specifically for the scalability. This solution has a more futuristic technology, as opposed to the old school kind of data warehousing. If people are interested in getting something that is more future-proof, then I would recommend this solution.
So far, we're comfortable with what we've seen. What we have configured is addressing our needs at the moment.
I would rate this solution an eight out of ten.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Buyer's Guide
VMware Tanzu Data Solutions
March 2025

Learn what your peers think about VMware Tanzu Data Solutions. Get advice and tips from experienced pros sharing their opinions. Updated: March 2025.
848,716 professionals have used our research since 2012.
VP of Software at a manufacturing company with 11-50 employees
Guaranteed message delivery, queuing, and low latency delivery are the most valuable features.
What is most valuable?
Guaranteed message delivery, queuing, and low latency delivery.
How has it helped my organization?
This allowed us to create a resilient network and independently scale various parts of the system dynamically as the business needs changed.
What needs improvement?
The biggest area we struggled with was operations troubleshooting. We were running a pretty big cluster and ended up with some random cluster failures that were difficult to troubleshoot. A good portion of these were self inflicted but occasionally the distributed database would end up corrupted.
For how long have I used the solution?
I have been using RabbitMQ for six years, from prototype to production.
What do I think about the stability of the solution?
We had a bit of an issue with stability. The usual initial cause would be a hiccup in IOPS in EC2 but then this would cascade into more instability in our main clusters.
What do I think about the scalability of the solution?
For the most part the product was very scalable. The only times we would have problems were usually related to Amazon hiccups that would cause the cluster to slow down.
How are customer service and technical support?
We did not use their technical support much.
Which solution did I use previously and why did I switch?
We evaluated a variety of message products and found that for the feature set RabbitMQ was the best.
How was the initial setup?
Initial setup was relatively simple and then we were able to grow into the product by using more of the features.
What's my experience with pricing, setup cost, and licensing?
We used the open source implementation and did not need to pay for support.
Which other solutions did I evaluate?
We looked at ZeroMQ, Kafka and Redis.
What other advice do I have?
You really need to have or train Erlang expertise. The Erlang tools will become the best way to troubleshoot misbehaving clusters.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Consultant at a financial services firm with 5,001-10,000 employees
The MPP element is crucial, so far as it allows us to query millions of rows at a time, at speed.
What is most valuable?
The MPP element is crucial, so far as it allows us to query millions of rows at a time, at speed.
How has it helped my organization?
The previous data warehouse was built in Oracle. One of the things which has improved in GreenPlum is that we can query millions of rows at speed, without creating lags. We’ve also built far more views; slowly changing dimensions can instantaneously update without creating the issue of having to rebuild tables to reflect new hierarchies, for example.
What needs improvement?
We found some issues with larger tables that have daily data appended, where after a while this seems to create lag in the query speed. This might just have to do with local knowledge rather than the product itself.
We have a table which is currently contains 27.6m rows and has a daily delta added to it of roughly 16.5k rows per day. While this isn’t particularly large, we have noticed the table begins to perform poorly when queried, in spite of having set up a VACUUM process to be performed weekly. It may be that the VACUUM process needs to be performed more frequently (like daily), but we’ve not yet found the optimal way of maintaining this particular table.
It’s worth saying that this is one table out of over 400 perfectly well performant tables and views in the same database. Hope that helps,
For how long have I used the solution?
I have used for approximately 30 months.
What was my experience with deployment of the solution?
I have not encountered any deployment, stability or scalability issues.
How are customer service and technical support?
I have not raised any service issues/tech queries, so I can’t really say.
Which solution did I use previously and why did I switch?
We used Oracle previously. We based our choice on expertise in our US operation, where we have a GreenPlum expert who provided some amazing use case examples to help us in our selection process.
What about the implementation team?
Implementation was done in-house.
What was our ROI?
Not within my area I’m afraid, but I understand that this was a very good fit from an ROI point of view
What other advice do I have?
Investigate whether this solution works for you. It is worth creating a rating matrix to compare other similar products, and it is very useful to look deeply at whether the third-generation MPP software might be a good fit.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Research Assistant at a university with 1,001-5,000 employees
We use it to distribute pieces of some large jobs to multiple machines.
What is most valuable?
Message queue, because it is easy to use, reliable, not a big load.
How has it helped my organization?
We are using it to distribute pieces of some large jobs to multiple machines, which improves performance several times.
What needs improvement?
Improve the ability to handle the large message load.
People usually use RabbitMQ as the lightweight messenger, if they have a large message load people are inclined to use Kafka. But at the beginning stage of most projects, the data is small, people do not need to use a Kafka type of messenger, they are more likely to use RabbitMQ. If RabbitMQ can handle the large message load and support ordered delivery, with the project growing, data bigger, people can still use RabbitMQ and wouldn't need to find another tool to use like Kafka which is much more convenient.
For how long have I used the solution?
Half a year.
What do I think about the stability of the solution?
Didn’t have issues.
What do I think about the scalability of the solution?
Didn’t have issues.
How is customer service and technical support?
Very good. 8/10.
How was the initial setup?
Simple. We followed the tutorial about RabbitMQ with Python.
What's my experience with pricing, setup cost, and licensing?
We are using it internally with a very small data load in the developing period, which is free right now.
Which other solutions did I evaluate?
Yes, I evaluated Kafka.
Kafka is more suitable to large amount events in order. RabbitMQ is more suitable to the related small amount of messages, which is my situation and I don’t care about the message order.
What other advice do I have?
RabbitMQ is a very easy to use and reliable message broker. If the work has a relatively small message load, RabbitMQ is the most robust and reliable choice.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Senior Director & Global Lead, Big Data Center of Excellence at a pharma/biotech company with 10,001+ employees
The loading and transformation of large data sets is valuable.
Valuable Features:
Processing speed – especially loading and transformation of large data sets.
Improvements to My Organization:
Before we implemented Greenplum, our weekly data loads (for third party marketing data sets) were taking over three days. (We also had some monthly data that could take up to 3 days to load and transform via Informatica.) After we implemented Greenplum, the loads were reduced to less than nine hours. Previously, we were receiving data early Wed a.m. and not getting out to the salesforce (if we were lucky) until noon on the following Monday. Now we get the data to the field early Friday mornings before they wake up.
Room for Improvement:
The Greenplum appliance itself has had some reliability issues, so it would be great if that could be improved in the next version. More critical, though, is that the latest devices are not backward compatible. i.e., We have to replace our entire environment to upgrade. That’s quite an expense. I would hope they could improve the upgrade roadmap in the future.
Implementation Team:
We have used EMC Consulting for some projects, and we have lots of EMC storage.
Other Advice:
If you can, do a benchmark with other MPP options including cloud alternatives. Although our Greenplum implementation was very successful (going on 4 years ago), I wish we had benchmarked against Teradata and Netezza (now IBM) at least. Today, I would consider not even buying hardware… just doing it all in the cloud.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
Engineer at a marketing services firm with 51-200 employees
We use Greenplum for production but not for development and testing because it consumes too much resources
We use the scrum methodology to manage our engineering work. And while this does give us the flexibility to quickly respond to changing business needs, it also means that our databases’ schemas are not set in stone. To accomodate frequent changes to the way we structure our data, we’ve adapted Ruby on Rails database migrations for use beyond our Rails apps.
Our first challenge was to find and extract the migration’s functionality from Rails so that we could use it without needing to bring in the other features of Rails we didn’t need. Thankfully, this turned out to be relatively simple because almost all of the functionality can already be accessed by Rake tasks, so it was just a matter of building a Rake file that contains the tasks we wanted from Rails. The only thing we had to add to get started was a Rake task for creating new migrations, since Rails creates them either automatically when creating new model objects or through the `rails generate migration` command.
We quickly ran into problems, though, because we use Greenplum, a DBMS built on Postgres that adds support for features that help accomodate big data. Unlike standard Postgres, Greenplum adds features, such as table partitions, not found in most DBMSes, but Rails is designed to be DBMS agnostic so you can easily switch between, say, Postgres and MySQL and SQLite without any more trouble than changing a single configuration file. So we decided to cut around Rails’s database abstractions and instead directly write SQL in our migrations and dump SQL structure files rather than Rails schema files.
Only this didn’t work because, to complicate matters further, although we use Greenplum in production and staging environments, for local testing and development we use Postgres because Greenplum has minimum requirements that ended up consuming too much of our workstations’ resources. So we ultimately ended up developing a hybrid solution that allows us to support Greenplum and Postgres in the same migrations.
The two main aspects of the solution are selective application of options when making changes to the database and keeping two separate dumps of the database, one for Greenplum with Greemplum-only syntax and one for Postgres with Greenplum-only syntax removed. For example, we rewrote the Rails `create_table` migration function to add support for Greenplum options like partitions, data distribution, and append-only tables, but then have the function ignore those options when running the migrations against our development Postgres databases. This allows us to use a single set of migrations on all of our databases, drastically simplifying what could have otherwise been a gnarly challenge.
By no means is our solution complete or perfect. It only supports those features of Greenplum that we use, and new features only get added in as we need them, plus there is some risk associated with keeping two separate schema dumps and the potential for them to get out-of-sync. Still, compared to maintaining a database with frequently changing schemas without the aid migrations, it’s lightyears better.
Disclosure: I am a real user, and this review is based on my own experience and opinions.
IT Software Engineer at a financial services firm with 10,001+ employees
Its valuable because of its high performance, integration with Spring Framework, easy installation, and configuration
What is most valuable?
- High performance
- Integration with Spring Framework
- Easy installation and configuration
How has it helped my organization?
We have been using GemFire for a Telco project, which we need process network data in real time and meanwhile access some reference data. GemFire has done a great job, as we have managed to process over 200,000 messages per second.
What needs improvement?
In build monitoring, the interface could be improved.
For how long have I used the solution?
18 months.
What do I think about the stability of the solution?
With version 8.1, we had some issues while we were querying data from memory, but it has been fixed in version 8.2, and after that we have never had problems.
What do I think about the scalability of the solution?
No, even when we had serious network problems, GemFire managed to recover.
How are customer service and technical support?
A nine out of 10.
Which solution did I use previously and why did I switch?
No, GemFire was always good enough for us.
How was the initial setup?
Initial setup was easy. In older versions, the user interface was not helpful, but it's improved lately.
What's my experience with pricing, setup cost, and licensing?
As a developer, I was never a part of this pricing decision. That's why I have no advice.
Which other solutions did I evaluate?
No, because we've started to use GemFire as part of a real-time intelligence platform.
What other advice do I have?
Trust GemFire and use the benefits of their strong documentation.
Disclosure: I am a real user, and this review is based on my own experience and opinions.

Buyer's Guide
Download our free VMware Tanzu Data Solutions Report and get advice and tips from experienced pros
sharing their opinions.
Updated: March 2025
Product Categories
Data Warehouse Database Development and Management Relational Databases Tools Message Queue (MQ) SoftwarePopular Comparisons
Oracle Exadata
Apache Hadoop
SAP BW4HANA
IBM Netezza Performance Server
Oracle Database Appliance
Microsoft Parallel Data Warehouse
Buyer's Guide
Download our free VMware Tanzu Data Solutions Report and get advice and tips from experienced pros
sharing their opinions.
Quick Links
Learn More: Questions:
- Looking for advice on how to migrate from Oracle Exadata to VMware Tanzu Greenplum
- What is the biggest difference between ActiveMQ and RabbitMQ?
- What is the biggest difference between IBM MQ and RabbitMQ?
- How does IBM MQ compare with VMware RabbitMQ?
- Oracle Exadata vs. HPE Vertica vs. EMC GreenPlum vs. IBM Netezza
- When evaluating Data Warehouse solutions, what aspect do you think is the most important to look for?
- At what point does a business typically invest in building a data warehouse?
- Is a data warehouse the best option to consolidate data into one location?
- What are the main differences between Data Lake and Data Warehouse?
- Infobright vs. Exadata vs. Teradata vs. SQL Server Data Warehouse- which is most compatible with front end tools?
Regarding the performance of few large tables, just a suggestion you can also try implementing the partitioning. By doing partitioning you can leverage the "swap partiton" while doing an insert and select the data for reporting based on your partitioning key.
Hope this helps