Try our new research platform with insights from 80,000+ expert users

Apache Flink vs Confluent comparison

 

Comparison Buyer's Guide

Executive SummaryUpdated on Dec 17, 2024

Review summaries and opinions

We asked business professionals to review the solutions they use. Here are some excerpts of what they said:
 

Categories and Ranking

Apache Flink
Ranking in Streaming Analytics
4th
Average Rating
7.8
Reviews Sentiment
6.7
Number of Reviews
19
Ranking in other categories
No ranking in other categories
Confluent
Ranking in Streaming Analytics
5th
Average Rating
8.2
Reviews Sentiment
6.3
Number of Reviews
25
Ranking in other categories
No ranking in other categories
 

Mindshare comparison

As of January 2026, in the Streaming Analytics category, the mindshare of Apache Flink is 12.3%, up from 11.6% compared to the previous year. The mindshare of Confluent is 6.8%, down from 8.7% compared to the previous year. It is calculated based on PeerSpot user engagement data.
Streaming Analytics Market Share Distribution
ProductMarket Share (%)
Apache Flink12.3%
Confluent6.8%
Other80.9%
Streaming Analytics
 

Featured Reviews

Aswini Atibudhi - PeerSpot reviewer
Distinguished AI Leader at Walmart Global Tech at Walmart
Enables robust real-time data processing but documentation needs refinement
Apache Flink is very powerful, but it can be challenging for beginners because it requires prior experience with similar tools and technologies, such as Kafka and batch processing. It's essential to have a clear foundation; hence, it can be tough for beginners. However, once they grasp the concepts and have examples or references, it becomes easier. Intermediate users who are integrating with Kafka or other sources may find it smoother. After setting up and understanding the concepts, it becomes quite stable and scalable, allowing for customization of jobs. Every software, including Apache Flink, has room for improvement as it evolves. One key area for enhancement is user-friendliness and the developer experience; improving documentation and API specifications is essential, as they can currently be verbose and complex. Debugging and local testing pose challenges for newcomers, particularly when learning about concepts such as time semantics and state handling. Although the APIs exist, they aren't intuitive enough. We also need to simplify operational procedures, such as developing tools and tuning Flink clusters, as these processes can be quite complex. Additionally, implementing one-click rollback for failures and improving state management during dynamic scaling while retaining the last states is vital, as the current large states pose scaling challenges.
PavanManepalli - PeerSpot reviewer
AVP - Sr Middleware Messaging Integration Engineer at Wells Fargo
Has supported streaming use cases across data centers and simplifies fraud analytics with SQL-based processing
I recommend that Confluent should improve its solution to keep up with competitors in the market, such as Solace and other upcoming tools such as NATS. Recently, there has been a lot of buzz about Confluent charging high fees while not offering features that match those of other tools. They need to improve in that direction by not only reducing costs but also providing better solutions for the problems customers face to avoid frustrations, whether through future enhancement requests or ensuring product stability. The cost should be worked on, and they should provide better solutions for customers. Solutions should focus on hierarchical topics; if a customer has different types of data and sources, they should be able to send them to the same place for analytics. Currently, Confluent requires everything to send to the same topic, which becomes very large and makes running analytics difficult. The hierarchy of topics should be improved. This part is available in MQ and other products such as Solace, but it is missing in Confluent, leading many in capital markets and trading to switch to Solace. In terms of stability, it is not the stability itself that needs improvement but rather the delivery semantics. Other products offer exactly-once delivery out of the box, whereas Confluent states it will offer this but lacks the knobs or levers for tuning configurations effectively. Confluent has hundreds of configurations that application teams must understand, which creates a gap. Users are often unaware of what values to set for better performance or to achieve exactly-once semantics, making it difficult to navigate through them. Delivery semantics also need to be worked on.

Quotes from Members

We asked business professionals to review the solutions they use. Here are some excerpts of what they said:
 

Pros

"This is truly a real-time solution."
"Allows us to process batch data, stream to real-time and build pipelines."
"The setup was not too difficult."
"Apache Flink provides faster and low-cost investment for me; I find it to have low hardware requirements, and it's faster with low code, meaning it's easy to understand for moving the streaming data."
"The ease of usage, even for complex tasks, stands out."
"Apache Flink is meant for low latency applications. You take one event opposite if you want to maintain a certain state. When another event comes and you want to associate those events together, in-memory state management was a key feature for us."
"With Flink, it provides out-of-the-box checkpointing and state management. It helps us in that way. When Storm used to restart, sometimes we would lose messages. With Flink, it provides guaranteed message processing, which helped us. It also helped us with maintenance or restarts."
"It is user-friendly and the reporting is good."
"One of the best features of Confluent is that it's very easy to search and have a live status with Jira."
"The biggest benefit of Confluent as a tool is that it is a distributed platform that provides more durability and stability."
"The most valuable is its capability to enhance the documentation process, particularly when creating software documentation."
"The design of the product is extremely well built and it is highly configurable."
"The documentation process is fast with the tool."
"The most valuable feature that we are using is the data replication between the data centers allowing us to configure a disaster recovery or software. However, is it's not mandatory to use and because most of the features that we use are from Apache Kafka, such as end-to-end encryption. Internally, we can develop our own kind of product or service from Apache Kafka."
"I find Confluent's Kafka Connectors and Kafka Streams invaluable for my use cases because they simplify real-time data processing and ETL tasks by providing reliable, pre-packaged connectors and tools."
"The most valuable feature of Confluent is the wide range of features provided. They're leading the market in this category."
 

Cons

"The solution could be more user-friendly."
"In terms of stability with Flink, it is something that you have to deal with every time. Stability is the number one problem that we have seen with Flink, and it really depends on the kind of problem that you're trying to solve."
"PyFlink is not as fully featured as Python itself, so there are some limitations to what you can do with it."
"Apache Flink should improve its data capability and data migration."
"The machine learning library is not very flexible."
"The TimeWindow feature is a bit tricky. The timing of the content and the windowing is a bit changed in 1.11. They have introduced watermarks. A watermark is basically associating every data with a timestamp. The timestamp could be anything, and we can provide the timestamp. So, whenever I receive a tweet, I can actually assign a timestamp, like what time did I get that tweet. The watermark helps us to uniquely identify the data. Watermarks are tricky if you use multiple events in the pipeline. For example, you have three resources from different locations, and you want to combine all those inputs and also perform some kind of logic. When you have more than one input screen and you want to collect all the information together, you have to apply TimeWindow all. That means that all the events from the upstream or from the up sources should be in that TimeWindow, and they were coming back. Internally, it is a batch of events that may be getting collected every five minutes or whatever timing is given. Sometimes, the use case for TimeWindow is a bit tricky. It depends on the application as well as on how people have given this TimeWindow. This kind of documentation is not updated. Even the test case documentation is a bit wrong. It doesn't work. Flink has updated the version of Apache Flink, but they have not updated the testing documentation. Therefore, I have to manually understand it. We have also been exploring failure handling. I was looking into changelogs for which they have posted the future plans and what are they going to deliver. We have two concerns regarding this, which have been noted down. I hope in the future that they will provide this functionality. Integration of Apache Flink with other metric services or failure handling data tools needs some kind of update or its in-depth knowledge is required in the documentation. We have a use case where we want to actually analyze or get analytics about how much data we process and how many failures we have. For that, we need to use Tomcat, which is an analytics tool for implementing counters. We can manage reports in the analyzer. This kind of integration is pretty much straightforward. They say that people must be well familiar with all the things before using this type of integration. They have given this complete file, which you can update, but it took some time. There is a learning curve with it, which consumed a lot of time. It is evolving to a newer version, but the documentation is not demonstrating that update. The documentation is not well incorporated. Hopefully, these things will get resolved now that they are implementing it. Failure is another area where it is a bit rigid or not that flexible. We never use this for scaling because complexity is very high in case of a failure. Processing and providing the scaled data back to Apache Flink is a bit challenging. They have this concept of offsetting, which could be simplified."
"The state maintains checkpoints and they use RocksDB or S3. They are good but sometimes the performance is affected when you use RocksDB for checkpointing."
"One way to improve Flink would be to enhance integration between different ecosystems. For example, there could be more integration with other big data vendors and platforms similar in scope to how Apache Flink works with Cloudera. Apache Flink is a part of the same ecosystem as Cloudera, and for batch processing it's actually very useful but for real-time processing there could be more development with regards to the big data capabilities amongst the various ecosystems out there."
"there is room for improvement in the visualization."
"Confluent has fallen behind in being the tool of the industry. It's taking second place to things such as Word and SharePoint and other office tools that are more dynamic and flexible than Confluent."
"It could have more integration with different platforms."
"One area we've identified that could be improved is the governance and access control to the Kafka topics. We've found some limitations, like a threshold of 10,000 rules per cluster, that make it challenging to manage access at scale if we have many different data sources."
"The product should integrate tools for incorporating diagrams like Lucidchart. It also needs to improve its formatting features. We also faced issues while granting permissions."
"I am not very impressed by Confluent. We continuously face issues, such as Kafka being down and slow responses from the support team."
"It would help if the knowledge based documents in the support portal could be available for public use as well."
"The formatting aspect within the page can be improved and more powerful."
 

Pricing and Cost Advice

"Apache Flink is open source so we pay no licensing for the use of the software."
"It's an open source."
"It's an open-source solution."
"The solution is open-source, which is free."
"This is an open-source platform that can be used free of charge."
"Regarding pricing, I think Confluent is a premium product, but it's hard for me to say definitively if it's overly expensive. We're still trying to understand if the features and reduced maintenance complexity justify the cost, especially as we scale our platform use."
"The solution is cheaper than other products."
"Confluence's pricing is quite reasonable, with a cost of around $10 per user that decreases as the number of users increases. Additionally, it's worth noting that for teams of up to 10 users, the solution is completely free."
"Confluent is highly priced."
"Confluent is an expensive solution."
"Confluent is expensive, I would prefer, Apache Kafka over Confluent because of the high cost of maintenance."
"You have to pay additional for one or two features."
"Confluent is an expensive solution as we went for a three contract and it was very costly for us."
report
Use our free recommendation engine to learn which Streaming Analytics solutions are best for your needs.
880,255 professionals have used our research since 2012.
 

Top Industries

By visitors reading reviews
Financial Services Firm
20%
Retailer
12%
Computer Software Company
10%
Manufacturing Company
6%
Financial Services Firm
17%
Computer Software Company
11%
Retailer
9%
Manufacturing Company
6%
 

Company Size

By reviewers
Large Enterprise
Midsize Enterprise
Small Business
By reviewers
Company SizeCount
Small Business5
Midsize Enterprise3
Large Enterprise12
By reviewers
Company SizeCount
Small Business6
Midsize Enterprise4
Large Enterprise16
 

Questions from the Community

What do you like most about Apache Flink?
The product helps us to create both simple and complex data processing tasks. Over time, it has facilitated integration and navigation across multiple data sources tailored to each client's needs. ...
What is your experience regarding pricing and costs for Apache Flink?
The solution is expensive. I rate the product’s pricing a nine out of ten, where one is cheap and ten is expensive.
What needs improvement with Apache Flink?
Apache could improve Apache Flink by providing more functionality, as they need to fully support data integration. The connectors are still very few for Apache Flink. There is a lack of functionali...
What do you like most about Confluent?
I find Confluent's Kafka Connectors and Kafka Streams invaluable for my use cases because they simplify real-time data processing and ETL tasks by providing reliable, pre-packaged connectors and to...
What is your experience regarding pricing and costs for Confluent?
They charge a lot for scaling, which makes it expensive.
What needs improvement with Confluent?
I recommend that Confluent should improve its solution to keep up with competitors in the market, such as Solace and other upcoming tools such as NATS. Recently, there has been a lot of buzz about ...
 

Comparisons

 

Also Known As

Flink
No data available
 

Overview

 

Sample Customers

LogRhythm, Inc., Inter-American Development Bank, Scientific Technologies Corporation, LotLinx, Inc., Benevity, Inc.
ING, Priceline.com, Nordea, Target, RBC, Tivo, Capital One, Chartboost
Find out what your peers are saying about Apache Flink vs. Confluent and other solutions. Updated: December 2025.
880,255 professionals have used our research since 2012.