Try our new research platform with insights from 80,000+ expert users
Abhik Ray - PeerSpot reviewer
Co-Founder at Quantic
Real User
Has good processing power and speed and is capable of handling large volumes of data and doing online analysis
Pros and Cons
  • "The most important feature is its ability to handle large volumes. Some of our customers have really large volumes, and it is capable of handling their data in terms of the core volume and daily incremental volume. So, its processing power and speed are most valuable."
  • "It requires a great deal of learning curve to understand. The overall Hadoop ecosystem has a large number of sub-products. There is ZooKeeper, and there are a whole lot of other things that are connected. In many cases, their functionalities are overlapping, and for a newcomer or our clients, it is very difficult to decide which of them to buy and which of them they don't really need. They require a consulting organization for it, which is good for organizations such as ours because that's what we do, but it is not easy for the end customers to gain so much knowledge and optimally use it."

What is our primary use case?

Its main use case is to create a data warehouse or data lake, which is a collection of data from multiple product processors used by a banking organization. They have core banking, which has savings accounts or deposits as one system, and they have a CRM or customer information system. They also have a credit card system. All of them are separate systems in most cases, but there is a linkage between the data. So, the main motivation is to consolidate all that data in one place and link it wherever required so that it acts as a single version of the truth, which is used for management reporting, regulatory reporting, and various forms of analyses.

We have done two or three projects with Hadoop, and we have taken the latest version available at that time. So far, it was deployed on-premises.

What is most valuable?

The most important feature is its ability to handle large volumes. Some of our customers have really large volumes, and it is capable of handling their data in terms of the core volume and daily incremental volume. So, its processing power and speed are most valuable.

Another feature that I like is online analysis. In some cases, data requires online analysis. We like using Hadoop for that.

What needs improvement?

It requires a great deal of learning curve to understand. The overall Hadoop ecosystem has a large number of sub-products. There is ZooKeeper, and there are a whole lot of other things that are connected. In many cases, their functionalities are overlapping, and for a newcomer or our clients, it is very difficult to decide which of them to buy and which of them they don't really need. They require a consulting organization for it, which is good for organizations such as ours because that's what we do, but it is not easy for the end customers to gain so much knowledge and optimally use it. However, when it comes to power, I have nothing to say. It is really good.

For how long have I used the solution?

We have been working with this solution for two and a half to three years.

Buyer's Guide
Apache Hadoop
January 2025
Learn what your peers think about Apache Hadoop. Get advice and tips from experienced pros sharing their opinions. Updated: January 2025.
831,265 professionals have used our research since 2012.

What do I think about the stability of the solution?

The core file system and the offline data ingestion are extremely stable. In my experience, there is a bit less stability during online data ingestion. When you have incremental online data, sometimes it stops or aborts before finishing. It is rare, but it does, but the offline data injection and the basic processing are very stable.

What do I think about the scalability of the solution?

Its scalability is very good. Most of our clients have used it on-prem. So, to a large extent, it is up to them to provide hardware for large data, which they have. Its scalability is linear. As long as the hardware is given to it, there are no complaints.

About 70% of its users are from a client's IT in terms of setting it up and providing support to make sure that the pipeline is there. Business users are about 30%. They are the people who use the analytics derived from the warehouse or data lake. Collectively, there are about 120 users. The size of the data is mostly in terms of the number of records it handles, which could be 30 or 40 million.

How are customer service and support?

We have not dealt with them too many times. I would rate them a four out of five. There are no complaints.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

Some of our clients are using Teradata, and some of them are using Hadoop.

How was the initial setup?

After the hardware is available, getting the environment and software up and running has taken us a minimum of a week or 10 days. Sometimes, it has taken us longer, but usually, this is what it takes at the minimum to get everything up. It includes the downloads and also setting it up and making things work together to start using it.

For the original deployment, because there are so many components and not everyone knows everything pretty well, we have seen that we had to deploy four or five people in various areas at the initial deployment stage. However, once it is running, one or two people are required for maintenance.

What was our ROI?

Different clients derive different levels of return based on the sophistication of the analytics that they derive out of it and how they use it. I don't know how much ROI they have got, but I can say that some clients have not got a decent ROI, but some of our clients are happy with it. It is very much client-dependent.

What's my experience with pricing, setup cost, and licensing?

We don't directly pay for it. Our clients pay for it, and they usually don't complain about the price. So, it is probably acceptable.

What other advice do I have?

I would rate it a nine out of ten because of the complexity, but technically, it is okay.

Which deployment model are you using for this solution?

On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
PeerSpot user
R&D Head, Big Data Adjunct Professor at SK Communications Co., Ltd.
Real User
Not dependent on third-party vendors
Pros and Cons
  • "We selected Apache Hadoop because it is not dependent on third-party vendors."
  • "Real-time data processing is weak. This solution is very difficult to run and implement."

What needs improvement?

Apache Hadoop's real-time data processing is weak and is not enough to satisfy our customers, so we may have to pick other products. We are continuously researching other solutions and other vendors.

Another weak point of this solution, technically speaking, is that it's very difficult to run and difficult to smoothly implement. Preparation and integration are important.

The integration of this solution with other data-related products and solutions, and having other functions, e.g. API connectivity, are what I want to see in the next release.

For how long have I used the solution?

We've started using Apache Hadoop since 2011.

Which solution did I use previously and why did I switch?

We selected Apache Hadoop because it is not dependent on third-party vendors. Previously, our main business unit was related to big vendors like IBM, Oracle, and EMC, etc. We wanted to have a competitive advantage in technology, so we selected the Apache project and used Apache open source.

What about the implementation team?

The solution was implemented through a local vendor team here in Korea.

Which other solutions did I evaluate?

We evaluated IBM, Oracle, and EMC solutions.

What other advice do I have?

My position in the company falls under the research and development of new technologies and solutions. I investigate, research, download, and read information and reports as part of my job.

Our company has a big data business division, and we propose, develop, and implement things which are related to big data projects. We are using Cloud Hadoop open source versions, distributed versions, and commercial Hadoop distributed versions. We propose all these versions to our customers from any industry.

Our focus is on the public sector. Big data is our strong point in Korea. Our company is the leader in big data technology, including infrastructure and visualization. This is a solution we provide to our customers. We are also in partnership with IBM. Our main focus is on Apache Hadoop.

We provide Apache Hadoop to our customers. I work for a systems integrator and technical consulting company.

Overall, our satisfaction with this solution is so-so. We continuously investigate new technologies and other solutions.

The Hadoop open source version was implemented in 95% of our company's customer base. Our remaining customers had the local vendor's Hadoop platform package implemented for them.

Our company is in the big data business. Before the big data business back in 1976, we implemented BI (business intelligence), DW (data warehouse), EIS, and DSS (decision support system), so we are in partnership with IBM.

I don't have advice for people looking into implementing this solution because I'm not in the business unit. I'm in the research field. My role is to plan new technology and provide consultation to our customers for big data projects in the early stages.

My rating for Apache Hadoop from a technical standpoint is eight out of ten.

Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
Buyer's Guide
Apache Hadoop
January 2025
Learn what your peers think about Apache Hadoop. Get advice and tips from experienced pros sharing their opinions. Updated: January 2025.
831,265 professionals have used our research since 2012.
Teodor Muraru - PeerSpot reviewer
Developer at Emag
Real User
Top 5
Helps to store and retrieve information
Pros and Cons
  • "Apache Hadoop is crucial in projects that save and retrieve data daily. Its valuable features are scalability and stability. It is easy to integrate with the existing infrastructure."

    What is our primary use case?

    The solution helps to store and retrieve information.

    What is most valuable?

    Apache Hadoop is crucial in projects that save and retrieve data daily. Its valuable features are scalability and stability. It is easy to integrate with the existing infrastructure. 

    For how long have I used the solution?

    I have been using the tool for a few years. 

    What do I think about the stability of the solution?

    I rate the tool's stability a nine out of ten. 

    How are customer service and support?

    I take support from the DevOps team. 

    What other advice do I have?

    I recommend the tool to others since it is good. 

    Which deployment model are you using for this solution?

    On-premises
    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    Flag as inappropriate
    PeerSpot user
    Anand Viswanath - PeerSpot reviewer
    Project Manager at Unimity Solutions
    Real User
    Top 5Leaderboard
    Offers reasonable integration features but needs to improve the setup process
    Pros and Cons
    • "The tool's stability is good."
    • "The load optimization capabilities of the product are an area of concern where improvements are required."

    What is our primary use case?

    I use the solution in my company for security purposes.

    In my company, we have intranet portals that we need to ensure are not accessible by outsiders. All the data that are within the internal applications is only accessible with valid credentials within the domain. In general, my company uses Apache Hadoop to secure our internal applications.

    What needs improvement?

    Tools like Apache Hadoop are knowledge-intensive in nature. Unlike other tools in the market currently, we cannot understand knowledge-intensive products straight away. To use Apache Hadoop, a person needs intensive knowledge, which is something that not everybody can get familiarized with in a straightforward manner. It would be beneficial if navigating through tools like Apache Hadoop is made user-friendly. For non-technical users, if the tool is made easy to navigate, it will be easier to use, and one may not have to depend on experts.

    The load optimization capabilities of the product are an area of concern where improvements are required.

    The complex setup phase can be made easier in the future.

    For how long have I used the solution?

    I have four years of experience with Apache Hadoop.

    What do I think about the stability of the solution?

    The tool's stability is good.

    What do I think about the scalability of the solution?

    I am not sure about the scalability features of the product.

    There are around 500 users of the product in my company.

    When there is a huge load or a huge number of people accessing the product simultaneously, there is a visible delay in the loading of pages.

    How was the initial setup?

    The product's initial setup phase is complex.

    I have not dealt with the setup phase straight away. I always like to rely on the infra person in my company who knows Apache Hadoop.

    The solution is deployed on the cloud.

    What about the implementation team?

    The product can be deployed with the help of the in-house infra team at my company.

    What other advice do I have?

    There was a scenario when the product was essential for my company's data analytics needs. Before my company makes any web solution available in production, we have prototypes and replicas of the application in lower environments. My company uses Apache Hadoop to ensure that the lower environments in which we operate are secure and accessible only by those people in our company with valid credentials.

    I suggest that those planning to use the product first understand the tool's features and capabilities and then choose the right configuration to avoid misconfigurations.

    The product's integration capabilities are good since I see that we have not faced any time outs or downtime in our company when using the tool.

    My company uses the tool to have security and the right availability, which means availability to the right people at the right time. So I think our expectation was met. The value we got from the tool was what we wanted in our company.

    My company started to use the tool expecting that it would offer security and ensure its availability to the right people at the right time. I believe that the tool was able to meet our company's expectations, so we got the value that we expected the product to deliver.

    I rate the tool a seven out of ten.

    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    PeerSpot user
    reviewer2324613 - PeerSpot reviewer
    Data Architect at a computer software company with 51-200 employees
    Real User
    Top 5
    Allows for customization and optimization of applications and performance using in-house resources but lacks community support
    Pros and Cons
    • "It's open-source, so it's very cost-effective."
    • "The main thing is the lack of community support. If you want to implement a new API or create a new file system, you won't find easy support."

    What is our primary use case?

    We work on Apache Hadoop for various customers. 

    What is most valuable?

    It's open-source, so it's very cost-effective. Apache Hadoop has its strengths. For example, in my previous organization, which was a small startup, we used it because it was cost-effective. 

    We only had to pay for the servers, and we could optimize applications and performance using our employees, which was especially cost-effective in India. So, human resources were the main investment, not software. 

    That was five years ago, though. In the last five years, I've mainly seen Redshift, Azure, and Oracle in the market.

    What needs improvement?

    The main thing is the lack of community support. If you want to implement a new API or create a new file system, you won't find easy support. 

    And then there's the server issue. You have to create and maintain servers on your own, which can be hectic. Sometimes, the configurations in the documentation don't work, and without a strong community to turn to, you can get stuck. That's where cloud services play a vital role.

    In future releases, the community needs to be improved a lot. We need a better community, and the documentation should be more accurate for the setup process.

    Sometimes, we face errors even when following the documentation for server setup and configuration. We need better support. 

    Even if we raise a ticket, it takes a long time to get addressed, and they don't offer online support. They ask for screenshots, which takes even more time. Instead of direct screensharing or hopping on a call. But it's free, so we can't complain too much.

    For how long have I used the solution?

    I've been working with Apache Hadoop for ten years. I started my career with Hadoop. I've worked with it at Infinia, Microsoft, and AWS, for a total of about eight years.

    What do I think about the stability of the solution?

    I would rate the stability a seven out of ten. There is room for improvement in performance.

    What do I think about the scalability of the solution?

    It can be scalable in certain cases. Typically, for startups or product-based companies with limited budgets during product development, Apache Hadoop is often the only viable option. They cannot afford the costs of other cloud-based systems, so Apache Hadoop plays a main role in those scenarios.

    Which solution did I use previously and why did I switch?

    For some customers, we use Oracle Autonomous Database. Now, I cannot compare Apache Hadoop with Oracle Autonomous Data Warehouse when it comes to value for money. They're not directly comparable.

    How was the initial setup?

    The initial setup is a hectic task. Configuring servers and nodes takes a long time. That's one of the big advantages of an Autonomous Data Warehouse. You can start implementing within half the time. 

    With Apache Hadoop, you have to wait for the setup, architecture, and data evaluation. But with Autonomous, those things are automated. It scales as you use more data, so you can focus on the business rather than infrastructure.

    What's my experience with pricing, setup cost, and licensing?

    We just use the free version.

    What other advice do I have?

    We can't use Apache Hadoop for everything, like storage and data errors. But we can use some tools that are native to Hadoop, like Kafka.

    For the current situation, I'd rate it a seven out of ten. 

    However, five years ago, I would have rated it a nine out of ten. Back then, I was working with it fully. But now we're used to working with cloud systems. Creating servers is more difficult nowadays.

    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    PeerSpot user
    reviewer1976262 - PeerSpot reviewer
    Credit & Fraud Risk Analyst at a financial services firm with 10,001+ employees
    Real User
    Has the ability to take a large amount of data and deliver the necessary splices and summary charts
    Pros and Cons
    • "Apache Hadoop can manage large amounts and volumes of data with relative ease, which is a feature that is beneficial."
    • "I mentioned it definitely, and this is probably the only feature we can improve a little bit because the terminal and coding screen on Hadoop is a little outdated, and it looks like the old C++ bio screen. If the UI and UX can be improved slightly, I believe it will go a long way toward increasing adoption and effectiveness."

    What is our primary use case?

    We use Apache Hadoop for analytics purposes.

    What is most valuable?

    The ability to take a lot of data and attempt to basically deliver the appropriate splices and summary chart is the most crucial function that I have discovered. 

    This stands in contrast to some of the other tools that are available, such as SQL and SAS, which are likely incapable of handling such a large volume of data. Even R, for instance, is unable to handle such data volumes. 

    Apache Hadoop can manage large amounts and volumes of data with relative ease, which is a feature that is beneficial.

    What needs improvement?

    In terms of processing speed, I believe that some of this software as well as the Hadoop-linked software can be better. While analyzing massive amounts of data, you also want it to happen quickly. Faster processing speed is definitely an area for improvement.

    I am not sure about the cloud's technical aspects, whether there are things that happen in the cloud architecture that essentially make it a little slow, but speed could be one. And, second, the Hadoop-linked programs and Hadoop-linked software that are available could do much more and much better in terms of UI and UX.

    I mentioned it definitely, and this is probably the only feature we can improve a little bit because the terminal and coding screen on Hadoop is a little outdated, and it looks like the old C++ bio screen. 

    If the UI and UX can be improved slightly, I believe it will go a long way toward increasing adoption and effectiveness.

    For how long have I used the solution?

    I have been using Apache Hadoop for six months.

    What do I think about the stability of the solution?

    It is far more stable than some of the other software that I have tried. It's also the current version of Hadoop software and is becoming increasingly more stable.

    When a new version is released, the subsequent ones are always more stable and easier to use.

    What do I think about the scalability of the solution?

    According to what I have seen in my current enterprise, once I joined the organization, it was fairly simple to have it for an employee, and this is true for everyone who's been onboarded in my own designation. I would imagine that it is fairly scalable across an enterprise.

    I am fairly certain that we have between 10 and 15,000 employees who use it.

    How are customer service and support?

    I have not had any direct experience with technical support.

    We have an in-house technical support team that handles it.

    Which solution did I use previously and why did I switch?

    I have since changed careers, I no longer use any automation tools, nor does my job need me to compare the capabilities of other tools.

    I am working with Risk Analytic tools. I work with data these days, therefore I use technologies like Hive, Shiny R, and other data-intensive programs.

    Shiny is a plugin that you can have on R. As a result of changing my profiles, I am now working in a position that is more data-centric and less focused on process automation.

    We currently have proprietary tools, a proprietary cloud software, therefore I don't really need to employ any external cloud vendors. Aside from that, I only use the third-party technologies I've already indicated, primarily Hadoop and R.

    This is one of the prime, one of the cornerstone software that we use. I have never been in a position to compare the like-for-like comparison with another software.

    How was the initial setup?

    As it is proprietary software for the enterprise that I am currently working on, I had no trouble setting it up.

    What's my experience with pricing, setup cost, and licensing?

    I am not sure about the price, but in terms of usability and utility of the software as a whole, I would rate it a three and a half to four out of five.

    Which other solutions did I evaluate?

    When I was a digital transformation consultant for my prior employer, I downloaded and read the reviews.

    It involved learning about workflow automation tools as well as process automation. I looked at a number of these platforms as part of that, but I have never actually used them.

    What other advice do I have?

    I would recommend this solution for data professionals who have to work hands-on with big data.

    For instance, if you work with smaller or more finite data sets, that is, data sets that do not keep updating themselves, I would most likely recommend R or even Excel, where you can do a lot of analysis. However, for data professionals who work with large amounts of data, I would strongly recommend Hadoop. It's a little more technical, but it does the job.

    I would rate Apache Hadoop an eight out of ten. I would like to see some improvements, but I appreciate the utility it provides.

    Which deployment model are you using for this solution?

    Public Cloud
    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    PeerSpot user
    Senior Associate at a financial services firm with 10,001+ employees
    Real User
    Relatively fast when reading data into other platforms but can't handle queries with insufficient memory
    Pros and Cons
    • "As compared to Hive on MapReduce, Impala on MPP returns results of SQL queries in a fairly short amount of time, and is relatively fast when reading data into other platforms like R."
    • "The key shortcoming is its inability to handle queries when there is insufficient memory. This limitation can be bypassed by processing the data in chunks."

    What is most valuable?

    Impala. As compared to Hive on MapReduce, Impala on MPP returns results of SQL queries in a fairly short amount of time, and is relatively fast when reading data into other platforms like R (for further data analysis) or QlikView (for data visualisation).

    How has it helped my organization?

    The quick access to data enabled more frequent data backed decisions.

    What needs improvement?

    The key shortcoming is its inability to handle queries when there is insufficient memory. This limitation can be bypassed by processing the data in chunks.

    For how long have I used the solution?

    Two-plus years.

    What do I think about the stability of the solution?

    Typically instability is experienced due to insufficient memory, either due to a large job being triggered or multiple concurrent small requests.

    What do I think about the scalability of the solution?

    No. This is by default a cluster-based setup and hence scaling is just a matter of adding on new data nodes.

    How are customer service and technical support?

    Not applicable to Cloudera. We have a separate onsite vendor to manage the cluster.

    Which solution did I use previously and why did I switch?

    No. Two years ago this was a new team and hence there were no legacy systems to speak of.

    How was the initial setup?

    Complex. Cloudera stack itself was insufficient. Integration with other tools like R and QlikView was required and in-house programs had to be built to create an automated data pipeline.

    What's my experience with pricing, setup cost, and licensing?

    Not much advice as pricing and licensing is handled at an enterprise level.

    However do take into consider that data storage and compute capacity scale differently and hence purchasing a "boxed" / 'all-in-one" solution (software and hardware) might not be the best idea.

    Which other solutions did I evaluate?

    Yes. Oracle Exadata and Teradata.

    What other advice do I have?

    Try open-source Hadoop first but be aware of greater implementation complexity. If open-source Hadoop is "too" complex, then consider a vendor packaged Hadoop solution like HortonWorks, Cloudera, etc.

    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    PeerSpot user
    PeerSpot user
    Senior Hadoop Engineer with 1,001-5,000 employees
    Vendor
    The heart of BigData

    What is most valuable?

    • Storage
    • Processing (cost efficient)

    How has it helped my organization?

    With the increase in data size for the business, this horizontal scalable appliance has answered every business question in terms of storage and processing. Hadoop ecosystem has not only provided a reliable distributed aggregation system but has also allowed room for analytics which has resulted in great data insights.

    What needs improvement?

    The Apache team is doing great job and releasing Hadoop versions much ahead of what we can think about. Every room for improvement is fixed as soon as a version is released by ASF. Currently, Apache Oozie 4.0.1 has some compatibility issues with Hadoop 2.5.2.

    For how long have I used the solution?

    2.5 years

    What was my experience with deployment of the solution?

    Not at all.

    What do I think about the stability of the solution?

    We did when we started initially with Hadoop 1.x, which did’t have HA, but now we don’t have any stability issue.

    What do I think about the scalability of the solution?

    Hadoop is known for its scalability. Yahoo stores approx. 455 PB in their Hadoop cluster.

    How are customer service and technical support?

    Customer Service:

    It depends on the Hadoop distributor. I would rate Hortonworks 9/10.

    Technical Support:

    I would rate Hortonworks 9/10.

    Which solution did I use previously and why did I switch?

    We previously used Netezza. We switched because our business required a highly scalable appliance like Hadoop.

    How was the initial setup?

    It's a bit complex in terms of build around for commodities, but soon it will ease up as the product matures.

    What about the implementation team?

    We used a vendor team who were 9/10.

    What was our ROI?

    Valuable storage and processing with a lower cost than previously.

    What's my experience with pricing, setup cost, and licensing?

    Best in pricing and licensing depends on the flavors, but remember it is only good if you have very large data set which cannot be handled by traditional RDBMS.

    Which other solutions did I evaluate?

    Cloud options.

    What other advice do I have?

    First, understand your business requirement; second, evaluate the traditional RDBMS scalability and capability, and finally, if you have reached to the tip of an iceberg (RDBMS) then yes, you definitely need an island (Hadoop) for your business. Feasibility checks are important and efficient for any business before you can take any crucial step. I would also say “Don’t always flow with stream of a river because some time it will lead you to a waterfall, so always research and analyze before you take a ride.”

    Disclosure: I am a real user, and this review is based on my own experience and opinions.
    PeerSpot user
    Buyer's Guide
    Download our free Apache Hadoop Report and get advice and tips from experienced pros sharing their opinions.
    Updated: January 2025
    Product Categories
    Data Warehouse
    Buyer's Guide
    Download our free Apache Hadoop Report and get advice and tips from experienced pros sharing their opinions.