Software Engineer at a aerospace/defense firm with 1,001-5,000 employees
Vendor
2016-12-31T03:25:54Z
Dec 31, 2016
It depends...what is your endgame ?
Hadoop these days mostly servers as a distributed clustering file system that specializes in storing very large files. If you are merely interested in writing software for distributed processing....Apache Spark, or NVIDIA CUDA are a much better choice....if you are interested in the distributed processing of large amounts of data, then the common practice is to use Apache Spark to write the code to process the data, and Hadoop for persistent file system storage.
Hadoop architectures refer to the various design patterns and configurations used to implement Hadoop, an open-source framework for distributed storage and processing of large datasets.
It depends...what is your endgame ?
Hadoop these days mostly servers as a distributed clustering file system that specializes in storing very large files. If you are merely interested in writing software for distributed processing....Apache Spark, or NVIDIA CUDA are a much better choice....if you are interested in the distributed processing of large amounts of data, then the common practice is to use Apache Spark to write the code to process the data, and Hadoop for persistent file system storage.
The enterprise readiness of the distribution.