Best Hadoop Online Training
The base Apache Hadoop framework is composed of the following modules:
Hadoop Common – contains libraries and utilities needed by other Hadoop modules;
Hadoop Distributed File System (HDFS) – a distributed file-system that stores data on commodity machines, providing very high aggregate bandwidth across the cluster;
Hadoop YARN – a platform responsible for managing computing resources in clusters and using them for scheduling users' applications; and
Hadoop MapReduce – an implementation of the MapReduce programming model for large-scale data processing.
Comments
Post a Comment