Apache Hadoop is a framework that allows for the distributed processing of large data sets across clusters of commodity computers using a simple programming model. It provides a distributed file system and map/reduce programming model. Hadoop can scale from a single server to thousands of machines, each offering local computation and storage.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as TXT, PDF, TXT or read online on Scribd
0 ratings0% found this document useful (0 votes)
33 views
Notes
Apache Hadoop is a framework that allows for the distributed processing of large data sets across clusters of commodity computers using a simple programming model. It provides a distributed file system and map/reduce programming model. Hadoop can scale from a single server to thousands of machines, each offering local computation and storage.