What is apache hadoop?

Apache Hadoop controls costs by storing data more affordably per terabyte than other platforms. Instead of thousands to tens of thousands per terabyte, Hadoop delivers compute and storage for hundreds of dollars per terabyte.
 
Hadoop is an open source Apache framework written in java allows for distributed processing large data sets on computer clusters via simple programming model.
 
Apache Hadoop is an open-source software framework used for distributed storage and processing of dataset of big data using the MapReduce programming model. It consists of computer clusters built from commodity hardware.
 
The latest release came out about a month ago. It has 30 new features with around 500 or more subtasks, hundreds and hundreds of bugfixes since 2.8.2 version.
 
The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using simple programming models. It is designed to scale up from single servers to thousands of machines, each offering local computation and storage.
 
Back
Top