Open Source Framework to Process Massive Datasets: Apache Hadoop
If you are planning on a data intensive product and want to avoid headache when the product would be transacting trillions of bytes, this is the right time to consider Apache Hadoop and make sure that you product would be built on a very flexible and clustered platform.
Apache Hadoop Common is an open source framework for running applications on large cluster built of commodity hardware. It provides applications both reliability and data motion by implementing a computational paradigm. Inspired by Google’s MapReduce and Google File System (GFS), Hadoop is contributed and used by global community of developers using Java language. Yahoo!, being the largest contributor and consumer, is using Hadoop extensively. Other major users are Amazon.com, IBM, Twitter, Stumbleupon, Ning, eBay, AOL, Apple and Image Shack.
Recent Comments