News

Two Google Fellows just published a paper in the latest issue of Communications of the ACM about MapReduce, the parallel programming model used to process more than 20 petabytes of data every day ...
Hadoop is the most significant concrete technology behind the so called 'Big Data' revolution. Hadoop combines an economical model for storing massive quantities of data - the Hadoop Distributed File ...
Hadoop Streaming allows developers to use virtually any programming language to create MapReduce jobs, but it’s a bit of a kludge. The MapReduce programming environment needs to be pluggable.