Here's the reality: Every sliver of data is going to land on HDFS as the most trusted and authoritative resource or 'record of truth.' It is the most cost-effective highly available storage mechanism available. Batch computing is here to stay, and Hadoop MapReduce will be a big part of it.
One might bet against Hadoop MapReduce, but betting against the Hadoop Filesystem as cheap storage built on commodity hardware that can serve large data in a highly available fashion is... misguided. Nothing else scales to 10,000 nodes per cluster and provides data locality (processor near disk spindles) so that data is accessible, or is even close.
Many systems will sit in front of Hadoop to do things other than batch computing, and many new types of distributed compute systems will sit on top of Hadoop and Zookeeper. Hadoop is here to stay.
MapReduce is too low level, and systems like Pig and Hive will continue to grow, improve and be the standard interface to working with Hadoop.
One might bet against Hadoop MapReduce, but betting against the Hadoop Filesystem as cheap storage built on commodity hardware that can serve large data in a highly available fashion is... misguided. Nothing else scales to 10,000 nodes per cluster and provides data locality (processor near disk spindles) so that data is accessible, or is even close.
Many systems will sit in front of Hadoop to do things other than batch computing, and many new types of distributed compute systems will sit on top of Hadoop and Zookeeper. Hadoop is here to stay.
MapReduce is too low level, and systems like Pig and Hive will continue to grow, improve and be the standard interface to working with Hadoop.