Ah!! The name is everywhere, carried with the wind. Apache Hadoop!!
The BIG DATA crunching platform!
We all know how alien it can be at start too! Phew!! :o
Its my personal experience, nearly 11 months before, I was trying to install HBase, I faced few issues! The problem was version compatibility. Ex: “HBase some x.version” with “Hadoop some y.version”.
This is a real issue because you will never know which package of what version blends well with the other, unless, someone has tested it. This testing again depends on the environment where they have set up and could be another issue.
There was a pressing demand for the management of distributions and then comes an open source project which attempts to create a fully integrated and tested Big Data management distribution, “Apache Bigtop“.
Goals of Apache Bigtop:
of all the sub-projects of Hadoop. This project aims at system as a whole, than the individual project.
I love the way Doug Cutting quoted in the Keynote, back then, wherein he expressed the similarity between Hadoop and Linux kernel,and the corresponding similarity between the big stack of Hadoop ( Hive, Hbase, Pig, Avro, etc.) and the fully operational operating systems with its distributions (RedHat, Ubuntu, Fedora, Debian etc.). This is an awesome analogy! :)
Life is made easy with Bigtop:
Bigtop Hadoop distribution artifacts won’t make you feel that you live in an alien world! After installing, you will get a chance to blend a Hadoop cluster in any mode, with the sub-projects of it. Its all for you to garnish next! :)
Setup Of Bigtop and Installing Hadoop:
It’s time to welcome all your packages home. [I also mean /home/..] ;)
I’ve tested on Ubuntu 11.04 and here goes a quick and easy installation process.
Step 1: Installing the GNU Privacy Guard key, a key management system to access all public key directories.
Step 3: Updating the apt cache
sudo apt-get update
Step 4: Checking in the artifacts
sudo apt-cache search hadoop
|Search in the apt cache|
Step 5: Set your JAVA_HOME
export $JAVA_HOME in ~/.bashrc
Step 6: Installing the complete Hadoop stack
sudo apt-get install hadoop\*
Step 1: Formatting the namendoe
sudo -u hdfs hadoop namenode -format
|Formatting the namenode|
Step 2: Starting the Namenode, Datanode, Jobtracker, Tasktracker of Hadoop
for i in hadoop-namenode hadoop-datanode hadoop-jobtracker hadoop-tasktracker ; do sudo service $i start ; done
Now, the cluster is up and running.
|Start all the services|
Step 3: Creating a new directory in hdfs
sudo -u hdfs hadoop fs -mkdir /user/bigtop
bigtop is the directory name in the user $USER
sudo -u hdfs hadoop fs -chown $USER /user/bigtop
|Create a directory in HDFS|
Step 4: List the directories in file system
hadoop fs -lsr /
Step 5: Running a sample pi example
hadoop jar /usr/lib/hadoop/hadoop-examples.jar pi 10 1000
|Running a sample program|
Enjoy with your cluster! :) We shall see what more blending could be done with Hadoop (with Hive, Hbase, etc.) in the next post! Until then, Happy Learning!! :):)
Reference: Hadoop Hangover : Introduction To Apache Bigtop and Playing With It (Installing Hadoop)! from our JCG partner Swathi V at the * Techie(S)pArK * blog.