Install Hadoop on multiple nodes using Ubuntu 15.10


We give a brief introduction of Hadoop in previous tutorial, for today we will learn to install Hadoop on multiple nodes, in demonstration scenario  we will be using Ubuntu 15.10 as Desktop, we will create 2 Slave or Data Nodes along with 1 Name node. Make sure you have shared ssh public keys with Data nodes and assign appropriate IP addresses, host name and other Hadoop  services (we will mention in tutorial) required to run Hadoop multiple cluster node.


we will be using Ubuntu 15.10 as 1 master node, 2 Slave/data nodes. hostname for namenode  will be masternode, datanodes  will have hostname slave1 and  slave2 respectively.

masternode IP address:

Slave1 IP Address:

Slave2 IP Address:


Instillation process is similar to previous tutorial except few changes. First of all let us configure master node .

Define hostname of Namenode

# vim /etc/hostname



Define hosts in /etc/hosts file

# vim /etc/hosts

Sample output       localhost    masternode    slave1    slave2

Configure Hadoop Services

# cd /usr/local/hadoop/etc/hadoop/

Edit hdfs-site.xml

# vim /usr/local/hadoop/etc/hadoop/hdfs-site.xml

File will look like below, change replication value to 3.


Make sure that you possess a namenode directory under /usr/local/hadoop

# mkdir -p /usr/local/hadoop/hadoopdata/hdfs/namenode
# sudo chown -R hadoop:hadoop /usr/local/hadoop/

Similarly edit yarn-site.xml, it  will look like below, make sure you have assigned hostname of masternode appropriately

# vim yarn-site.xml

Sample output


Make sure core-site.xml have appropriate hostname



Create a file named slaves under /usr/local/hadoop/etc/hadoop directory  and assign hostnames of datanodes

# vim /usr/local/hadoop/etc/hadoop/slaves

Put following entries


Similarly create file named mastersunder same directory hierarchy

# vim /usr/local/hadoop/etc/hadoop/masters

Enter following


We have a working master node at this stage, let us create 2 slave nodes. We created two  clone virtual machines using VirtualBox, first clone is slave1 and second cone is slave2, as this machine is clone of Masternode so we will be having all of the hadoop configuration files (.xml) in ready to use form.


Similarity create another clone for slave2 datanode.

Change IP address to


Change hostname to slave1 and reboot the system. Replete the process for another VirtualBox Clone which will be used as slave2,assign IP address to slave2.

Name we have one NameNode (masternode) with IP address and two datanodes (slave1, slave2).


Now switch back to master node and share ssh rsa keys with slave1 and slave2, so that there is no need for ssh passwords.

# ssh-keygen -t rsa
# ssh hadoop@ "chmod 755 .ssh; chmod 640 .ssh/authorized_keys"

# cat .ssh/ | ssh hadoop@ 'cat >> .ssh/authorized_keys'

# ssh hadoop@ "chmod 755 .ssh; chmod 640 .ssh/authorized_keys"

Reboot all three systems to make sure all things are going smooth.

Edit hdfs-site.xml file of slave1 and slave2 data nodes make sure you have following entries


Create  /usr/local/hadoop/hadoopdata/hdfs/datanode directory  on both data nodes

# mkdir -p /usr/local/hadoop/hadoopdata/hdfs/datanode
# chown -R hadoop:hadoop /usr/local/hadoop/

Go to Masternode and run start node services

# cd /usr/local/hadoop/sbin && ls


Run all node services

# ./


We can see that both of datanodes (slave1, slave2) are working properly.

Run jps command on  Masternode

# jps

Sample output

8499 SecondaryNameNode
8922 Jps
8650 ResourceManager

Swith to Slave1 and run jps command again

# ssh hadoop@slave1

# jps

Sample output, datanode is working

4373 DataNode
4499 NodeManager
4671 Jps

Similarly in slave2 datanode is working perfectly


Multinode  Hadoop Cluster installation process is over at that  stage.

Open browser and type <change IP addr in your scenario>

Taht Selection_017

Thats it! Have Fun!!