This post describes the process that I followed to install Hortontworks HDP 3.1.0 on a cluster with three VMWare virtual Hosts. This process followes three major steps: 1) set up the cluster environemnt; 2) set up a local repository for both Ambari and HDP stacks; 3) install HDP stacks through Ambari server. You can follow this process to install other versions, you can check product versions through Hortonworks support matrix:

Table of Content

Virtual Nodes Info

First thing I did is set up three virtual hosts in VMWare with following info:

Host Name Host IP Configuration
hadoop-master 4 x cpu, 16 GB Ram, 100 GB disk
hadoop-node-1 4 x cpu, 16 GB Ram, 100 GB disk
hadoop-node-2 4 x cpu, 16 GB Ram, 100 GB disk

for each host, yum has been setup.

Prepare the Cluster Environment

0. Setting Proxy

If you’re behind proxy, then many repositories, including yum repository, are accessed through proxy servers. For each host, you can set up proxy server info through adding

export https_proxy=https://<your.https-proxy.address>:<port#>
export http_proxy=http://<your.http-proxy.address>:<port#>

to the /root/.bashrc. And run source .bashrc to refresh.

1. Make sure the run level is multi-user text mode.

Run following command to check the run level:

systemctl get-default

It is expected to see the response of If not, run following command to change:

systemctl set-default

2. Check and set hostnames.

  • For each host in the cluster, confirm that the hostname is set to a FQDN name by running following command:
hostname -f

This should return a a fully qualified domain name (FQDN), which has a format like You can use hostname command to set the hostname on each host in the cluster. For example:

  • Edit Network Configuration File. For each host in the cluster, open the network configuration file through
vi /etc/sysconfig/network

Modify the HOSTNAME property to FQDN:

  • Edit the hosts file. Open the hosts file on each host in the cluster through vi /etc/hosts add a line to each file. For example: hadoop-master hadoop-node-1 hadoop-node-2

After made these changes. It needs reboot through running reboot.

3. Set up password-less SSH:

  • Login to the hadoop-master host with root user and generate SSH keys using ssh-keygen -t rsa. Press enter for all prompts and accept all default values.
  • Run following command to copy ssh identification for localhost. Enter password when prompted for the password:
ssh-copy-id localhost

Then run command ssh hadoop-master to make sure no password needed.

For each of other hosts in the cluster:

  • Copy the SSH file from hadoop-master to every other hosts in the cluster, for example:
scp -pr /root/.ssh
  • Upload the generated to the root’s .ssh directory as a file with name authorized_keys, for example:
 cat .ssh/ | ssh 'cat >> .ssh/authorized_keys'
  • Set permissions for .ssh directory and authorized_keys file:
ssh; chmod 700 .ssh; chmod 640 .ssh/authorized_keys

From the hadoop-master host, run following commands in sequence, to make sure inter-node connection using SSH without password:

ssh hadoop-node-1
ssh hadoop-node-2
ssh hadoop-master
ssh hadoop-node-2
ssh hadoop-node-1
ssh hadoop-master

4. Enable NTP

Run following commands on each host to install and enable NTP service:

yum install -y ntp
systemctl enable ntpd
systemctl start ntpd

After that run timedatectl status and look for following lines to verfiy that NTP is running

NTP enabled: yes
NTP synchronized: yes

To synchronze:

  • stop ntp serivce: systemctl stop ntpd
  • add server your.ntp.server.address to the /etc/ntp.conf ‘s servers part.
  • Force time synchronize: ntpdate your.ntp.server.address
  • Restart ntp: systemctl start ntpd
  • Run systemctl enable ntpdate to make sure running the ntpdate at boot time.

5. Configuring Firewall

Run following commands to disable firewall on each host in the cluster:

systemctl disable firewalld
service firewalld stop

Run systemctl status firewalld to make sure firewall is disabled.

6. Disable SElinux

For each host in the cluster, change SELINUX value from enhancing to disabled in /etc/selinux/config

7. Install wget

Install wget on all nodes:

yum install -y wget

Set up a Local Repository for Ambari and HDP Stack

1. Create and start an HTTP sever on the master host:

yum install -y httpd
service httpd restart
chkconfig httpd on

Make sure there is the repository /var/www/htmlhas been created on the host.

2. Set up the local repository

  • Download the tarball files for Ambari and HDP stacks through following commands:




The URL’s can be obtained from Ambari Repositories and HDP 3.1.0 Repositories

  • Untar and copy the files to /var/www/html/. For example, tar zxvf ambari- -C /var/www/html/ Then record the base URL’s which are needed for installing the cluster:


  1. make sure you can browser in the web browser;
  2. The path where you can see the repodata directory.

Installing Ambari Server and Agent

1. Download Ambari Repository

  • Login to the hadoop-master host as root
  • Check the repository URL from Ambari Repository Links
  • Download Ambari repository file to the directory/etc/yum.repos.d/, through following command:
wget -nv -O /etc/yum.repos.d/ambari.repo
  • Edit the ambari.repo file and change the baseurl and gpgkey to the local repository obtained above.
  • Run yum repolist to confirm that the repository has been configured successfully. You should see ambari- on the list. See Download Ambari Repository for more information.

2. Install Ambari Server

Install the Ambari server on the master node through command:

yum install -y ambari-server

See Install Ambari Server for more information.

3. Set up Ambari Server

Set -Dhttp.proxyHost=<yourProxyHost> -Dhttp.proxyPort=<yourProxyPort> -Dhttps.proxyHost=<yourProxyHost> -Dhttps.proxyPort=<yourProxyPort> in the file /var/lib/ambari-server/ Run following command on the Ambari server host to start the setup process:

ambari-server setup

See Set Up Ambari Server for more information.

Install, Configure and Deploy the Cluster

1. Start the Ambari Server

ambari-server start

After the server starts successfully, you can login to the server with default user/name admin/admin at:

2. Installing HDP through Installation Wizard

Follow the steps of the Wizard intall HDP:

  • Step 0 Get Started: give a name to your cluster, for example, InfoSecHadoop

  • Step 1 Select Version: Select HDP-3.1, Use Local Repository. Delete all other OS, leave readhat7 only. Copy the Base URL to the places.

  • Step 2 Install Options:


  • Step 3 Confirm Hosts: it will automatically do the regiestration with the setting in Step 2.

  • Step 4 Choose Services: choose basic ones, you can add more later.

  • Step 5 Assign Masters: keep default

  • Step 6: Select all for the Client option.

Error 1: emtpty HDP Url:

Error 2: Requires: libtirpc-devel: Run following commands on all hosts:

subscription-manager repos --enable=rhel-7-server-optional-rpms
yum install -y libtirpc-devel

Error 3 Hive install failed because of mysql-connector-java.jar due to HTTP error: HTTP Error 404: Not Found Run following commands on Ambari server:

yum install -y mysql-connector-java
ls -al /usr/share/java/mysql-connector-java.jar
cd /var/lib/ambari-server/resources/
ln -s /usr/share/java/mysql-connector-java.jar mysql-connector-java.jar

Error 4 Empty Baseurl for Public Repository (No solution, might be proxy issue):

Error 5 Ambari Files View - Service hdfs check failed: Solved: In order to fix it you should try creating a new “File View” instance by clicking on the “Create Instance” button on the File View. You can choose the default options to create the view instance (if it is not kerberized)


Official HDP 3.10 installation documentation:

Apache Ambari Installation Document:

Check Hortonworks Support Matrix to make sure product versions:

Using yum with a Proxy server