0% found this document useful (0 votes)
49 views5 pages

Hadoop Installation Final

Download as odt, pdf, or txt
Download as odt, pdf, or txt
Download as odt, pdf, or txt
You are on page 1/ 5

Hadoop Installation

Copy below installation file from http://www-eu.apache.org/dist/hadoop/common/hadoop-2.6.5/

hadoop-2.6.5.tar.gz
1. to home folder ( Login is same as India domain login) .

2. Extract tar file


tar -xvf hadoop-2.6.5.tar.gz

3. Generate Key

• [xxxxxx@01hw744052 hadoop-2.6.5]$ ssh-keygen


Generating public/private rsa key pair.
Enter file in which to save the key (/home/xxxxxx/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/xxxxxx/.ssh/id_rsa.
Your public key has been saved in /home/xxxxxx/.ssh/id_rsa.pub.
The key fingerprint is:
ee:d9:c8:a7:80:cc:6b:d8:d7:c6:47:b6:62:b2:00:ad xxxxxx@01hw744052.tcs-mobility.com
The key's randomart image is:
+--[ RSA 2048]----+
| |
| |
| |
| . |
| .. S |
| =.. o |
| Eo= .o.o . |
| . ooo+*++ |
| ....=*=. |
+-----------------+
1. Setup passwordless ssh
[xxxxxx@01hw744052 hadoop-2.6.5]$ ssh-copy-id -i empid@localhost
The authenticity of host 'localhost (<no hostip for proxy command>)' can't be established.
RSA key fingerprint is 04:f4:d1:1c:62:ad:1f:29:71:5c:50:bd:93:49:a9:68.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added 'localhost' (RSA) to the list of known hosts.
xxxxxx@localhost's password:
Now try logging into the machine, with "ssh 'xxxxxx@localhost'", and check in:

.ssh/authorized_keys
to make sure we haven't added extra keys that you weren't expecting.
2. Verify passwordless ssh

[xxxxxx@01hw744052 hadoop-2.6.5]$ ssh empid@localhost

3. Locate java path to update JAVA_HOME variable using


locate jps
or
which java

eg:
[aauser@dc2100-r1-node1 ~]$ which java
/usr/bin/java
[aauser@dc2100-r1-node1 ~]$ ll /usr/bin/java
lrwxrwxrwx 1 root root 22 Aug 3 18:31 /usr/bin/java -> /etc/alternatives/java
[aauser@dc2100-r1-node1 ~]$ ll /etc/alternatives/java
lrwxrwxrwx 1 root root 39 Aug 3 18:31 /etc/alternatives/java -> /usr/java/jdk1.7.0_67-cloudera/bin/java

Here, Java home is /usr/java/jdk1.7.0_67-cloudera/

4. Update JAVA_HOME in /etc/hadoop/hadoop-env.sh


export JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera/

5. Set the PATH and CLASSPATH variables appropriately in the .bashrc file in your home folder.

export HADOOP_HOME=/home/<empid>/ hadoop-2.6.5

export JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera/
export HADOOP_PREFIX=$HADOOP_HOME

export HADOOP_HDFS_HOME=$HADOOP_HOME/libexec

export CLASSPATH= /home/<empid>/ hadoop-2.6.5


/share/hadoop/hdfs:/home/<empid>/ hadoop-2.6.5
/share/hadoop/hdfs/lib:/home/<empid>/ hadoop-2.6.5
/share/hadoop/common:/home/<empid>/ hadoop-2.6.5/share/hadoop/common/lib

export PATH=$PATH:$HADOOP_HOME/bin:$JAVA_HOME/bin

export HADOOP_INSTALL=$HADOOP_HOME
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export YARN_HOME=$HADOOP_HOME
export HADOOP_CONF_DIR=$HADOOP_HOME
export HADOOP_PREFIX=$HADOOP_HOME
export HADOOP_LIBEXEC_DIR=$HADOOP_HOME/libexec
export JAVA_LIBRARY_PATH=$HADOOP_HOME/lib/native:$JAVA_LIBRARY_PATH
export HADOOP_CONF_DIR=$HADOOP_PREFIX/etc/hadoop

export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native

export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib"

export HADOOP_HOME=/home/1099404/hadoop-2.3.0-cdh5.1.0

6. Set the hadoop env variables : source .bashrc

7. Edit the core-site.xml( in /etc/hadoop) file as follows


<property>
<name>fs.default.name</name>
<value>hdfs://127.0.0.1:9000</value>
</property>
<property>
<name>fs.tmp.dir</name>
<value>/home/<empid>/hadoop-2.6.5/tmp</value>
</property>
NB: Make sure directory configured as fs.tmp.dir are created.

8. Edit the hdfs-site.xml ( in /etc/hadoop)


//Update the path to reflect your installation home directory

<property>
<name>dfs.name.dir</name>
<value>/home/<empid>/hadoop-2.6.5/dfs/nn</value>
</property>
<property>
<name>dfs.data.dir</name>
<value>/home/<empid>/hadoop-2.6.5/dfs/dn</value>
</property>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
NB: Make sure the directories configured as dfs.name.dir and dfs.data.dir are created.

9. Edit the yarn-site.xml ( in /etc/hadoop)

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>

10. Edit mapred-site.xml ( in /etc/hadoop)

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>

11. Format the name node


hadoop namenode -format

note: If there is an error in this step check following commands . If the commands are showing
error then contact IS team.

I. Hostname -i
[xxxxxx@01hw744052]$ hostname -i
it will give 127.0.1.1 if everything is ok

ii. hostname
[xxxxxx@01hw744052]$ hostname
eg:01hw664400.ln.india.tcs.com

12. Starting your single-node cluster

To start all the daemons run start-all.sh [present in sbin folder of installation dir]

13. To view the started daemons , run jps

14. To see files in HDFS

hadoop fs -ls /

//Verify the working

15. Create a directory in HDFS

hadoop fs -mkdir /input

16. Copy files to HDFS

hadoop fs -copyFromLocal <localpath-input-file> /input/

17. Run a sample MapReduce job


//Copy local data to HDFS .Run the MapReduce job. Now, we actually run the WordCount example
job.
*abc is input file name

hadoop jar /home/<empid>/ hadoop-2.6.5/share/hadoop/mapreduce1/hadoop-examples-2.3.0-mr1-


cdh5.1.0.jar wordcount hdfs://127.0.0.1:9000/input/abc hdfs://localhost:9000/output

18. Output is as shown below

[xxxxxx@01hw743932 ~]$ hadoop fs -ls /output1/


15/05/22 12:35:48 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your
platform... using builtin-java classes where applicable
Found 2 items
-rw-r--r-- 1 xxxxxx supergroup 0 2015-05-22 12:35 /output1/_SUCCESS -rw-r--r-- 1 xxxxxx
supergroup 11 2015-05-22 12:35 /output1/part-r-00000

[xxxxxx@01hw743932 ~]$ hadoop fs -cat /output1/part-r-00000


15/05/22 12:36:07 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your
platform... using builtin-java classes where applicable
Leo 4
hi 2

19. To copy files from HDFS onto local system, use below command. (Optional)
hadoop fs -copyToLocal /output1/part-r-00000 .

You might also like