一、安裝準備
- 創建hadoop賬號
- 更改ip
安裝Java 更改/etc/profile 配置環境變量
export $JAVA_HOME=/usr/java/jdk1.7.0_71
修改host文件域名
172.16.133.149 hadoop101 172.16.133.150 hadoop102 172.16.133.151 hadoop103
- 安裝ssh 配置無密碼登錄
解壓hadoop
/hadoop/hadoop-2.6.2
二、修改conf下面的配置文件
依次修改hadoop-env.sh、core-site.xml、hdfs-site.xml、mapred-site.xml、yarn-site.xml和slaves文件。
1.hadoop-env.sh
`#添加JAVA_HOME:`
`export JAVA_HOME=/usr/java/jdk1.7.0_71`
2.core-site.xml
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/hadoop/hadoop-2.6.2/hdfs/tmp</value>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://hadoop101:9000</value>
</property>
</configuration>
3.hdfs-site.xml
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:///hadoop/hadoop-2.6.2/hdfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>file:///hadoop/hadoop-2.6.2/hdfs/data</value>
</property>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>hadoop101:9001</value>
</property>
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>
</configuration>
4.mapred-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
5.yarn-site.xml
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop101</value>
</property>
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>
</configuration>
6.slaves
hadoop102
hadoop103
7.最後,將整個hadoop-2.6.2文件夾及其子文件夾使用scp複製到兩臺Slave(hadoop102、hadoop103)的相同目錄中:
scp -r \hadoop\hadoop-2.6.2\ hadoop@hadoop102:\hadoop\
scp -r \hadoop\hadoop-2.6.2\ hadoop@hadoop103:\hadoop\
三、啓動運行Hadoop(進入hadoop文件夾下)
格式化NameNode
dfs namenode -format
啓動Namenode、SecondaryNameNode和DataNode
[hadoop@hadoop101]$ start-dfs.sh
啓動ResourceManager和NodeManager
[hadoop@hadoop101]$ start-yarn.sh
最終運行結果
四、測試Hadoop
測試HDFS
瀏覽器輸入http://<-NameNode主機名或IP->:50070
測試ResourceManager
瀏覽器輸入http://<-ResourceManager所在主機名或IP->:8088