spark--环境搭建--2.Hadoop2.4.1集群搭建

在spark1上操作

1. 安装Hadoop

$ cd /usr/local

$ tar -zxvf hadoop-2.4.1.tar.gz

$ mv hadoop-2.4.1 hadoop

$ vi ~/.bashrc

export HADOOP_HOME=/usr/local/hadoop
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

$ source .bashrc

2. Hadoop配置

$ cd /usr/local/hadoop/etc/hadoop

$ vi core-site.xml

<property>
 <name>fs.default.name</name>
 <value>hdfs://spark1:9000</value>
</property>

$ mkdir /usr/local/data/

$ vi hdfs-site.xml

<property>
 <name>dfs.name.dir</name>
 <value>/usr/local/data/namenode</value>
</property>
<property>
 <name>dfs.data.dir</name>
 <value>/usr/local/data/datanode</value>
</property>
<property>
 <name>dfs.tmp.dir</name>
 <value>/usr/local/data/tmp</value>
</property>
<property>
 <name>dfs.replication</name>
 <value>3</value>
</property>

$ vi mapred-site.xml.template

<property>
 <name>mapreduce.framework.name</name>
 <value>yarn</value>
</property>

$ vi yarn-site.xml

<property>
 <name>yarn.resourcemanager.hostname</name>
 <value>spark1</value>
</property>
<property>
 <name>yarn.nodemanager.aux-services</name>
 <value>mapreduce_shuffle</value>
</property>

$ vi slaves

spark1
spark2
spark3

$ scp -r hadoop root@spark2:/usr/local/

$ scp ~/.bashrc root@spark3:~/

 

在spark2和spark3的 /usr/local/ 目录下创建 data 目录, 以及source .bashrc

$ mkdir /usr/local/data

$ source .bashrc

 

在spark1上执行

# 格式化namenode
$ hdfs namenode -format
# 启动集群
$ start-dfs.sh
# 查看启动情况
$ jps
spark1
1424 SecondaryNameNode
1324 DataNode
1218 NameNode
1554 Jps
spark2
1619 Jps
1555 DataNode
spark3
1473 DataNode
1537 Jps
# 在浏览器打开  http://spark1:50070/

 

# 启动yarn集群
$ start-yarn.sh
$ jps
spark1
ResourceManager
NodeManager
spark2
NodeManager
spark3
NodeManager
# 在浏览器打开  http://spark1:8088/

 

posted @ 2018-03-18 14:27  _殇j  阅读(195)  评论(0编辑  收藏  举报