在hadoop2.4集群环境下安装spark
来源:互联网 发布:mac 配置ant 编辑:程序博客网 时间:2024/04/26 13:55
1、下载spark最新版:http://spark.apache.org/downloads.html
下载对应的Scala版本:http://www.scala-lang.org/download/
2、解压文件:
root@Mhadoop:/home/hadoop/Downloads# tar -zxf spark-1.3.1-bin-hadoop2.4.tgz
3、创建spark目录:
root@Mhadoop:/home/hadoop/Downloads# mkdir /usr/local/spark
root@Mhadoop:/home/hadoop/Downloads# mv spark-1.3.1-bin-hadoop2.4 /usr/local/spark/
root@Mhadoop:/home/hadoop/Downloads# cd /usr/local/spark/
root@Mhadoop:/usr/local/spark# ls
spark-1.3.1-bin-hadoop2.4
4、同样安装Scala;
root@Mhadoop:/home/hadoop/Downloads# tar -zxf ./scala-2.11.6.tgz
root@Mhadoop:/home/hadoop/Downloads# ls
scala-2.11.6 scala-2.11.6.tgz spark-1.3.1-bin-hadoop2.4.tgz
root@Mhadoop:/home/hadoop/Downloads# mkdir /usr/lib/scala
root@Mhadoop:/home/hadoop/Downloads# mv ./scala-2.11.6 /usr/lib/scala/
5、配置环境变量:
geidt ~/.bashrc
加入以下内容:
export PATH=$PATH:$HADOOP_HOME/bin:$SCALA_HOME/bin:$SPARK_HOME/bin
export SCALA_HOME=/usr/lib/scala/scala-2.11.6
export SPARK_HOME=/usr/local/spark/spark-1.3.1-bin-hadoop2.4
使配置生效:source ~/.bashrc
6、配置spark:
进入conf目录,修改spark-env.sh
export JAVA_HOME=/usr/lib/jvm/java-7-oracle
export HADOOP_CONF_DIR=/usr/local/hadoop/etc/hadoop
export SPARK_MASTER_IP=Mhadoop
export SPARK_WORKER_MEMORY=1g
export SCALA_HOME=/usr/lib/scala/scala-2.11.6
export SPARK_HOME=/usr/local/spark/spark-1.3.1-bin-hadoop2.4
修改slaves文件;
Mhadoop
Shadoop1
Shadoop2
7、配置slave节点:
copy文件到slave节点:
sudo scp scala-2.11.6.tgz spark-1.3.1-bin-hadoop2.4.tgz hadoop@Shadoop1:/home/hadoop/Downloads/
sudo scp scala-2.11.6.tgz spark-1.3.1-bin-hadoop2.4.tgz hadoop@Shadoop2:/home/hadoop/Downloads/
在两个slave节点做同样配置
hadoop@Shadoop1:~$ sudo mkdir /usr/lib/scala
[sudo] password for hadoop:
hadoop@Shadoop1:~$ sudo mkdir /usr/local/spark
hadoop@Shadoop1:~$ cd /home/hadoop/Downloads/
hadoop@Shadoop1:~/Downloads$ ls
scala-2.11.6.tgz spark-1.3.1-bin-hadoop2.4.tgz
hadoop@Shadoop1:~/Downloads$ tar -zxf scala-2.11.6.tgz
hadoop@Shadoop1:~/Downloads$ tar -zxf spark-1.3.1-bin-hadoop2.4.tgz
hadoop@Shadoop1:~/Downloads$ sudo mv scala-2.11.6 /usr/lib/scala/
hadoop@Shadoop1:~/Downloads$ sudo mv spark-1.3.1-bin-hadoop2.4 /usr/local/spark/
复制环境信息到slave节点:
hadoop@Mhadoop:~/Downloads$ sudo scp /home/hadoop/.bashrc hadoop@Shadoop1:/home/hadoop
hadoop@shadoop1's password:
.bashrc 100% 3972 3.9KB/s 00:00
hadoop@Mhadoop:~/Downloads$ sudo scp /home/hadoop/.bashrc hadoop@Shadoop2:/home/hadoop
hadoop@shadoop2's password:
.bashrc 100% 3972 3.9KB/s 00:00
hadoop@Mhadoop:~/Downloads$ sudo scp /usr/local/spark/spark-1.3.1-bin-hadoop2.4/conf/* hadoop@Shadoop1:/usr/local/spark/spark-1.3.1-bin-hadoop2.4/conf
hadoop@shadoop1's password:
fairscheduler.xml.template 100% 303 0.3KB/s 00:00
log4j.properties.template 100% 620 0.6KB/s 00:00
metrics.properties.template 100% 5371 5.3KB/s 00:00
slaves 100% 97 0.1KB/s 00:00
slaves~ 100% 97 0.1KB/s 00:00
slaves.template 100% 80 0.1KB/s 00:00
spark-defaults.conf.template 100% 507 0.5KB/s 00:00
spark-env.sh 100% 3482 3.4KB/s 00:00
spark-env.sh~ 100% 3482 3.4KB/s 00:00
spark-env.sh.template 100% 3217 3.1KB/s 00:00
hadoop@Mhadoop:~/Downloads$ sudo scp /usr/local/spark/spark-1.3.1-bin-hadoop2.4/conf/* hadoop@Shadoop2:/usr/local/spark/spark-1.3.1-bin-hadoop2.4/conf
hadoop@shadoop2's password:
fairscheduler.xml.template 100% 303 0.3KB/s 00:00
log4j.properties.template 100% 620 0.6KB/s 00:00
metrics.properties.template 100% 5371 5.3KB/s 00:00
slaves 100% 97 0.1KB/s 00:00
slaves~ 100% 97 0.1KB/s 00:00
slaves.template 100% 80 0.1KB/s 00:00
spark-defaults.conf.template 100% 507 0.5KB/s 00:00
spark-env.sh 100% 3482 3.4KB/s 00:00
spark-env.sh~ 100% 3482 3.4KB/s 00:00
spark-env.sh.template 100% 3217 3.1KB/s 00:00
8、启动hadoop和spark:
首先启动hadoop:用jps查看
hadoop@Mhadoop:/usr/local/hadoop/sbin$ jps
4316 ResourceManager
4167 SecondaryNameNode
4596 Jps
3861 NameNode
再进入spark的sbin目录,输入命令:./start-all.sh
注意必须标明是当前目录下的,否则执行的可能是hadoop的sbin目录下的;
再用jps查看:
hadoop@Mhadoop:/usr/local/spark/spark-1.3.1-bin-hadoop2.4/sbin$ jps
6280 Master
4316 ResourceManager
4167 SecondaryNameNode
6571 Worker
3861 NameNode
6625 Jps
多了Master和Worker进程。(slave节点都多了worker进程)
9、大功告成,进入8080端口
- 在hadoop2.4集群环境下安装spark
- Ubuntu14(64位) 集群环境下安装Hadoop2.4
- Hadoop2.2.0集群在RHEL6.2下的安装实战
- Hadoop2.2.0集群在RHEL6.2下的安装实战
- hadoop2.7.1在centos6.5下集群安装配置
- spark学习1--centOS7.2下基于hadoop2.7.3的spark2.0集群环境搭建
- 【Spark】Spark2.0.2在Hadoop2.7.3上的安装 二(集群安装)
- Hadoop2.2集群安装配置-Spark集群安装部署
- hadoop2.2 在伪分布式环境下安装hbase
- Spark 1.2 集群环境安装
- ubuntu14.04下hadoop2.2.0集群安装
- Ubuntu14.04下hadoop2.6.0集群安装
- 在集群上安装 hadoop2.5.2
- hadoop集群下spark安装
- 在集群上安装spark
- Linux安装Spark集群(CentOS7+Spark2.1.1+Hadoop2.8.0)
- Linux安装Spark集群(CentOS7+Spark2.1.1+Hadoop2.8.0)
- (自总结详细资料)如何在CentOS7下安装hadoop2.8分布式集群
- 淺談Rails 4 中Strong Parameters機制
- 巧克力爱好者匿名用例图 小组成员:张猛 李轩 张雨 谷雨 韩蕊
- 常见面试之机器学习算法思想简单梳理(进阶版)
- 邮票分你一半
- 第六届蓝桥杯java试题-九数组分数
- 在hadoop2.4集群环境下安装spark
- protocol、category和继承的区别
- 经典算法题之Happy Number
- android组件之service
- Nova 是如何统计 OpenStack 计算资源
- uva 划分成回文串 11584 题解
- 开启一个新的终端并执行特定的命令
- 面试题45:圆圈中最后剩下的数字
- SQLite数据库