Spark On Yarn集群环境搭建

来源:互联网 发布:知乎 北影 编辑:程序博客网 时间:2024/06/05 01:01

一、Scala安装

  1. 下载scala安装包,地址:http://www.scala-lang.org/download/
  2. 配置环境变量
    下载完成后,解压到指定的目录下,在/etc/profile文件中配置环境变量:

    export SCALA_HOME=/usr/local/jiang/scala-2.10.6export PATH=$PATH:$SCALA_HOME/bin
  3. 验证scala的安装

    [root@logsrv03 etc]# scala -versionScala code runner version 2.10.6 -- Copyright 2002-2013, LAMP/EPFL

二、安装spark

  1. 下载spark安装包,地址:http://spark.apache.org/downloads.html
    这里写图片描述
  2. 下载完成后,拷贝到指定的目录下:/usr/local/jiang/,然后解压:

    tar -zxvf spark-1.6.0-bin-hadoop2.6.tgz 
  3. 解压后在配置中添加java环境变量、scala环境变量、hadoop环境变量等conf/spark-env.sh

    # set scala environmentexport SCALA_HOME=/usr/local/jiang/scala-2.10.6# set java environmentexport JAVA_HOME=/usr/local/jdk1.7.0_71# set hadoopexport HADOOP_HOME=/usr/local/jiang/hadoop-2.7.1export HADOOP_CONF_DIR=$HADOOP_HOME/etc/hadoop# set sparkSPARK_MASTER_IP=logsrv03SPARK_LOCAL_DIRS=/usr/local/jiang/spark-1.6.0SPARK_DRIVER_MEMORY=1G
  4. 配置从机conf/slaves

    logsrv02logsrv04

    当然这里配置的是主机名,所以在/etc/hosts中一定要添加主机名和ip的映射,不然没法识别的:

    127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4::1         localhost localhost.localdomain localhost6 localhost6.localdomain6172.17.6.141 logsrv01172.17.6.142 logsrv02172.17.6.149 logsrv04172.17.6.148 logsrv03172.17.6.150 logsrv05172.17.6.159 logsrv08172.17.6.160 logsrv09172.17.6.161 logsrv10172.17.6.164 logtest01172.17.6.165 logtest02172.17.6.166 logtest03172.30.2.193 devops172302193172.30.2.194 devops172302194172.30.2.195 devops172302195
  5. 将配置好的spark-1.6.0-bin-hadoop2.6文件远程拷贝到相对应的从机中:

    [root@logsrv03 jiang]# scp -r spark-1.6.0-bin-hadoop2.6 root@logsrv02:/usr/local/jiang/[root@logsrv03 jiang]# scp -r spark-1.6.0-bin-hadoop2.6 root@logsrv04:/usr/local/jiang/
  6. 启动集群

    [root@logsrv03 spark-1.6.0-bin-hadoop2.6]# sbin/start-all.sh 
  7. 启动完成后,查看主从机的进程:
    主机:

    [root@logsrv03 spark-1.6.0-bin-hadoop2.6]# jps 25325 NameNode23973 Master17643 ResourceManager25523 SecondaryNameNode28839 Jps

    从机:

    [root@logsrv02 spark-1.6.0-bin-hadoop2.6]# jps744 Worker4406 Jps2057 DataNode2170 NodeManager

    三、安装完成后,可以查看spark的UI:
    这里写图片描述
    运行wordcout例子:
    命令:

 ./bin/spark-submit \ --name JavaKafkaWordCount \ --master spark://logsrv03:7077 \ --executor-memory 1G \ --class examples.streaming.JavaKafkaWordCount \ log_spark-0.0.1-SNAPSHOT.jar 172.17.6.142:2181,172.17.6.148:2181,172.17.6.149:2181 11 log-topic 5

运行后spark的UI:
这里写图片描述
这里写图片描述

0 1
原创粉丝点击