spark和zeppelin实践一:安装hadoop篇

来源:互联网 发布:pcb开料软件 编辑:程序博客网 时间:2024/05/29 18:39

一、安装JDK

1.7 JDK下载地址:http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html

下载后安装

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. rpm -ivh jdk-8u112-linux-x64.rpm  

设置JDK环境变量

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. export JAVA_HOME=/usr/java/jdk1.8.0_112  
  2. export CLASSPATH=$JAVA_HOME/lib/tools.jar    
  3. export PATH=$JAVA_HOME/bin:$PATH    

二、安装Hadoop

1、DNS绑定
vi /etc/hosts,增加一行内容,如下(这里我的Master节点IP设置的为192.168.80.100):

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. 192.168.80.100 IMM-SJJ01-Server18  

2、SSH的免密码登录

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. cd /home/data/.ssh  
  2. ssh-keygen -t rsa  
  3. cat id_rsa.pub >> authorized_keys  

3、安装Hadoop

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. #http://hadoop.apache.org/releases.html  
  2. #wget http://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz  
  3.   
  4. cd /home/game/soft  
  5. tar zxvf hadoop-2.7.3.tar.gz  
  6. ln -s /home/game/soft/hadoop-2.7.3 /home/game/soft/hadoop  

#4、配置

1) 设置Hadoop环境变量

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. vim ~/.bash_profile 或 /etc/profile  
  2. export HADOOP_HOME=/home/game/soft/hadoop  
  3. export PATH=$HADOOP_HOME/bin:$PATH  
  4.   
  5. echo $HADOOP_HOME  

2)修改hadoop-env.sh

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. vim $HADOOP_HOME/etc/hadoop/hadoop-env.sh  
  2. export JAVA_HOME=${JAVA_HOME} 改为  
  3. export JAVA_HOME=/usr/java/jdk1.8.0_112  
3)修改/etc/hosts

4)修改core-site.xml

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. cd $HADOOP_HOME  
  2. cp ./share/doc/hadoop/hadoop-project-dist/hadoop-common/core-default.xml ./etc/hadoop/core-site.xml  
  3. cp ./share/doc/hadoop/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml ./etc/hadoop/hdfs-site.xml  
  4. cp ./share/doc/hadoop/hadoop-yarn/hadoop-yarn-common/yarn-default.xml ./etc/hadoop/yarn-site.xml  
  5. cp ./share/doc/hadoop/hadoop-mapreduce-client/hadoop-mapreduce-client-core/mapred-default.xml ./etc/hadoop/mapred-site.xml  
  6.   
  7.   
  8. vim $HADOOP_HOME/etc/hadoop/core-site.xml  
  9. <property>  
  10.   <name>fs.default.name</name>  
  11.   <value>hdfs://192.168.80.100:19000</value>  
  12. </property>  
  13. <property>  
  14.   <name>hadoop.tmp.dir</name>  
  15.   <value>/home/game/hadoop/tmp</value>  
  16. </property>  
5)修改配置hdfs-site.xml

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. <property>  
  2.   <name>dfs.namenode.rpc-address</name>  
  3.   <value>192.168.80.100:19001</value>  
  4. </property>  
  5.   
  6. <property>  
  7.   <name>dfs.namenode.http-address</name>  
  8.   <value>0.0.0.0:10070</value>  
  9. </property>  
6)修改mapred-site.xml

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. cp mapred-site.xml.template mapred-site.xml  
  2. <property>  
  3.   <name>mapreduce.framework.name</name>  
  4.   <value>yarn</value>  
  5. </property>  

7)修改yarn-site.xml

[plain] view plain copy
 在CODE上查看代码片派生到我的代码片
  1. <property>  
  2.   <description>The http address of the RM web application.</description>  
  3.   <name>yarn.resourcemanager.webapp.address</name>  
  4.   <value>${yarn.resourcemanager.hostname}:18088</value>  
  5. </property>  

5、启动

1)格式化NameNode

cd $HADOOP_HOME/bin
./hdfs namenode -format

#2)启动hdfs
/home/game/soft/hadoop/sbin/start-dfs.sh
jps查看是否启动成功
16704 DataNode
16545 NameNode
16925 SecondaryNameNode

hdfs dfs -ls hdfs://192.168.80.100:19001/


#3) 启动yarn
/home/game/hadoop-2.7.3/sbin/start-yarn.sh
[game@IM-SJ01-Server18 sbin]$ jps
17427 NodeManager
19668 ResourceManager


yarn node -list
yarn node -status


#4)页面显示
192.168.80.100:10070
192.168.80.100:18088


#6、上传测试
hadoop fs -mkdir -p hdfs://192.168.80.100:19001/test/
hadoop fs -copyFromLocal ./test.txt hdfs://192.168.80.100:19001/test/
hadoop fs -ls hdfs://192.168.80.100:19001/


hadoop fs -put /opt/program/userall20140828 hdfs://localhost:9000/tmp/tvbox/
0 0
原创粉丝点击