Hadoop集群(CHD4)实践之 (4) Oozie搭建
来源:互联网 发布:sqlserver 创建临时表 编辑:程序博客网 时间:2024/06/05 06:54
目录结构
Hadoop集群(CDH4)实践之 (0) 前言
Hadoop集群(CDH4)实践之 (1) Hadoop(HDFS)搭建
Hadoop集群(CDH4)实践之 (2) HBase&Zookeeper搭建
Hadoop集群(CDH4)实践之 (3) Hive搭建
Hadoop集群(CHD4)实践之 (4) Oozie搭建
Hadoop集群(CHD4)实践之 (5) Sqoop安装
本文内容
Hadoop集群(CHD4)实践之 (4) Oozie搭建
参考资料
http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Installation-Guide/CDH4-Installation-Guide.html
环境准备
OS: CentOS 6.4 x86_64
Servers:
hadoop-master: 172.17.20.230 内存10G
- namenode
- hbase-master
hadoop-secondary: 172.17.20.234 内存10G
- secondarybackupnamenode,jobtracker
- hive-server,hive-metastore
- oozie
hadoop-node-1: 172.17.20.231 内存10G sudo yum install hbase-regionserver
- datanode,tasktracker
- hbase-regionserver,zookeeper-server
hadoop-node-2: 172.17.20.232 内存10G
- datanode,tasktracker
- hbase-regionserver,zookeeper-server
hadoop-node-3: 172.17.20.233 内存10G
- datanode,tasktracker
- hbase-regionserver,zookeeper-server
对以上角色做一些简单的介绍:
namenode - 整个HDFS的命名空间管理服务
secondarynamenode - 可以看做是namenode的冗余服务
jobtracker - 并行计算的job管理服务
datanode - HDFS的节点服务
tasktracker - 并行计算的job执行服务
hbase-master - Hbase的管理服务
hbase-regionServer - 对Client端插入,删除,查询数据等提供服务
zookeeper-server - Zookeeper协作与配置管理服务
hive-server - Hive的管理服务
hive-metastore - Hive的元存储,用于对元数据进行类型检查与语法分析
oozie - Oozie是一种Java Web应用程序,用于工作流的定义和管理
本文定义的规范,避免在配置多台服务器上产生理解上的混乱:
以下操作都只需要在 Oozie 所在主机,即 hadoop-secondary 上执行。
1. 安装前的准备
Hadoop集群(CDH4)实践之 (3) Hive搭建
2. 安装Oozie
$ sudo yum install oozie oozie-client
3. 创建Oozie数据库
$ mysql -uroot -phiveserver
1
mysql> create database oozie;
2
mysql> grant all privileges on oozie.* to
'oozie'
@
'localhost'
identified by
'oozie'
;
3
mysql> grant all privileges on oozie.* to
'oozie'
@
'%'
identified by
'oozie'
;
4
mysql>
exit
;
4.配置oozie-site.xml
$ sudo vim /etc/oozie/conf/oozie-site.xml
001
<?
xml
version
=
"1.0"
?>
002
<
configuration
>
003
<
property
>
004
<
name
>oozie.service.ActionService.executor.ext.classes</
name
>
005
<
value
>
006
org.apache.oozie.action.email.EmailActionExecutor,
007
org.apache.oozie.action.hadoop.HiveActionExecutor,
008
org.apache.oozie.action.hadoop.ShellActionExecutor,
009
org.apache.oozie.action.hadoop.SqoopActionExecutor,
010
org.apache.oozie.action.hadoop.DistcpActionExecutor
011
</
value
>
012
</
property
>
013
<
property
>
014
<
name
>oozie.service.SchemaService.wf.ext.schemas</
name
>
015
<
value
>shell-action-0.1.xsd,shell-action-0.2.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,hive-action-0.3.xsd,hive-action-0.4.xsd,hive-action-0.5.xsd,sqoop-action-0.2.xsd,sqoop-action-0.3.xsd,ssh-action-0.1.xsd,ssh-action-0.2.xsd,distcp-action-0.1.xsd</
value
>
016
</
property
>
017
<
property
>
018
<
name
>oozie.system.id</
name
>
019
<
value
>oozie-${user.name}</
value
>
020
</
property
>
021
<
property
>
022
<
name
>oozie.systemmode</
name
>
023
<
value
>NORMAL</
value
>
024
</
property
>
025
<
property
>
026
<
name
>oozie.service.AuthorizationService.security.enabled</
name
>
027
<
value
>false</
value
>
028
</
property
>
029
<
property
>
030
<
name
>oozie.service.PurgeService.older.than</
name
>
031
<
value
>30</
value
>
032
</
property
>
033
<
property
>
034
<
name
>oozie.service.PurgeService.purge.interval</
name
>
035
<
value
>3600</
value
>
036
</
property
>
037
<
property
>
038
<
name
>oozie.service.CallableQueueService.queue.size</
name
>
039
<
value
>10000</
value
>
040
</
property
>
041
<
property
>
042
<
name
>oozie.service.CallableQueueService.threads</
name
>
043
<
value
>10</
value
>
044
</
property
>
045
<
property
>
046
<
name
>oozie.service.CallableQueueService.callable.concurrency</
name
>
047
<
value
>3</
value
>
048
</
property
>
049
<
property
>
050
<
name
>oozie.service.coord.normal.default.timeout
051
</
name
>
052
<
value
>120</
value
>
053
</
property
>
054
055
<
property
>
056
<
name
>oozie.db.schema.name</
name
>
057
<
value
>oozie</
value
>
058
</
property
>
059
<
property
>
060
<
name
>oozie.service.JPAService.create.db.schema</
name
>
061
<
value
>true</
value
>
062
</
property
>
063
064
<
property
>
065
<
name
>oozie.service.JPAService.jdbc.driver</
name
>
066
<
value
>com.mysql.jdbc.Driver</
value
>
067
</
property
>
068
<
property
>
069
<
name
>oozie.service.JPAService.jdbc.url</
name
>
070
<
value
>jdbc:mysql://localhost:3306/oozie</
value
>
071
</
property
>
072
<
property
>
073
<
name
>oozie.service.JPAService.jdbc.username</
name
>
074
<
value
>oozie</
value
>
075
</
property
>
076
<
property
>
077
<
name
>oozie.service.JPAService.jdbc.password</
name
>
078
<
value
>oozie</
value
>
079
</
property
>
080
081
<
property
>
082
<
name
>oozie.service.JPAService.pool.max.active.conn</
name
>
083
<
value
>10</
value
>
084
</
property
>
085
086
<
property
>
087
<
name
>oozie.service.HadoopAccessorService.kerberos.enabled</
name
>
088
<
value
>false</
value
>
089
</
property
>
090
<
property
>
091
<
name
>local.realm</
name
>
092
<
value
>LOCALHOST</
value
>
093
</
property
>
094
<
property
>
095
<
name
>oozie.service.HadoopAccessorService.keytab.file</
name
>
096
<
value
>${user.home}/oozie.keytab</
value
>
097
</
property
>
098
<
property
>
099
<
name
>oozie.service.HadoopAccessorService.kerberos.principal</
name
>
100
<
value
>${user.name}/localhost@${local.realm}</
value
>
101
</
property
>
102
<
property
>
103
<
name
>oozie.service.HadoopAccessorService.jobTracker.whitelist</
name
>
104
<
value
> </
value
>
105
</
property
>
106
<
property
>
107
<
name
>oozie.service.HadoopAccessorService.nameNode.whitelist</
name
>
108
<
value
> </
value
>
109
</
property
>
110
111
<
property
>
112
<
name
>oozie.service.HadoopAccessorService.hadoop.configurations</
name
>
113
<
value
>*=/etc/hadoop/conf</
value
>
114
</
property
>
115
<
property
>
116
<
name
>oozie.service.WorkflowAppService.system.libpath</
name
>
117
<
value
>/user/${user.name}/share/lib</
value
>
118
</
property
>
119
120
<
property
>
121
<
name
>use.system.libpath.for.mapreduce.and.pig.jobs</
name
>
122
<
value
>false</
value
>
123
</
property
>
124
125
<
property
>
126
<
name
>oozie.authentication.type</
name
>
127
<
value
>simple</
value
>
128
</
property
>
129
<
property
>
130
<
name
>oozie.authentication.token.validity</
name
>
131
<
value
>36000</
value
>
132
</
property
>
133
<
property
>
134
<
name
>oozie.authentication.signature.secret</
name
>
135
<
value
>oozie</
value
>
136
</
property
>
137
138
<
property
>
139
<
name
>oozie.authentication.cookie.domain</
name
>
140
<
value
></
value
>
141
</
property
>
142
143
<
property
>
144
<
name
>oozie.authentication.simple.anonymous.allowed</
name
>
145
<
value
>true</
value
>
146
</
property
>
147
148
<
property
>
149
<
name
>oozie.authentication.kerberos.principal</
name
>
150
<
value
>HTTP/localhost@${local.realm}</
value
>
151
</
property
>
152
153
<
property
>
154
<
name
>oozie.authentication.kerberos.keytab</
name
>
155
<
value
>${oozie.service.HadoopAccessorService.keytab.file}</
value
>
156
</
property
>
157
158
<
property
>
159
<
name
>oozie.authentication.kerberos.name.rules</
name
>
160
<
value
>DEFAULT</
value
>
161
</
property
>
162
163
<
property
>
164
<
name
>oozie.service.ProxyUserService.proxyuser.oozie.hosts</
name
>
165
<
value
>*</
value
>
166
</
property
>
167
168
<
property
>
169
<
name
>oozie.service.ProxyUserService.proxyuser.oozie.groups</
name
>
170
<
value
>*</
value
>
171
</
property
>
172
173
<
property
>
174
<
name
>oozie.service.ProxyUserService.proxyuser.hue.hosts</
name
>
175
<
value
>*</
value
>
176
</
property
>
177
<
property
>
178
<
name
>oozie.service.ProxyUserService.proxyuser.hue.groups</
name
>
179
<
value
>*</
value
>
180
</
property
>
181
182
<
property
>
183
<
name
>oozie.action.mapreduce.uber.jar.enable</
name
>
184
<
value
>true</
value
>
185
</
property
>
186
<
property
>
187
<
name
>oozie.service.HadoopAccessorService.supported.filesystems</
name
>
188
<
value
>hdfs,viewfs</
value
>
189
</
property
>
190
</
configuration
>
5. 配置Oozie Web Console
$ cd /tmp/
$ wget http://archive.cloudera.com/gplextras/misc/ext-2.2.zip
$ cd /var/lib/oozie/
$ sudo unzip /tmp/ext-2.2.zip
$ cd ext-2.2/
$ sudo -u hdfs hadoop fs -mkdir /user/oozie
$ sudo -u hdfs hadoop fs -chown oozie:oozie /user/oozie
6. 配置Oozie ShareLib
$ mkdir /tmp/ooziesharelib
$ cd /tmp/ooziesharelib
$ tar xzf /usr/lib/oozie/oozie-sharelib.tar.gz
$ sudo -u oozie hadoop fs -put share /user/oozie/share
$ sudo -u oozie hadoop fs -ls /user/oozie/share
$ sudo -u oozie hadoop fs -ls /user/oozie/share/lib
$ sudo -u oozie hadoop fs -put /usr/lib/hive/lib/hbase.jar /user/oozie/share/lib/hive/
$ sudo -u oozie hadoop fs -put /usr/lib/hive/lib/zookeeper.jar /user/oozie/share/lib/hive/
$ sudo -u oozie hadoop fs -put /usr/lib/hive/lib/hive-hbase-handler-0.10.0-cdh4.5.0.jar /user/oozie/share/lib/hive/
$ sudo -u oozie hadoop fs -put /usr/lib/hive/lib/guava-11.0.2.jar /user/oozie/share/lib/hive/
$ sudo ln -s /usr/share/java/mysql-connector-java.jar /var/lib/oozie/mysql-connector-java.jar
7. 启动Oozie
$ sudo service oozie start
8. 访问Oozie Web Console
http://hadoop-secondary:11000/oozie
9. 至此,Oozie的搭建就已经完成。
- Hadoop集群(CHD4)实践之 (4) Oozie搭建
- hadoop之oozie配置
- Hadoop之--集群环境搭建
- hadoop之zookeeper集群搭建
- Hadoop入门之集群搭建
- Hadoop集群搭建之搭建DNS
- Hadoop集群搭建之二(测试hadoop集群)
- Hadoop CDH4.4集群搭建
- 搭建Hadoop2.0.0CHD4.4.0
- hadoop集群搭建(hadoop)
- HADOOP: 搭建hadoop集群
- Redis集群搭建之Redis Cluster实践
- hadoop之HBase集群搭建及使用
- hadoop集群搭建之NFS服务
- Hadoop之Hbase------>Hbase集群搭建
- Hadoop分布式 之 Hbase集群搭建
- Hadoop集群搭建之三 配置文件说明
- hadoop集群搭建之hive安装
- 基于CAS实现单点登录(SSO):工作原理
- 我的JavaWeb学习2
- 让页面在打开时自动刷新
- iOS中一个图展的实现
- Web编码乱码解决原理方案
- Hadoop集群(CHD4)实践之 (4) Oozie搭建
- php中的$_GET如何获取带有井号“#”的参数
- 郭美美的经济学
- Windows Azure Mobile Service - Data Sync lessons learned
- Lua的function、closure和upvalue
- Hello my blog on CSDN
- KVC 和KVO机制
- 使用C++ ostringstream来格式化字符串输出
- LeetCode 086 Partition List