hadoop运维之jobtracker无故停止服务
来源:互联网 发布:数据恢复大师怎么样 编辑:程序博客网 时间:2024/05/29 08:38
今天下午同事在使用hive提交查询时,抛出执行错误:
于是打开jobtracker的管理页面,发现正在运行的job数目为零,tasktracker心跳正常,这一异常现象让我觉得jobtracker可能是停止服务了(一般很少出现集群的运行job数为零的情况),于是手动提交了一个mapred任务进行测试,运行错误信息如下:
- 12/07/03 18:07:22 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.EOFException
- 12/07/03 18:07:22 INFO hdfs.DFSClient: Abandoning block blk_-1772232086636991
- 458_5671628
- 12/07/03 18:07:28 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.EOFException
- 12/07/03 18:07:28 INFO hdfs.DFSClient: Abandoning block blk_-2108024038073283869_5671629
- 12
- /07/03 18:07:34 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink as 192.168.1.25:50010
- 12/07/03 18:07:34 INFO hdfs.DFSClient: Abandoning
- block blk_-6674020380591432013_5671629
- 12/07/03 18:07:40 INFO hdfs.DFSClient: Exception in createBlockOutputStream java.io.IOException: Bad connect ack with firstBadLink as 192.168.1.26:50010
- 12/07/0
- 3 18:07:40 INFO hdfs.DFSClient: Abandoning block blk_-3788726859662311832_5671629
- 12/07/03 18:07:46 WARN hdfs.DFSClient: DataStreamer Exception: java.io.IOException: Unable to create new block.
- at or
- g.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:3002)
- at org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2255)
- at org.apache.hadoop.hd
- fs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2446)
- 12/07/03 18:07:46 WARN hdfs.DFSClient: Error Recovery for block blk_-3788726859662311832_5671629 bad datanode[2] nodes == null
- 12/07
- /03 18:07:46 WARN hdfs.DFSClient: Could not get block locations. Source file "/tmp/hadoop-hadoop/mapred/staging/hadoop/.staging/job_201206270914_17301/job.jar" - Aborting...
- 2012-07-03 18:07:27,316 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* NameSystem.allocateBlock: /tmp/hadoop-hadoop/mapred/staging/
- hadoop/.staging/job_201206270914_17301/job.jar. blk_-2108024038073283869_5671629
于是再仔细查看问题发生时datanode上的日志,发现这么一条日志信息:
- 2012-07-03 18:07:10,274 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(192.168.1.25:50010, storageID=DS
- -841642307-50010-1324273874581, infoPort=50075, ipcPort=50020):DataXceiver
- java.io.IOException: xceiverCount 257 exceeds the limit of concurrent xcievers 256
- <property>
- <name>dfs.datanode.max.xcievers</name>
- <value>256</value>
- </property>
好了,问题找到了,只要找机会修改集群所有datanode节点的配置,将dfs.datanode.max.xcievers参数修改大一些即可。
0 0
- hadoop运维之jobtracker无故停止服务
- hadoop运维之jobtracker无故停止服务
- hadoop - jobTracker自动停止问题
- Hadoop源码之JobTracker
- Hadoop源码之JobTracker
- Hadoop JobTracker之ExpireLaunchingTasks
- Hadoop之JobTracker源码分析
- Hadoop之Jobtracker启动过程
- Hadoop之JobTracker,TaskTracker,Scheduler
- hadoop之JobTracker,TaskTracker,hadoop调度器
- hadoop之JobTracker,TaskTracker,hadoop调度器
- hadoop作业调度详解之JobTracker
- TOMCAT无故停止的下策
- Hadoop JobTracker 分析
- hadoop–JobTracker 相关
- 【Hadoop】JobTracker Restart Recovery
- Hadoop JobTracker和TaskTracker
- hadoop jobtracker与tasktracker
- It's like there's train in your mouth.
- 酒吧空间怎样量身打造
- cisco ios有关接口下配置ipv6地址需要注意的问题
- 献给那些浮躁的职场人(转)
- Android APK反编译详解(附图)
- hadoop运维之jobtracker无故停止服务
- Ubuntu安装nginx php5-fpm mysql (lnmp环境搭建)
- 威士忌诞生源由
- MFC绑定控件与变量
- 链表c++实现
- WLAN用户接入流程(转)
- KTV工作,KTV上班時常用術語有哪些?
- Linux下查看网卡的驱动程序
- GCD编程