hadoop conf erros
来源:互联网 发布:关于网络的力量的段子 编辑:程序博客网 时间:2024/06/14 13:42
start-all.sh> ./sbin/start-all.sh This script is Deprecated. Instead use start-dfs.sh and start-yarn.shStarting namenodes on [localhost]localhost: starting namenode, logging to /usr/bin/hadoop-2.2.0/logs/hadoop-usergrid-namenode-sornora-localdomain.outlocalhost: starting datanode, logging to /usr/bin/hadoop-2.2.0/logs/hadoop-usergrid-datanode-sornora-localdomain.outStarting secondary namenodes [0.0.0.0]0.0.0.0: starting secondarynamenode, logging to /usr/bin/hadoop-2.2.0/logs/hadoop-usergrid-secondarynamenode-sornora-localdomain.outstarting yarn daemonsstarting resourcemanager, logging to /usr/bin/hadoop-2.2.0/logs/yarn-usergrid-resourcemanager-sornora-localdomain.outlocalhost: starting nodemanager, logging to /usr/bin/hadoop-2.2.0/logs/yarn-usergrid-nodemanager-sornora-localdomain.out
Cluster Summary
There are no datanodes in the cluster.
http://blog.csdn.net/greensurfer/article/details/7618660
HTTP ERROR 500
Problem accessing /nn_browsedfscontent.jsp. Reason:
Can't browse the DFS since there are no live nodes available to redirect to.
Caused by:
java.io.IOException: Can't browse the DFS since there are no live nodes available to redirect to.at org.apache.hadoop.hdfs.server.namenode.NamenodeJspHelper.redirectToRandomDataNode(NamenodeJspHelper.java:646)at org.apache.hadoop.hdfs.server.namenode.nn_005fbrowsedfscontent_jsp._jspService(nn_005fbrowsedfscontent_jsp.java:70)at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:98)at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)at org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:511)at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1221)at org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter.doFilter(StaticUserWebFilter.java:109)at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)at org.apache.hadoop.http.HttpServer$QuotingInputFilter.doFilter(HttpServer.java:1081)at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)at org.apache.hadoop.http.NoCacheFilter.doFilter(NoCacheFilter.java:45)at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)at org.apache.hadoop.http.NoCacheFilter.doFilter(NoCacheFilter.java:45)at org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)at org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399)at org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)at org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182)at org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766)at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450)at org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230)at org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)at org.mortbay.jetty.Server.handle(Server.java:326)at org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542)at org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928)at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549)at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212)at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404)at org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410)at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582)
2.
is not in the sudoers file 解决(转)
解决方案:首需要切换到root身份
$su -
(注意有- ,这和su是不同的,在用命令"su"的时候只是切换到root,但没有把root的环境变量传过去,还是当前用户的环境变量,用"su -"命令将环境变量也一起带过去,就象和root登录一样)
然后
$visudo //切记,此处没有vi和sudo之间没有空格
1、移动光标,到最后一行
2、按a,进入append模式
3、输入
your_user_name ALL=(ALL) ALL
4、按Esc
5、输入“:w”(保存文件)
6、输入“:q”(退出)
这样就把自己加入了sudo组,可以使用sudo命令了。
3.
hadoop启动错误 JAVA_HOME is not set and could not be found
Starting namenodes on []
localhost: Error: JAVA_HOME is not set and could not be found.localhost: Error: JAVA_HOME is not set and could not be found.
...
starting yarn daemons
starting resourcemanager, logging to /home/lihanhui/open-source/hadoop-2.1.0-beta/logs/yarn-admin-resourcemanager-localhost.out
localhost: Error: JAVA_HOME is not set and could not be found.
直接命令行执行export JAVA_HOME=/PATH/TO/JDK无法解决问题
grep命令发现libexec/hadoop-config.sh文件中有错误提示中的黑体字,于是在对应位置前直接设置export JAVA_HOME=/PATH/TO/JDK
错误消失。
4.
Hadoop conf 下的 masters 和 slaves 文件的作用是什么?
http://blog.163.com/ly_89/blog/static/186902299201242151129443/
从字面意思上来看,masters 是主要的,应该记录的是 namenode 的 IP 或者是域名。但是文件的名称确实有点误导人,它主要记录运行辅助 namenode 的所有机器。slaves 文件记录了运行 datanode 和 tasktracker 的所有机器。用户也可以改变 hadoop-env.sh 的 HADOOP_SLAVES 项的值,将 slaves 文件放在其他地方。这两个文件无需分发到各个工作节点,因为只有运行在 namenode 或 jobtracker 上的控制脚本能使用这些文件。不过,分发了也不会有什么影响。
5.
http://blog.csdn.net/meng_tianshi/article/details/6784962
http://www.infoq.com/cn/articles/hadoop-config-tip
http://blog.csdn.net/guxch/article/details/7786831
http://wiki.apache.org/hadoop/HowToConfigure
http://gbif.blogspot.com/2011/01/setting-up-hadoop-cluster-part-1-manual.html
6.
Let me add a bit more to kkrugler's answer:
There're three HDFS properties which contain hadoop.tmp.dir
in their values
dfs.name.dir
: directory where namenode stores its metadata, with default value${hadoop.tmp.dir}/dfs/name
.dfs.data.dir
: directory where HDFS data blocks are stored, with default value${hadoop.tmp.dir}/dfs/data
.fs.checkpoint.dir
: directory where secondary namenode store its checkpoints, default value is${hadoop.tmp.dir}/dfs/namesecondary
.
This is why you saw the /mnt/hadoop-tmp/hadoop-${user.name}
in your HDFS after formatting namenode.
- hadoop conf erros
- org.apache.hadoop.conf
- Erros in building service
- Hadoop 七:Hadoop-Hdfs源码 conf包
- hadoop 出现FATAL conf.Configuration: error parsing conf file,异常
- "hadoop fs -conf " does not work
- #org.apache.hadoop.conf.Configuration.addDeprecations
- Hadoop源码详解 org.apache.hadoop.conf (配置类) (二)
- 1.hadoop学习之org.apache.hadoop.conf
- .conf
- Hadoop集群安装 (2) 配置conf/core-site.xml
- Hadoop集群安装 (3) 配置conf/hdfs-site.xml
- Hadoop 学习笔记:DistributedCache.addCacheFile(URI, conf)的用法
- Code analysis of Hadoop v0.1.0 (2) conf package
- java:5: error: package org.apache.hadoop.conf does not
- hadoop源码解析2 - conf包中Configuration.java解析
- org.apache.hadoop.conf.Configuration.getFloat(String arg0, float arg1)
- hadoop有关conf和mapreduce包的存放位置
- 关于sar的一个问题: Invalid system activity file
- java学习笔记_类建立对象初始化过程分析
- java webservice 结构
- Android Handler的使用!!!
- 颜色选择器(一)
- hadoop conf erros
- 开发者必备,超实用的PHP代码片段!
- 自定义排序准则,自定义数据类型
- Redis 默认配置精简及注释说明
- Oracle 用户权限
- USACO
- MyBatis连接SQLServer数据库
- ACM 进阶学习第一课----素数相关(2)
- C++字符数组(定义和初始化、赋值和引用)