spark-streaming:Could not compute split, block not found
来源:互联网 发布:龙族名句知乎 编辑:程序博客网 时间:2024/05/18 02:10
14/10/07 18:10:27 WARN scheduler.TaskSetManager: Lost task 45.0 in stage 12.0 (TID 129, domU-12-31-39-04-60-07.compute-1.internal): java.lang.Exception: Could not compute split, block input-0-1412705397200 not found
1278 org.apache.spark.rdd.BlockRDD.compute(BlockRDD.scala:51)
1279 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1280 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1281 org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)
1282 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1283 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1284 org.apache.spark.rdd.FlatMappedRDD.compute(FlatMappedRDD.scala:33)
1285 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1286 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1287 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
1288 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
1289 org.apache.spark.scheduler.Task.run(Task.scala:54)
1290 org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
1291 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
1292 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
1278 org.apache.spark.rdd.BlockRDD.compute(BlockRDD.scala:51)
1279 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1280 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1281 org.apache.spark.rdd.MappedRDD.compute(MappedRDD.scala:31)
1282 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1283 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1284 org.apache.spark.rdd.FlatMappedRDD.compute(FlatMappedRDD.scala:33)
1285 org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
1286 org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
1287 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
1288 org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
1289 org.apache.spark.scheduler.Task.run(Task.scala:54)
1290 org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177)
1291 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
1292 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
1293 java.lang.Thread.run(Thread.java:745)
spark-streaming程序在跑时有时会报错如上。
原因如下:
JavaReceiverInputDStream<SparkFlumeEvent> flumeStream = FlumeUtils.createStream(jssc, hostIp, port);
用上面的方法创建DStream时,默认采用的storageLevel是MEMROY_ONLY_SER,就是接收到的数据仅存于内存。
那么如果我们启动内存driver-memory和executor-memory(主要是executor-memory)设置小了,那么spark会自动丢弃放不下的数据,那么当需要计算该数据时自然报错说找不到数据块。解决办法是使用MEMORY_AND_DISK_SER级别来存日志,当内存不够时直接写磁盘或者增大executor-memory的大小。
0 0
- spark-streaming:Could not compute split, block not found
- could not found taximedia.apk
- 【笔记】Spark Streaming's Kafka libraries not found in class path 处理
- Page not found The requested page could not be found.
- [转]解决journal block not found错误
- [转]解决journal block not found错误
- The JRE could not be found 解决办法
- The JRE could not be found
- `DllImport' could not be found 错误
- `DllImport' could not be found 错误
- The specified module could not be found
- The specified module could not be found
- Jwplayer播放器:Could not load plugins: File not found
- JAVA_HOME is not set and could not be found.
- hadoop2 JAVA_HOME is not set and could not be found
- JAVA_HOME is not set and could not be found
- Hibernate:关于Column '*' not found和could not deserialize
- Error:JAVA_HOME is not set and could not be found
- imagesmagick 详细安装
- 零售系统软件架构---数据架构之数据分片
- 自动化脚本中运到的一些简单算法
- 一次php进程诡异退出的排查过程
- Android ActionBar应用实战,高仿微信主界面的设计
- spark-streaming:Could not compute split, block not found
- Best Time to Buy and Sell Stock III Leetcode Python
- 如何在SharePoint 2013 添加Ribbon菜单
- net-snmp centOS下的安装配置
- [解题报告]hdoj1069(动态规划)
- JAVA 数据结构之HashMap构建通用树实例
- centos下编译安装node.js、并测试简单demo
- HTML5学习第三天---20150126
- shell脚本中用getopts处理用户输入选项和参数