Spark错误异常-sparkMaster GC overhead limit exceeded

来源:互联网 发布:网络教学app有哪些 编辑:程序博客网 时间:2024/05/17 20:33

Spark在以standalone模式运行一段时间户总会出现Spark Master GC overhead limit exceeded异常

16/09/20 05:42:24 ERROR ActorSystemImpl: Uncaught fatal error from thread [sparkMaster-akka.actor.default-dispatcher-6] shutting down ActorSystem [sparkMaster]java.lang.OutOfMemoryError: GC overhead limit exceeded

Spark master的内存被GC释放不掉,主要是因为随着时间的推移会缓存大量的对象Web UI中需要显示的已完成的Applications对象,默认设置的缓存数据是50,而启动的spark master进程内存默认设置的大小1G   -Xms1g -Xmx1g

spark.history.retainedApplications50在配置文件spark-default.conf中修改保存为一个合适的数量

spark.eventLog.enabled              truespark.eventLog.dir                  hdfs://namespace/tmp/spark/eventsspark.serializer                    org.apache.spark.serializer.KryoSerializerspark.history.fs.logDirectory       hdfs://namespace/tmp/spark/eventsspark.history.ui.port               18080spark.history.retainedApplications  20spark.kafka.metadata.broker.list    kafka1:9092,kafka2:9092,kafka3:9092spark.flume.listener.port           44445spark.executor.extraJavaOptions     -XX:HeapDumpPath=/data0/spark/temp/dump


0 0