解决Consider increasing spark.rpc.message.maxSize

来源:互联网 发布:淘宝虚假交易新规 编辑:程序博客网 时间:2024/06/06 03:23

apache.spark.SparkException: Job aborted due to stage failure: Serialized task 32:5 was 204136673 bytes, which exceeds max allowed: spark.rpc.message.maxSize (134217728 bytes).Consider increasing spark.rpc.message.maxSize or using broadcast variables for large values.
用sc.parallelize(data,slices)时,如果data数据过大,易出现该问题
解决:增加spark.rpc.message.maxSize,该值默认大小128M
提交任务是加:–conf spark.rpc.message.maxSize=512