快上网专注成都网站设计 成都网站制作 成都网站建设
成都网站建设公司服务热线:028-86922220

网站建设知识

十年网站开发经验 + 多家企业客户 + 靠谱的建站团队

量身定制 + 运营维护+专业推广+无忧售后,网站问题一站解决

k8s临时内存-创新互联

如何查看CPU总占用率?
top -bn 1 -i -c
sar -P 0 -u 1 5

创新互联专业为企业提供茫崖网站建设、茫崖做网站、茫崖网站设计、茫崖网站制作等企业网站建设、网页设计与制作、茫崖企业网站模板建站服务,十余年茫崖做网站经验,不只是建网站,更提供有价值的思路和整体网络服务。

I had a similar error. My analysis:

Pods on a same k8s node share the ephemeral storage, which (if no special configuration was used) is used by spark to store temp data of spark jobs (disk spillage and shuffle data). The amount of ephemeral storage of a node is basically the size of the available storage in your k8s node.

If some executor pods use up all of the ephemeral storage of a node, other pods will fail when they try to write data to ephemeral storage. In your case the failing pod is the driver pod, but it could have been any other pods on that node. In my case it was an executor that failed with a similar error message.

I would try to optimize the spark code first before changing the deployment configuration.

  • reduce disk spillage, shuffle write
  • split transforms if possible
  • and increase the amount of executors as the last resource :)

If you know upfront the amount of storage required in each executor, maybe you can try to set the resources requests (and not limits) for ephemeral storage to right amount.

你是否还在寻找稳定的海外服务器提供商?创新互联www.cdcxhl.cn海外机房具备T级流量清洗系统配攻击溯源,准确流量调度确保服务器高可用性,企业级服务器适合批量采购,新人活动首月15元起,快前往官网查看详情吧


当前文章:k8s临时内存-创新互联
本文路径:http://6mz.cn/article/dsphdg.html

其他资讯