spark1.2.0
Property Name | Default | Meaning |
spark.yarn.applicationMaster.waitTries | 10 | ApplicationMaster 鏈接Spark master和SparkContext初始化的嘗試次數 |
spark.yarn.submit.file.replication | 3 | 上傳到HDFS上的Spark jar、app jar登文件的備份數 |
spark.yarn.preserve.staging.files | false | spark任務結束后是否刪除上傳的Spark jar、app jar等文件 |
spark.yarn.scheduler.heartbeat.interval-ms | 5000 | Spark application master向YARN ResourceManager發送心跳的時間間隔 |
spark.yarn.max.executor.failures | numExecutors * 2 最小是3 |
executor 失敗多少次就標記此應用的運行失敗 |
park.yarn.historyServer.address | 空 | 默認無,是可選項,格式為host.com:port,不帶http://,是spark歷史服務器的地址 |
spark.yarn.dist.archives | 空 | Comma separated list of archives to be extracted into the working directory of each executor |
spark.yarn.dist.files | 空 | Comma-separated list of files to be placed in the working directory of each executor. |
spark.yarn.executor.memoryOverhead | executorMemory * 0.07, with minimum of 384 |
The amount of off heap memory (in megabytes) to be allocated per executor. This is memory that accounts for things like VM overheads, interned strings, other native overheads, etc. This tends to grow with the executor size (typically 6-10%). |
spark.yarn.driver.memoryOverhead | driverMemory * 0.07 with minimum of 384 |
The amount of off heap memory (in megabytes) to be allocated per driver. This is memory that accounts for things like VM overheads, interned strings, other native overheads, etc. This tends to grow with the container size (typically 6-10%). |
spark.yarn.queue | default | 應用提交給yarn隊列的名字 |
spark.yarn.jar | 空 | spark jar包的的路徑,默認使用本地spark目錄中的,也可以放到HDFS上 |
spark.yarn.access.namenodes | 空 | 有安全認證的HDFS namenode的地址如:spark.yarn.access.namenodes=hdfs://nn1.com:8032,hdfs://nn2.com:8032 |
spark.yarn.appMasterEnv.[EnvironmentVariableName] | 空 | 環境變量的設置 |
spark.yarn.containerLauncherMaxThreads | 25 | application master啟動executor container的線程最大數量 |