Skip to main content

环境参数模版

shell/python

环境参数模版

## 每个worker所占内存,比如512m 
# dtscript.worker.memory=512m

## 每个worker所占的cpu核的数量
# dtscript.worker.cores=1

## worker数量
# dtscript.worker.num=1

## 是否独占机器节点
# dtscript.worker.exclusive=false

## 任务优先级, 值越小,优先级越高,范围:1-1000
job.priority=10

## 指定work运行节点,需要注意不要写ip应填写对应的hostname
# dtscript.worker.nodes=

## 指定work运行机架
# dtscript.worker.racks=

## 日志级别可选ALL, DEBUG, ERROR, FATAL, INFO, OFF, TRACE, WARN
logLevel=INFO

控制台参数模版

dtscript.java.opts=-Dfile.encoding=UTF-8
dtscript.am.memory=512m
dtscript.am.cores=1
dtscript.worker.memory=512m
dtscript.worker.cores=1
dtscript.worker.num=1
container.staging.dir=/dtInsight/dtscript/staging
dtscript.container.heartbeat.interval=10000
dtscript.container.heartbeat.timeout=120000
dtscript.python2.path=/data/miniconda2/bin/python2
dtscript.python3.path=/data/miniconda3/bin/python3

参数配置项说明

参数项默认值说明是否必填
dtscript.java.opts-Dfile.encoding=UTF-8dtscript container jvm扩展参数
dtscript.am.memory512mam container使用的内存量
dtscript.am.cores1am container使用的cpu核数
dtscript.worker.memory512mwork container使用的内存量
dtscript.worker.cores1work container使用的cpu核数
dtscript.worker.num1work container实例数量
container.staging.dir/dtInsight/dtscript/staging任务临时文件路径
dtscript.container.heartbeat.interval10000am和work之间的心跳间隔,单位毫秒
dtscript.container.heartbeat.timeout120000am和work之间的心跳超时时间,单位毫秒
dtscript.python2.path/data/miniconda2/bin/python2python2.x二进制可执行文件地址
dtscript.python3.path/data/miniconda3/bin/python3python3.x二进制可执行文件地址

spark/pyspark/spark sql 版本:2.1

环境参数模版

## Driver程序使用的CPU核数,默认为1
# driver.cores=1

## Driver程序使用内存大小,默认512m
# driver.memory=512m

## 对Spark每个action结果集大小的限制,最少是1M,若设为0则不限制大小。
## 若Job结果超过限制则会异常退出,若结果集限制过大也可能造成OOM问题,默认1g
# driver.maxResultSize=1g

## 启动的executor的数量,默认为1
executor.instances=1

## 每个executor使用的CPU核数,默认为1
executor.cores=1

## 每个executor内存大小,默认512m
executor.memory=512m

## 任务优先级, 值越小,优先级越高,范围:1-1000
job.priority=10

## spark 日志级别可选ALL, DEBUG, ERROR, FATAL, INFO, OFF, TRACE, WARN
# logLevel = INFO

## spark中所有网络交互的最大超时时间
# spark.network.timeout=120s

## executor的OffHeap内存,和spark.executor.memory配置使用
# spark.yarn.executor.memoryOverhead

控制台参数模版

# 主要
spark.submit.deployMode=cluster
spark.yarn.maxAppAttempts=4
sparkPythonExtLibPath=hdfs:///dtInsight/pythons/pyspark.zip,hdfs:///dtInsight/pythons/py4j-0.10.7-src.zip
sparkSqlProxyPath=hdfs:///dtInsight/user/spark/client/spark-sql-proxy.jar
sparkYarnArchive=hdfs:///dtInsight/sparkjars/jars
spark.resources.dir=hdfs:///dtInsight/spark
yarnAccepterTaskNumber=3
spark.speculation=true

# 资源
spark.executor.cores=1
spark.executor.memory=512m
spark.executor.instances=1
spark.cores.max=1

# 网络
spark.network.timeout=700s
spark.rpc.askTimeout=600s
spark.executor.heartbeatInterval=10s

# sql
spark.sql.crossJoin.enabled=true

# 事件日志
spark.eventLog.compress=false
spark.eventLog.dir=hdfs:///tmp/spark-yarn-logs
spark.eventLog.enabled=true

# JVM
spark.driver.extraJavaOptions=-Dfile.encoding=UTF-8
spark.executor.extraJavaOptions=-Dfile.encoding=UTF-8

# 环境变量
spark.yarn.appMasterEnv.PYSPARK_DRIVER_PYTHON=/data/miniconda2/bin/python3
spark.yarn.appMasterEnv.PYSPARK_PYTHON=/data/miniconda2/bin/python3

# 安全
spark.yarn.security.credentials.hive.enabled=true

参数配置项说明

参数项默认值说明是否必填
spark.cores.max1standalone模式下任务最大能申请的cpu核数
spark.driver.extraJavaOptions-Dfile.encoding=utf-8spark driver的jvm扩展参数
spark.executor.extraJavaOptions-Dfile.encoding=utf-8spark executor的jvm扩展参数
spark.eventLog.compressfalse是否对spark事件日志进行压缩
spark.eventLog.dirhdfs:///tmp/spark-yarn-logsspark事件日志存放路径
spark.eventLog.enabledtrue是否记录 spark 事件日志
spark.sql.crossJoin.enabledtrue开启笛卡尔积join
spark.executor.cores1每个executor可以使用的cpu核数
spark.executor.heartbeatInterval10sdriver和executor之间心跳时间间隔
spark.executor.instances1executor实例数
spark.executor.memory512m每个executor可以使用的内存量
spark.network.timeout700s所有组件间网络通信超时时间
spark.rpc.askTimeout600sRPC 请求操作在超时之前等待的持续时间
spark.speculationtruespark任务推测行为
spark.submit.deployModeclusterspark任务部署模式
spark.yarn.appMasterEnv.PYSPARK_DRIVER_PYTHON/data/miniconda2/bin/python3driver中用于执行pyspark任务的python二进制可执行文件路径
spark.yarn.appMasterEnv.PYSPARK_PYTHON/data/miniconda2/bin/python3用于执行pyspark任务的python二进制可执行文件路径
spark.yarn.maxAppAttempts4spark driver最大尝试次数, 默认为yarn上yarn.resourcemanager.am.max-attempts配置的值注:如果spark.yarn.maxAppAttempts配置的大于yarn.resourcemanager.am.max-attempts则无效
spark.yarn.security.credentials.hive.enabledtrue开启kerberos场景下是否获取hive 票据

自定义参数项说明

参数项默认值说明是否必填
sparkPythonExtLibPathhdfs:///dtInsight/pythons/pyspark.zip,hdfs://ns1/dtInsight/pythons/py4j-0.10.7-src.zip远程存储系统上pyspark.zip和py4j-0.10.7-src.zip的路径注:pyspark.zip和py4j-0.10.7-src.zip在$SPARK_HOME/python/lib路径下获取
sparkSqlProxyPathhdfs:///dtInsight/user/spark/client/spark-sql-proxy.jar远程存储系统上spark-sql-proxy.jar路径注:spark-sql-proxy.jar是用来执行spark sql的jar包
sparkYarnArchivehdfs:///dtInsight/sparkjars/jars远程存储系统上spark jars的路径
yarnAccepterTaskNumber3允许yarn上同时存在状态为accepter的任务数量,当达到这个值后会禁止任务提交
spark.resources.dirhdfs:///dtInsight/spark远程存储系统上hadoopconf,kerberos,sparksqlproxy文件存放位置。

spark/pyspark/spark sql 版本:2.4

## Driver程序使用的CPU核数,默认为1
# spark.driver.cores=1

## Driver程序使用内存大小,默认1g
# spark.driver.memory=1g

## 对Spark每个action结果集大小的限制,最少是1M,若设为0则不限制大小。
## 若Job结果超过限制则会异常退出,若结果集限制过大也可能造成OOM问题,默认1g
# spark.driver.maxResultSize=1g

## 启动的executor的数量,默认为1
# spark.executor.instances=1

## 每个executor使用的CPU核数,默认为1
# spark.executor.cores=1

## 每个executor内存大小,默认1g
# spark.executor.memory=1g

## 任务优先级, 值越小,优先级越高,范围:1-1000
job.priority=10

## spark 日志级别可选ALL, DEBUG, ERROR, FATAL, INFO, OFF, TRACE, WARN
# logLevel = INFO

## spark中所有网络交互的最大超时时间
# spark.network.timeout=120s

## executor的OffHeap内存,和spark.executor.memory配置使用
# spark.yarn.executor.memoryOverhead=

## 设置spark sql shuffle分区数,默认200
# spark.sql.shuffle.partitions=200

## 开启spark推测行为,默认false
# spark.speculation=false

控制台参数模版

# 主要
spark.submit.deployMode=cluster
spark.yarn.maxAppAttempts=4
sparkPythonExtLibPath=hdfs:///dtInsight/spark240/pythons/pyspark.zip,hdfs:///dtInsight/spark240/pythons/py4j-0.10.7-src.zip
sparkSqlProxyPath=hdfs:///dtInsight/spark240/client/spark-sql-proxy.jar
sparkYarnArchive=hdfs:///dtInsight/spark240/jars
spark.resources.dir=hdfs:///dtInsight/spark
yarnAccepterTaskNumber=3

# 资源
spark.executor.cores=1
spark.executor.memory=1g
spark.executor.instances=1

# 网络
spark.network.timeout=700s
spark.rpc.askTimeout=600s
spark.executor.heartbeatInterval=10s

# sql
spark.sql.crossJoin.enabled=true

# 事件日志
spark.eventLog.compress=false
spark.eventLog.dir=hdfs:///dtInsight/spark240/eventlogs
spark.eventLog.enabled=true

# JVM
spark.driver.extraJavaOptions=-Dfile.encoding=UTF-8
spark.executor.extraJavaOptions=-Dfile.encoding=UTF-8

# 环境变量
spark.yarn.appMasterEnv.PYSPARK_PYTHON=/data/anaconda3/bin/python3

# 安全
spark.yarn.security.credentials.hive.enabled=true
spark.ranger.enabled=false
spark.sql.extensions=org.apache.spark.ranger.security.api.RangerSparkSQLExtension

# metrics
metrics.prometheus.server.host=
metrics.prometheus.server.port=9090

metrics.prometheus.sink.pushgateway.host=
metrics.prometheus.sink.pushgateway.port=9091

metrics.prometheus.sink.pushgateway.class.instance=*
metrics.prometheus.sink.pushgateway.class=org.apache.spark.metrics.sink.PrometheusPushGatewaySink

metrics.prometheus.sink.pushgateway.protocol.instance=*
metrics.prometheus.sink.pushgateway.protocol=http

metrics.prometheus.sink.pushgateway.period.instance=*
metrics.prometheus.sink.pushgateway.period=5

metrics.prometheus.sink.pushgateway.enable-dropwizard-collector.instance=*
metrics.prometheus.sink.pushgateway.enable-dropwizard-collector=true

metrics.prometheus.sink.pushgateway.enable-hostname.instance=*
metrics.prometheus.sink.pushgateway.enable-hostname=true

metrics.prometheus.source.jvm.class.instance=*
metrics.prometheus.source.jvm.class=org.apache.spark.metrics.source.JvmSource

参数配置项说明

参数项默认值说明是否必填
spark.driver.extraJavaOptions-Dfile.encoding=UTF-8spark driver的jvm扩展参数
spark.executor.extraJavaOptions-Dfile.encoding=UTF-8spark executor的jvm扩展参数
spark.eventLog.compressfalse是否对spark事件日志进行压缩
spark.eventLog.dirhdfs:///dtInsight/spark240/eventlogsspark事件日志存放路径
spark.eventLog.enabledtrue是否记录 spark 事件日志
spark.executor.cores1每个executor可以使用的cpu核数
spark.executor.heartbeatInterval10sdriver和executor之间心跳时间间隔
spark.sql.crossJoin.enabledtrue是否开启笛卡尔积join
spark.ranger.enabledfalse是否开启ranger
spark.sql.extensionsorg.apache.spark.ranger.security.api.RangerSparkSQLExtensionranger所需相关扩展类类名
spark.executor.instances1executor实例数
spark.executor.memory1g每个executor可以使用的内存量
spark.network.timeout700s所有组件间网络通信超时时间
spark.rpc.askTimeout600sRPC 请求操作在超时之前等待的持续时间
spark.submit.deployModeclusterspark任务部署模式
spark.yarn.appMasterEnv.PYSPARK_PYTHON/data/anaconda3/bin/python3用于执行pyspark任务的python二进制可执行文件路径
spark.yarn.maxAppAttempts4spark driver最大尝试次数, 默认为yarn上yarn.resourcemanager.am.max-attempts配置的值注:如果spark.yarn.maxAppAttempts配置的大于yarn.resourcemanager.am.max-attempts则无效
spark.yarn.security.credentials.hive.enabledtrue开启kerberos场景下是否获取hive 票据
metrics.prometheus.server.host默认值为空prometheus host 参考值:flink03
metrics.prometheus.server.port9090prometheus port
metrics.prometheus.sink.pushgateway.host默认值为空pushgateway host 参考值:flink03
metrics.prometheus.sink.pushgateway.port9091pushgateway port
metrics.prometheus.sink.pushgateway.class.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.sink.pushgateway.classorg.apache.spark.metrics.sink.PrometheusPushGatewaySink声明用于将指标push到PushGatewaySink的类
metrics.prometheus.sink.pushgateway.protocol.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.sink.pushgateway.protocolhttppushgateway的地址协议
metrics.prometheus.sink.pushgateway.period.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.sink.pushgateway.period5指标sink到pushgateway的时间间隔
metrics.prometheus.sink.pushgateway.enable-dropwizard-collector.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.sink.pushgateway.enable-dropwizard-collectortrue开启dropwizard-collector来收集Spark App metrics
metrics.prometheus.sink.pushgateway.enable-hostname.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.sink.pushgateway.enable-hostnametrue开启URI地址以主机名的形式展现
metrics.prometheus.source.jvm.class.instance*取值为master、worker、executor、driver、applications,也可以取值为 * *,** 代表所有的instance。
metrics.prometheus.source.jvm.classorg.apache.spark.metrics.source.JvmSource开启JvmSource,收集各个instance的jvm信息

自定义参数项说明

参数项默认值说明是否必填
sparkPythonExtLibPathhdfs:///dtInsight/spark240/pythons/pyspark.zip,hdfs:///dtInsight/spark240/pythons/py4j-0.10.7-src.zip远程存储系统上pyspark.zip和py4j-0.10.7-src.zip的路径注:pyspark.zip和py4j-0.10.7-src.zip在$SPARK_HOME/python/lib路径下获取
sparkSqlProxyPathhdfs:///dtInsight/spark240/client/spark-sql-proxy.jar远程存储系统上spark-sql-proxy.jar路径注:spark-sql-proxy.jar是用来执行spark sql的jar包
sparkYarnArchivehdfs:///dtInsight/spark240/jars远程存储系统上spark jars的路径
yarnAccepterTaskNumber3允许yarn上同时存在状态为accepter的任务数量,当达到这个值后会禁止任务提交
spark.resources.dirhdfs:///dtInsight/spark远程存储系统上hadoopconf,kerberos,sparksqlproxy文件存放位置。

数据同步 1.12

环境参数模版

#==============================================================================
# 公共
#==============================================================================
## jobManager配置的内存大小,默认1024(单位M)
# jobmanager.memory.mb=1024

## taskManager配置的内存大小,默认1024(单位M)
# taskmanager.memory.mb=1024

## taskManager 对应 slot的数量
slots=1

## sql任务并发度设置
sql.env.parallelism=1

## 时间窗口类型(ProcessingTime或者EventTime)
time.characteristic=ProcessingTime

## 窗口提前触发时间,单位为秒(填写正整数即可)
# early.trigger=1


#==============================================================================
# 高可用
#==============================================================================

# The high-availability mode. Possible options are 'NONE' or 'zookeeper'.
#
# high-availability: zookeeper

# The path where metadata for master recovery is persisted. While ZooKeeper stores
# the small ground truth for checkpoint and leader election, this location stores
# the larger objects, like persisted dataflow graphs.
#
# Must be a durable file system that is accessible from all nodes
# (like HDFS, S3, Ceph, nfs, ...)
#
# high-availability.storageDir: hdfs:///flink/ha/

# The list of ZooKeeper quorum peers that coordinate the high-availability
# setup. This must be a list of the form:
# "host1:clientPort,host2:clientPort,..." (default clientPort: 2181)
#
# high-availability.zookeeper.quorum: localhost:2181


# ACL options are based on https://zookeeper.apache.org/doc/r3.1.2/zookeeperProgrammers.html#sc_BuiltinACLSchemes
# It can be either "creator" (ZOO_CREATE_ALL_ACL) or "open" (ZOO_OPEN_ACL_UNSAFE)
# The default value is "open" and it can be changed to "creator" if ZK security is enabled
#
# high-availability.zookeeper.client.acl: open

#==============================================================================
# 容错和checkpointing
#==============================================================================
## checkpoint 外存的清理动作
## true(任务结束之后删除checkpoint外部存储信息)
## false(任务结束之后保留checkpoint外部存储信息)
sql.checkpoint.cleanup.mode=false

## ttl状态控制
## 最小过期时间,大于0的整数,如1d、1h(dD:天,hH:小时,mM:分钟,ss:秒)
# sql.ttl.min=1h
## 最大过期时间,大于0的整数,如2d、2h(dD:天,hH:小时,mM:分钟,ss:秒),需同时设置最小时间,且比最小时间大5分钟
# sql.ttl.max=2h

## 生成checkpoint时间间隔(以毫秒为单位),默认:5分钟,注释掉该选项会关闭checkpoint生成
flink.checkpoint.interval=300000

## 设置checkpoint生成超时(以毫秒为单位),默认:10分钟
sql.checkpoint.timeout=600000

## 任务出现故障的时候一致性处理,可选参数EXACTLY_ONCE,AT_LEAST_ONCE;默认为EXACTLY_ONCE
# sql.checkpoint.mode=EXACTLY_ONCE

## 最大并发生成 checkpoint 数量,默认:1 次
# sql.max.concurrent.checkpoints=1


#==============================================================================
# Rest & web frontend
#==============================================================================

# The port to which the REST client connects to. If rest.bind-port has
# not been specified, then the server will bind to this port as well.
#
#rest.port: 8081

# The address to which the REST client will connect to
#
#rest.address: 0.0.0.0

# Port range for the REST and web server to bind to.
#
#rest.bind-port: 8080-8090

# The address that the REST & web server binds to
#
#rest.bind-address: 0.0.0.0

# Flag to specify whether job submission is enabled from the web-based
# runtime monitor. Uncomment to disable.

#web.submit.enable: false

# Flag to specify whether job cancellation is enabled from the web-based
# runtime monitor. Uncomment to disable.

#web.cancel.enable: false


#==============================================================================
# 高级
#==============================================================================

## logLevel: error,debug,info(默认),warn
logLevel=info

## Watermark发送周期,单位毫秒
# autoWatermarkInterval=200

## 设置输出缓冲区的最大刷新时间频率(毫秒)
# sql.buffer.timeout.millis=100

## 任务优先级, 值越小,优先级越高,范围:1-1000
job.priority=10


## 异步访问维表是否开启连接池共享,开启则 1.一个tm上多个task共享该池, 2.一个tm上多个url相同的维表单/多个task共享该池 (默认false)
# async.side.clientShare=false
## 连接池中连接的个数,上面参数为true才生效(默认5)
# async.side.poolSize=5


#==============================================================================
# 安全
#==============================================================================
## kafka kerberos相关参数
## security.kerberos.login.use-ticket-cache=true
## security.kerberos.login.contexts=Client,KafkaClient
## security.kerberos.login.keytab=/opt/keytab/kafka.keytab
## security.kerberos.login.principal=kafka@HADOOP.COM
## zookeeper.sasl.service-name=zookeeper
## zookeeper.sasl.login-context-name=Client


#==============================================================================
# ZK 安全
#==============================================================================

# Override below configuration to provide custom ZK service name if configured
# zookeeper.sasl.service-name: zookeeper

# The configuration below must match one of the values set in "security.kerberos.login.contexts"
# zookeeper.sasl.login-context-name: Client





控制台参数模版

#==============================================================================
# 公共
#==============================================================================
jobmanager.memory.process.size: 1600m
taskmanager.memory.process.size: 2048m
taskmanager.numberOfTaskSlots: 1
slotmanager.number-of-slots.max: 10
#==============================================================================
# 高可用
#==============================================================================
high-availability: ZOOKEEPER
high-availability.storageDir: hdfs://ns1/dtInsight/flink112/ha
high-availability.zookeeper.path.root: /flink112
high-availability.zookeeper.quorum 172.16.23.25

#==============================================================================
# metric 监控
#==============================================================================
prometheusHost: 172.16.23.25
prometheusPort: 9090
metrics.reporter.promgateway.class: org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporter
metrics.reporter.promgateway.deleteOnShutdown: true
metrics.reporter.promgateway.host: 172.16.23.25
metrics.reporter.promgateway.jobName: 112job
metrics.reporter.promgateway.port: 9091
metrics.reporter.promgateway.randomJobNameSuffix: true

#==============================================================================
# 容错和checkpointing
#==============================================================================
state.backend: RocksDB
state.backend.incremental: true
state.checkpoints.dir: hdfs://ns1/dtInsight/flink112/checkpoints
state.checkpoints.num-retained: 11
state.savepoints.dir: hdfs://ns1/dtInsight/flink1 12/savepoints
restart-strategy: none
execution.checkpointing.externalized-checkpoint-retention: RETAIN_ON_CANCELLATION


#==============================================================================
# 高级
#==============================================================================
classloader.resolve-order: parent-first
jobmanager.archive.fs.dir: hdfs://ns1/dtInsight/flink112/completed-jobs

#==============================================================================
# JVM 参数
#==============================================================================
env.java.opts: XX:+UseConcMarkSweepGC -XX:+CMSParallelRemarkEnabled -XX:+CMSIncrementalMode -XX:+CMSIncrementalP

#==============================================================================
# Resource Orchestration Frameworks(Yarn)
#==============================================================================
yarn.application-attempt-failures-validity-interval: 3600000
yarn.application-attempts: 3
yarn.application.queue: b

#==============================================================================
# 自定义参数
#==============================================================================
checkSubmitJobGraphInterval: 60
classloader.dtstack-cache: true
clusterMode: session
flinkLibDir: /data/112_flinkplugin/lib
flinkSessionName: batchsession112
flinkxDistDir: /data/112_flinkplugin/flinkxplugins_test
remoteFlinkLibDir: /data/112_flinkplugin/lib
remoteFlinkxDistDir:/data/112_flinkplugin/flinkxplugins_test
monitorAcceptedApp: false
pluginLoadMode: shipfile
yarnAccepterTaskNumber: 3
queue: b
sessionRetryNum: 5
sessionStartAuto: true

参数配置项说明

公共参数

参数项默认值案例值说明是否必填
jobmanager.memory.process.size(none)1024mJobManager 总内存(master)
taskmanager.memory.process.size(none)2048mTaskManager 总内存(slaves)
taskmanager.numberOfTaskSlots11单个 TaskManager 可以运行的并行算子或用户函数实例的数量。
slotmanager.number-of-slots.max(none)10flink session允许的最大slot数

高可用

参数项默认值案例值说明是否必填
high-availability(none)ZOOKEEPERflink ha类型
high-availability.zookeeper.quorum(none)kudu1:2181,kudu2:2181,kudu3:2181zookeeper地址,当ha选择是zookeeper时必填
high-availability.zookeeper.path.root(none)/flink110ha节点路径,
high-availability.storageDir(none)hdfs://ns1/dtInsight/flink110/haha元数据存储路径

metric 监控

参数项默认值案例值说明是否必填
metrics.reporter.promgateway.class(none)org.apache.flink.metrics.prometheus.PrometheusPushGatewayReporter用来推送指标类
metrics.reporter.promgateway.host172.16.23.25promgateway地址
metrics.reporter.promgateway.port(none)9091promgateway端口
metrics.reporter.promgateway.deleteOnShutdown(none)true任务结束后是否删除指标
metrics.reporter.promgateway.jobName(none)110job指标任务名
metrics.reporter.promgateway.randomJobNameSuffix(none)true是否在任务名上添加随机值

容错和checkpointing

参数项默认值案例值说明是否必填
state.backend(none)RocksDB状态后端
state.backend.incremental(none)true是否开启增量
state.checkpoints.dir(none)hdfs://ns1/dtInsight/flink110/checkpointscheckpoint路径地址
state.checkpoints.num-retained111checkpoint保存个数
state.savepoints.dir(none)hdfs://ns1/dtInsight/flink110/savepointssavepoint路径
execution.checkpointing.externalized-checkpoint-retentionRETAIN_ON_CANCELLATIONRETAIN_ON_CANCELLATIONcheckpoint保留策略外部化配置

高级

参数项默认值案例值说明是否必填
akka.ask.timeout10s60sakka 调用的超时时间
akka.tcp.timeout20s60stcp 连接的超时时间
classloader.resolve-orderperjob默认为child-firstsession默认为(none)(none)child-first类加载模式
jobmanager.archive.fs.dir(none)hdfs://ns1/dtInsight/flink110/completed-jobs任务结束后任务信息存储路径

JVM 参数

参数项默认值案例值说明是否必填
env.java.opts(none)-XX:+UseConcMarkSweepGC -XX:+CMSParallelRemarkEnabled -XX:+CMSIncrementalMode -XX:+CMSIncrementalPacing -XX:MaxMetaspaceSize=300m -Dfile.encoding=UTF-8jvm参数
env.java.opts.taskmanager(none)-agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=9751debug tm 的 jvm参数

Yarn

参数项默认值案例值说明是否必填
yarn.application-attempt-failures-validity-interval100003600000以毫秒为单位的时间窗口,它定义了重新启动 AM 时应用程序尝试失败的次数。不在此窗口范围内的故障不予考虑。将此值设置为 -1 以便全局计数。
yarn.application-attempts(none)3ApplicationMaster 重新启动的次数。默认情况下,该值将设置为 1。如果启用了高可用性,则默认值为 2。重启次数也受 YARN 限制(通过 yarn.resourcemanager.am.max-attempts 配置)。注意整个 Flink 集群会重启,YARN Client 会失去连接。
yarn.application.queuebyarn 队列名称

自定义参数

参数项默认值案例值说明是否必填
checkSubmitJobGraphInterval(none)60session check间隔(60 * 10s)
clusterMode(none)perjob任务执行模式:perjob, session, standalone
flinkLibDir(none)/data/112_flinkplugin/libflink lib path
flinkxDistDir(none)/data/112_flinkplugin/flinkxplugins_testflinkx plugins父级本地目录
remoteFlinkLibDir(none)/data/112_flinkplugin/libflink lib 远程路径
remoteFlinkxDistDir(none)/data/112_flinkplugin/flinkxplugins_testflinkx plugins父级远程目录
pluginLoadModeclasspathshipfile插件加载类型
prometheusHost(none)172.16.23.25prometheus地址,平台端使用
prometheusPort90909090prometheus,平台端使用
classloader.dtstack-cachetrue是否缓存classloader
sessionStartAutotruetrue是否允许engine启动flink session
submitTimeout55单位分钟,任务提交超时时间
queue(none)byarn队列
flinkSessionName(none)batchsession112yarn session名称
monitorAcceptedApp(none)false是否监控yarn accepted状态任务
yarnAccepterTaskNumber(none)3允许yarn accepter任务数量,达到这个值后不允许任务提交

HiveSQL

## 指定mapreduce在yarn上的任务名称,默认为任务名称,可以重复
#hiveconf:mapreduce.job.name=

## 指定mapreduce运行的队列,默认走控制台配置的queue
# hiveconf:mapreduce.job.queuename=default_queue_name

## hivevar配置,用户自定义变量
#hivevar:ageParams=30