Flink 配置文件 对于管理员来说,差不多经常调整的就只有 conf 下的flink-conf.yaml : 经过初步的调整,大约有以下模块的参数(未优化)

Licensed to the Apache Software Foundation (ASF) under one

or more contributor license agreements. See the NOTICE file

distributed with this work for additional information

regarding copyright ownership. The ASF licenses this file

to you under the Apache License, Version 2.0 (the

"License"); you may not use this file except in compliance

with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software

distributed under the License is distributed on an "AS IS" BASIS,

WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

See the License for the specific language governing permissions and

limitations under the License.

==============================================================================

Jobmanager && TaskManager 的基本参数

==============================================================================

jobmanager.rpc.address: flink-master

jobmanager.rpc.port: 6123
jobmanager.heap.size: 32768m
jobstore.cache-size: 52428800
taskmanager.heap.size: 92160m
taskmanager.numberOfTaskSlots: 30
parallelism.default: 3
io.tmp.dirs: /data1/data/flink/local
taskmanager.network.numberOfBuffers: 4096
taskmanager.memory.fraction: 0.7
taskmanager.memory.preallocate: false
taskmanager.debug.memory.startLogThread: 8
taskmanager.debug.memory.logIntervalMs: 3600000
blob.fetch.retries: 50
blob.fetch.num-concurrent: 200
blob.fetch.backlog: 500
task.cancellation-interval: 30000
taskmanager.exit-on-fatal-akka-error: false
taskkmanager.log.path:

==============================================================================

High Availability && and checkpointing

=============================================================================

high-availability: ZOOKEEPER
high-availability.cluster-id: /default_flink_dc
high-availability.zookeeper.quorum: VECS01455:2181,VECS01456:2181,VECS01457:2181,VECS01458:2181,VECS01459:2181
high-availability.storageDir: hdfs://flashHadoop/flink/recovery
high-availability.zookeeper.path.root: /flink
high-availability.zookeeper.path.checkpoint-counter: /checkpoint-counter
high-availability.zookeeper.path.checkpoints: /checkpoints
high-availability.zookeeper.path.jobgraphs: /jobgraphs
high-availability.zookeeper.path.latch: /leaderlatch
high-availability.zookeeper.path.leader: /leader
high-availability.zookeeper.path.mesos-workers: /mesos-workers
high-availability.zookeeper.path.running-registry: /running_job_registry/
high-availability.zookeeper.client.connection-timeout: 18000
high-availability.zookeeper.client.max-retry-attempts: 10
high-availability.zookeeper.client.retry-wait: 5000
high-availability.zookeeper.client.session-timeout: 60000

==============================================================================

HDFS and Fault tolerance and checkpointing

==============================================================================

fs.default-scheme: hdfs://flashHadoop/flink/data
fs.output.always-create-directory: false
fs.overwrite-files: false

state.backend: filesystem
state.backend.async: true
state.backend.fs.memory-threshold: 4096
state.backend.incremental: false
state.backend.local-recovery: false
state.checkpoints.dir: hdfs://flashHadoop/flink/checkpoints_data
state.checkpoints.num-retained: 1
state.savepoints.dir: hdfs://flashHadoop/flink/savepoints_data

==============================================================================

Web Frontend

==============================================================================

web.address: 0.0.0.0
web.access-control-allow-origin: "*"
web.backpressure.cleanup-interval: 600000
web.backpressure.delay-between-samples: 50
web.backpressure.num-samples: 100
web.backpressure.refresh-interval: 60000
web.checkpoints.history: 20
web.history: 20
web.log.path: /data1/data/flink/web/log
web.refresh-interval: 3000
web.submit.enable: true
web.timeout: 10000
web.tmpdir: /data1/data/flink/web/tmp

==============================================================================

Flink Cluster Security Configuration

==============================================================================

test is not set Security

==============================================================================

HistoryServer Config

==============================================================================

historyserver.web.address: 0.0.0.0
historyserver.web.port: 8082
historyserver.archive.fs.dir: hdfs://flashHadoop/flink/completed_jobs/
historyserver.archive.fs.refresh-interval: 2000
historyserver.web.tmpdir: /data1/data/flink/tmp

==============================================================================

==============================================================================

Environment 角色环境 申明

==============================================================================

env.java.opts: -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+UnlockExperimentalVMOptions -XX:+G1SummarizeConcMark -XX:MaxGCPauseMillis=100 -XX:-ResizePLAB -XX:+ParallelRefProcEnabled -XX:ParallelGCThreads=24 -XX:ConcGCThreads=16 -XX:G1HeapWastePercent=3 -XX:InitiatingHeapOccupancyPercent=35 -XX:G1MixedGCLiveThresholdPercent=85

env.java.opts.historyserver: -Xmx16g -Xms16g -Xmn4g -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+UnlockExperimentalVMOptions -XX:+G1SummarizeConcMark -XX:MaxGCPauseMillis=100 -XX:-ResizePLAB -XX:+ParallelRefProcEnabled -XX:+AlwaysPreTouch -XX:ParallelGCThreads=24 -XX:ConcGCThreads=16 -XX:G1HeapWastePercent=3 -XX:InitiatingHeapOccupancyPercent=35 -XX:G1MixedGCLiveThresholdPercent=85 -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:/data1/data/flink/gc/gc-historyserver-hostname.log

env.java.opts.jobmanager: -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+UnlockExperimentalVMOptions -XX:+G1SummarizeConcMark -XX:MaxGCPauseMillis=100 -XX:-ResizePLAB -XX:+ParallelRefProcEnabled -XX:ParallelGCThreads=24 -XX:ConcGCThreads=16 -XX:G1HeapWastePercent=3 -XX:InitiatingHeapOccupancyPercent=35 -XX:G1MixedGCLiveThresholdPercent=85 -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:/data1/data/flink/gc/gc-jobmanager-hostname.log

env.java.opts.taskmanager: -XX:+UseG1GC -XX:+UnlockDiagnosticVMOptions -XX:+UnlockExperimentalVMOptions -XX:+G1SummarizeConcMark -XX:MaxGCPauseMillis=100 -XX:-ResizePLAB -XX:+ParallelRefProcEnabled -XX:ParallelGCThreads=24 -XX:ConcGCThreads=16 -XX:G1HeapWastePercent=3 -XX:InitiatingHeapOccupancyPercent=35 -XX:G1MixedGCLiveThresholdPercent=85 -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xloggc:/data1/data/flink/gc/gc-taskmanager-hostname.log

env.log.dir: /data1/data/flink/log

env.log.max: 1

flink-conf.yaml的更多相关文章

  1. flink部署操作-flink standalone集群安装部署

    flink集群安装部署 standalone集群模式 必须依赖 必须的软件 JAVA_HOME配置 flink安装 配置flink 启动flink 添加Jobmanager/taskmanager 实 ...

  2. flink Standalone Cluster

    Requirements Software Requirements Flink runs on all UNIX-like environments, e.g. Linux, Mac OS X, a ...

  3. Flink集群部署

    部署方式 一般来讲有三种方式: Local Standalone Flink On Yarn/Mesos/K8s… 单机模式 参考上一篇Flink从入门到放弃(入门篇2)-本地环境搭建&构建第 ...

  4. Flink(二)CentOS7.5搭建Flink1.6.1分布式集群

    一. Flink的下载 安装包下载地址:http://flink.apache.org/downloads.html  ,选择对应Hadoop的Flink版本下载 [admin@node21 soft ...

  5. Flink的安装配置

    一. Flink的下载 安装包下载地址:http://flink.apache.org/downloads.html  ,选择对应Hadoop的Flink版本下载 [admin@node21 soft ...

  6. 一张图轻松掌握 Flink on YARN 应用启动全流程(上)

    Flink 支持 Standalone 独立部署和 YARN.Kubernetes.Mesos 等集群部署模式,其中 YARN 集群部署模式在国内的应用越来越广泛.Flink 社区将推出 Flink ...

  7. kerberos系列之flink认证配置

    大数据安全系列的其它文章 https://www.cnblogs.com/bainianminguo/p/12548076.html-----------安装kerberos https://www. ...

  8. 分享一个Flink checkpoint失败的问题和解决办法

    本文来自: PerfMa技术社区 PerfMa(笨马网络)官网 接触Flink一段时间了,遇到了一些问题,其中有一个checkpoint失败导致作业重启的问题,遇到了很多次,重启之后一般也能恢复正常, ...

  9. flink集群模式安装配置

    一.手动下载安装包 wget http://mirrors.tuna.tsinghua.edu.cn/apache/flink/flink-1.6.1/flink-1.6.1-bin-hadoop27 ...

  10. Linux使用docker安装flink

    配置文件不是原始的 修改过,可以根据自己的配置文件来 jobmanager配置 flink-conf.yaml ############################################ ...

随机推荐

  1. STL源码剖析-vector

    STL(Standard Template Library) C++标准模板库,acm选手一定对它不陌生,除了算法,那几乎是“吃饭的家伙了”.我们使用库函数非常方便,且非常高效(相对于自己实现来说). ...

  2. 27号华为笔试(三道ac两道)

    三道题目case:100,100,0: 三个题目: 前两个都全部ac了,第三题没时间: 记录一下大概的思路: 第一题 主要通过Java中的字符串处理函数:然后控制字符串输入格式: 卡bug的点: 1: ...

  3. Chromium被用于Microsoft Edge与ChakraCore的未来【译】

    注:英语不好,力求大概能懂.持笔人是:Limin Zhu,好像是中国人,但是没有提供中文版本. 大家好,ChakraCore的朋友们: 昨天,微软公布,Microsoft Edge桌面浏览器采用Chr ...

  4. JMeter 接口测试-if控制器

    JMeter 接口测试-if控制器 ​ 使用场景: 当业务场景是用户登录才能支付操作, 不登录点击支付, 页面会跳转到登录页面. 对于接口也是这样, 一个接口需要执行前, 需要有前提条件, 比如0状态 ...

  5. 本地Windows环境Dubbo搭建测试

    Dubbo介绍 Dubbo[]是一个分布式服务框架,致力于提供高性能和透明化的RPC远程服务调用方案,以及SOA服务治理方案. 其核心部分包含: 远程通讯: 提供对多种基于长连接的NIO框架抽象封装, ...

  6. Centos6搭建vsftpd

    CentOS 6.5下安装Vsftp,虚拟用户一.安装:1.安装Vsftpd服务相关部件:[root@localhost ~]# yum install vsftpd*Loaded plugins: ...

  7. 简单的纯js三级联动

    参考这个  日尼禾尔  二级联动 写了三级联动 <!DOCTYPE html> <html> <head> <meta charset="UTF-8 ...

  8. [20190416]exclusive latch测试脚本.txt

    [20190416]exclusive latch测试脚本.txt --//昨天做了shared latch的测试脚本,今天完善exclusive latch测试脚本,上个星期的测试我是手工执行的.- ...

  9. Redis笔记-集群搭建

    Redis单机版搭建上一篇已经基本介绍了,下面讨论Redis集群搭建方案和示例. 1.关于Redis常用的集群方案(三种): a.一主多从,如一个Master.两个Slave b.薪火相传,即集群中的 ...

  10. .net core下Redis帮助类

      0.引入.net core环境下Redis的NuGet包,StackExchange.Redis,现目前最新的2.0.519. 帮助类Code: using System; using Syste ...