##flink 1.7.2 安裝
需要java環境
下載地址 https://flink.apache.org/downloads.html
#1、單機版
#創建用戶flink
useradd flink -d /home/flink
echo "flink123" | passwd flink --stdin
#解壓
tar -zxvf flink-1.7.2-bin-hadoop26-scala_2.11.tgz
#啟動
cd flink-1.7.2/bin/ && ./start-cluster.sh
#測試
1、web頁面查看:ip:8081 例如:http://192.168.88.132:8081
2、運行實例wordcount程序:cd flink-1.7.2/bin/ && ./flink run ../examples/batch/WordCount.jar
3、jps 查看進程
#2、獨立集群版
#准備機器 /etc/hosts
192.168.88.130 lgh
192.168.88.131 lgh1
192.168.88.132 lgh2
#創建用戶flink(所有機器)
useradd flink -d /home/flink
echo "flink123" | passwd flink --stdin
#ssh免密登錄(在192.168.88.130,指定一台操作)
su - flink
ssh-keygen -t rsa
ssh-copy-id 192.168.88.131
ssh-copy-id 192.168.88.132
#解壓
tar -zxvf flink-1.7.2-bin-hadoop26-scala_2.11.tgz
cd flink-1.7.2/conf
#修改配置文件
#1、masters
192.168.88.130:8081
#2、slaves
192.168.88.131
192.168.88.132
#3、flink-conf.yaml
cat flink-conf.yaml | grep -v ^# | grep -v "^$"
jobmanager.rpc.address: 192.168.88.130
jobmanager.rpc.port: 6123
env.java.home: /usr/java/default
jobmanager.heap.size: 1024m
taskmanager.heap.size: 1024m
taskmanager.numberOfTaskSlots: 1
parallelism.default: 1
rest.port: 8081
#分發flink
scp -r flink-1.7.2 flink@192.168.88.131/home/flink
scp -r flink-1.7.2 flink@192.168.88.132/home/flink
#啟動
在主節點上:cd flink-1.7.2/bin/ && ./start-cluster.sh
#測試
1、web頁面查看:ip:8081 例如:http://192.168.88.132:8081
2、運行實例wordcount程序:cd flink-1.7.2/bin/ && ./flink run ../examples/batch/WordCount.jar
3、jps 查看
#3、基於yarn的集群版(前提是安裝好了hadoop)
hadoop的安裝:可參考https://www.cnblogs.com/zsql/p/10736420.html
在獨立集群的基礎上:新增如下
#配置hadoop的環境變量 /etc/profile 或者 ~/.bashrc
export HADOOP_HOME=/apps/opt/cloudera/parcels/CDH/lib/hadoop
export HADOOP_CONF_DIR=/etc/hadoop/conf
export YARN_CONF_DIR=/etc/hadoop/conf
然后使用source命令使配置文件生效
#分發flink
scp -r flink-1.7.2 flink@192.168.88.131/home/flink
scp -r flink-1.7.2 flink@192.168.88.132/home/flink
#啟動集群和yarn-session
cd flink-1.7.2 && ./start-cluster.sh
cd flink-1.7.2 && nohup ./bin/yarn-session.sh &
#測試
1、jps 查看進程
2、執行程序:./bin/flink run -m yarn-cluster -yn 2 ./examples/batch/WordCount.jar
然后去hadoop的yarn頁面查看作業,ip:8088
#4、高可用版(安裝zookeeper,或者修改conf/zoo.cfg,推薦安裝zookeeper)
zookeeper的安裝可以參考:https://www.cnblogs.com/zsql/p/10736420.html
在基於yarn版本集群的基礎上修改如下配置:
#1、masters
192.168.88.130:8081
192.168.88.131:8082
#2、flink-conf.yaml (一定要注意空格..踩過坑)
jobmanager.rpc.address: 192.168.88.130
jobmanager.rpc.port: 6123
env.java.home: /usr/java/default
jobmanager.heap.size: 1024m
taskmanager.heap.size: 1024m
taskmanager.numberOfTaskSlots: 1
parallelism.default: 1
high-availability: zookeeper
high-availability.zookeeper.path.root:/user/flink/root
high-availability.storageDir: hdfs:///user/flink/ha/ #該目錄flink用戶一定要有權限
high-availability.zookeeper.quorum: 192.168.88.130:2181,192.168.88.131:2181,192.168.88.132:2181
rest.port: 8081
#分發flink
scp -r flink-1.7.2 flink@192.168.88.131/home/flink
scp -r flink-1.7.2 flink@192.168.88.132/home/flink
#測試如上yarn集群
新增測試,kill掉一個主節點進程,看是否能跑作業