- kubernetes部署單節點redis: https://www.cnblogs.com/zisefeizhu/p/14282299.html
Redis 介紹
• Redis代表REmote DIctionary Server是一種開源的內存中數據存儲,通常用作數據庫,緩存或消息代理。它可以存儲和操作高級數據類型,例如列表,地圖,集合和排序集合。
• 由於Redis接受多種格式的密鑰,因此可以在服務器上執行操作,從而減少了客戶端的工作量。
• 它僅將磁盤用於持久性,而將數據完全保存在內存中。
• Redis是一種流行的數據存儲解決方案,並被GitHub,Pinterest,Snapchat,Twitter,StackOverflow,Flickr等技術巨頭所使用。
為什么要用Redis
• 它的速度非常快。它是用ANSI C編寫的,並且可以在POSIX系統上運行,例如Linux,Mac OS X和Solaris。
• Redis通常被排名為最流行的鍵/值數據庫和最流行的與容器一起使用的NoSQL數據庫。
• 其緩存解決方案減少了對雲數據庫后端的調用次數。
• 應用程序可以通過其客戶端API庫對其進行訪問。
• 所有流行的編程語言都支持Redis。
• 它是開源且穩定的。
什么是Redis Cluster集群
• Redis Cluster是一組Redis實例,旨在通過對數據庫進行分區來擴展數據庫,從而使其更具彈性。
• 群集中的每個成員(無論是主副本還是輔助副本)都管理哈希槽的子集。如果主機無法訪問,則其從機將升級為主機。在由三個主節點組成的最小Redis群集中,每個主節點都有一個從節點(以實現最小的故障轉移),每個主節點都分配有一個介於0到16,383之間的哈希槽范圍。節點A包含從0到5000的哈希槽,節點B從5001到10000,節點C從10001到16383。
• 群集內部的通信是通過內部總線進行的,使用協議傳播有關群集的信息或發現新節點。
在Kubernetes中部署Redis Cluster集群過程記錄
在Kubernetes中部署Redis集群面臨挑戰,因為每個Redis實例都依賴於一個配置文件,該文件可以跟蹤其他集群實例及其角色。為此,我們需要結合使用StatefulSets控制器和PersistentVolumes持久化存儲。
StatefulSet的設計原理模型:
• 拓撲狀態:
應用的多個實例之間不是完全對等的關系,這個應用實例的啟動必須按照某些順序啟動,比如應用的主節點 A 要先於從節點 B 啟動。而如果你把 A 和 B 兩個Pod刪除掉,他們再次被創建出來是也必須嚴格按照這個順序才行,並且,新創建出來的Pod,必須和原來的Pod的網絡標識一樣,這樣原先的訪問者才能使用同樣的方法,訪問到這個新的Pod
• 存儲狀態:
應用的多個實例分別綁定了不同的存儲數據.對於這些應用實例來說,Pod A第一次讀取到的數據,和隔了十分鍾之后再次讀取到的數據,應該是同一份,哪怕在此期間Pod A被重新創建過.一個數據庫應用的多個存儲實例。
存儲卷
了解statefulset狀態后,應該知道要為數據准備一個存儲卷了,創建方式有靜態方式和動態方式,靜態方式就是手動創建PV、PVC,然后POD進行進行調用即可。這里使用動態NFS作為掛載卷,需要部署NFS動態StorageClass
1、使用NFS配置StatefulSet的動態持久化存儲
1)在NFS服務器端(172.16.60.238)通過nfs創建Redis Cluster集群的共享目錄
[root@k8s-harbor01 ~]# mkdir -p /data/storage/k8s/redis
2)創建nfs的rbac
[root@k8s-master01 ~]# mkdir -p /opt/k8s/k8s_project/redis
[root@k8s-master01 ~]# cd /opt/k8s/k8s_project/redis
[root@k8s-master01 redis]# vim nfs-rbac.yaml
---
apiVersion: v1
kind: ServiceAccount
metadata:
name: nfs-provisioner
namespace: wiseco
---
kind: ClusterRole
apiVersion: rbac.authorization.k8s.io/v1
metadata:
name: nfs-provisioner-runner
namespace: wiseco
rules:
- apiGroups: [""]
resources: ["persistentvolumes"]
verbs: ["get", "list", "watch", "create", "delete"]
- apiGroups: [""]
resources: ["persistentvolumeclaims"]
verbs: ["get", "list", "watch", "update"]
- apiGroups: ["storage.k8s.io"]
resources: ["storageclasses"]
verbs: ["get", "list", "watch"]
- apiGroups: [""]
resources: ["events"]
verbs: ["watch", "create", "update", "patch"]
- apiGroups: [""]
resources: ["services", "endpoints"]
verbs: ["get","create","list", "watch","update"]
- apiGroups: ["extensions"]
resources: ["podsecuritypolicies"]
resourceNames: ["nfs-provisioner"]
verbs: ["use"]
---
kind: ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1
metadata:
name: run-nfs-provisioner
subjects:
- kind: ServiceAccount
name: nfs-provisioner
namespace: wiseco
roleRef:
kind: ClusterRole
name: nfs-provisioner-runner
apiGroup: rbac.authorization.k8s.io
創建並查看
[root@k8s-master01 redis]# kubectl apply -f nfs-rbac.yaml
serviceaccount/nfs-provisioner created
clusterrole.rbac.authorization.k8s.io/nfs-provisioner-runner created
clusterrolebinding.rbac.authorization.k8s.io/run-nfs-provisioner created
[root@k8s-master01 redis]# kubectl get sa -n wiseco|grep nfs
nfs-provisioner 1 24s
[root@k8s-master01 redis]# kubectl get clusterrole -n wiseco|grep nfs
nfs-provisioner-runner 2021-02-04T02:21:11Z
[root@k8s-master01 redis]# kubectl get clusterrolebinding -n wiseco|grep nfs
run-nfs-provisioner ClusterRole/nfs-provisioner-runner 34s
3)創建redis cluster集群的storageclass
[root@k8s-master01 redis]# ll
total 4
-rw-r--r-- 1 root root 1216 Feb 4 10:20 nfs-rbac.yaml
[root@k8s-master01 redis]# vim redis-nfs-class.yaml
apiVersion: storage.k8s.io/v1beta1
kind: StorageClass
metadata:
name: redis-nfs-storage
namespace: wiseco
provisioner: redis/nfs
reclaimPolicy: Retain
創建並查看
[root@k8s-master01 redis]# kubectl apply -f redis-nfs-class.yaml
storageclass.storage.k8s.io/redis-nfs-storage created
[root@k8s-master01 redis]# kubectl get sc -n wiseco
NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE
redis-nfs-storage redis/nfs Retain Immediate false
4)創建redis cluster集群的nfs-client-provisioner
[root@k8s-master01 redis]# ll
total 8
-rw-r--r-- 1 root root 1216 Feb 4 10:20 nfs-rbac.yaml
-rw-r--r-- 1 root root 155 Feb 4 10:24 redis-nfs-class.yaml
[root@k8s-master01 redis]# vim redis-nfs.yml
apiVersion: apps/v1
kind: Deployment
metadata:
name: redis-nfs-client-provisioner
namespace: wiseco
spec:
replicas: 1
selector:
matchLabels:
app: redis-nfs-client-provisioner
strategy:
type: Recreate
template:
metadata:
labels:
app: redis-nfs-client-provisioner
spec:
serviceAccount: nfs-provisioner
containers:
- name: redis-nfs-client-provisioner
image: registry.cn-hangzhou.aliyuncs.com/open-ali/nfs-client-provisioner
imagePullPolicy: IfNotPresent
volumeMounts:
- name: nfs-client-root
mountPath: /persistentvolumes
env:
- name: PROVISIONER_NAME
value: redis/nfs
- name: NFS_SERVER
value: 172.16.60.238
- name: NFS_PATH
value: /data/storage/k8s/redis
volumes:
- name: nfs-client-root
nfs:
server: 172.16.60.238
path: /data/storage/k8s/redis
創建並查看
[root@k8s-master01 redis]# kubectl apply -f redis-nfs.yml
deployment.apps/redis-nfs-client-provisioner created
[root@k8s-master01 redis]# kubectl get pods -n wiseco|grep nfs
redis-nfs-client-provisioner-58b46549dd-h87gg 1/1 Running 0 40s
2、部署Redis Cluster集群
本案例部署采用的namespace命名空間是wiseco
1)准備image鏡像
redis-trib.rb工具可以去redis源碼中拷貝一個到當前目錄,然后構建鏡像。
[root@k8s-master01 redis]# pwd
/opt/k8s/k8s_project/redis
[root@k8s-master01 redis]# ll
total 12
-rw-r--r-- 1 root root 1216 Feb 4 15:31 nfs-rbac.yaml
-rw-r--r-- 1 root root 155 Feb 4 15:32 redis-nfs-class.yaml
-rw-r--r-- 1 root root 1006 Feb 4 15:32 redis-nfs.yml
[root@k8s-master01 redis]# mkdir image
[root@k8s-master01 redis]# cd image
[root@k8s-master01 image]# ll
total 64
-rw-r--r-- 1 root root 191 Feb 4 18:14 Dockerfile
-rwxr-xr-x 1 root root 60578 Feb 4 15:49 redis-trib.rb
[root@k8s-master01 image]# cat Dockerfile
FROM redis:4.0.11
RUN apt-get update -y
RUN apt-get install -y ruby \
rubygems
RUN apt-get clean all
RUN gem install redis
RUN apt-get install dnsutils -y
COPY redis-trib.rb /usr/local/bin/
創建鏡像並上傳到Harbor倉庫
[root@k8s-master01 image]# docker build -t 172.16.60.238/wiseco/redis:4.0.11 .
[root@k8s-master01 image]# docker push 172.16.60.238/wiseco/redis:4.0.11
2)創建configmap
redis配置文件使用configmap方式進行掛載,如果將配置封裝到docker image中的話,俺么每次修改配置就需要重新docker build。個人覺得比較麻煩,所以使用configmap方式掛載配置。
[root@k8s-master01 redis]# pwd
/opt/k8s/k8s_project/redis
[root@k8s-master01 redis]# ll
total 12
drwxr-xr-x 2 root root 45 Feb 4 18:14 image
-rw-r--r-- 1 root root 1216 Feb 4 15:31 nfs-rbac.yaml
-rw-r--r-- 1 root root 155 Feb 4 15:32 redis-nfs-class.yaml
-rw-r--r-- 1 root root 1006 Feb 4 15:32 redis-nfs.yml
[root@k8s-master01 redis]# mkdir conf
[root@k8s-master01 redis]# cd conf/
[root@k8s-master01 conf]# vim redis-configmap.yaml
apiVersion: v1
kind: ConfigMap
metadata:
name: redis-cluster
namespace: wiseco
data:
fix-ip.sh: |
#!/bin/sh
CLUSTER_CONFIG="/data/nodes.conf"
if [ -f ${CLUSTER_CONFIG} ]; then
if [ -z "${POD_IP}" ]; then
echo "Unable to determine Pod IP address!"
exit 1
fi
echo "Updating my IP to ${POD_IP} in ${CLUSTER_CONFIG}"
sed -i.bak -e '/myself/ s/[0-9]\{1,3\}\.[0-9]\{1,3\}\.[0-9]\{1,3\}\.[0-9]\{1,3\}/'${POD_IP}'/' ${CLUSTER_CONFIG}
fi
exec "$@"
redis.conf: |
cluster-enabled yes
cluster-config-file /data/nodes.conf
cluster-node-timeout 10000
protected-mode no
daemonize no
pidfile /var/run/redis.pid
port 6379
tcp-backlog 511
bind 0.0.0.0
timeout 3600
tcp-keepalive 1
loglevel verbose
logfile /data/redis.log
databases 16
save 900 1
save 300 10
save 60 10000
stop-writes-on-bgsave-error yes
rdbcompression yes
rdbchecksum yes
dbfilename dump.rdb
dir /data
#requirepass yl123456
appendonly yes
appendfilename "appendonly.aof"
appendfsync everysec
no-appendfsync-on-rewrite no
auto-aof-rewrite-percentage 100
auto-aof-rewrite-min-size 64mb
lua-time-limit 20000
slowlog-log-slower-than 10000
slowlog-max-len 128
#rename-command FLUSHALL ""
latency-monitor-threshold 0
notify-keyspace-events ""
hash-max-ziplist-entries 512
hash-max-ziplist-value 64
list-max-ziplist-entries 512
list-max-ziplist-value 64
set-max-intset-entries 512
zset-max-ziplist-entries 128
zset-max-ziplist-value 64
hll-sparse-max-bytes 3000
activerehashing yes
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 256mb 64mb 60
client-output-buffer-limit pubsub 32mb 8mb 60
hz 10
aof-rewrite-incremental-fsync yes
需要注意:fix-ip.sh 腳本的作用用於當redis集群某pod重建后Pod IP發生變化,在/data/nodes.conf中將新的Pod IP替換原Pod IP。不然集群會出問題。
創建和查看
[root@k8s-master01 conf]# kubectl apply -f redis-configmap.yaml
[root@k8s-master01 conf]# kubectl get cm -n wiseco|grep redis
redis-cluster 2 8m55s
3)准備StatefulSet
volumeClaimTemplates 用於StatefulSet控制器場景:
[root@k8s-master01 redis]# pwd
/opt/k8s/k8s_project/redis
[root@k8s-master01 redis]# ll
total 12
drwxr-xr-x 2 root root 34 Feb 4 18:52 conf
drwxr-xr-x 2 root root 45 Feb 4 18:14 image
-rw-r--r-- 1 root root 1216 Feb 4 15:31 nfs-rbac.yaml
-rw-r--r-- 1 root root 155 Feb 4 15:32 redis-nfs-class.yaml
-rw-r--r-- 1 root root 1006 Feb 4 15:32 redis-nfs.yml
[root@k8s-master01 redis]# mkdir deploy
[root@k8s-master01 redis]# cd deploy/
[root@k8s-master01 deploy]# cat redis-cluster.yml
---
apiVersion: v1
kind: Service
metadata:
namespace: wiseco
name: redis-cluster
spec:
clusterIP: None
ports:
- port: 6379
targetPort: 6379
name: client
- port: 16379
targetPort: 16379
name: gossip
selector:
app: redis-cluster
---
apiVersion: apps/v1
kind: StatefulSet
metadata:
namespace: wiseco
name: redis-cluster
spec:
serviceName: redis-cluster
podManagementPolicy: OrderedReady
replicas: 6
selector:
matchLabels:
app: redis-cluster
template:
metadata:
labels:
app: redis-cluster
spec:
containers:
- name: redis
image: 172.16.60.238/wiseco/redis:4.0.11
ports:
- containerPort: 6379
name: client
- containerPort: 16379
name: gossip
command: ["/etc/redis/fix-ip.sh", "redis-server", "/etc/redis/redis.conf"]
env:
- name: POD_IP
valueFrom:
fieldRef:
fieldPath: status.podIP
volumeMounts:
- name: conf
mountPath: /etc/redis/
readOnly: false
- name: data
mountPath: /data
readOnly: false
volumes:
- name: conf
configMap:
name: redis-cluster
defaultMode: 0755
volumeClaimTemplates:
- metadata:
name: data
annotations:
volume.beta.kubernetes.io/storage-class: "redis-nfs-storage"
spec:
accessModes:
- ReadWriteMany
resources:
requests:
storage: 10Gi
創建並查看
[root@k8s-master01 deploy]# kubectl apply -f redis-cluster.yml
[root@k8s-master01 deploy]# kubectl get pods -n wiseco|grep redis-cluster
redis-cluster-0 1/1 Running 0 10m
redis-cluster-1 1/1 Running 0 10m
redis-cluster-2 1/1 Running 0 10m
redis-cluster-3 1/1 Running 0 10m
redis-cluster-4 1/1 Running 0 9m35s
redis-cluster-5 1/1 Running 0 9m25s
[root@k8s-master01 deploy]# kubectl get svc -n wiseco|grep redis-cluster
redis-cluster ClusterIP None <none> 6379/TCP,16379/TCP 10m
查看PV、PVC
[root@k8s-master01 deploy]# kubectl get pv
NAME CAPACITY ACCESS MODES RECLAIM POLICY STATUS CLAIM STORAGECLASS REASON AGE
pvc-20bcb3be-90e1-4354-bd11-4f442a3bd562 10Gi RWX Delete Bound wiseco/data-redis-cluster-0 redis-nfs-storage 19m
pvc-3b53a31b-9a53-4bd4-93ff-2cf9fed551de 10Gi RWX Delete Bound wiseco/data-redis-cluster-2 redis-nfs-storage 12m
pvc-43c0cba2-54a9-4416-afb6-8b7730a199dc 10Gi RWX Delete Bound wiseco/data-redis-cluster-1 redis-nfs-storage 12m
pvc-66daade5-1b97-41ce-a9e0-4cf88d63894d 10Gi RWX Delete Terminating wiseco/data-redis-cluster-5 redis-nfs-storage 11m
pvc-dd62a086-1802-446a-9f9d-35620f7f0b4a 10Gi RWX Delete Bound wiseco/data-redis-cluster-4 redis-nfs-storage 11m
pvc-e5aa9802-b983-471c-a7da-32eebc497610 10Gi RWX Delete Bound wiseco/data-redis-cluster-3 redis-nfs-storage 12m
[root@k8s-master01 deploy]# kubectl get pvc -n wiseco
NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE
data-redis-cluster-0 Bound pvc-20bcb3be-90e1-4354-bd11-4f442a3bd562 10Gi RWX redis-nfs-storage 19m
data-redis-cluster-1 Bound pvc-43c0cba2-54a9-4416-afb6-8b7730a199dc 10Gi RWX redis-nfs-storage 12m
data-redis-cluster-2 Bound pvc-3b53a31b-9a53-4bd4-93ff-2cf9fed551de 10Gi RWX redis-nfs-storage 12m
data-redis-cluster-3 Bound pvc-e5aa9802-b983-471c-a7da-32eebc497610 10Gi RWX redis-nfs-storage 12m
data-redis-cluster-4 Bound pvc-dd62a086-1802-446a-9f9d-35620f7f0b4a 10Gi RWX redis-nfs-storage 11m
data-redis-cluster-5 Bound pvc-66daade5-1b97-41ce-a9e0-4cf88d63894d 10Gi RWX redis-nfs-storage 11m
4)查看NFS共享存儲
NFS服務器(172.16.60.238),查看共享目錄/data/storage/k8s/redis
[root@k8s-harbor01 redis]# pwd
/data/storage/k8s/redis
[root@k8s-harbor01 redis]# ll
total 0
drwxrwxrwx 2 root root 63 Feb 4 18:59 wiseco-data-redis-cluster-0-pvc-20bcb3be-90e1-4354-bd11-4f442a3bd562
drwxrwxrwx 2 root root 63 Feb 4 18:59 wiseco-data-redis-cluster-1-pvc-43c0cba2-54a9-4416-afb6-8b7730a199dc
drwxrwxrwx 2 root root 63 Feb 4 18:59 wiseco-data-redis-cluster-2-pvc-3b53a31b-9a53-4bd4-93ff-2cf9fed551de
drwxrwxrwx 2 root root 63 Feb 4 19:00 wiseco-data-redis-cluster-3-pvc-e5aa9802-b983-471c-a7da-32eebc497610
drwxrwxrwx 2 root root 63 Feb 4 19:00 wiseco-data-redis-cluster-4-pvc-dd62a086-1802-446a-9f9d-35620f7f0b4a
drwxrwxrwx 2 root root 63 Feb 4 19:00 wiseco-data-redis-cluster-5-pvc-66daade5-1b97-41ce-a9e0-4cf88d63894d
[root@k8s-harbor01 redis]# ls ./*
./wiseco-data-redis-cluster-0-pvc-20bcb3be-90e1-4354-bd11-4f442a3bd562:
appendonly.aof nodes.conf redis.log
./wiseco-data-redis-cluster-1-pvc-43c0cba2-54a9-4416-afb6-8b7730a199dc:
appendonly.aof nodes.conf redis.log
./wiseco-data-redis-cluster-2-pvc-3b53a31b-9a53-4bd4-93ff-2cf9fed551de:
appendonly.aof nodes.conf redis.log
./wiseco-data-redis-cluster-3-pvc-e5aa9802-b983-471c-a7da-32eebc497610:
appendonly.aof nodes.conf redis.log
./wiseco-data-redis-cluster-4-pvc-dd62a086-1802-446a-9f9d-35620f7f0b4a:
appendonly.aof nodes.conf redis.log
./wiseco-data-redis-cluster-5-pvc-66daade5-1b97-41ce-a9e0-4cf88d63894d:
appendonly.aof nodes.conf redis.log
3、初始化Redis Cluster集群
接下來是形成Redis Cluster集群,運行以下命令並鍵入yes以接受配置。
集群形式:前三個節點成為主節點,后三個節點成為從節點。
需要注意:
redis-trib.rb必須使用ip進行初始化redis集群,使用域名會報如下錯誤:*******/redis/client.rb:126:in `call’: ERR Invalid node address specified: redis-cluster-0.redis-headless.sts-app.svc.cluster.local:6379 (Redis::CommandError)
這里進行Redis Cluster集群初始化的命令:
以下命令並鍵入yes以接受配置。前三個節點成為主節點,后三個節點成為從節點。
kubectl exec -it redis-cluster-0 -n wiseco -- redis-trib.rb create --replicas 1 $(kubectl get pods -l app=redis-cluster -n wiseco -o jsonpath='{range.items[*]}{.status.podIP}:6379 ')
先獲取Redis Cluster集群的6個節點Pod的ip地址
[root@k8s-master01 redis]# kubectl get pods -n wiseco -o wide|grep redis-cluster
redis-cluster-0 1/1 Running 0 4h34m 172.30.217.83 k8s-node04 <none> <none>
redis-cluster-1 1/1 Running 0 4h34m 172.30.85.217 k8s-node01 <none> <none>
redis-cluster-2 1/1 Running 0 4h34m 172.30.135.181 k8s-node03 <none> <none>
redis-cluster-3 1/1 Running 0 4h34m 172.30.58.251 k8s-node02 <none> <none>
redis-cluster-4 1/1 Running 0 4h33m 172.30.85.216 k8s-node01 <none> <none>
redis-cluster-5 1/1 Running 0 4h33m 172.30.217.82 k8s-node04 <none> <none>
[root@k8s-master01 redis]# kubectl get pods -l app=redis-cluster -n wiseco -o jsonpath='{range.items[*]}{.status.podIP}:6379 '
172.30.217.83:6379 172.30.85.217:6379 172.30.135.181:6379 172.30.58.251:6379 172.30.85.216:6379 172.30.217.82:6379
這里特別注意一下:
上面命令最后一個單引號前面一定要有空格!!
因為接下來進行Redis Cluster集群初始化的時候,集群節點間的ip+port之間要通過空格隔開。
[root@k8s-master01 redis]# kubectl exec -it redis-cluster-0 -n wiseco -- redis-trib.rb create --replicas 1 $(kubectl get pods -l app=redis-cluster -n wiseco -o jsonpath='{range.items[*]}{.status.podIP}:6379 ')
>>> Creating cluster
>>> Performing hash slots allocation on 6 nodes...
Using 3 masters:
172.30.217.83:6379
172.30.85.217:6379
172.30.135.181:6379
Adding replica 172.30.58.251:6379 to 172.30.217.83:6379
Adding replica 172.30.85.216:6379 to 172.30.85.217:6379
Adding replica 172.30.217.82:6379 to 172.30.135.181:6379
M: e5a3154a17131075f35fb32953b8cf8d6cfc7df0 172.30.217.83:6379
slots:0-5460 (5461 slots) master
M: 961398483262f505a115957e7e4eda7ff3e64900 172.30.85.217:6379
slots:5461-10922 (5462 slots) master
M: 2d1440e37ea4f4e9f6d39d240367deaa609d324d 172.30.135.181:6379
slots:10923-16383 (5461 slots) master
S: 0d7bf40bf18d474509116437959b65551cd68b03 172.30.58.251:6379
replicates e5a3154a17131075f35fb32953b8cf8d6cfc7df0
S: 8cbf699a850c0dafe51524127a594fdbf0a27784 172.30.85.216:6379
replicates 961398483262f505a115957e7e4eda7ff3e64900
S: 2987a33f4ce2e412dcc11c1c1daa2538591cd930 172.30.217.82:6379
replicates 2d1440e37ea4f4e9f6d39d240367deaa609d324d
Can I set the above configuration? (type 'yes' to accept): yes
>>> Nodes configuration updated
>>> Assign a different config epoch to each node
>>> Sending CLUSTER MEET messages to join the cluster
Waiting for the cluster to join......
>>> Performing Cluster Check (using node 172.30.217.83:6379)
M: e5a3154a17131075f35fb32953b8cf8d6cfc7df0 172.30.217.83:6379
slots:0-5460 (5461 slots) master
M: 961398483262f505a115957e7e4eda7ff3e64900 172.30.85.217:6379
slots:5461-10922 (5462 slots) master
M: 2d1440e37ea4f4e9f6d39d240367deaa609d324d 172.30.135.181:6379
slots:10923-16383 (5461 slots) master
M: 0d7bf40bf18d474509116437959b65551cd68b03 172.30.58.251:6379
slots: (0 slots) master
replicates e5a3154a17131075f35fb32953b8cf8d6cfc7df0
M: 8cbf699a850c0dafe51524127a594fdbf0a27784 172.30.85.216:6379
slots: (0 slots) master
replicates 961398483262f505a115957e7e4eda7ff3e64900
M: 2987a33f4ce2e412dcc11c1c1daa2538591cd930 172.30.217.82:6379
slots: (0 slots) master
replicates 2d1440e37ea4f4e9f6d39d240367deaa609d324d
[OK] All nodes agree about slots configuration.
>>> Check for open slots...
>>> Check slots coverage...
[OK] All 16384 slots covered.
通過上面初始化信息,可以看出集群關系:
redis-cluster-0是master節點,redis-cluster-3是它的從節點。
redis-cluster-1是master節點,redis-cluster-4是它的從節點。
redis-cluster-2是master節點,redis-cluster-5是它的從節點。
4、驗證Redis Cluster集群部署
[root@k8s-master01 redis]# kubectl exec -it redis-cluster-0 -n wiseco -- redis-cli cluster info
cluster_state:ok
cluster_slots_assigned:16384
cluster_slots_ok:16384
cluster_slots_pfail:0
cluster_slots_fail:0
cluster_known_nodes:6
cluster_size:3
cluster_current_epoch:6
cluster_my_epoch:1
cluster_stats_messages_ping_sent:130
cluster_stats_messages_pong_sent:137
cluster_stats_messages_sent:267
cluster_stats_messages_ping_received:132
cluster_stats_messages_pong_received:130
cluster_stats_messages_meet_received:5
cluster_stats_messages_received:267
[root@k8s-master01 redis]# for x in $(seq 0 5); do echo "redis-cluster-$x"; kubectl exec redis-cluster-$x -n wiseco -- redis-cli role; echo; done
redis-cluster-0
master
168
172.30.58.251
6379
168
redis-cluster-1
master
168
172.30.85.216
6379
168
redis-cluster-2
master
182
172.30.217.82
6379
168
redis-cluster-3
slave
172.30.217.83
6379
connected
182
redis-cluster-4
slave
172.30.85.217
6379
connected
168
redis-cluster-5
slave
172.30.135.181
6379
connected
182
注:之前寫過一篇 在k8s上部署單節點redis, 今天發現散盡浮華
大佬 寫了一篇 集群版 的,在其基礎上進行再升級。