Redis集羣是Redis提供的分佈式數據庫方案,集羣經過分片(sharding)來進行數據共享,並提供複製和故障轉移功能。node
一個Redis集羣一般由多個節點(node)組成,在剛開始的時候,每一個節點都是相互獨立的,它們都處於一個只包含本身的集羣當中,要組建一個真正可工做的集羣,咱們必須將各個獨立的節點鏈接起來,構成一個包含多個節點的集羣。git
下載配置文件:https://raw.githubusercontent.com/antirez/redis/5.0/redis.conf
github
調整 CLUSTER 節點配置redis
# 開啓cluster集羣 cluster-enabled yes # 集羣配置文件 cluster-config-file nodes-6379.conf # 集羣節點超時 cluster-node-timeout 15000
├── conf │ ├── redis.conf │ └── sentinel.conf ├── redis │ ├── data_6379 │ ├── data_6380 │ ├── data_6381 │ ├── data_6382 │ ├── data_6383 │ └── data_6384 └── scripts ├── cluster.sh ├── run.sh └── sentinel.sh
#!/usr/bin/env bash set -e # 腳本當前目錄 cPath=$(cd $(dirname "$0") || exit; pwd) # 根目錄 dirPath=$(dirname "$cPath") # 獲取端口 port="$1" if [[ ! "$port" ]]; then port=6379 fi # 建立數據目錄 mkdir -p "$dirPath"/redis/data_"$port" # 刪除已啓動服務 containerId=$(docker ps -a | grep "redis_$port" | awk -F' ' '{print $1}') if [[ "$containerId" ]]; then docker rm -f ${containerId} > /dev/null fi # 啓動服務 containerName=redis_"$port" docker run -itd --privileged=true -p "$port":6379 --name ${containerName} \ -v="$dirPath"/conf/redis.conf:/etc/redis/redis.conf \ -v="$dirPath"/redis/data_"$port":/data \ redis \ redis-server /etc/redis/redis.conf > /dev/null # 獲取容器IP地址 dockerIp=$(docker inspect -f "{{.NetworkSettings.IPAddress}}" "$containerName") # 獲取容器啓動狀態 isRunning=$(docker inspect -f "{{.State.Running}}" "$containerName") if [[ "$isRunning" == "true" ]]; then echo "容器:$containerName - IP:$dockerIp - 啓動成功" fi
#!/usr/bin/env bash set -e # 腳本當前目錄 cPath=$(cd $(dirname "$0") || exit; pwd) # 啓動集羣數量 num="$1" if [[ ! "$num" ]]; then num=6 fi sPort=6378 for((i=1;i<=$num;i++)); do sh ${cPath}/run.sh $(($sPort+$i)) done
執行腳本文件,默認建立6個節點docker
sh scripts/cluster.sh
容器:redis_6379 - IP:172.17.0.2 - 啓動成功 容器:redis_6380 - IP:172.17.0.3 - 啓動成功 容器:redis_6381 - IP:172.17.0.4 - 啓動成功 容器:redis_6382 - IP:172.17.0.5 - 啓動成功 容器:redis_6383 - IP:172.17.0.6 - 啓動成功 容器:redis_6384 - IP:172.17.0.7 - 啓動成功
執行 docker ps
確實是否啓動成功數據庫
root@DESKTOP-Q13EI52:~/docker-config/redis# docker ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES c0601df1a456 redis "docker-entrypoint.s…" 27 seconds ago Up 26 seconds 0.0.0.0:6384->6379/tcp redis_6384 6fecf70465b8 redis "docker-entrypoint.s…" 27 seconds ago Up 26 seconds 0.0.0.0:6383->6379/tcp redis_6383 1af15e90b7a0 redis "docker-entrypoint.s…" 28 seconds ago Up 27 seconds 0.0.0.0:6382->6379/tcp redis_6382 6c495f31a5df redis "docker-entrypoint.s…" 28 seconds ago Up 28 seconds 0.0.0.0:6381->6379/tcp redis_6381 e54fd9fd0550 redis "docker-entrypoint.s…" 29 seconds ago Up 28 seconds 0.0.0.0:6380->6379/tcp redis_6380 be92ad2f7046 redis "docker-entrypoint.s…" 29 seconds ago Up 29 seconds 0.0.0.0:6379->6379/tcp redis_6379
到此爲止,6個獨立集羣節點建立完畢,目前還沒法正常工做。bash
此處能夠跳過,本人是爲了省事
獲取容器爲redis_
開始全部的容器IP地址app
docker inspect -f "{{.NetworkSettings.IPAddress}}:6379" `docker ps | grep redis_ | awk -F' ' '{print $1}'` | sort |xargs | sed 's/ /, /g' # 返回結果 # 172.17.0.2:6379, 172.17.0.3:6379, 172.17.0.4:6379, 172.17.0.5:6379, 172.17.0.6:6379, 172.17.0.7:6379
初次建立集羣執行tcp
./redis-cli --cluster create 172.17.0.2:6379, 172.17.0.3:6379, 172.17.0.4:6379, 172.17.0.5:6379, 172.17.0.6:6379, 172.17.0.7:6379 --cluster-replicas 1
輸出結果分佈式
licas 1 >>> Performing hash slots allocation on 6 nodes... Master[0] -> Slots 0 - 5460 Master[1] -> Slots 5461 - 10922 Master[2] -> Slots 10923 - 16383 Adding replica 172.17.0.6:6379 to 172.17.0.2:6379 Adding replica 172.17.0.7:6379 to 172.17.0.3:6379 Adding replica 172.17.0.5:6379 to 172.17.0.4:6379 M: e8da1fef656984de3ec2a677edc8d9c48d01cd95 172.17.0.2:6379 slots:[0-5460] (5461 slots) master M: 68b925ab0fbbc1a632c1754587fb6dad3fa14c91 172.17.0.3:6379 slots:[5461-10922] (5462 slots) master M: 0a46ab2f6d176738b55fe699c2df1c34f8200d06 172.17.0.4:6379 slots:[10923-16383] (5461 slots) master S: bd3064ad5297dfc258e9236943455c589be8b2a3 172.17.0.5:6379 replicates 0a46ab2f6d176738b55fe699c2df1c34f8200d06 S: f1d8c897882d29e6538b1158525493b3b782289a 172.17.0.6:6379 replicates e8da1fef656984de3ec2a677edc8d9c48d01cd95 S: 619e1cb52f39e07b321719b77fc3631fa6293cef 172.17.0.7:6379 replicates 68b925ab0fbbc1a632c1754587fb6dad3fa14c91 Can I set the above configuration? (type 'yes' to accept):
輸入:yes,將平均分配槽位
>>> Nodes configuration updated >>> Assign a different config epoch to each node >>> Sending CLUSTER MEET messages to join the cluster Waiting for the cluster to join ..... >>> Performing Cluster Check (using node 172.17.0.2:6379) M: e8da1fef656984de3ec2a677edc8d9c48d01cd95 172.17.0.2:6379 slots:[0-5460] (5461 slots) master 1 additional replica(s) S: f1d8c897882d29e6538b1158525493b3b782289a 172.17.0.6:6379 slots: (0 slots) slave replicates e8da1fef656984de3ec2a677edc8d9c48d01cd95 S: bd3064ad5297dfc258e9236943455c589be8b2a3 172.17.0.5:6379 slots: (0 slots) slave replicates 0a46ab2f6d176738b55fe699c2df1c34f8200d06 M: 0a46ab2f6d176738b55fe699c2df1c34f8200d06 172.17.0.4:6379 slots:[10923-16383] (5461 slots) master 1 additional replica(s) S: 619e1cb52f39e07b321719b77fc3631fa6293cef 172.17.0.7:6379 slots: (0 slots) slave replicates 68b925ab0fbbc1a632c1754587fb6dad3fa14c91 M: 68b925ab0fbbc1a632c1754587fb6dad3fa14c91 172.17.0.3:6379 slots:[5461-10922] (5462 slots) master 1 additional replica(s) [OK] All nodes agree about slots configuration. >>> Check for open slots... >>> Check slots coverage... [OK] All 16384 slots covered.
經過客戶端鏈接
redis-cli -c <端口>
執行命令:cluster info
127.0.0.1:6379> cluster info cluster_state:ok cluster_slots_assigned:16384 cluster_slots_ok:16384 cluster_slots_pfail:0 cluster_slots_fail:0 cluster_known_nodes:6 cluster_size:3 cluster_current_epoch:6 cluster_my_epoch:1 cluster_stats_messages_ping_sent:104 cluster_stats_messages_pong_sent:120 cluster_stats_messages_sent:224 cluster_stats_messages_ping_received:115 cluster_stats_messages_pong_received:104 cluster_stats_messages_meet_received:5 cluster_stats_messages_received:224
看到:cluster_state:ok
說明集羣已能夠正常工做
客戶端控制檯:cluster help
127.0.0.1:6379> cluster help 1) CLUSTER <subcommand> arg arg ... arg. Subcommands are: 2) ADDSLOTS <slot> [slot ...] -- Assign slots to current node. 3) BUMPEPOCH -- Advance the cluster config epoch. 4) COUNT-failure-reports <node-id> -- Return number of failure reports for <node-id>. 5) COUNTKEYSINSLOT <slot> - Return the number of keys in <slot>. 6) DELSLOTS <slot> [slot ...] -- Delete slots information from current node. 7) FAILOVER [force|takeover] -- Promote current replica node to being a master. 8) FORGET <node-id> -- Remove a node from the cluster. 9) GETKEYSINSLOT <slot> <count> -- Return key names stored by current node in a slot. 10) FLUSHSLOTS -- Delete current node own slots information. 11) INFO - Return information about the cluster. 12) KEYSLOT <key> -- Return the hash slot for <key>. 13) MEET <ip> <port> [bus-port] -- Connect nodes into a working cluster. 14) MYID -- Return the node id. 15) NODES -- Return cluster configuration seen by node. Output format: 16) <id> <ip:port> <flags> <master> <pings> <pongs> <epoch> <link> <slot> ... <slot> 17) REPLICATE <node-id> -- Configure current node as replica to <node-id>. 18) RESET [hard|soft] -- Reset current node (default: soft). 19) SET-config-epoch <epoch> - Set config epoch of current node. 20) SETSLOT <slot> (importing|migrating|stable|node <node-id>) -- Set slot state. 21) REPLICAS <node-id> -- Return <node-id> replicas. 22) SAVECONFIG - Force saving cluster configuration on disk. 23) SLOTS -- Return information about slots range mappings. Each range is made of: 24) start, end, master and replicas IP addresses, ports and ids
查看客戶端提供的集羣相關命令:redis-cli --cluster help
Cluster Manager Commands: create host1:port1 ... hostN:portN --cluster-replicas <arg> check host:port --cluster-search-multiple-owners info host:port fix host:port --cluster-search-multiple-owners reshard host:port --cluster-from <arg> --cluster-to <arg> --cluster-slots <arg> --cluster-yes --cluster-timeout <arg> --cluster-pipeline <arg> --cluster-replace rebalance host:port --cluster-weight <node1=w1...nodeN=wN> --cluster-use-empty-masters --cluster-timeout <arg> --cluster-simulate --cluster-pipeline <arg> --cluster-threshold <arg> --cluster-replace add-node new_host:new_port existing_host:existing_port --cluster-slave --cluster-master-id <arg> del-node host:port node_id call host:port command arg arg .. arg set-timeout host:port milliseconds import host:port --cluster-from <arg> --cluster-copy --cluster-replace