Redis的主从集群环境搭建

2023年 4月 29日 40.6k 0

三台主机搭建Redis的三对主从服务器集群环境准备 host1:192.168.1.9:6379 192.168.1.9:6380 host2:192.168.1.106:6379 192.168.1.106:6380 host3:192.168.1.110:6379 192.168.1.110:6380 注意: (1)在建立redis的cluster环境时

三台主机搭建Redis的三对主从服务器集群环境准备

host1:192.168.1.9:6379                    192.168.1.9:6380host2:192.168.1.106:6379                    192.168.1.106:6380host3:192.168.1.110:6379                    192.168.1.110:6380注意:(1)在建立redis的cluster环境时必须清空所有redis服务的所有key-value数据,没有任何数据(2)每个 redis node 节点采用相同的硬件配置、相同的密码

1.分别开启三台主机的6379和6380两个端口,需给redis配置两个独立的配置文件,以host1为例

1)给redis配置6379监听端口

    [root@localhost ~]# vim /app/redis/etc/redis.conf         ……    bind 192.168.1.9  #绑定ip        ……    port 6379  #绑定6379端口    ……    cluster-enabled yes  #开启redis的集群功能    ……    cluster-config-file nodes-6379.conf  #开启自动创建集群配置文件    ……

2)给redis配置6380监听端口

    [root@localhost ~]# cp/app/redis/etc/redis.conf  /app/redis/etc/redis.6380.conf     [root@localhost ~]# vim /app/redis/etc/redis.6380.conf    ……    bind 192.168.1.9  #绑定ip        ……    port 6380  #绑定6380端口    ……    cluster-enabled yes  #开启redis的集群功能    ……    cluster-config-file nodes-6379.conf  #开启自动创建集群配置文件    ……    :%s/6379/6380/g  #配置文件底部实现全局替换,将所有6379替换为6380

2.三台主机redis配置文件设置完成后再启动redis服务

1)同时开启两个redis进程服务

[root@localhost ~]#  redis-server /app/redis/etc/redis.6380.conf && redis-server /app/redis/etc/redis.conf

2)查看redis的6379和6380两个端口是否处于监听状态

[root@localhost ~]# ss -tnlpState      Recv-Q Send-Q                  Local Address:Port                      Peer Address:Port              LISTEN    0      511                            192.168.1.9:6379                            *:*                   users:(("redis-server",pid=7492,fd=6))LISTEN    0      511                            192.168.1.9:6380                            *:*                  users:(("redis-server",pid=7487,fd=6))LISTEN    0      511                            192.168.1.9:16379                          *:*                   users:(("redis-server",pid=7492,fd=8))LISTEN    0      511                            192.168.1.9:16380                          *:*                  users:(("redis-server",pid=7487,fd=8))

3.将创建集群的命令redis-trib.rb复制到/usr/bin下,此命令暂时不可用,还需编译安装ruby以及安装redis模块

[root@localhost ruby-2.5.5]# cp redis-trib.rb /usr/bin

4.安装编译ruby工具包时编译环境

yum ×××tall -y vim lrzsz tree screen psmisc lsof tcpdump wget  ntpdate  gcc gcc-c++ glibc glibc-devel pcre pcre-devel openssl  openssl-devel systemd-devel net-tools iotop bc  zip unzip zlib-devel bash-completion nfs-utils automake libxml2  libxml2-devel libxslt libxslt-devel perl perl-ExtUtils-Embed

5.编译安装ruby工具包(因yum安装的ruby版本太低,不能满足gem安装redis模块所依赖的ruby版本)

[root@localhost ~]# cd /data/ruby/[root@localhost ruby]# tar xf ruby-2.5.5.tar.gz[root@localhost ruby]# cd ruby-2.5.5/[root@localhost ruby-2.5.5]#  ./configure && make –j 4 && make ×××tall

6.安装 rubygems包

[root@localhost ruby-2.5.5]# yum ×××tall -y rubygems

7.分别给ruby命令和gem命令创建环境变量路径软连接

[root@localhost ruby-2.5.5]# ln -sv /data/ruby/ruby-2.5.5/bin/gem /usr/bin/[root@localhost ruby-2.5.5]# ln -sv/data/ruby/ruby-2.5.5/ruby /usr/bin/

8.gem安装redis模块

    [root@localhost ruby-2.5.5]# gem ×××tall redis -y   

9.修改redis模块登录集群环境的密码

[root@localhost ~]# vim /usr/local/lib/ruby/gems/2.5.0/gems/redis-4.1.2/lib/redis/client.rb#frozen_string_literal: true

require_relative "errors"require "socket"require "cgi"

class Redis    class Client

        DEFAULTS = {            :url => lambda { ENV["REDIS_URL"] },            :scheme => "redis",            :host => "127.0.0.1",            :port => 6379,            :path => nil,            :timeout => 5.0,            :password => 123456, #登录密码改为123456            :db => 0,            :driver => nil,            ……

10.使用redis-trib.rb命令创建集群环境,会自动分配主从服务

[root@localhost ~]# redis-trib.rb create --replicas 1 192.168.1.9:6379  192.168.1.9:6380  192.168.1.106:6379  192.168.1.106:6380  192.168.1.110:6379  192.168.1.110:6380

>>> Creating cluster>>> Performing hash slots allocation on 6 nodes...Using 3 masters:192.168.1.9:6379192.168.1.106:6379192.168.1.110:6379Adding replica 192.168.1.106:6380 to 192.168.1.9:6379Adding replica 192.168.1.110:6380 to 192.168.1.106:6379Adding replica 192.168.1.9:6380 to 192.168.1.110:6379M: eed2e22136cbdca6770a46bbb2e137ab693dd16b 192.168.1.9:6379    slots:0-5460 (5461 slots) master#主服务:M#主服务id:eed2e22136cbdca6770a46bbb2e137ab693dd16b#主服务ip和端口:192.168.1.9:6379#分得的槽位区间:0-5460#一共分得的槽:5460

S: 8efbcf7fdd4675c77199a2a1206f0209ac2255f3 192.168.1.9:6380    replicates e387a0ba7c95d0c27d7e28a9b57d23117711eadc#从服务:S#从服务id:8efbcf7fdd4675c77199a2a1206f0209ac2255f3#从服务ip和端口:192.168.1.9:6380#从服务所属的主服务id:e387a0ba7c95d0c27d7e28a9b57d23117711eadc

M: c922b4bf56f0086609fd4fb23d987df0a77bec22 192.168.1.106:6379    slots:5461-10922 (5462 slots) masterS: a4fd89d79cdd27698bc394134b2df25b63ddb4c5 192.168.1.106:6380    replicates eed2e22136cbdca6770a46bbb2e137ab693dd16bM: e387a0ba7c95d0c27d7e28a9b57d23117711eadc 192.168.1.110:6379    slots:10923-16383 (5461 slots) masterS: 34549b777963b16e65125def8d9a8e50e27ed2a4 192.168.1.110:6380    replicates c922b4bf56f0086609fd4fb23d987df0a77bec22Can I set the above configuration? (type 'yes' to accept): yes>>> Nodes configuration updated>>> Assign a different config epoch to each node>>> Sending CLUSTER MEET messages to join the clusterWaiting for the cluster to join..>>> Performing Cluster Check (using node 192.168.1.9:6379)M: eed2e22136cbdca6770a46bbb2e137ab693dd16b 192.168.1.9:6379    slots:0-5460 (5461 slots) master    1 additional replica(s)S: a4fd89d79cdd27698bc394134b2df25b63ddb4c5 192.168.1.106:6380    slots: (0 slots) slave    replicates eed2e22136cbdca6770a46bbb2e137ab693dd16bS: 8efbcf7fdd4675c77199a2a1206f0209ac2255f3 192.168.1.9:6380    slots: (0 slots) slave    replicates e387a0ba7c95d0c27d7e28a9b57d23117711eadcM: c922b4bf56f0086609fd4fb23d987df0a77bec22 192.168.1.106:6379    slots:5461-10922 (5462 slots) master    1 additional replica(s)M: e387a0ba7c95d0c27d7e28a9b57d23117711eadc 192.168.1.110:6379    slots:10923-16383 (5461 slots) master    1 additional replica(s)S: 34549b777963b16e65125def8d9a8e50e27ed2a4 192.168.1.110:6380    slots: (0 slots) slave    replicates c922b4bf56f0086609fd4fb23d987df0a77bec22[OK] All nodes agree about slots configuration.>>> Check for open slots...>>> Check slots coverage...[OK] All 16384 slots covered.

11.查看slave从服务的连接状态,可以看到还未建立主从连接

[root@localhost ~]# redis-cli -h 192.168.1.106 -p 6380192.168.1.106:6380> auth 123456OK192.168.1.106:6380> info  replication#Replicationrole:slavemaster_host:192.168.1.9master_port:6379master_link_status:down  #未与主服务建立连接master_last_io_seconds_ago:-1master_sync_in_progress:0slave_repl_offset:1master_link_down_since_seconds:1560643385slave_priority:100slave_read_only:1connected_slaves:0master_replid:84865c623e15576c50c80a46ee16845b80b872d8master_replid2:0000000000000000000000000000000000000000master_repl_offset:0second_repl_offset:-1repl_backlog_active:0repl_backlog_size:1048576repl_backlog_first_byte_offset:0repl_backlog_histlen:0

11.三台salve从服务分别手动输入主从连接认证密码

1)输入认证密码

192.168.1.106:6380> config set masterauth 123456OK

2)再次查看salve服务连接状态,可以看到为up表明主从连接成功

192.168.1.106:6380> info  replication#Replicationrole:slavemaster_host:192.168.1.9master_port:6379master_link_status:up  #主从服务已建立连接master_last_io_seconds_ago:1master_sync_in_progress:0slave_repl_offset:0slave_priority:100slave_read_only:1connected_slaves:0master_replid:df9da60e308938e7a817ca08b20b58c248ad409dmaster_replid2:0000000000000000000000000000000000000000master_repl_offset:0second_repl_offset:-1repl_backlog_active:1repl_backlog_size:1048576repl_backlog_first_byte_offset:1repl_backlog_histlen:0

12.实验总结

1) 实现了三台主机搭建3对主从服务器的集群环境,实现冗余可用和较高的并发量;

2)即可允许其中一台主机宕机,被宕掉的一个主服务会被它的从服替代,从服务会被提升为新的主服务,从而不影响另外两台主机服务器的正常工作,也不会丢失数据;

3)减少了主机的数量从而节约成本。

相关文章

Oracle如何使用授予和撤销权限的语法和示例
Awesome Project: 探索 MatrixOrigin 云原生分布式数据库
下载丨66页PDF,云和恩墨技术通讯(2024年7月刊)
社区版oceanbase安装
Oracle 导出CSV工具-sqluldr2
ETL数据集成丨快速将MySQL数据迁移至Doris数据库

发布评论