zookeeper02/data:/data \ -v /root/zookeeper02/datalog:/datalog \ zookeeper:3.4.14 docker run -d --name zookeeper 重启三个容器 或者 先停止再启动. # 重启容器 docker restart zookeeper-1 zookeeper-2 zookeeper-3 # 停止容器 docker stop zookeeper -1 zookeeper-2 zookeeper-3 # 启动容器 docker start zookeeper-1 zookeeper-2 zookeeper-3 7.
-3/conf/zoo_sample.cfg /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg 5、 配置每一个Zookeeper 的dataDir -3/conf/zoo.cfg vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg clientPort=2184 dataDir=/usr /local/zookeeper-cluster/zookeeper-3/data 3 配置集群 1、在每个zookeeper的 data 目录下创建一个 myid 文件,内容分别是1、2、3 。 -3 vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg server.1=192.168.58.150:2881:3881 server /usr/local/zookeeper-cluster/zookeeper-3/bin/zkServer.sh start 把2号服务器停掉 /usr/local/zookeeper-cluster/
-3/conf/ zoo_sample.cfg /usr/local/zookeeper-cluster/zookeeper-3/conf/ zoo.cfg 5、 配置每一个Zookeeper 的 -3/conf/zoo.cfg vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg 内容: clientPort=2184 dataDir=/usr/local/zookeeper-cluster/zookeeper-3/data 3、 配置集群 3.1、在每个zookeeper的 data 目录下创建一个 myid 文件 -3 vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg 内容: server.1=192.168.58.150:2881:3881 /usr/local/zookeeper-cluster/zookeeper-3/bin/ zkServer.sh stop 观察1号和2号 /usr/local/zookeeper-cluster
value: "1" - name: ZOO_SERVERS value: 0.0.0.0:2888:3888,zookeeper-2:2888:3888,zookeeper -3:2888:3888 --- kind: Deployment apiVersion: apps/v1 metadata: name: zookeeper-3 namespace: rcmd spec: replicas: 1 selector: matchLabels: app: zookeeper-3 template: metadata: labels: app: zookeeper-3 spec: containers: - name: zookeeper-3 image: -3 namespace: rcmd labels: app: zookeeper-3 spec: ports: - name: client port: 2181
zookeeper-cluster/zookeeper-1 /usr/local/zookeeper-cluster/zookeeper-2 /usr/local/zookeeper-cluster/zookeeper -3/conf/zoo_sample.cfg /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg (5) 配置每一个Zookeeper 的dataDir -3/conf/zoo.cfg vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg clientPort=2183 dataDir=/usr /local/zookeeper-cluster/zookeeper-3/data 4.4 配置集群 (1)在每个zookeeper的 data 目录下创建一个 myid 文件,内容分别是1、2、3 。 /usr/local/zookeeper-cluster/zookeeper-3/bin/zkServer.sh start /usr/local/zookeeper-cluster/zookeeper
zookeeper-cluster/zookeeper-1 /usr/local/zookeeper-cluster/zookeeper-2 /usr/local/zookeeper-cluster/zookeeper -3/conf/zoo_sample.cfg /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg (5) 配置每一个Zookeeper 的dataDir -3/conf/zoo.cfg vim /usr/local/zookeeper-cluster/zookeeper-3/conf/zoo.cfg clientPort=2183 dataDir=/usr /local/zookeeper-cluster/zookeeper-3/data 5.4)配置集群 (1)在每个zookeeper的 data 目录下创建一个 myid 文件,内容分别是1、2、3 。 /usr/local/zookeeper-cluster/zookeeper-3/bin/zkServer.sh start /usr/local/zookeeper-cluster/zookeeper
: 1 ZOO_SERVERS: server.1=0.0.0.0:2888:3888;2181 server.2=zookeeper-2:2888:3888;2181 server.3=zookeeper : 2 ZOO_SERVERS: server.1=zookeeper-1:2888:3888;2181 server.2=0.0.0.0:2888:3888;2181 server.3=zookeeper -3: image: zookeeper restart: always hostname: zookeeper-3 ports: - 2183:2181 kafka-manager:latest restart: always links: - zookeeper-1 - zookeeper-2 - zookeeper kfk3 ports: - 9000:9000 environment: ZK_HOSTS: zookeeper-1:2181,zookeeper-2:2181,zookeeper
value: "1" - name: ZOO_SERVERS value: 0.0.0.0:2888:3888,zookeeper-2:2888:3888,zookeeper value: "2" - name: ZOO_SERVERS value: zookeeper-1:2888:3888,0.0.0.0:2888:3888,zookeeper -3:2888:3888 --- kind: Deployment apiVersion: apps/v1 metadata: name: zookeeper-3 namespace: rcmd spec: replicas: 1 selector: matchLabels: app: zookeeper-3 template: metadata: labels: app: zookeeper-3 spec: containers: - name: zookeeper image: bitnami
192.168.88.113 zookeeper-1(codis-proxy-1) 192.168.88.114 zookeeper-2(codis-proxy-2) 192.168.88.115 zookeeper 2811,19000 hostname:zookeeper-2 apps: zookeeper2, codis_proxy_2 prots:2811,19000 hostname:zookeeper [root@zookeeper-3 ~]# echo 3 > /data/zookeeper/data/myid 5,启动zookeeper 启动顺序zookeeper-1>zookeeper-2>zookeeper 配置codis_proxy ( zookeeper-1、zookeeper-2、zookeeper-3 机器上配置) 配置codis_proxy_1 ( zookeeper-1 机器上配置) cd / 修改start_proxy.sh,启动codis-proxy服务 ( 在zookeeper-1、zookeeper-2、zookeeper-3上配置) zookeeper-1上(其他上面就是codis_proxy
zookeeper-deployment-3 namespace: ms spec: replicas: 1 selector: matchLabels: app: zookeeper -3 name: zookeeper-3 template: metadata: labels: app: zookeeper-3 name : zookeeper-3 spec: containers: - name: zoo3 image: uhub.service.ucloud.cn/metersphere selector: app: zookeeper-2 --- apiVersion: v1 kind: Service metadata: name: zoo3 labels: app: zookeeper port: 2888 protocol: TCP - name: leader port: 3888 protocol: TCP selector: app: zookeeper
zookeeper-cluster/zookeeper-1 /usr/local/zookeeper-cluster/zookeeper-2 /usr/local/zookeeper-cluster/zookeeper zookeeper-cluster/zookeeper-2[root@localhost ~]# cp -r zookeeper-3.4.6 /usr/local/zookeeper-cluster/zookeeper clientPort=2182dataDir=/usr/local/zookeeper-cluster/zookeeper-2/data 修改/usr/local/zookeeper-cluster/zookeeper -3/conf/zoo.cfg clientPort=2183dataDir=/usr/local/zookeeper-cluster/zookeeper-3/data 3配置集群 (1)在每个zookeeper
下载解压 解压到3个目录(模拟3台zk server): /home/hadoop/zookeeper-1 /home/hadoop/zookeeper-2 /home/hadoop/zookeeper server.1=localhost:2287:3387 server.2=localhost:2288:3388 server.3=localhost:2289:3389 /home/hadoop/zookeeper hadoop/zookeeper-1/bin/zkServer.sh start /home/hadoop/zookeeper-2/bin/zkServer.sh start /home/hadoop/zookeeper
zookeeper-cluster/zookeeper-1 /usr/local/zookeeper-cluster/zookeeper-2 /usr/local/zookeeper-cluster/zookeeper zookeeper-cluster/zookeeper-2[root@localhost ~]# cp -r zookeeper-3.4.6 /usr/local/zookeeper-cluster/zookeeper clientPort=2182dataDir=/usr/local/zookeeper-cluster/zookeeper-2/data 修改/usr/local/zookeeper-cluster/zookeeper -3/conf/zoo.cfg clientPort=2183dataDir=/usr/local/zookeeper-cluster/zookeeper-3/data 2.2.3配置集群 (1)在每个
cfg # vim conf/zoo-2.cfg dataDir=/tmp/zookeeper-2 clientPort=2182 # vim conf/zoo-2.cfg dataDir=/tmp/zookeeper 写入当前实例的server id,即1.2.3 # cd /tmp/zookeeper-1 # vim myid 1 # cd /tmp/zookeeper-2 # vim myid 2 # cd /tmp/zookeeper
</property> <property> <name>ha.zookeeper.quorum</name> <value>zookeeper-1:2181,zookeeper-2:2181,zookeeper <property> <name>yarn.resourcemanager.zk-address</name> <value>zookeeper-1:2181,zookeeper-2:2181,zookeeper 4lw.commands.whitelist=mntr 增加节点: server.1=zookeeper-1:2888:3888 server.2=zookeeper-2:2888:3888 server.3=zookeeper journalnode-1 resourcemanager-1 10.0.36.169 zookeeper-2 namenode-2 journalnode-2 resourcemanager-2 10.0.41.11 zookeeper
cfg # vim conf/zoo-2.cfg dataDir=/tmp/zookeeper-2 clientPort=2182 # vim conf/zoo-2.cfg dataDir=/tmp/zookeeper 写入当前实例的server id,即1.2.3 # cd /tmp/zookeeper-1 # vim myid 1 # cd /tmp/zookeeper-2 # vim myid 2 # cd /tmp/zookeeper
value> </property> <property> <name>hadoop.zk.address</name> <value>zookeeper-1:2181,zookeeper-2:2181,zookeeper
zookeeper-cluster/zookeeper-1 /usr/local/zookeeper-cluster/zookeeper-2 /usr/local/zookeeper-cluster/zookeeper zookeeper-cluster/zookeeper-2[root@localhost ~]# cp -r zookeeper-3.4.6 /usr/local/zookeeper-cluster/zookeeper clientPort=2182dataDir=/usr/local/zookeeper-cluster/zookeeper-2/data 修改/usr/local/zookeeper-cluster/zookeeper -3/conf/zoo.cfg clientPort=2183dataDir=/usr/local/zookeeper-cluster/zookeeper-3/data 2.2.3配置集群 在每个zookeeper
2184 Step5:标识Server ID 创建四个文件夹/usr/local/data/zookeeper-1,/usr/local/data/zookeeper-2,/usr/local/data/zookeeper local/data/zookeeper-1 vim myid 1 cd /usr/local/data/zookeeper-2 vim myid 2 cd /usr/local/data/zookeeper
2184 Step5:标识Server ID 创建四个文件夹/usr/local/data/zookeeper-1,/usr/local/data/zookeeper-2,/usr/local/data/zookeeper local/data/zookeeper-1 vim myid 1 cd /usr/local/data/zookeeper-2 vim myid 2 cd /usr/local/data/zookeeper