我有一个由8个minio节点组成的集群。每个节点都有一个1.6 - 1.9 TB的驱动器。节点6的空闲空间小于1兆字节,而其余节点的空闲空间约为200 TB至1TB。是否有任何方法触发已使用资源的再平衡?
Minio是通过docker运行的,下面是堆栈缺陷:
version: '3.7'
services:
minio1:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio1
volumes:
- /opt/iqdata:/data
ports:
- "9001:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio1==true
minio2:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio2
volumes:
- /opt/iqdata:/data
ports:
- "9002:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio2==true
minio3:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio3
volumes:
- /opt/iqdata:/data
ports:
- "9003:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio3==true
minio4:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
hostname: minio4
volumes:
- /opt/iqdata:/data
ports:
- "9004:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio4==true
minio5:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio5
volumes:
- /opt/iqdata:/data
ports:
- "9005:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio5==true
minio6:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio6
volumes:
- /opt/iqdata:/data
ports:
- "9006:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio6==true
minio7:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio7
volumes:
- /opt/iqdata:/data
ports:
- "9007:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio7==true
minio8:
image: minio/minio:RELEASE.2021-03-17T02-33-02Z
network_mode: "host"
hostname: minio8
volumes:
- /opt/iqdata:/data
ports:
- "9008:9000"
command: server http://minio{1...8}/data
secrets:
- secret_key
- access_key
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3
deploy:
replicas: 1
restart_policy:
delay: 10s
placement:
constraints:
- node.labels.minio8==true
secrets:
secret_key:
external: true
access_key:
external: true发布于 2022-02-25 02:44:26
MinIO不执行任何再平衡。如果磁盘空间不足( https://docs.min.io/docs/distributed-minio-quickstart-guide.html ),它支持扩展集群(通过添加新的磁盘池)。
但是,您的情况似乎是异常的--可能是应用程序中的错误(可能创建了同一个对象的太多版本),或者是MinIO中的错误--这种情况不应该发生。如果没有更多的信息,就很难确定这个问题。
请知道您可以在https://slack.min.io/ 24/7/365找到我们。如果你有商业问题,请联系我们的hello@min.io或我们在https://min.io/pricing?action=talk-to-us的专家聊天功能。
https://stackoverflow.com/questions/69302154
复制相似问题