可见 docker.service 是依赖于 docker.socket 的,但是并不必要开启 Tip: 其实上面的步骤完成,就已经能保证docker会开机启动,原因是它依赖的 docker.socket 虽然本身设定为不要开机启动,但开机时会被systemctl检查然后触发启动以支持 docker.service 的运行 可以用上面方法也将 docker.socket 设为开机启动(但这一步不是非常必要 ) [root@h103 ~]# systemctl list-unit-files| grep docker docker.service enabled docker.socket disabled [root@h103 ~]# systemctl enable docker.socket /docker.socket.
~]# docker -H :4000 ps Cannot connect to the Docker daemon. Is the docker daemon running on this host? [root@docker ~]# [root@docker ~]# docker -H :4000 info Cannot connect to the Docker daemon. Is the docker daemon running on this host? [root@docker ~]# docker -H 192.168.100.104:4000 info Containers: 11 Images: 17 Server Version: swarm
@docker ~]# pip install -U docker-compose==1.24.1 检查docker compose版本: [root@docker ~]# docker-compose ~]# mkdir /usr/local/docker-compose [root@docker ~]# cd /usr/local/docker-compose/ [root@docker docker-compose docker-compose]# mkdir test [root@docker docker-compose]# cd test/ [root@docker test]# vi docker-compose.yml ]# docker volume ls DRIVER VOLUME NAME local 4ae86d65d3c9d303fbe30f4cdd57b70d874b2f2d8bbc9d1c7edb3e1fb5bf1cb2 local c257d582258d30274aae0e411c3d39d2ce9c70c8c771fe702da0d9c20dbfa9cd local wordpress_db_data
原因是证书中指定的主机名为 docker-registry 而不是 h104 [root@h103 ~]# vim /etc/hosts [root@h103 ~]# grep docker-registry /etc/hosts 192.168.100.104 h104 docker-registry [root@h103 ~]# docker push docker-registry:5000/ubuntu systemctl start docker [root@h103 ~]# docker push docker-registry:5000/ubuntu The push refers to a repository f15ce52fc004: Pushed latest: digest: sha256:45d78ef16a9e6199ffbbc78f71c2c6ef6647f3be6b9721fe3f1b08d6e3fcf6b3 Pulling from ubuntu Digest: sha256:45d78ef16a9e6199ffbbc78f71c2c6ef6647f3be6b9721fe3f1b08d6e3fcf6b3
下载镜像 docker pull harisekhon/hbase 运行容器 docker run -ti harisekhon/hbase当你执行成功之后直接进入hbase shell,慢慢享用吧 使用hbase 退出容器后,下次再进入可以使用 docker exec -ti hbase1 /hbase/bin/hbase shell 创建表,第一个参数是表名,第二个参数是列簇名 create '
参考链接 Run Elasticsearch locally | Elastic Docs 环境 操作系统:Ubuntu 24.04 LTS 容器:Docker 28.2.2 ES 版本:最新版(9.0.3 ) 安装 1)安装 docker【已安装可跳过】 参考:Ubuntu 安装 docker-CSDN博客 2)启动 elasticsearch 若是出现报错请往 Q&A 看 # 通过 curl 下载启动脚本 并进行执行 curl -fsSL https://elastic.co/start-local | sh 3)查看是否启动成功 docker ps 开启远程访问【可选】 默认是只能本地访问 1)切换到 elastic-start-local 目录 这个目录就在你执行 curl -fsSL https://elastic.co/start-local | sh 的目录中, cd elastic-start-local 2)修改 docker-compose.yml vim docker-compose.yml 删除(有两处需要删除) 127.0.0.1: 删除后: 3)执行停止脚本 .
randrepeat=boolForrandomIOworkloads,seedthegeneratorinapredictablewaysothatresultsarerepeatableacrossrepetitions.Defaultstotrue.randseed=intSeedtherandomnumbergeneratorsbasedonthisseedvalue,tobeabletocontrolwhatsequenceofoutputisbeinggenerated.Ifnotset,therandomsequencedependsontherandrepeatsetting.fallocate=strWhetherpre-allocationisperformedwhenlayingdownfiles.Acceptedvaluesare:noneDonotpre-allocatespaceposixPre-allocateviaposix_fallocate()keepPre-allocateviafallocate()withFALLOC_FL_KEEP_SIZEset0Backward-compatiblealiasfor'none'1Backward-compatiblealiasfor'posix'Maynotbeavailableonallsupportedplatforms.'keep'isonlyavailableonLinux.IfusingZFSonSolaristhismustbesetto'none'becauseZFSdoesn't support it. Default: 'posix'. fadvise_hint=bool By default, fio will use fadvise() to advise the kernel on what IO patterns it is likely to issue. Sometimes you want to test specific IO patterns without telling the kernel about it, in which case you can disable this option. If set, fio will use POSIX_FADV_SEQUENTIAL for sequential IO and POSIX_FADV_RANDOM for random IO. fadvise_stream=int Notify the kernel what write stream ID to place these writes under. Only supported on Linux. Note, this option may change going forward. size=int The total size of file io for this job. Fio will run until this many bytes has been transferred, unless runtime is limited by other options (such as 'runtime', for instance, or increased/decreased by 'io_size'). Unless specific nrfiles and filesize options are given, fio will divide this size between the available files specified by the job. If not set, fio will use the full size of the given files or devices. If the files do not exist, size must be given. It is also possible to give size as a percentage between 1 and 100. If size=20% is given, fio will use 20% of the full size of the given files or devices. io_size=int io_limit=int Normally fio operates within the region set by 'size', which means that the 'size' option sets both the region and size of IO to
添加 Redis Server Group 准备配置文件 [root@h102 codis]# grep -v "^#" extern/redis-2.8.21/redis.conf | grep -v "^$" > /etc/codis/redis_s.conf [root@h102 codis]# grep -v "^#" extern/redis-2.8.21/redis.conf | grep -v "^$" > /etc/codis/redis2.conf [root@h102 codis
解决方法 : 安装 pcre.x86_64 和 pcre-devel.x86_64 软件包
分离数据库 无法分离 main 和 temp 数据库 sqlite> .databases seq name file --- --------------- ---------------------------------------------------------- 0 main /usr/local/sqlite3.11/bin/
stats查看状态statsSTAT pid 8917STAT uptime 1183STAT time 1442996635STAT version 1.4.24STAT libevent 1.4.13-stableSTAT pointer_size 64STAT rusage_user 0.014997STAT rusage_system 0.105983STAT curr_connections 12STAT total_connections 14STAT connection_structures
打开本地的 tcp udp 5000端口 标记类型为 syslog filter中作判断如果类型是 syslog 拆分解析信息 添加 received_at received_from 字段 使用 syslog_pri { } 来处理 定义 syslog_timestamp 的格式 输出到ES 以 rubydebug 的格式输出到终端 启动 Logstash [root@h102 etc]# /opt/logstash/bin/logstash -f logstash-syslog.conf Setti
刷新目录过期时间 [root@docker ~]# date +%s ;curl http://127.0.0.1:2379/v2/keys/testdir -XPUT -d ttl=10 -d dir "expiration":"2016-02-02T06:23:25.962070512Z","ttl":10,"modifiedIndex":85,"createdIndex":85}} [root@docker ,"expiration":"2016-02-02T06:23:25.962070512Z","ttl":6,"modifiedIndex":85,"createdIndex":85}} [root@docker ,"expiration":"2016-02-02T06:23:25.962070512Z","ttl":4,"modifiedIndex":85,"createdIndex":85}} [root@docker ,"expiration":"2016-02-02T06:23:32.096351254Z","ttl":8,"modifiedIndex":85,"createdIndex":85}} [root@docker
upstream 是nginx 负载均衡的主要模块,它提供了一个简单方法来轮询后端的服务器
error: /root/unrealircd/conf/unrealircd.conf:378: set::cloak-keys: (key 2) Keys should be mixed a-zA-Z0-9, error: /root/unrealircd/conf/unrealircd.conf:379: set::cloak-keys: (key 3) Keys should be mixed a-zA-Z0-9, IRCd configuration failed to pass testing 分别为: 1.要求修改操作员角色的用户名和密码 2.cloak-keys 必须得相互不一样,并且由 a-zA-Z0-9
日志信息 =INFO REPORT==== 23-Oct-2015::16:52:50 === Starting RabbitMQ 3.5.6 on Erlang R14B04 Copyright (C) 2007-2015 Pivotal Software, Inc. Licensed under the MPL. See http://www.rabbitmq.com/ =INFO REPORT==== 23-Oct-2015::16:52:50 === node : rabbi
# 7.传递参数 ENTRYPOINT ["pytest"] # 8.默认显示帮助信息 CMD ["--help"] Dockerfile配置详细解释 FROM:先拉取python3.6.8作为基础镜像 这个点代表上下文目录的路径,就是当前路径,我这里是/root/docker_run docker build -t jkc_pytest:v1 . [root@jkc docker-run]# docker build -t jkc_pytest:v1 . more-itertools>=4.0.0 Downloading https://pypi.doubanio.com/packages/66/58/63146600d8c35188f2c4de3f1e2fe77a54e9c9ac99ac3b652f09e042f637 ]# 运行过程中可以看到按步骤运行,如:Step 1/8 运行完成后,可以通过docker images查看生成的镜像 [root@jkc docker-run]# docker images REPOSITORY
我们还可以使用 DNS API 结合 tag 来过滤服务 [root@h104 ~]# dig @127.0.0.1 -p 8600 rails.web.service.consul ; <<>> DiG 9.9.4-RedHat-9.9.4-29.el7_2.1 <<>> @127.0.0.1 -p 8600 rails.web.service.consul ; (1 server found) ;; global options: +cmd ;; Got answer: ;; ->>HEADER<
启动 查看配置 [root@h102 conf]# grep -v "#" /usr/local/nginx/conf/nginx.conf | grep -v "^$" worker_processes 1; events { worker_connections 1024; } http { include mime.types; default_type application/octet-stream; sendfile on;
启动haproxy [root@h102 ~]# /usr/local/haproxy/sbin/haproxy -f /etc/haproxy/transparent_proxy.cfg ... ... ... ---------- [root@h102 ~]# netstat -ant | grep 1234 tcp 0 0 0.0.0.0:1234 0.0.0.0:* LISTEN [root@h1