主機名 | ip地址及磁盤 | 部署組件 |
---|---|---|
node01 | 192.168.1.181(三塊硬盤) | Mon、mgr、rgs、osd |
node02 | 192.168.1.182(三塊硬盤) | Mon、mgr、rgs、osd |
node03 | 192.168.1.183(三塊硬盤) | Mon、mgr、rgs、osd |
node0一、node0二、node03節點進行以下配置:node
#添加hosts解析; python
[root@node01 ~]# cat >/etc/hosts<<EOF 127.0.0.1 localhost localhost.localdomain 192.168.1.181 node01 192.168.1.182 node02 192.168.1.183 node03 EOF linux
#臨時關閉selinux和防火牆; web
[root@node01 ~]# sed -i '/SELINUX/s/enforcing/disabled/g' /etc/sysconfig/selinux json
[root@node01 ~]# setenforce 0bootstrap
[root@node01 ~]# systemctl stop firewalld.service api
[root@node01 ~]# systemctl disable firewalld.service babel
#同步節點時間less
[root@node01 ~]# yum install ntpdate -y dom
[root@node01 ~]# ntpdate pool.ntp.org
#配置epel yum源
[root@node01 ~]# yum install -y epel-release
#使用阿里開源鏡像提供的epel源
[root@node01 ~]# cd /etc/yum.repos.d
[root@node01 ~]# wget -O /etc/yum.repos.d/epel-7.repo http://mirrors.aliyun.com/repo/epel-7.repo
#配置ceph yum源
[root@node01 ~]# echo "[noarch]
name=ceph noarch
baseurl=https://mirrors.aliyun.com/ceph/rpm-nautilus/el7/noarch
enabled=1
gpgcheck=0
[x86_64]
name=ceph x86_64
baseurl=https://mirrors.aliyun.com/ceph/rpm-nautilus/el7/x86_64
enabled=1
gpgcheck=0" > /etc/yum.repos.d/ceph.repo
[root@node01 ~]# yum -y install ceph-deploy python-setuptools
[root@node01 ~]# yum install ceph-deploy python-setuptools Loaded plugins: fastestmirror, langpacks Repository epel is listed more than once in the configuration Repository epel-debuginfo is listed more than once in the configuration Repository epel-source is listed more than once in the configuration Loading mirror speeds from cached hostfile * base: mirrors.aliyun.com * extras: mirrors.aliyun.com * updates: mirror.bit.edu.cn base | 3.6 kB 00:00:00 epel | 4.7 kB 00:00:00 extras | 2.9 kB 00:00:00 noarch | 1.5 kB 00:00:00 updates | 2.9 kB 00:00:00 x86_64 | 1.5 kB 00:00:00 (1/3): epel/x86_64/group_gz | 95 kB 00:00:00 (2/3): epel/x86_64/updateinfo | 1.0 MB 00:00:00 (3/3): epel/x86_64/primary_db | 6.9 MB 00:00:01 Package python-setuptools-0.9.8-7.el7.noarch already installed and latest version Resolving Dependencies --> Running transaction check ---> Package ceph-deploy.noarch 0:2.0.1-0 will be installed --> Finished Dependency Resolution Dependencies Resolved ======================================================================================= Package Arch Version Repository Size ======================================================================================= Installing: ceph-deploy noarch 2.0.1-0 noarch 286 k Transaction Summary ======================================================================================= Install 1 Package Total download size: 286 k Installed size: 1.2 M Is this ok [y/d/N]: y Downloading packages: ceph-deploy-2.0.1-0.noarch.rpm | 286 kB 00:00:00 Running transaction check Running transaction test Transaction test succeeded Running transaction Installing : ceph-deploy-2.0.1-0.noarch 1/1 Verifying : ceph-deploy-2.0.1-0.noarch 1/1 Installed: ceph-deploy.noarch 0:2.0.1-0 Complete! [root@node01 ~]# [root@node01 ~]# rpm -qa|grep ceph-deploy ceph-deploy-2.0.1-0.noarch [root@node01 ~]# rpm -qa|grep python-setuptools python-setuptools-0.9.8-7.el7.noarch [root@node01 ~]#
[root@node01 ~]# for host in node{02..03};do scp -r /etc/yum.repos.d/* $host:/etc/yum.repos.d/ ;done
[root@node01 ~]# for host in node{02..03};do scp -r /etc/yum.repos.d/* $host:/etc/yum.repos.d;done The authenticity of host 'node02 (192.168.1.182)' can't be established. ECDSA key fingerprint is SHA256:V1WcZifyTKUtNITjoIv89siGUFa/aRKqhVbuSZo44IA. ECDSA key fingerprint is MD5:cb:08:52:c5:43:7c:fe:77:88:7e:cb:e3:4b:1b:26:ab. Are you sure you want to continue connecting (yes/no)? yes Warning: Permanently added 'node02,192.168.1.182' (ECDSA) to the list of known hosts. root@node02's password: Permission denied, please try again. root@node02's password: CentOS-Base.repo 100% 1664 2.1MB/s 00:00 CentOS-CR.repo 100% 1309 1.5MB/s 00:00 CentOS-Debuginfo.repo 100% 649 741.5KB/s 00:00 CentOS-fasttrack.repo 100% 314 562.5KB/s 00:00 CentOS-Media.repo 100% 630 1.0MB/s 00:00 CentOS-Sources.repo 100% 1331 2.3MB/s 00:00 CentOS-Vault.repo 100% 5701 8.8MB/s 00:00 ceph.repo 100% 231 492.4KB/s 00:00 epel-7.repo 100% 664 1.1MB/s 00:00 epel.repo 100% 951 2.0MB/s 00:00 epel-testing.repo 100% 1050 1.9MB/s 00:00 The authenticity of host 'node03 (192.168.1.183)' can't be established. ECDSA key fingerprint is SHA256:+J7KpBTku/CAcO+EjEyryEF0BuUaaDlKwxpimaJY0Qk. ECDSA key fingerprint is MD5:b0:88:a3:db:50:1a:bd:6b:6b:0a:cf:7a:a2:81:38:c7. Are you sure you want to continue connecting (yes/no)? yes Warning: Permanently added 'node03,192.168.1.183' (ECDSA) to the list of known hosts. root@node03's password: CentOS-Base.repo 100% 1664 2.0MB/s 00:00 CentOS-CR.repo 100% 1309 1.4MB/s 00:00 CentOS-Debuginfo.repo 100% 649 555.0KB/s 00:00 CentOS-fasttrack.repo 100% 314 338.7KB/s 00:00 CentOS-Media.repo 100% 630 892.6KB/s 00:00 CentOS-Sources.repo 100% 1331 2.1MB/s 00:00 CentOS-Vault.repo 100% 5701 7.3MB/s 00:00 ceph.repo 100% 231 325.5KB/s 00:00 epel-7.repo 100% 664 1.0MB/s 00:00 epel.repo 100% 951 1.6MB/s 00:00 epel-testing.repo 100% 1050 1.9MB/s 00:00 [root@node01 ~]#
[root@node01 ~]# for host in node{01..03};do ssh -l root $host yum install ceph ceph-radosgw -y ;done
[root@node01 ~]# ceph -v
[root@node01 ~]# rpm -qa |grep ceph
[root@node01 ~]# for host in node{01..03};do ssh -l root $host yum install ceph ceph-radosgw -y;done The authenticity of host 'node01 (192.168.1.181)' can't be established. ECDSA key fingerprint is SHA256:QfJ3r+TUqEFfG/D3a8YYN5twm0IT1vCI5OtIm28GX+M. ECDSA key fingerprint is MD5:1f:36:26:8d:d8:63:81:9f:d6:30:a2:ed:ba:fb:db:ba. Are you sure you want to continue connecting (yes/no)? yes Warning: Permanently added 'node01,192.168.1.181' (ECDSA) to the list of known hosts. root@node01's password: Loaded plugins: fastestmirror, langpacks Loading mirror speeds from cached hostfile Repository epel is listed more than once in the configuration Repository epel-debuginfo is listed more than once in the configuration Repository epel-source is listed more than once in the configuration * base: mirrors.aliyun.com * extras: mirrors.aliyun.com * updates: mirror.bit.edu.cn Resolving Dependencies --> Running transaction check ---> Package ceph.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: ceph-mon = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-osd = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mds = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mgr = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 ---> Package ceph-radosgw.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: librgw2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-selinux = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-base = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados.so.2(LIBRADOS_14.2.0)(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: liboath.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: libceph-common.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librabbitmq.so.4()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Running transaction check ---> Package ceph-base.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: libcephfs2 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: librbd1 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-common = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: libleveldb.so.1()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: liblttng-ust.so.0()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 ---> Package ceph-mds.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-mgr.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-pecan for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cherrypy for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-werkzeug for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-bcrypt for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 ---> Package ceph-mon.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-osd.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-selinux.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: selinux-policy-base >= 3.13.1-252.el7_7.6 for package: 2:ceph-selinux-14.2.12-0.el7.x86_64 ---> Package liboath.x86_64 0:2.6.2-1.el7 will be installed ---> Package librabbitmq.x86_64 0:0.8.0-2.el7 will be installed ---> Package librados2.x86_64 1:10.2.5-4.el7 will be updated ---> Package librados2.x86_64 2:14.2.12-0.el7 will be an update ---> Package librgw2.x86_64 2:14.2.12-0.el7 will be installed --> Running transaction check ---> Package ceph-common.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-rgw = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper1 = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-ceph-argparse = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rbd = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cephfs = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rados = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-prettytable for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace-ctf.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: librdkafka.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 ---> Package leveldb.x86_64 0:1.12.0-11.el7 will be installed ---> Package libcephfs2.x86_64 2:14.2.12-0.el7 will be installed ---> Package librbd1.x86_64 1:10.2.5-4.el7 will be updated ---> Package librbd1.x86_64 2:14.2.12-0.el7 will be an update ---> Package lttng-ust.x86_64 0:2.4.1-4.el7 will be installed --> Processing Dependency: liburcu-bp.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 --> Processing Dependency: liburcu-cds.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 ---> Package python-cherrypy.noarch 0:3.2.2-4.el7 will be installed ---> Package python-pecan.noarch 0:0.4.5-2.el7 will be installed --> Processing Dependency: python-mako >= 0.4.0 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-simplegeneric >= 0.8 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webob >= 1.2 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webtest >= 1.3.1 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-singledispatch for package: python-pecan-0.4.5-2.el7.noarch ---> Package python-werkzeug.noarch 0:0.9.1-2.el7 will be installed ---> Package python2-bcrypt.x86_64 0:3.1.6-2.el7 will be installed --> Processing Dependency: python2-six for package: python2-bcrypt-3.1.6-2.el7.x86_64 ---> Package selinux-policy-targeted.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 will be an update --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Running transaction check ---> Package libbabeltrace.x86_64 0:1.2.4-3.el7 will be installed ---> Package libradosstriper1.x86_64 2:14.2.12-0.el7 will be installed ---> Package librdkafka.x86_64 0:0.11.5-1.el7 will be installed ---> Package python-ceph-argparse.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-cephfs.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-mako.noarch 0:0.8.1-2.el7 will be installed --> Processing Dependency: python-markupsafe for package: python-mako-0.8.1-2.el7.noarch --> Processing Dependency: python-beaker for package: python-mako-0.8.1-2.el7.noarch ---> Package python-prettytable.noarch 0:0.7.2-3.el7 will be installed ---> Package python-rados.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rbd.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rgw.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-simplegeneric.noarch 0:0.8-7.el7 will be installed ---> Package python-singledispatch.noarch 0:3.4.0.2-2.el7 will be installed ---> Package python-webob.noarch 0:1.2.3-7.el7 will be installed ---> Package python-webtest.noarch 0:1.3.4-6.el7 will be installed ---> Package python2-six.noarch 0:1.9.0-0.el7 will be installed ---> Package selinux-policy.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy.noarch 0:3.13.1-266.el7_8.1 will be an update ---> Package userspace-rcu.x86_64 0:0.7.16-1.el7 will be installed --> Running transaction check ---> Package python-beaker.noarch 0:1.5.4-10.el7 will be installed --> Processing Dependency: python-paste for package: python-beaker-1.5.4-10.el7.noarch ---> Package python-markupsafe.x86_64 0:0.11-10.el7 will be installed --> Running transaction check ---> Package python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 will be installed --> Processing Dependency: python-tempita for package: python-paste-1.7.5.1-9.20111221hg1498.el7.noarch --> Running transaction check ---> Package python-tempita.noarch 0:0.5.1-6.el7 will be installed --> Finished Dependency Resolution Dependencies Resolved ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: ceph x86_64 2:14.2.12-0.el7 x86_64 3.0 k ceph-radosgw x86_64 2:14.2.12-0.el7 x86_64 5.8 M Installing for dependencies: ceph-base x86_64 2:14.2.12-0.el7 x86_64 5.4 M ceph-common x86_64 2:14.2.12-0.el7 x86_64 19 M ceph-mds x86_64 2:14.2.12-0.el7 x86_64 1.9 M ceph-mgr x86_64 2:14.2.12-0.el7 x86_64 1.7 M ceph-mon x86_64 2:14.2.12-0.el7 x86_64 4.3 M ceph-osd x86_64 2:14.2.12-0.el7 x86_64 16 M ceph-selinux x86_64 2:14.2.12-0.el7 x86_64 21 k leveldb x86_64 1.12.0-11.el7 epel 161 k libbabeltrace x86_64 1.2.4-3.el7 epel 147 k libcephfs2 x86_64 2:14.2.12-0.el7 x86_64 500 k liboath x86_64 2.6.2-1.el7 epel 51 k librabbitmq x86_64 0.8.0-2.el7 base 37 k libradosstriper1 x86_64 2:14.2.12-0.el7 x86_64 362 k librdkafka x86_64 0.11.5-1.el7 epel 368 k librgw2 x86_64 2:14.2.12-0.el7 x86_64 5.1 M lttng-ust x86_64 2.4.1-4.el7 epel 176 k python-beaker noarch 1.5.4-10.el7 base 80 k python-ceph-argparse x86_64 2:14.2.12-0.el7 x86_64 36 k python-cephfs x86_64 2:14.2.12-0.el7 x86_64 112 k python-cherrypy noarch 3.2.2-4.el7 base 422 k python-mako noarch 0.8.1-2.el7 base 307 k python-markupsafe x86_64 0.11-10.el7 base 25 k python-paste noarch 1.7.5.1-9.20111221hg1498.el7 base 866 k python-pecan noarch 0.4.5-2.el7 epel 255 k python-prettytable noarch 0.7.2-3.el7 base 37 k python-rados x86_64 2:14.2.12-0.el7 x86_64 200 k python-rbd x86_64 2:14.2.12-0.el7 x86_64 195 k python-rgw x86_64 2:14.2.12-0.el7 x86_64 78 k python-simplegeneric noarch 0.8-7.el7 epel 12 k python-singledispatch noarch 3.4.0.2-2.el7 epel 18 k python-tempita noarch 0.5.1-6.el7 base 33 k python-webob noarch 1.2.3-7.el7 base 202 k python-webtest noarch 1.3.4-6.el7 base 102 k python-werkzeug noarch 0.9.1-2.el7 extras 562 k python2-bcrypt x86_64 3.1.6-2.el7 epel 39 k python2-six noarch 1.9.0-0.el7 epel 2.9 k userspace-rcu x86_64 0.7.16-1.el7 epel 73 k Updating for dependencies: librados2 x86_64 2:14.2.12-0.el7 x86_64 3.4 M librbd1 x86_64 2:14.2.12-0.el7 x86_64 1.7 M selinux-policy noarch 3.13.1-266.el7_8.1 updates 497 k selinux-policy-targeted noarch 3.13.1-266.el7_8.1 updates 7.0 M Transaction Summary ================================================================================ Install 2 Packages (+37 Dependent packages) Upgrade ( 4 Dependent packages) Total size: 77 M Total download size: 70 M Downloading packages: No Presto metadata available for x86_64 -------------------------------------------------------------------------------- Total 4.3 MB/s | 70 MB 00:16 Running transaction check Running transaction test Transaction test succeeded Running transaction Installing : leveldb-1.12.0-11.el7.x86_64 1/47 Installing : liboath-2.6.2-1.el7.x86_64 2/47 Installing : librabbitmq-0.8.0-2.el7.x86_64 3/47 Installing : python-webob-1.2.3-7.el7.noarch 4/47 Installing : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 5/47 Installing : python-webtest-1.3.4-6.el7.noarch 6/47 Installing : python-tempita-0.5.1-6.el7.noarch 7/47 Installing : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 8/47 Installing : python-beaker-1.5.4-10.el7.noarch 9/47 Installing : python-cherrypy-3.2.2-4.el7.noarch 10/47 Installing : python-werkzeug-0.9.1-2.el7.noarch 11/47 Installing : librdkafka-0.11.5-1.el7.x86_64 12/47 Installing : python-markupsafe-0.11-10.el7.x86_64 13/47 Installing : python-mako-0.8.1-2.el7.noarch 14/47 Installing : python-singledispatch-3.4.0.2-2.el7.noarch 15/47 Updating : selinux-policy-3.13.1-266.el7_8.1.noarch 16/47 Updating : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 17/47 Installing : libbabeltrace-1.2.4-3.el7.x86_64 18/47 Installing : python-simplegeneric-0.8-7.el7.noarch 19/47 Installing : python-pecan-0.4.5-2.el7.noarch 20/47 Installing : userspace-rcu-0.7.16-1.el7.x86_64 21/47 Installing : lttng-ust-2.4.1-4.el7.x86_64 22/47 Updating : 2:librados2-14.2.12-0.el7.x86_64 23/47 Installing : 2:python-rados-14.2.12-0.el7.x86_64 24/47 Installing : 2:librgw2-14.2.12-0.el7.x86_64 25/47 Installing : 2:libcephfs2-14.2.12-0.el7.x86_64 26/47 Updating : 2:librbd1-14.2.12-0.el7.x86_64 27/47 Installing : 2:python-rbd-14.2.12-0.el7.x86_64 28/47 Installing : 2:python-cephfs-14.2.12-0.el7.x86_64 29/47 Installing : 2:python-rgw-14.2.12-0.el7.x86_64 30/47 Installing : 2:libradosstriper1-14.2.12-0.el7.x86_64 31/47 Installing : python-prettytable-0.7.2-3.el7.noarch 32/47 Installing : 2:ceph-common-14.2.12-0.el7.x86_64 33/47 Installing : 2:ceph-base-14.2.12-0.el7.x86_64 34/47 Installing : 2:ceph-selinux-14.2.12-0.el7.x86_64 35/47 Installing : 2:ceph-mon-14.2.12-0.el7.x86_64 36/47 Installing : 2:ceph-osd-14.2.12-0.el7.x86_64 37/47 Installing : 2:ceph-mds-14.2.12-0.el7.x86_64 38/47 Installing : python2-six-1.9.0-0.el7.noarch 39/47 Installing : python2-bcrypt-3.1.6-2.el7.x86_64 40/47 Installing : 2:ceph-mgr-14.2.12-0.el7.x86_64 41/47 Installing : 2:ceph-14.2.12-0.el7.x86_64 42/47 Installing : 2:ceph-radosgw-14.2.12-0.el7.x86_64 43/47 Cleanup : selinux-policy-targeted-3.13.1-229.el7.noarch 44/47 Cleanup : 1:librbd1-10.2.5-4.el7.x86_64 45/47 Cleanup : selinux-policy-3.13.1-229.el7.noarch 46/47 Cleanup : 1:librados2-10.2.5-4.el7.x86_64 47/47 Verifying : liboath-2.6.2-1.el7.x86_64 1/47 Verifying : python2-six-1.9.0-0.el7.noarch 2/47 Verifying : 2:python-rbd-14.2.12-0.el7.x86_64 3/47 Verifying : python-prettytable-0.7.2-3.el7.noarch 4/47 Verifying : leveldb-1.12.0-11.el7.x86_64 5/47 Verifying : python2-bcrypt-3.1.6-2.el7.x86_64 6/47 Verifying : userspace-rcu-0.7.16-1.el7.x86_64 7/47 Verifying : python-simplegeneric-0.8-7.el7.noarch 8/47 Verifying : 2:python-cephfs-14.2.12-0.el7.x86_64 9/47 Verifying : libbabeltrace-1.2.4-3.el7.x86_64 10/47 Verifying : python-mako-0.8.1-2.el7.noarch 11/47 Verifying : 2:ceph-selinux-14.2.12-0.el7.x86_64 12/47 Verifying : selinux-policy-3.13.1-266.el7_8.1.noarch 13/47 Verifying : 2:librgw2-14.2.12-0.el7.x86_64 14/47 Verifying : 2:ceph-mon-14.2.12-0.el7.x86_64 15/47 Verifying : python-pecan-0.4.5-2.el7.noarch 16/47 Verifying : 2:ceph-radosgw-14.2.12-0.el7.x86_64 17/47 Verifying : python-singledispatch-3.4.0.2-2.el7.noarch 18/47 Verifying : python-markupsafe-0.11-10.el7.x86_64 19/47 Verifying : 2:ceph-common-14.2.12-0.el7.x86_64 20/47 Verifying : 2:libcephfs2-14.2.12-0.el7.x86_64 21/47 Verifying : librdkafka-0.11.5-1.el7.x86_64 22/47 Verifying : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 23/47 Verifying : lttng-ust-2.4.1-4.el7.x86_64 24/47 Verifying : 2:ceph-osd-14.2.12-0.el7.x86_64 25/47 Verifying : python-werkzeug-0.9.1-2.el7.noarch 26/47 Verifying : 2:python-rados-14.2.12-0.el7.x86_64 27/47 Verifying : 2:libradosstriper1-14.2.12-0.el7.x86_64 28/47 Verifying : 2:ceph-base-14.2.12-0.el7.x86_64 29/47 Verifying : 2:ceph-mgr-14.2.12-0.el7.x86_64 30/47 Verifying : python-cherrypy-3.2.2-4.el7.noarch 31/47 Verifying : python-beaker-1.5.4-10.el7.noarch 32/47 Verifying : 2:python-rgw-14.2.12-0.el7.x86_64 33/47 Verifying : 2:librados2-14.2.12-0.el7.x86_64 34/47 Verifying : 2:ceph-mds-14.2.12-0.el7.x86_64 35/47 Verifying : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 36/47 Verifying : python-tempita-0.5.1-6.el7.noarch 37/47 Verifying : python-webtest-1.3.4-6.el7.noarch 38/47 Verifying : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 39/47 Verifying : librabbitmq-0.8.0-2.el7.x86_64 40/47 Verifying : 2:librbd1-14.2.12-0.el7.x86_64 41/47 Verifying : python-webob-1.2.3-7.el7.noarch 42/47 Verifying : 2:ceph-14.2.12-0.el7.x86_64 43/47 Verifying : 1:librbd1-10.2.5-4.el7.x86_64 44/47 Verifying : selinux-policy-3.13.1-229.el7.noarch 45/47 Verifying : selinux-policy-targeted-3.13.1-229.el7.noarch 46/47 Verifying : 1:librados2-10.2.5-4.el7.x86_64 47/47 Installed: ceph.x86_64 2:14.2.12-0.el7 ceph-radosgw.x86_64 2:14.2.12-0.el7 Dependency Installed: ceph-base.x86_64 2:14.2.12-0.el7 ceph-common.x86_64 2:14.2.12-0.el7 ceph-mds.x86_64 2:14.2.12-0.el7 ceph-mgr.x86_64 2:14.2.12-0.el7 ceph-mon.x86_64 2:14.2.12-0.el7 ceph-osd.x86_64 2:14.2.12-0.el7 ceph-selinux.x86_64 2:14.2.12-0.el7 leveldb.x86_64 0:1.12.0-11.el7 libbabeltrace.x86_64 0:1.2.4-3.el7 libcephfs2.x86_64 2:14.2.12-0.el7 liboath.x86_64 0:2.6.2-1.el7 librabbitmq.x86_64 0:0.8.0-2.el7 libradosstriper1.x86_64 2:14.2.12-0.el7 librdkafka.x86_64 0:0.11.5-1.el7 librgw2.x86_64 2:14.2.12-0.el7 lttng-ust.x86_64 0:2.4.1-4.el7 python-beaker.noarch 0:1.5.4-10.el7 python-ceph-argparse.x86_64 2:14.2.12-0.el7 python-cephfs.x86_64 2:14.2.12-0.el7 python-cherrypy.noarch 0:3.2.2-4.el7 python-mako.noarch 0:0.8.1-2.el7 python-markupsafe.x86_64 0:0.11-10.el7 python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 python-pecan.noarch 0:0.4.5-2.el7 python-prettytable.noarch 0:0.7.2-3.el7 python-rados.x86_64 2:14.2.12-0.el7 python-rbd.x86_64 2:14.2.12-0.el7 python-rgw.x86_64 2:14.2.12-0.el7 python-simplegeneric.noarch 0:0.8-7.el7 python-singledispatch.noarch 0:3.4.0.2-2.el7 python-tempita.noarch 0:0.5.1-6.el7 python-webob.noarch 0:1.2.3-7.el7 python-webtest.noarch 0:1.3.4-6.el7 python-werkzeug.noarch 0:0.9.1-2.el7 python2-bcrypt.x86_64 0:3.1.6-2.el7 python2-six.noarch 0:1.9.0-0.el7 userspace-rcu.x86_64 0:0.7.16-1.el7 Dependency Updated: librados2.x86_64 2:14.2.12-0.el7 librbd1.x86_64 2:14.2.12-0.el7 selinux-policy.noarch 0:3.13.1-266.el7_8.1 selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 Complete! root@node02's password: Loaded plugins: fastestmirror, langpacks, product-id, search-disabled-repos, : subscription-manager This system is not registered with an entitlement server. You can use subscription-manager to register. Repository epel is listed more than once in the configuration Repository epel-debuginfo is listed more than once in the configuration Repository epel-source is listed more than once in the configuration Loading mirror speeds from cached hostfile * base: mirrors.aliyun.com * extras: mirrors.bfsu.edu.cn * updates: mirror.bit.edu.cn Resolving Dependencies --> Running transaction check ---> Package ceph.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: ceph-mon = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-osd = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mds = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mgr = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 ---> Package ceph-radosgw.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: librgw2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-selinux = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-base = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados.so.2(LIBRADOS_14.2.0)(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: liboath.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: libceph-common.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librabbitmq.so.4()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Running transaction check ---> Package ceph-base.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: libcephfs2 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: librbd1 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-common = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: libleveldb.so.1()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: liblttng-ust.so.0()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 ---> Package ceph-mds.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-mgr.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-pecan for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cherrypy for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-werkzeug for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-bcrypt for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 ---> Package ceph-mon.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-osd.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-selinux.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: selinux-policy-base >= 3.13.1-252.el7_7.6 for package: 2:ceph-selinux-14.2.12-0.el7.x86_64 ---> Package liboath.x86_64 0:2.6.2-1.el7 will be installed ---> Package librabbitmq.x86_64 0:0.8.0-2.el7 will be installed ---> Package librados2.x86_64 1:10.2.5-4.el7 will be updated ---> Package librados2.x86_64 2:14.2.12-0.el7 will be an update ---> Package librgw2.x86_64 2:14.2.12-0.el7 will be installed --> Running transaction check ---> Package ceph-common.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-rgw = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper1 = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-ceph-argparse = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rbd = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cephfs = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rados = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-prettytable for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace-ctf.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: librdkafka.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 ---> Package leveldb.x86_64 0:1.12.0-11.el7 will be installed ---> Package libcephfs2.x86_64 2:14.2.12-0.el7 will be installed ---> Package librbd1.x86_64 1:10.2.5-4.el7 will be updated ---> Package librbd1.x86_64 2:14.2.12-0.el7 will be an update ---> Package lttng-ust.x86_64 0:2.4.1-4.el7 will be installed --> Processing Dependency: liburcu-bp.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 --> Processing Dependency: liburcu-cds.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 ---> Package python-cherrypy.noarch 0:3.2.2-4.el7 will be installed ---> Package python-pecan.noarch 0:0.4.5-2.el7 will be installed --> Processing Dependency: python-mako >= 0.4.0 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-simplegeneric >= 0.8 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webob >= 1.2 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webtest >= 1.3.1 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-singledispatch for package: python-pecan-0.4.5-2.el7.noarch ---> Package python-werkzeug.noarch 0:0.9.1-2.el7 will be installed ---> Package python2-bcrypt.x86_64 0:3.1.6-2.el7 will be installed --> Processing Dependency: python2-six for package: python2-bcrypt-3.1.6-2.el7.x86_64 ---> Package selinux-policy-targeted.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 will be an update --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Running transaction check ---> Package libbabeltrace.x86_64 0:1.2.4-3.el7 will be installed ---> Package libradosstriper1.x86_64 2:14.2.12-0.el7 will be installed ---> Package librdkafka.x86_64 0:0.11.5-1.el7 will be installed ---> Package python-ceph-argparse.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-cephfs.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-mako.noarch 0:0.8.1-2.el7 will be installed --> Processing Dependency: python-markupsafe for package: python-mako-0.8.1-2.el7.noarch --> Processing Dependency: python-beaker for package: python-mako-0.8.1-2.el7.noarch ---> Package python-prettytable.noarch 0:0.7.2-3.el7 will be installed ---> Package python-rados.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rbd.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rgw.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-simplegeneric.noarch 0:0.8-7.el7 will be installed ---> Package python-singledispatch.noarch 0:3.4.0.2-2.el7 will be installed ---> Package python-webob.noarch 0:1.2.3-7.el7 will be installed ---> Package python-webtest.noarch 0:1.3.4-6.el7 will be installed ---> Package python2-six.noarch 0:1.9.0-0.el7 will be installed ---> Package selinux-policy.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy.noarch 0:3.13.1-266.el7_8.1 will be an update ---> Package userspace-rcu.x86_64 0:0.7.16-1.el7 will be installed --> Running transaction check ---> Package python-beaker.noarch 0:1.5.4-10.el7 will be installed --> Processing Dependency: python-paste for package: python-beaker-1.5.4-10.el7.noarch ---> Package python-markupsafe.x86_64 0:0.11-10.el7 will be installed --> Running transaction check ---> Package python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 will be installed --> Processing Dependency: python-tempita for package: python-paste-1.7.5.1-9.20111221hg1498.el7.noarch --> Running transaction check ---> Package python-tempita.noarch 0:0.5.1-6.el7 will be installed --> Finished Dependency Resolution Dependencies Resolved ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: ceph x86_64 2:14.2.12-0.el7 x86_64 3.0 k ceph-radosgw x86_64 2:14.2.12-0.el7 x86_64 5.8 M Installing for dependencies: ceph-base x86_64 2:14.2.12-0.el7 x86_64 5.4 M ceph-common x86_64 2:14.2.12-0.el7 x86_64 19 M ceph-mds x86_64 2:14.2.12-0.el7 x86_64 1.9 M ceph-mgr x86_64 2:14.2.12-0.el7 x86_64 1.7 M ceph-mon x86_64 2:14.2.12-0.el7 x86_64 4.3 M ceph-osd x86_64 2:14.2.12-0.el7 x86_64 16 M ceph-selinux x86_64 2:14.2.12-0.el7 x86_64 21 k leveldb x86_64 1.12.0-11.el7 epel 161 k libbabeltrace x86_64 1.2.4-3.el7 epel 147 k libcephfs2 x86_64 2:14.2.12-0.el7 x86_64 500 k liboath x86_64 2.6.2-1.el7 epel 51 k librabbitmq x86_64 0.8.0-2.el7 base 37 k libradosstriper1 x86_64 2:14.2.12-0.el7 x86_64 362 k librdkafka x86_64 0.11.5-1.el7 epel 368 k librgw2 x86_64 2:14.2.12-0.el7 x86_64 5.1 M lttng-ust x86_64 2.4.1-4.el7 epel 176 k python-beaker noarch 1.5.4-10.el7 base 80 k python-ceph-argparse x86_64 2:14.2.12-0.el7 x86_64 36 k python-cephfs x86_64 2:14.2.12-0.el7 x86_64 112 k python-cherrypy noarch 3.2.2-4.el7 base 422 k python-mako noarch 0.8.1-2.el7 base 307 k python-markupsafe x86_64 0.11-10.el7 base 25 k python-paste noarch 1.7.5.1-9.20111221hg1498.el7 base 866 k python-pecan noarch 0.4.5-2.el7 epel 255 k python-prettytable noarch 0.7.2-3.el7 base 37 k python-rados x86_64 2:14.2.12-0.el7 x86_64 200 k python-rbd x86_64 2:14.2.12-0.el7 x86_64 195 k python-rgw x86_64 2:14.2.12-0.el7 x86_64 78 k python-simplegeneric noarch 0.8-7.el7 epel 12 k python-singledispatch noarch 3.4.0.2-2.el7 epel 18 k python-tempita noarch 0.5.1-6.el7 base 33 k python-webob noarch 1.2.3-7.el7 base 202 k python-webtest noarch 1.3.4-6.el7 base 102 k python-werkzeug noarch 0.9.1-2.el7 extras 562 k python2-bcrypt x86_64 3.1.6-2.el7 epel 39 k python2-six noarch 1.9.0-0.el7 epel 2.9 k userspace-rcu x86_64 0.7.16-1.el7 epel 73 k Updating for dependencies: librados2 x86_64 2:14.2.12-0.el7 x86_64 3.4 M librbd1 x86_64 2:14.2.12-0.el7 x86_64 1.7 M selinux-policy noarch 3.13.1-266.el7_8.1 updates 497 k selinux-policy-targeted noarch 3.13.1-266.el7_8.1 updates 7.0 M Transaction Summary ================================================================================ Install 2 Packages (+37 Dependent packages) Upgrade ( 4 Dependent packages) Total download size: 77 M Downloading packages: No Presto metadata available for x86_64 No Presto metadata available for updates -------------------------------------------------------------------------------- Total 6.2 MB/s | 77 MB 00:12 Running transaction check Running transaction test Transaction test succeeded Running transaction Installing : leveldb-1.12.0-11.el7.x86_64 1/47 Installing : liboath-2.6.2-1.el7.x86_64 2/47 Installing : librabbitmq-0.8.0-2.el7.x86_64 3/47 Installing : python-webob-1.2.3-7.el7.noarch 4/47 Installing : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 5/47 Installing : python-webtest-1.3.4-6.el7.noarch 6/47 Installing : python-tempita-0.5.1-6.el7.noarch 7/47 Installing : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 8/47 Installing : python-beaker-1.5.4-10.el7.noarch 9/47 Installing : python-cherrypy-3.2.2-4.el7.noarch 10/47 Installing : python-werkzeug-0.9.1-2.el7.noarch 11/47 Installing : librdkafka-0.11.5-1.el7.x86_64 12/47 Installing : python-markupsafe-0.11-10.el7.x86_64 13/47 Installing : python-mako-0.8.1-2.el7.noarch 14/47 Installing : python-singledispatch-3.4.0.2-2.el7.noarch 15/47 Updating : selinux-policy-3.13.1-266.el7_8.1.noarch 16/47 Updating : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 17/47 Installing : libbabeltrace-1.2.4-3.el7.x86_64 18/47 Installing : python-simplegeneric-0.8-7.el7.noarch 19/47 Installing : python-pecan-0.4.5-2.el7.noarch 20/47 Installing : userspace-rcu-0.7.16-1.el7.x86_64 21/47 Installing : lttng-ust-2.4.1-4.el7.x86_64 22/47 Updating : 2:librados2-14.2.12-0.el7.x86_64 23/47 Installing : 2:python-rados-14.2.12-0.el7.x86_64 24/47 Installing : 2:librgw2-14.2.12-0.el7.x86_64 25/47 Installing : 2:libcephfs2-14.2.12-0.el7.x86_64 26/47 Updating : 2:librbd1-14.2.12-0.el7.x86_64 27/47 Installing : 2:python-rbd-14.2.12-0.el7.x86_64 28/47 Installing : 2:python-cephfs-14.2.12-0.el7.x86_64 29/47 Installing : 2:python-rgw-14.2.12-0.el7.x86_64 30/47 Installing : 2:libradosstriper1-14.2.12-0.el7.x86_64 31/47 Installing : python-prettytable-0.7.2-3.el7.noarch 32/47 Installing : 2:ceph-common-14.2.12-0.el7.x86_64 33/47 Installing : 2:ceph-base-14.2.12-0.el7.x86_64 34/47 Installing : 2:ceph-selinux-14.2.12-0.el7.x86_64 35/47 Installing : 2:ceph-mon-14.2.12-0.el7.x86_64 36/47 Installing : 2:ceph-osd-14.2.12-0.el7.x86_64 37/47 Installing : 2:ceph-mds-14.2.12-0.el7.x86_64 38/47 Installing : python2-six-1.9.0-0.el7.noarch 39/47 Installing : python2-bcrypt-3.1.6-2.el7.x86_64 40/47 Installing : 2:ceph-mgr-14.2.12-0.el7.x86_64 41/47 Installing : 2:ceph-14.2.12-0.el7.x86_64 42/47 Installing : 2:ceph-radosgw-14.2.12-0.el7.x86_64 43/47 Cleanup : selinux-policy-targeted-3.13.1-229.el7.noarch 44/47 Cleanup : 1:librbd1-10.2.5-4.el7.x86_64 45/47 Cleanup : selinux-policy-3.13.1-229.el7.noarch 46/47 Cleanup : 1:librados2-10.2.5-4.el7.x86_64 47/47 Verifying : liboath-2.6.2-1.el7.x86_64 1/47 Verifying : python2-six-1.9.0-0.el7.noarch 2/47 Verifying : 2:python-rbd-14.2.12-0.el7.x86_64 3/47 Verifying : python-prettytable-0.7.2-3.el7.noarch 4/47 Verifying : leveldb-1.12.0-11.el7.x86_64 5/47 Verifying : python2-bcrypt-3.1.6-2.el7.x86_64 6/47 Verifying : userspace-rcu-0.7.16-1.el7.x86_64 7/47 Verifying : python-simplegeneric-0.8-7.el7.noarch 8/47 Verifying : 2:python-cephfs-14.2.12-0.el7.x86_64 9/47 Verifying : libbabeltrace-1.2.4-3.el7.x86_64 10/47 Verifying : python-mako-0.8.1-2.el7.noarch 11/47 Verifying : 2:ceph-selinux-14.2.12-0.el7.x86_64 12/47 Verifying : selinux-policy-3.13.1-266.el7_8.1.noarch 13/47 Verifying : 2:librgw2-14.2.12-0.el7.x86_64 14/47 Verifying : 2:ceph-mon-14.2.12-0.el7.x86_64 15/47 Verifying : python-pecan-0.4.5-2.el7.noarch 16/47 Verifying : 2:ceph-radosgw-14.2.12-0.el7.x86_64 17/47 Verifying : python-singledispatch-3.4.0.2-2.el7.noarch 18/47 Verifying : python-markupsafe-0.11-10.el7.x86_64 19/47 Verifying : 2:ceph-common-14.2.12-0.el7.x86_64 20/47 Verifying : 2:libcephfs2-14.2.12-0.el7.x86_64 21/47 Verifying : librdkafka-0.11.5-1.el7.x86_64 22/47 Verifying : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 23/47 Verifying : lttng-ust-2.4.1-4.el7.x86_64 24/47 Verifying : 2:ceph-osd-14.2.12-0.el7.x86_64 25/47 Verifying : python-werkzeug-0.9.1-2.el7.noarch 26/47 Verifying : 2:python-rados-14.2.12-0.el7.x86_64 27/47 Verifying : 2:libradosstriper1-14.2.12-0.el7.x86_64 28/47 Verifying : 2:ceph-base-14.2.12-0.el7.x86_64 29/47 Verifying : 2:ceph-mgr-14.2.12-0.el7.x86_64 30/47 Verifying : python-cherrypy-3.2.2-4.el7.noarch 31/47 Verifying : python-beaker-1.5.4-10.el7.noarch 32/47 Verifying : 2:python-rgw-14.2.12-0.el7.x86_64 33/47 Verifying : 2:librados2-14.2.12-0.el7.x86_64 34/47 Verifying : 2:ceph-mds-14.2.12-0.el7.x86_64 35/47 Verifying : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 36/47 Verifying : python-tempita-0.5.1-6.el7.noarch 37/47 Verifying : python-webtest-1.3.4-6.el7.noarch 38/47 Verifying : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 39/47 Verifying : librabbitmq-0.8.0-2.el7.x86_64 40/47 Verifying : 2:librbd1-14.2.12-0.el7.x86_64 41/47 Verifying : python-webob-1.2.3-7.el7.noarch 42/47 Verifying : 2:ceph-14.2.12-0.el7.x86_64 43/47 Verifying : 1:librbd1-10.2.5-4.el7.x86_64 44/47 Verifying : selinux-policy-3.13.1-229.el7.noarch 45/47 Verifying : selinux-policy-targeted-3.13.1-229.el7.noarch 46/47 Verifying : 1:librados2-10.2.5-4.el7.x86_64 47/47 Installed: ceph.x86_64 2:14.2.12-0.el7 ceph-radosgw.x86_64 2:14.2.12-0.el7 Dependency Installed: ceph-base.x86_64 2:14.2.12-0.el7 ceph-common.x86_64 2:14.2.12-0.el7 ceph-mds.x86_64 2:14.2.12-0.el7 ceph-mgr.x86_64 2:14.2.12-0.el7 ceph-mon.x86_64 2:14.2.12-0.el7 ceph-osd.x86_64 2:14.2.12-0.el7 ceph-selinux.x86_64 2:14.2.12-0.el7 leveldb.x86_64 0:1.12.0-11.el7 libbabeltrace.x86_64 0:1.2.4-3.el7 libcephfs2.x86_64 2:14.2.12-0.el7 liboath.x86_64 0:2.6.2-1.el7 librabbitmq.x86_64 0:0.8.0-2.el7 libradosstriper1.x86_64 2:14.2.12-0.el7 librdkafka.x86_64 0:0.11.5-1.el7 librgw2.x86_64 2:14.2.12-0.el7 lttng-ust.x86_64 0:2.4.1-4.el7 python-beaker.noarch 0:1.5.4-10.el7 python-ceph-argparse.x86_64 2:14.2.12-0.el7 python-cephfs.x86_64 2:14.2.12-0.el7 python-cherrypy.noarch 0:3.2.2-4.el7 python-mako.noarch 0:0.8.1-2.el7 python-markupsafe.x86_64 0:0.11-10.el7 python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 python-pecan.noarch 0:0.4.5-2.el7 python-prettytable.noarch 0:0.7.2-3.el7 python-rados.x86_64 2:14.2.12-0.el7 python-rbd.x86_64 2:14.2.12-0.el7 python-rgw.x86_64 2:14.2.12-0.el7 python-simplegeneric.noarch 0:0.8-7.el7 python-singledispatch.noarch 0:3.4.0.2-2.el7 python-tempita.noarch 0:0.5.1-6.el7 python-webob.noarch 0:1.2.3-7.el7 python-webtest.noarch 0:1.3.4-6.el7 python-werkzeug.noarch 0:0.9.1-2.el7 python2-bcrypt.x86_64 0:3.1.6-2.el7 python2-six.noarch 0:1.9.0-0.el7 userspace-rcu.x86_64 0:0.7.16-1.el7 Dependency Updated: librados2.x86_64 2:14.2.12-0.el7 librbd1.x86_64 2:14.2.12-0.el7 selinux-policy.noarch 0:3.13.1-266.el7_8.1 selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 Complete! root@node03's password: Loaded plugins: fastestmirror, langpacks, product-id, search-disabled-repos, : subscription-manager This system is not registered with an entitlement server. You can use subscription-manager to register. Repository epel is listed more than once in the configuration Repository epel-debuginfo is listed more than once in the configuration Repository epel-source is listed more than once in the configuration Loading mirror speeds from cached hostfile * base: mirrors.aliyun.com * extras: mirrors.huaweicloud.com * updates: mirror.bit.edu.cn Resolving Dependencies --> Running transaction check ---> Package ceph.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: ceph-mon = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-osd = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mds = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-mgr = 2:14.2.12-0.el7 for package: 2:ceph-14.2.12-0.el7.x86_64 ---> Package ceph-radosgw.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: librgw2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-selinux = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-base = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados2 = 2:14.2.12-0.el7 for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librados.so.2(LIBRADOS_14.2.0)(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: liboath.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: libceph-common.so.0()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Processing Dependency: librabbitmq.so.4()(64bit) for package: 2:ceph-radosgw-14.2.12-0.el7.x86_64 --> Running transaction check ---> Package ceph-base.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: libcephfs2 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: librbd1 = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: ceph-common = 2:14.2.12-0.el7 for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: libleveldb.so.1()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 --> Processing Dependency: liblttng-ust.so.0()(64bit) for package: 2:ceph-base-14.2.12-0.el7.x86_64 ---> Package ceph-mds.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-mgr.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-pecan for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cherrypy for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-werkzeug for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 --> Processing Dependency: python-bcrypt for package: 2:ceph-mgr-14.2.12-0.el7.x86_64 ---> Package ceph-mon.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-osd.x86_64 2:14.2.12-0.el7 will be installed ---> Package ceph-selinux.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: selinux-policy-base >= 3.13.1-252.el7_7.6 for package: 2:ceph-selinux-14.2.12-0.el7.x86_64 ---> Package liboath.x86_64 0:2.6.2-1.el7 will be installed ---> Package librabbitmq.x86_64 0:0.8.0-2.el7 will be installed ---> Package librados2.x86_64 1:10.2.5-4.el7 will be updated ---> Package librados2.x86_64 2:14.2.12-0.el7 will be an update ---> Package librgw2.x86_64 2:14.2.12-0.el7 will be installed --> Running transaction check ---> Package ceph-common.x86_64 2:14.2.12-0.el7 will be installed --> Processing Dependency: python-rgw = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper1 = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-ceph-argparse = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rbd = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-cephfs = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-rados = 2:14.2.12-0.el7 for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: python-prettytable for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace-ctf.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: librdkafka.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libbabeltrace.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 --> Processing Dependency: libradosstriper.so.1()(64bit) for package: 2:ceph-common-14.2.12-0.el7.x86_64 ---> Package leveldb.x86_64 0:1.12.0-11.el7 will be installed ---> Package libcephfs2.x86_64 2:14.2.12-0.el7 will be installed ---> Package librbd1.x86_64 1:10.2.5-4.el7 will be updated ---> Package librbd1.x86_64 2:14.2.12-0.el7 will be an update ---> Package lttng-ust.x86_64 0:2.4.1-4.el7 will be installed --> Processing Dependency: liburcu-bp.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 --> Processing Dependency: liburcu-cds.so.1()(64bit) for package: lttng-ust-2.4.1-4.el7.x86_64 ---> Package python-cherrypy.noarch 0:3.2.2-4.el7 will be installed ---> Package python-pecan.noarch 0:0.4.5-2.el7 will be installed --> Processing Dependency: python-mako >= 0.4.0 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-simplegeneric >= 0.8 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webob >= 1.2 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-webtest >= 1.3.1 for package: python-pecan-0.4.5-2.el7.noarch --> Processing Dependency: python-singledispatch for package: python-pecan-0.4.5-2.el7.noarch ---> Package python-werkzeug.noarch 0:0.9.1-2.el7 will be installed ---> Package python2-bcrypt.x86_64 0:3.1.6-2.el7 will be installed --> Processing Dependency: python2-six for package: python2-bcrypt-3.1.6-2.el7.x86_64 ---> Package selinux-policy-targeted.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 will be an update --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Processing Dependency: selinux-policy = 3.13.1-266.el7_8.1 for package: selinux-policy-targeted-3.13.1-266.el7_8.1.noarch --> Running transaction check ---> Package libbabeltrace.x86_64 0:1.2.4-3.el7 will be installed ---> Package libradosstriper1.x86_64 2:14.2.12-0.el7 will be installed ---> Package librdkafka.x86_64 0:0.11.5-1.el7 will be installed ---> Package python-ceph-argparse.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-cephfs.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-mako.noarch 0:0.8.1-2.el7 will be installed --> Processing Dependency: python-markupsafe for package: python-mako-0.8.1-2.el7.noarch --> Processing Dependency: python-beaker for package: python-mako-0.8.1-2.el7.noarch ---> Package python-prettytable.noarch 0:0.7.2-3.el7 will be installed ---> Package python-rados.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rbd.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-rgw.x86_64 2:14.2.12-0.el7 will be installed ---> Package python-simplegeneric.noarch 0:0.8-7.el7 will be installed ---> Package python-singledispatch.noarch 0:3.4.0.2-2.el7 will be installed ---> Package python-webob.noarch 0:1.2.3-7.el7 will be installed ---> Package python-webtest.noarch 0:1.3.4-6.el7 will be installed ---> Package python2-six.noarch 0:1.9.0-0.el7 will be installed ---> Package selinux-policy.noarch 0:3.13.1-229.el7 will be updated ---> Package selinux-policy.noarch 0:3.13.1-266.el7_8.1 will be an update ---> Package userspace-rcu.x86_64 0:0.7.16-1.el7 will be installed --> Running transaction check ---> Package python-beaker.noarch 0:1.5.4-10.el7 will be installed --> Processing Dependency: python-paste for package: python-beaker-1.5.4-10.el7.noarch ---> Package python-markupsafe.x86_64 0:0.11-10.el7 will be installed --> Running transaction check ---> Package python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 will be installed --> Processing Dependency: python-tempita for package: python-paste-1.7.5.1-9.20111221hg1498.el7.noarch --> Running transaction check ---> Package python-tempita.noarch 0:0.5.1-6.el7 will be installed --> Finished Dependency Resolution Dependencies Resolved ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: ceph x86_64 2:14.2.12-0.el7 x86_64 3.0 k ceph-radosgw x86_64 2:14.2.12-0.el7 x86_64 5.8 M Installing for dependencies: ceph-base x86_64 2:14.2.12-0.el7 x86_64 5.4 M ceph-common x86_64 2:14.2.12-0.el7 x86_64 19 M ceph-mds x86_64 2:14.2.12-0.el7 x86_64 1.9 M ceph-mgr x86_64 2:14.2.12-0.el7 x86_64 1.7 M ceph-mon x86_64 2:14.2.12-0.el7 x86_64 4.3 M ceph-osd x86_64 2:14.2.12-0.el7 x86_64 16 M ceph-selinux x86_64 2:14.2.12-0.el7 x86_64 21 k leveldb x86_64 1.12.0-11.el7 epel 161 k libbabeltrace x86_64 1.2.4-3.el7 epel 147 k libcephfs2 x86_64 2:14.2.12-0.el7 x86_64 500 k liboath x86_64 2.6.2-1.el7 epel 51 k librabbitmq x86_64 0.8.0-2.el7 base 37 k libradosstriper1 x86_64 2:14.2.12-0.el7 x86_64 362 k librdkafka x86_64 0.11.5-1.el7 epel 368 k librgw2 x86_64 2:14.2.12-0.el7 x86_64 5.1 M lttng-ust x86_64 2.4.1-4.el7 epel 176 k python-beaker noarch 1.5.4-10.el7 base 80 k python-ceph-argparse x86_64 2:14.2.12-0.el7 x86_64 36 k python-cephfs x86_64 2:14.2.12-0.el7 x86_64 112 k python-cherrypy noarch 3.2.2-4.el7 base 422 k python-mako noarch 0.8.1-2.el7 base 307 k python-markupsafe x86_64 0.11-10.el7 base 25 k python-paste noarch 1.7.5.1-9.20111221hg1498.el7 base 866 k python-pecan noarch 0.4.5-2.el7 epel 255 k python-prettytable noarch 0.7.2-3.el7 base 37 k python-rados x86_64 2:14.2.12-0.el7 x86_64 200 k python-rbd x86_64 2:14.2.12-0.el7 x86_64 195 k python-rgw x86_64 2:14.2.12-0.el7 x86_64 78 k python-simplegeneric noarch 0.8-7.el7 epel 12 k python-singledispatch noarch 3.4.0.2-2.el7 epel 18 k python-tempita noarch 0.5.1-6.el7 base 33 k python-webob noarch 1.2.3-7.el7 base 202 k python-webtest noarch 1.3.4-6.el7 base 102 k python-werkzeug noarch 0.9.1-2.el7 extras 562 k python2-bcrypt x86_64 3.1.6-2.el7 epel 39 k python2-six noarch 1.9.0-0.el7 epel 2.9 k userspace-rcu x86_64 0.7.16-1.el7 epel 73 k Updating for dependencies: librados2 x86_64 2:14.2.12-0.el7 x86_64 3.4 M librbd1 x86_64 2:14.2.12-0.el7 x86_64 1.7 M selinux-policy noarch 3.13.1-266.el7_8.1 updates 497 k selinux-policy-targeted noarch 3.13.1-266.el7_8.1 updates 7.0 M Transaction Summary ================================================================================ Install 2 Packages (+37 Dependent packages) Upgrade ( 4 Dependent packages) Total download size: 77 M Downloading packages: No Presto metadata available for x86_64 No Presto metadata available for updates -------------------------------------------------------------------------------- Total 3.0 MB/s | 77 MB 00:25 Running transaction check Running transaction test Transaction test succeeded Running transaction Installing : leveldb-1.12.0-11.el7.x86_64 1/47 Installing : liboath-2.6.2-1.el7.x86_64 2/47 Installing : librabbitmq-0.8.0-2.el7.x86_64 3/47 Installing : python-webob-1.2.3-7.el7.noarch 4/47 Installing : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 5/47 Installing : python-webtest-1.3.4-6.el7.noarch 6/47 Installing : python-tempita-0.5.1-6.el7.noarch 7/47 Installing : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 8/47 Installing : python-beaker-1.5.4-10.el7.noarch 9/47 Installing : python-cherrypy-3.2.2-4.el7.noarch 10/47 Installing : python-werkzeug-0.9.1-2.el7.noarch 11/47 Installing : librdkafka-0.11.5-1.el7.x86_64 12/47 Installing : python-markupsafe-0.11-10.el7.x86_64 13/47 Installing : python-mako-0.8.1-2.el7.noarch 14/47 Installing : python-singledispatch-3.4.0.2-2.el7.noarch 15/47 Updating : selinux-policy-3.13.1-266.el7_8.1.noarch 16/47 Updating : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 17/47 Installing : libbabeltrace-1.2.4-3.el7.x86_64 18/47 Installing : python-simplegeneric-0.8-7.el7.noarch 19/47 Installing : python-pecan-0.4.5-2.el7.noarch 20/47 Installing : userspace-rcu-0.7.16-1.el7.x86_64 21/47 Installing : lttng-ust-2.4.1-4.el7.x86_64 22/47 Updating : 2:librados2-14.2.12-0.el7.x86_64 23/47 Installing : 2:python-rados-14.2.12-0.el7.x86_64 24/47 Installing : 2:librgw2-14.2.12-0.el7.x86_64 25/47 Installing : 2:libcephfs2-14.2.12-0.el7.x86_64 26/47 Updating : 2:librbd1-14.2.12-0.el7.x86_64 27/47 Installing : 2:python-rbd-14.2.12-0.el7.x86_64 28/47 Installing : 2:python-cephfs-14.2.12-0.el7.x86_64 29/47 Installing : 2:python-rgw-14.2.12-0.el7.x86_64 30/47 Installing : 2:libradosstriper1-14.2.12-0.el7.x86_64 31/47 Installing : python-prettytable-0.7.2-3.el7.noarch 32/47 Installing : 2:ceph-common-14.2.12-0.el7.x86_64 33/47 Installing : 2:ceph-base-14.2.12-0.el7.x86_64 34/47 Installing : 2:ceph-selinux-14.2.12-0.el7.x86_64 35/47 Installing : 2:ceph-mon-14.2.12-0.el7.x86_64 36/47 Installing : 2:ceph-osd-14.2.12-0.el7.x86_64 37/47 Installing : 2:ceph-mds-14.2.12-0.el7.x86_64 38/47 Installing : python2-six-1.9.0-0.el7.noarch 39/47 Installing : python2-bcrypt-3.1.6-2.el7.x86_64 40/47 Installing : 2:ceph-mgr-14.2.12-0.el7.x86_64 41/47 Installing : 2:ceph-14.2.12-0.el7.x86_64 42/47 Installing : 2:ceph-radosgw-14.2.12-0.el7.x86_64 43/47 Cleanup : selinux-policy-targeted-3.13.1-229.el7.noarch 44/47 Cleanup : 1:librbd1-10.2.5-4.el7.x86_64 45/47 Cleanup : selinux-policy-3.13.1-229.el7.noarch 46/47 Cleanup : 1:librados2-10.2.5-4.el7.x86_64 47/47 Verifying : liboath-2.6.2-1.el7.x86_64 1/47 Verifying : python2-six-1.9.0-0.el7.noarch 2/47 Verifying : 2:python-rbd-14.2.12-0.el7.x86_64 3/47 Verifying : python-prettytable-0.7.2-3.el7.noarch 4/47 Verifying : leveldb-1.12.0-11.el7.x86_64 5/47 Verifying : python2-bcrypt-3.1.6-2.el7.x86_64 6/47 Verifying : userspace-rcu-0.7.16-1.el7.x86_64 7/47 Verifying : python-simplegeneric-0.8-7.el7.noarch 8/47 Verifying : 2:python-cephfs-14.2.12-0.el7.x86_64 9/47 Verifying : libbabeltrace-1.2.4-3.el7.x86_64 10/47 Verifying : python-mako-0.8.1-2.el7.noarch 11/47 Verifying : 2:ceph-selinux-14.2.12-0.el7.x86_64 12/47 Verifying : selinux-policy-3.13.1-266.el7_8.1.noarch 13/47 Verifying : 2:librgw2-14.2.12-0.el7.x86_64 14/47 Verifying : 2:ceph-mon-14.2.12-0.el7.x86_64 15/47 Verifying : python-pecan-0.4.5-2.el7.noarch 16/47 Verifying : 2:ceph-radosgw-14.2.12-0.el7.x86_64 17/47 Verifying : python-singledispatch-3.4.0.2-2.el7.noarch 18/47 Verifying : python-markupsafe-0.11-10.el7.x86_64 19/47 Verifying : 2:ceph-common-14.2.12-0.el7.x86_64 20/47 Verifying : 2:libcephfs2-14.2.12-0.el7.x86_64 21/47 Verifying : librdkafka-0.11.5-1.el7.x86_64 22/47 Verifying : python-paste-1.7.5.1-9.20111221hg1498.el7.noarch 23/47 Verifying : lttng-ust-2.4.1-4.el7.x86_64 24/47 Verifying : 2:ceph-osd-14.2.12-0.el7.x86_64 25/47 Verifying : python-werkzeug-0.9.1-2.el7.noarch 26/47 Verifying : 2:python-rados-14.2.12-0.el7.x86_64 27/47 Verifying : 2:libradosstriper1-14.2.12-0.el7.x86_64 28/47 Verifying : 2:ceph-base-14.2.12-0.el7.x86_64 29/47 Verifying : 2:ceph-mgr-14.2.12-0.el7.x86_64 30/47 Verifying : python-cherrypy-3.2.2-4.el7.noarch 31/47 Verifying : python-beaker-1.5.4-10.el7.noarch 32/47 Verifying : 2:python-rgw-14.2.12-0.el7.x86_64 33/47 Verifying : 2:librados2-14.2.12-0.el7.x86_64 34/47 Verifying : 2:ceph-mds-14.2.12-0.el7.x86_64 35/47 Verifying : selinux-policy-targeted-3.13.1-266.el7_8.1.noarch 36/47 Verifying : python-tempita-0.5.1-6.el7.noarch 37/47 Verifying : python-webtest-1.3.4-6.el7.noarch 38/47 Verifying : 2:python-ceph-argparse-14.2.12-0.el7.x86_64 39/47 Verifying : librabbitmq-0.8.0-2.el7.x86_64 40/47 Verifying : 2:librbd1-14.2.12-0.el7.x86_64 41/47 Verifying : python-webob-1.2.3-7.el7.noarch 42/47 Verifying : 2:ceph-14.2.12-0.el7.x86_64 43/47 Verifying : 1:librbd1-10.2.5-4.el7.x86_64 44/47 Verifying : selinux-policy-3.13.1-229.el7.noarch 45/47 Verifying : selinux-policy-targeted-3.13.1-229.el7.noarch 46/47 Verifying : 1:librados2-10.2.5-4.el7.x86_64 47/47 Installed: ceph.x86_64 2:14.2.12-0.el7 ceph-radosgw.x86_64 2:14.2.12-0.el7 Dependency Installed: ceph-base.x86_64 2:14.2.12-0.el7 ceph-common.x86_64 2:14.2.12-0.el7 ceph-mds.x86_64 2:14.2.12-0.el7 ceph-mgr.x86_64 2:14.2.12-0.el7 ceph-mon.x86_64 2:14.2.12-0.el7 ceph-osd.x86_64 2:14.2.12-0.el7 ceph-selinux.x86_64 2:14.2.12-0.el7 leveldb.x86_64 0:1.12.0-11.el7 libbabeltrace.x86_64 0:1.2.4-3.el7 libcephfs2.x86_64 2:14.2.12-0.el7 liboath.x86_64 0:2.6.2-1.el7 librabbitmq.x86_64 0:0.8.0-2.el7 libradosstriper1.x86_64 2:14.2.12-0.el7 librdkafka.x86_64 0:0.11.5-1.el7 librgw2.x86_64 2:14.2.12-0.el7 lttng-ust.x86_64 0:2.4.1-4.el7 python-beaker.noarch 0:1.5.4-10.el7 python-ceph-argparse.x86_64 2:14.2.12-0.el7 python-cephfs.x86_64 2:14.2.12-0.el7 python-cherrypy.noarch 0:3.2.2-4.el7 python-mako.noarch 0:0.8.1-2.el7 python-markupsafe.x86_64 0:0.11-10.el7 python-paste.noarch 0:1.7.5.1-9.20111221hg1498.el7 python-pecan.noarch 0:0.4.5-2.el7 python-prettytable.noarch 0:0.7.2-3.el7 python-rados.x86_64 2:14.2.12-0.el7 python-rbd.x86_64 2:14.2.12-0.el7 python-rgw.x86_64 2:14.2.12-0.el7 python-simplegeneric.noarch 0:0.8-7.el7 python-singledispatch.noarch 0:3.4.0.2-2.el7 python-tempita.noarch 0:0.5.1-6.el7 python-webob.noarch 0:1.2.3-7.el7 python-webtest.noarch 0:1.3.4-6.el7 python-werkzeug.noarch 0:0.9.1-2.el7 python2-bcrypt.x86_64 0:3.1.6-2.el7 python2-six.noarch 0:1.9.0-0.el7 userspace-rcu.x86_64 0:0.7.16-1.el7 Dependency Updated: librados2.x86_64 2:14.2.12-0.el7 librbd1.x86_64 2:14.2.12-0.el7 selinux-policy.noarch 0:3.13.1-266.el7_8.1 selinux-policy-targeted.noarch 0:3.13.1-266.el7_8.1 Complete! [root@node01 ~]# [root@node01 ~]# ceph -v ceph version 14.2.12 (2f3caa3b8b3d5c5f2719a1e9d8e7deea5ae1a5c6) nautilus (stable) [root@node01 ~]# [root@node01 ~]# rpm -qa|grep ceph ceph-base-14.2.12-0.el7.x86_64 ceph-mgr-14.2.12-0.el7.x86_64 ceph-selinux-14.2.12-0.el7.x86_64 ceph-radosgw-14.2.12-0.el7.x86_64 ceph-mds-14.2.12-0.el7.x86_64 python-ceph-argparse-14.2.12-0.el7.x86_64 python-cephfs-14.2.12-0.el7.x86_64 ceph-osd-14.2.12-0.el7.x86_64 ceph-deploy-2.0.1-0.noarch libcephfs2-14.2.12-0.el7.x86_64 ceph-14.2.12-0.el7.x86_64 ceph-common-14.2.12-0.el7.x86_64 ceph-mon-14.2.12-0.el7.x86_64 [root@node01 ~]# [root@node02 ~]# ceph -v ceph version 14.2.12 (2f3caa3b8b3d5c5f2719a1e9d8e7deea5ae1a5c6) nautilus (stable) [root@node02 ~]# [root@node03 ~]# rpm -qa|grep ceph ceph-mon-14.2.12-0.el7.x86_64 ceph-common-14.2.12-0.el7.x86_64 ceph-selinux-14.2.12-0.el7.x86_64 ceph-radosgw-14.2.12-0.el7.x86_64 python-ceph-argparse-14.2.12-0.el7.x86_64 python-cephfs-14.2.12-0.el7.x86_64 ceph-osd-14.2.12-0.el7.x86_64 libcephfs2-14.2.12-0.el7.x86_64 ceph-base-14.2.12-0.el7.x86_64 ceph-14.2.12-0.el7.x86_64 ceph-mgr-14.2.12-0.el7.x86_64 ceph-mds-14.2.12-0.el7.x86_64 [root@node02 ~]# [root@node03 ~]# ceph -v ceph version 14.2.12 (2f3caa3b8b3d5c5f2719a1e9d8e7deea5ae1a5c6) nautilus (stable) [root@node03 ~]# [root@node03 ~]# rpm -qa|grep ceph ceph-mon-14.2.12-0.el7.x86_64 ceph-common-14.2.12-0.el7.x86_64 ceph-selinux-14.2.12-0.el7.x86_64 ceph-radosgw-14.2.12-0.el7.x86_64 python-ceph-argparse-14.2.12-0.el7.x86_64 python-cephfs-14.2.12-0.el7.x86_64 ceph-osd-14.2.12-0.el7.x86_64 libcephfs2-14.2.12-0.el7.x86_64 ceph-base-14.2.12-0.el7.x86_64 ceph-14.2.12-0.el7.x86_64 ceph-mgr-14.2.12-0.el7.x86_64 ceph-mds-14.2.12-0.el7.x86_64 [root@node03 ~]#
[root@node01 ~]# mkdir -p /cluster
[root@node01 ~]# cd /cluster/
[root@node01 cluster]# ceph-deploy new node01 node02 node03
[root@node01 ~]# mkdir -p /cluster [root@node01 ~]# [root@node01 ~]# cd /cluster/ [root@node01 ~]# [root@node01 cluster]# ceph-deploy new node01 node02 node03 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy new node01 node02 node03 [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] func : <function new at 0x7f2c2dfbd7d0> [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f2c2d736dd0> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] ssh_copykey : True [ceph_deploy.cli][INFO ] mon : ['node01', 'node02', 'node03'] [ceph_deploy.cli][INFO ] public_network : None [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] cluster_network : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] fsid : None [ceph_deploy.new][DEBUG ] Creating new cluster named ceph [ceph_deploy.new][INFO ] making sure passwordless SSH succeeds [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /usr/sbin/ip link show [node01][INFO ] Running command: /usr/sbin/ip addr show [node01][DEBUG ] IP addresses found: [u'192.168.1.181'] [ceph_deploy.new][DEBUG ] Resolving host node01 [ceph_deploy.new][DEBUG ] Monitor node01 at 192.168.1.181 [ceph_deploy.new][INFO ] making sure passwordless SSH succeeds [node02][DEBUG ] connected to host: node01 [node02][INFO ] Running command: ssh -CT -o BatchMode=yes node02 [ceph_deploy.new][WARNIN] could not connect via SSH [ceph_deploy.new][INFO ] creating a passwordless id_rsa.pub key file [ceph_deploy.new][DEBUG ] connected to host: node01 [ceph_deploy.new][INFO ] Running command: ssh-keygen -t rsa -N -f /root/.ssh/id_rsa [ceph_deploy.new][DEBUG ] Generating public/private rsa key pair. [ceph_deploy.new][DEBUG ] Your identification has been saved in /root/.ssh/id_rsa. [ceph_deploy.new][DEBUG ] Your public key has been saved in /root/.ssh/id_rsa.pub. [ceph_deploy.new][DEBUG ] The key fingerprint is: [ceph_deploy.new][DEBUG ] SHA256:q7J5xVb9xGPEw1SPQVlkndreYjqHZYUPSfpr6+h/ojU root@node01 [ceph_deploy.new][DEBUG ] The key's randomart image is: [ceph_deploy.new][DEBUG ] +---[RSA 2048]----+ [ceph_deploy.new][DEBUG ] | =+=B| [ceph_deploy.new][DEBUG ] | B=o| [ceph_deploy.new][DEBUG ] | . =++.| [ceph_deploy.new][DEBUG ] | . o.O..| [ceph_deploy.new][DEBUG ] | .S. =.=.| [ceph_deploy.new][DEBUG ] | +. B.o| [ceph_deploy.new][DEBUG ] | o. *Eo | [ceph_deploy.new][DEBUG ] | .... +o=..| [ceph_deploy.new][DEBUG ] | o+. .+B++ | [ceph_deploy.new][DEBUG ] +----[SHA256]-----+ [ceph_deploy.new][INFO ] will connect again with password prompt root@node02's password: [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][WARNIN] .ssh/authorized_keys does not exist, will skip adding keys root@node02's password: root@node02's password: [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /usr/sbin/ip link show [node02][INFO ] Running command: /usr/sbin/ip addr show [node02][DEBUG ] IP addresses found: [u'172.17.93.1', u'192.168.1.182'] [ceph_deploy.new][DEBUG ] Resolving host node02 [ceph_deploy.new][DEBUG ] Monitor node02 at 192.168.1.182 [ceph_deploy.new][INFO ] making sure passwordless SSH succeeds [node03][DEBUG ] connected to host: node01 [node03][INFO ] Running command: ssh -CT -o BatchMode=yes node03 [ceph_deploy.new][WARNIN] could not connect via SSH [ceph_deploy.new][INFO ] will connect again with password prompt root@node03's password: [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][WARNIN] .ssh/authorized_keys does not exist, will skip adding keys root@node03's password: root@node03's password: [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /usr/sbin/ip link show [node03][INFO ] Running command: /usr/sbin/ip addr show [node03][DEBUG ] IP addresses found: [u'192.168.1.183', u'172.17.12.1'] [ceph_deploy.new][DEBUG ] Resolving host node03 [ceph_deploy.new][DEBUG ] Monitor node03 at 192.168.1.183 [ceph_deploy.new][DEBUG ] Monitor initial members are ['node01', 'node02', 'node03'] [ceph_deploy.new][DEBUG ] Monitor addrs are ['192.168.1.181', '192.168.1.182', '192.168.1.183'] [ceph_deploy.new][DEBUG ] Creating a random mon key... [ceph_deploy.new][DEBUG ] Writing monitor keyring to ceph.mon.keyring... [ceph_deploy.new][DEBUG ] Writing initial config to ceph.conf... [root@node01 cluster]#
[root@node01 cluster]# ssh-keygen
[root@node01 cluster]# ssh-copy-id node01
[root@node01 cluster]# ssh-copy-id node02
[root@node01 cluster]# ssh-copy-id node03
[root@node01 cluster]# ssh-keygen Generating public/private rsa key pair. Enter file in which to save the key (/root/.ssh/id_rsa): /root/.ssh/id_rsa already exists. Overwrite (y/n)? y Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /root/.ssh/id_rsa. Your public key has been saved in /root/.ssh/id_rsa.pub. The key fingerprint is: SHA256:b+18tga3Z9OxJdIbwJRo4yi9qUu8NB4ZBcoMKwfDtyA root@node01 The key's randomart image is: +---[RSA 2048]----+ |+ . . . . | |E= * . . + o | |o = = ..+ + | | o . ..o . o | | ..So o | | . oo. .o =..| | B. o .+ ==| | +.+ . o *o+| | +. o+.+.| +----[SHA256]-----+ [root@node01 cluster]# [root@node01 cluster]# ssh-copy-id node01 /usr/bin/ssh-copy-id: INFO: Source of key(s) to be installed: "/root/.ssh/id_rsa.pub" /usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed /usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed -- if you are prompted now it is to install the new keys root@node01's password: Number of key(s) added: 1 Now try logging into the machine, with: "ssh 'node01'" and check to make sure that only the key(s) you wanted were added. [root@node01 cluster]# [root@node01 cluster]# ssh-copy-id node02 /usr/bin/ssh-copy-id: INFO: Source of key(s) to be installed: "/root/.ssh/id_rsa.pub" /usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed /usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed -- if you are prompted now it is to install the new keys root@node02's password: Number of key(s) added: 1 Now try logging into the machine, with: "ssh 'node02'" and check to make sure that only the key(s) you wanted were added. [root@node01 cluster]# [root@node01 cluster]# ssh-copy-id node03 /usr/bin/ssh-copy-id: INFO: Source of key(s) to be installed: "/root/.ssh/id_rsa.pub" /usr/bin/ssh-copy-id: INFO: attempting to log in with the new key(s), to filter out any that are already installed /usr/bin/ssh-copy-id: INFO: 1 key(s) remain to be installed -- if you are prompted now it is to install the new keys root@node03's password: Number of key(s) added: 1 Now try logging into the machine, with: "ssh 'node03'" and check to make sure that only the key(s) you wanted were added. [root@node01 cluster]#
[root@node01 cluster]# ceph-deploy mon create-initial
[root@node01 cluster]# ceph-deploy admin node01 node02 node03
[root@node01 cluster]# ceph -s
[root@node01 cluster]# ceph-deploy mon create-initial [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mon create-initial [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] subcommand : create-initial [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f96f1d85950> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] func : <function mon at 0x7f96f2611d70> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] keyrings : None [ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts node01 node02 node03 [ceph_deploy.mon][DEBUG ] detecting platform for host node01 ... [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [ceph_deploy.mon][INFO ] distro info: CentOS Linux 7.6.1810 Core [node01][DEBUG ] determining if provided host has same hostname in remote [node01][DEBUG ] get remote short hostname [node01][DEBUG ] deploying mon to node01 [node01][DEBUG ] get remote short hostname [node01][DEBUG ] remote hostname: node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][DEBUG ] create the mon path if it does not exist [node01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-node01/done [node01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-node01/done [node01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-node01.mon.keyring [node01][DEBUG ] create the monitor keyring file [node01][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i node01 --keyring /var/lib/ceph/tmp/ceph-node01.mon.keyring --setuser 167 --setgroup 167 [node01][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-node01.mon.keyring [node01][DEBUG ] create a done file to avoid re-doing the mon deployment [node01][DEBUG ] create the init path if it does not exist [node01][INFO ] Running command: systemctl enable ceph.target [node01][INFO ] Running command: systemctl enable ceph-mon@node01 [node01][WARNIN] Created symlink from /etc/systemd/system/ceph-mon.target.wants/ceph-mon@node01.service to /usr/lib/systemd/system/ceph-mon@.service. [node01][INFO ] Running command: systemctl start ceph-mon@node01 [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [node01][DEBUG ] ******************************************************************************** [node01][DEBUG ] status for monitor: mon.node01 [node01][DEBUG ] { [node01][DEBUG ] "election_epoch": 0, [node01][DEBUG ] "extra_probe_peers": [ [node01][DEBUG ] { [node01][DEBUG ] "addrvec": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.182:3300", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v2" [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.182:6789", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v1" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addrvec": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.183:3300", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v2" [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.183:6789", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v1" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] } [node01][DEBUG ] ], [node01][DEBUG ] "feature_map": { [node01][DEBUG ] "mon": [ [node01][DEBUG ] { [node01][DEBUG ] "features": "0x3ffddff8ffecffff", [node01][DEBUG ] "num": 1, [node01][DEBUG ] "release": "luminous" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] "features": { [node01][DEBUG ] "quorum_con": "0", [node01][DEBUG ] "quorum_mon": [], [node01][DEBUG ] "required_con": "0", [node01][DEBUG ] "required_mon": [] [node01][DEBUG ] }, [node01][DEBUG ] "monmap": { [node01][DEBUG ] "created": "2020-10-31 23:08:25.227025", [node01][DEBUG ] "epoch": 0, [node01][DEBUG ] "features": { [node01][DEBUG ] "optional": [], [node01][DEBUG ] "persistent": [] [node01][DEBUG ] }, [node01][DEBUG ] "fsid": "c90f3c76-20f0-4091-a66b-74ec0e6f4ec8", [node01][DEBUG ] "min_mon_release": 0, [node01][DEBUG ] "min_mon_release_name": "unknown", [node01][DEBUG ] "modified": "2020-10-31 23:08:25.227025", [node01][DEBUG ] "mons": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.181:6789/0", [node01][DEBUG ] "name": "node01", [node01][DEBUG ] "public_addr": "192.168.1.181:6789/0", [node01][DEBUG ] "public_addrs": { [node01][DEBUG ] "addrvec": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.181:3300", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v2" [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addr": "192.168.1.181:6789", [node01][DEBUG ] "nonce": 0, [node01][DEBUG ] "type": "v1" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] "rank": 0 [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addr": "0.0.0.0:0/1", [node01][DEBUG ] "name": "node02", [node01][DEBUG ] "public_addr": "0.0.0.0:0/1", [node01][DEBUG ] "public_addrs": { [node01][DEBUG ] "addrvec": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "0.0.0.0:0", [node01][DEBUG ] "nonce": 1, [node01][DEBUG ] "type": "v1" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] "rank": 1 [node01][DEBUG ] }, [node01][DEBUG ] { [node01][DEBUG ] "addr": "0.0.0.0:0/2", [node01][DEBUG ] "name": "node03", [node01][DEBUG ] "public_addr": "0.0.0.0:0/2", [node01][DEBUG ] "public_addrs": { [node01][DEBUG ] "addrvec": [ [node01][DEBUG ] { [node01][DEBUG ] "addr": "0.0.0.0:0", [node01][DEBUG ] "nonce": 2, [node01][DEBUG ] "type": "v1" [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] "rank": 2 [node01][DEBUG ] } [node01][DEBUG ] ] [node01][DEBUG ] }, [node01][DEBUG ] "name": "node01", [node01][DEBUG ] "outside_quorum": [ [node01][DEBUG ] "node01" [node01][DEBUG ] ], [node01][DEBUG ] "quorum": [], [node01][DEBUG ] "rank": 0, [node01][DEBUG ] "state": "probing", [node01][DEBUG ] "sync_provider": [] [node01][DEBUG ] } [node01][DEBUG ] ******************************************************************************** [node01][INFO ] monitor: mon.node01 is running [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [ceph_deploy.mon][DEBUG ] detecting platform for host node02 ... [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [ceph_deploy.mon][INFO ] distro info: CentOS Linux 7.6.1810 Core [node02][DEBUG ] determining if provided host has same hostname in remote [node02][DEBUG ] get remote short hostname [node02][DEBUG ] deploying mon to node02 [node02][DEBUG ] get remote short hostname [node02][DEBUG ] remote hostname: node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][DEBUG ] create the mon path if it does not exist [node02][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-node02/done [node02][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-node02/done [node02][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-node02.mon.keyring [node02][DEBUG ] create the monitor keyring file [node02][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i node02 --keyring /var/lib/ceph/tmp/ceph-node02.mon.keyring --setuser 167 --setgroup 167 [node02][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-node02.mon.keyring [node02][DEBUG ] create a done file to avoid re-doing the mon deployment [node02][DEBUG ] create the init path if it does not exist [node02][INFO ] Running command: systemctl enable ceph.target [node02][INFO ] Running command: systemctl enable ceph-mon@node02 [node02][WARNIN] Created symlink from /etc/systemd/system/ceph-mon.target.wants/ceph-mon@node02.service to /usr/lib/systemd/system/ceph-mon@.service. [node02][INFO ] Running command: systemctl start ceph-mon@node02 [node02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node02.asok mon_status [node02][DEBUG ] ******************************************************************************** [node02][DEBUG ] status for monitor: mon.node02 [node02][DEBUG ] { [node02][DEBUG ] "election_epoch": 0, [node02][DEBUG ] "extra_probe_peers": [ [node02][DEBUG ] { [node02][DEBUG ] "addrvec": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.181:3300", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v2" [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.181:6789", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v1" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addrvec": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.183:3300", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v2" [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.183:6789", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v1" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] } [node02][DEBUG ] ], [node02][DEBUG ] "feature_map": { [node02][DEBUG ] "mon": [ [node02][DEBUG ] { [node02][DEBUG ] "features": "0x3ffddff8ffecffff", [node02][DEBUG ] "num": 1, [node02][DEBUG ] "release": "luminous" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] "features": { [node02][DEBUG ] "quorum_con": "0", [node02][DEBUG ] "quorum_mon": [], [node02][DEBUG ] "required_con": "0", [node02][DEBUG ] "required_mon": [] [node02][DEBUG ] }, [node02][DEBUG ] "monmap": { [node02][DEBUG ] "created": "2020-10-31 23:08:28.614711", [node02][DEBUG ] "epoch": 0, [node02][DEBUG ] "features": { [node02][DEBUG ] "optional": [], [node02][DEBUG ] "persistent": [] [node02][DEBUG ] }, [node02][DEBUG ] "fsid": "c90f3c76-20f0-4091-a66b-74ec0e6f4ec8", [node02][DEBUG ] "min_mon_release": 0, [node02][DEBUG ] "min_mon_release_name": "unknown", [node02][DEBUG ] "modified": "2020-10-31 23:08:28.614711", [node02][DEBUG ] "mons": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.182:6789/0", [node02][DEBUG ] "name": "node02", [node02][DEBUG ] "public_addr": "192.168.1.182:6789/0", [node02][DEBUG ] "public_addrs": { [node02][DEBUG ] "addrvec": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.182:3300", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v2" [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addr": "192.168.1.182:6789", [node02][DEBUG ] "nonce": 0, [node02][DEBUG ] "type": "v1" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] "rank": 0 [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addr": "0.0.0.0:0/1", [node02][DEBUG ] "name": "node01", [node02][DEBUG ] "public_addr": "0.0.0.0:0/1", [node02][DEBUG ] "public_addrs": { [node02][DEBUG ] "addrvec": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "0.0.0.0:0", [node02][DEBUG ] "nonce": 1, [node02][DEBUG ] "type": "v1" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] "rank": 1 [node02][DEBUG ] }, [node02][DEBUG ] { [node02][DEBUG ] "addr": "0.0.0.0:0/2", [node02][DEBUG ] "name": "node03", [node02][DEBUG ] "public_addr": "0.0.0.0:0/2", [node02][DEBUG ] "public_addrs": { [node02][DEBUG ] "addrvec": [ [node02][DEBUG ] { [node02][DEBUG ] "addr": "0.0.0.0:0", [node02][DEBUG ] "nonce": 2, [node02][DEBUG ] "type": "v1" [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] "rank": 2 [node02][DEBUG ] } [node02][DEBUG ] ] [node02][DEBUG ] }, [node02][DEBUG ] "name": "node02", [node02][DEBUG ] "outside_quorum": [ [node02][DEBUG ] "node02" [node02][DEBUG ] ], [node02][DEBUG ] "quorum": [], [node02][DEBUG ] "rank": 0, [node02][DEBUG ] "state": "probing", [node02][DEBUG ] "sync_provider": [] [node02][DEBUG ] } [node02][DEBUG ] ******************************************************************************** [node02][INFO ] monitor: mon.node02 is running [node02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node02.asok mon_status [ceph_deploy.mon][DEBUG ] detecting platform for host node03 ... [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [ceph_deploy.mon][INFO ] distro info: CentOS Linux 7.6.1810 Core [node03][DEBUG ] determining if provided host has same hostname in remote [node03][DEBUG ] get remote short hostname [node03][DEBUG ] deploying mon to node03 [node03][DEBUG ] get remote short hostname [node03][DEBUG ] remote hostname: node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][DEBUG ] create the mon path if it does not exist [node03][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-node03/done [node03][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-node03/done [node03][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-node03.mon.keyring [node03][DEBUG ] create the monitor keyring file [node03][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i node03 --keyring /var/lib/ceph/tmp/ceph-node03.mon.keyring --setuser 167 --setgroup 167 [node03][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-node03.mon.keyring [node03][DEBUG ] create a done file to avoid re-doing the mon deployment [node03][DEBUG ] create the init path if it does not exist [node03][INFO ] Running command: systemctl enable ceph.target [node03][INFO ] Running command: systemctl enable ceph-mon@node03 [node03][WARNIN] Created symlink from /etc/systemd/system/ceph-mon.target.wants/ceph-mon@node03.service to /usr/lib/systemd/system/ceph-mon@.service. [node03][INFO ] Running command: systemctl start ceph-mon@node03 [node03][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node03.asok mon_status [node03][DEBUG ] ******************************************************************************** [node03][DEBUG ] status for monitor: mon.node03 [node03][DEBUG ] { [node03][DEBUG ] "election_epoch": 1, [node03][DEBUG ] "extra_probe_peers": [ [node03][DEBUG ] { [node03][DEBUG ] "addrvec": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.181:3300", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v2" [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.181:6789", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v1" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addrvec": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.182:3300", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v2" [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.182:6789", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v1" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] } [node03][DEBUG ] ], [node03][DEBUG ] "feature_map": { [node03][DEBUG ] "mon": [ [node03][DEBUG ] { [node03][DEBUG ] "features": "0x3ffddff8ffecffff", [node03][DEBUG ] "num": 1, [node03][DEBUG ] "release": "luminous" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] "features": { [node03][DEBUG ] "quorum_con": "0", [node03][DEBUG ] "quorum_mon": [], [node03][DEBUG ] "required_con": "0", [node03][DEBUG ] "required_mon": [] [node03][DEBUG ] }, [node03][DEBUG ] "monmap": { [node03][DEBUG ] "created": "2020-10-31 23:08:31.811660", [node03][DEBUG ] "epoch": 0, [node03][DEBUG ] "features": { [node03][DEBUG ] "optional": [], [node03][DEBUG ] "persistent": [] [node03][DEBUG ] }, [node03][DEBUG ] "fsid": "c90f3c76-20f0-4091-a66b-74ec0e6f4ec8", [node03][DEBUG ] "min_mon_release": 0, [node03][DEBUG ] "min_mon_release_name": "unknown", [node03][DEBUG ] "modified": "2020-10-31 23:08:31.811660", [node03][DEBUG ] "mons": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.182:6789/0", [node03][DEBUG ] "name": "node02", [node03][DEBUG ] "public_addr": "192.168.1.182:6789/0", [node03][DEBUG ] "public_addrs": { [node03][DEBUG ] "addrvec": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.182:3300", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v2" [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.182:6789", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v1" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] "rank": 0 [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.183:6789/0", [node03][DEBUG ] "name": "node03", [node03][DEBUG ] "public_addr": "192.168.1.183:6789/0", [node03][DEBUG ] "public_addrs": { [node03][DEBUG ] "addrvec": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.183:3300", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v2" [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "192.168.1.183:6789", [node03][DEBUG ] "nonce": 0, [node03][DEBUG ] "type": "v1" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] "rank": 1 [node03][DEBUG ] }, [node03][DEBUG ] { [node03][DEBUG ] "addr": "0.0.0.0:0/1", [node03][DEBUG ] "name": "node01", [node03][DEBUG ] "public_addr": "0.0.0.0:0/1", [node03][DEBUG ] "public_addrs": { [node03][DEBUG ] "addrvec": [ [node03][DEBUG ] { [node03][DEBUG ] "addr": "0.0.0.0:0", [node03][DEBUG ] "nonce": 1, [node03][DEBUG ] "type": "v1" [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] "rank": 2 [node03][DEBUG ] } [node03][DEBUG ] ] [node03][DEBUG ] }, [node03][DEBUG ] "name": "node03", [node03][DEBUG ] "outside_quorum": [], [node03][DEBUG ] "quorum": [], [node03][DEBUG ] "rank": 1, [node03][DEBUG ] "state": "electing", [node03][DEBUG ] "sync_provider": [] [node03][DEBUG ] } [node03][DEBUG ] ******************************************************************************** [node03][INFO ] monitor: mon.node03 is running [node03][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node03.asok mon_status [ceph_deploy.mon][INFO ] processing monitor mon.node01 [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [ceph_deploy.mon][WARNIN] mon.node01 monitor is not yet in quorum, tries left: 5 [ceph_deploy.mon][WARNIN] waiting 5 seconds before retrying [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [ceph_deploy.mon][WARNIN] mon.node01 monitor is not yet in quorum, tries left: 4 [ceph_deploy.mon][WARNIN] waiting 10 seconds before retrying [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [ceph_deploy.mon][WARNIN] mon.node01 monitor is not yet in quorum, tries left: 3 [ceph_deploy.mon][WARNIN] waiting 10 seconds before retrying [node01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node01.asok mon_status [ceph_deploy.mon][INFO ] mon.node01 monitor has reached quorum! [ceph_deploy.mon][INFO ] processing monitor mon.node02 [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node02.asok mon_status [ceph_deploy.mon][INFO ] mon.node02 monitor has reached quorum! [ceph_deploy.mon][INFO ] processing monitor mon.node03 [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.node03.asok mon_status [ceph_deploy.mon][INFO ] mon.node03 monitor has reached quorum! [ceph_deploy.mon][INFO ] all initial monitors are running and have formed quorum [ceph_deploy.mon][INFO ] Running gatherkeys... [ceph_deploy.gatherkeys][INFO ] Storing keys in temp directory /tmp/tmp05thDR [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] get remote short hostname [node01][DEBUG ] fetch remote file [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --admin-daemon=/var/run/ceph/ceph-mon.node01.asok mon_status [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-node01/keyring auth get client.admin [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-node01/keyring auth get client.bootstrap-mds [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-node01/keyring auth get client.bootstrap-mgr [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-node01/keyring auth get client.bootstrap-osd [node01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-node01/keyring auth get client.bootstrap-rgw [ceph_deploy.gatherkeys][INFO ] Storing ceph.client.admin.keyring [ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-mds.keyring [ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-mgr.keyring [ceph_deploy.gatherkeys][INFO ] keyring 'ceph.mon.keyring' already exists [ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-osd.keyring [ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-rgw.keyring [ceph_deploy.gatherkeys][INFO ] Destroy temp directory /tmp/tmp05thDR [root@node01 cluster]# [root@node01 cluster]# ceph-deploy admin node01 node02 node03 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy admin node01 node02 node03 [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f73aa83ef38> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] client : ['node01', 'node02', 'node03'] [ceph_deploy.cli][INFO ] func : <function admin at 0x7f73ab2ddb90> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to node01 [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to node02 [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to node03 [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 3m) mgr: no daemons active osd: 0 osds: 0 up, 0 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 ~]#
[root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdb
[root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdc
[root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdd
[root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdb
[root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdc
[root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdd
[root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdb
[root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdc
[root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdd
[root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdb [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node01 --data /dev/sdb [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f0a97390f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node01 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7f0a97c102a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdb [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][WARNIN] osd keyring does not exist yet, creating one [node01][DEBUG ] create a keyring file [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 8ffe5922-9235-4740-a711-f79594c517e8 [node01][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b /dev/sdb [node01][WARNIN] stdout: Physical volume "/dev/sdb" successfully created. [node01][WARNIN] stdout: Volume group "ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b" successfully created [node01][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-8ffe5922-9235-4740-a711-f79594c517e8 ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b [node01][WARNIN] stdout: Logical volume "osd-block-8ffe5922-9235-4740-a711-f79594c517e8" created. [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b/osd-block-8ffe5922-9235-4740-a711-f79594c517e8 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2 [node01][WARNIN] Running command: /usr/bin/ln -s /dev/ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b/osd-block-8ffe5922-9235-4740-a711-f79594c517e8 /var/lib/ceph/osd/ceph-0/block [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap [node01][WARNIN] stderr: 2020-11-12 15:34:25.726 7fb6f6b94700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node01][WARNIN] 2020-11-12 15:34:25.726 7fb6f6b94700 -1 AuthRegistry(0x7fb6f0065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node01][WARNIN] stderr: got monmap epoch 2 [node01][WARNIN] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-0/keyring --create-keyring --name osd.0 --add-key AQCA5axf7j2qNxAACwvKjsqkOOyxCL1gbDAI7Q== [node01][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-0/keyring [node01][WARNIN] added entity osd.0 auth(key=AQCA5axf7j2qNxAACwvKjsqkOOyxCL1gbDAI7Q==) [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ [node01][WARNIN] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid 8ffe5922-9235-4740-a711-f79594c517e8 --setuser ceph --setgroup ceph [node01][WARNIN] stderr: 2020-11-12 15:34:26.334 7f7872156a80 -1 bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid [node01][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdb [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 [node01][WARNIN] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b/osd-block-8ffe5922-9235-4740-a711-f79594c517e8 --path /var/lib/ceph/osd/ceph-0 --no-mon-config [node01][WARNIN] Running command: /usr/bin/ln -snf /dev/ceph-15aeb760-677d-47b4-8f8a-cce9bb39d84b/osd-block-8ffe5922-9235-4740-a711-f79594c517e8 /var/lib/ceph/osd/ceph-0/block [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 [node01][WARNIN] Running command: /usr/bin/systemctl enable ceph-volume@lvm-0-8ffe5922-9235-4740-a711-f79594c517e8 [node01][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-8ffe5922-9235-4740-a711-f79594c517e8.service to /usr/lib/systemd/system/ceph-volume@.service. [node01][WARNIN] Running command: /usr/bin/systemctl enable --runtime ceph-osd@0 [node01][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@0.service to /usr/lib/systemd/system/ceph-osd@.service. [node01][WARNIN] Running command: /usr/bin/systemctl start ceph-osd@0 [node01][WARNIN] --> ceph-volume lvm activate successful for osd ID: 0 [node01][WARNIN] --> ceph-volume lvm create successful for: /dev/sdb [node01][INFO ] checking OSD status... [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node01 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 37m) mgr: no daemons active osd: 1 osds: 1 up (since 46s), 1 in (since 46s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdc [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node01 --data /dev/sdc [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f0f90954f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node01 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7f0f911d42a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdc [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdc [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdc [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 2ab77d24-d999-4fd3-871c-3003b65cc718 [node01][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-777aa516-f115-405a-b9c0-dfdc40adf94f /dev/sdc [node01][WARNIN] stdout: Physical volume "/dev/sdc" successfully created. [node01][WARNIN] stdout: Volume group "ceph-777aa516-f115-405a-b9c0-dfdc40adf94f" successfully created [node01][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718 ceph-777aa516-f115-405a-b9c0-dfdc40adf94f [node01][WARNIN] stdout: Logical volume "osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718" created. [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-1 [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-777aa516-f115-405a-b9c0-dfdc40adf94f/osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3 [node01][WARNIN] Running command: /usr/bin/ln -s /dev/ceph-777aa516-f115-405a-b9c0-dfdc40adf94f/osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718 /var/lib/ceph/osd/ceph-1/block [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-1/activate.monmap [node01][WARNIN] stderr: 2020-11-12 15:37:24.629 7fd2330b7700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node01][WARNIN] 2020-11-12 15:37:24.629 7fd2330b7700 -1 AuthRegistry(0x7fd22c065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node01][WARNIN] stderr: got monmap epoch 2 [node01][WARNIN] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-1/keyring --create-keyring --name osd.1 --add-key AQAz5qxfWfy+NhAAxIbsmTCoACrMwJPXdSLRjQ== [node01][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-1/keyring [node01][WARNIN] added entity osd.1 auth(key=AQAz5qxfWfy+NhAAxIbsmTCoACrMwJPXdSLRjQ==) [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/keyring [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/ [node01][WARNIN] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 1 --monmap /var/lib/ceph/osd/ceph-1/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-1/ --osd-uuid 2ab77d24-d999-4fd3-871c-3003b65cc718 --setuser ceph --setgroup ceph [node01][WARNIN] stderr: 2020-11-12 15:37:25.208 7f5d9b58ca80 -1 bluestore(/var/lib/ceph/osd/ceph-1/) _read_fsid unparsable uuid [node01][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdc [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 [node01][WARNIN] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-777aa516-f115-405a-b9c0-dfdc40adf94f/osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718 --path /var/lib/ceph/osd/ceph-1 --no-mon-config [node01][WARNIN] Running command: /usr/bin/ln -snf /dev/ceph-777aa516-f115-405a-b9c0-dfdc40adf94f/osd-block-2ab77d24-d999-4fd3-871c-3003b65cc718 /var/lib/ceph/osd/ceph-1/block [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-1/block [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 [node01][WARNIN] Running command: /usr/bin/systemctl enable ceph-volume@lvm-1-2ab77d24-d999-4fd3-871c-3003b65cc718 [node01][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-1-2ab77d24-d999-4fd3-871c-3003b65cc718.service to /usr/lib/systemd/system/ceph-volume@.service. [node01][WARNIN] Running command: /usr/bin/systemctl enable --runtime ceph-osd@1 [node01][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@1.service to /usr/lib/systemd/system/ceph-osd@.service. [node01][WARNIN] Running command: /usr/bin/systemctl start ceph-osd@1 [node01][WARNIN] --> ceph-volume lvm activate successful for osd ID: 1 [node01][WARNIN] --> ceph-volume lvm create successful for: /dev/sdc [node01][INFO ] checking OSD status... [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node01 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 39m) mgr: no daemons active osd: 2 osds: 2 up (since 7s), 2 in (since 7s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node01 --data /dev/sdd [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node01 --data /dev/sdd [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fdcb9828f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node01 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7fdcba0a82a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdd [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdd [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [node01][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdd [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 [node01][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b /dev/sdd [node01][WARNIN] stdout: Physical volume "/dev/sdd" successfully created. [node01][WARNIN] stdout: Volume group "ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b" successfully created [node01][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b [node01][WARNIN] stdout: Logical volume "osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3" created. [node01][WARNIN] Running command: /usr/bin/ceph-authtool --gen-print-key [node01][WARNIN] Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-2 [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b/osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-4 [node01][WARNIN] Running command: /usr/bin/ln -s /dev/ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b/osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 /var/lib/ceph/osd/ceph-2/block [node01][WARNIN] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-2/activate.monmap [node01][WARNIN] stderr: 2020-11-12 15:38:18.115 7fe2492a1700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node01][WARNIN] 2020-11-12 15:38:18.115 7fe2492a1700 -1 AuthRegistry(0x7fe244065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node01][WARNIN] stderr: got monmap epoch 2 [node01][WARNIN] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-2/keyring --create-keyring --name osd.2 --add-key AQBp5qxfS5w4FhAAvNbazD8pqbHG3ytO5fQDiw== [node01][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-2/keyring [node01][WARNIN] added entity osd.2 auth(key=AQBp5qxfS5w4FhAAvNbazD8pqbHG3ytO5fQDiw==) [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/keyring [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/ [node01][WARNIN] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 2 --monmap /var/lib/ceph/osd/ceph-2/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-2/ --osd-uuid 5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 --setuser ceph --setgroup ceph [node01][WARNIN] stderr: 2020-11-12 15:38:19.727 7ffa6d39fa80 -1 bluestore(/var/lib/ceph/osd/ceph-2/) _read_fsid unparsable uuid [node01][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdd [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 [node01][WARNIN] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b/osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 --path /var/lib/ceph/osd/ceph-2 --no-mon-config [node01][WARNIN] Running command: /usr/bin/ln -snf /dev/ceph-ddaf1857-a390-4ffa-a21b-bfba478e6c0b/osd-block-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 /var/lib/ceph/osd/ceph-2/block [node01][WARNIN] Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-2/block [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /dev/dm-4 [node01][WARNIN] Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 [node01][WARNIN] Running command: /usr/bin/systemctl enable ceph-volume@lvm-2-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3 [node01][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-2-5f153e7a-85a7-446d-a7ec-ef2efec2b3a3.service to /usr/lib/systemd/system/ceph-volume@.service. [node01][WARNIN] Running command: /usr/bin/systemctl enable --runtime ceph-osd@2 [node01][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@2.service to /usr/lib/systemd/system/ceph-osd@.service. [node01][WARNIN] Running command: /usr/bin/systemctl start ceph-osd@2 [node01][WARNIN] --> ceph-volume lvm activate successful for osd ID: 2 [node01][WARNIN] --> ceph-volume lvm create successful for: /dev/sdd [node01][INFO ] checking OSD status... [node01][DEBUG ] find the location of an executable [node01][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node01 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 40m) mgr: no daemons active osd: 3 osds: 3 up (since 7s), 3 in (since 7s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdb [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node02 --data /dev/sdb [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7ff2fe446f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node02 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7ff2fecc62a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdb [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][WARNIN] osd keyring does not exist yet, creating one [node02][DEBUG ] create a keyring file [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 381b6597-3d41-41bd-8304-cc485c348b52 [node02][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb /dev/sdb [node02][WARNIN] stdout: Physical volume "/dev/sdb" successfully created. [node02][WARNIN] stdout: Volume group "ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb" successfully created [node02][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-381b6597-3d41-41bd-8304-cc485c348b52 ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb [node02][WARNIN] stdout: Logical volume "osd-block-381b6597-3d41-41bd-8304-cc485c348b52" created. [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-3 [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb/osd-block-381b6597-3d41-41bd-8304-cc485c348b52 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-2 [node02][WARNIN] Running command: /bin/ln -s /dev/ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb/osd-block-381b6597-3d41-41bd-8304-cc485c348b52 /var/lib/ceph/osd/ceph-3/block [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-3/activate.monmap [node02][WARNIN] stderr: 2020-11-12 15:39:13.792 7f902f899700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node02][WARNIN] 2020-11-12 15:39:13.792 7f902f899700 -1 AuthRegistry(0x7f9028065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node02][WARNIN] stderr: got monmap epoch 2 [node02][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-3/keyring --create-keyring --name osd.3 --add-key AQCg5qxfQuI4KRAAAL0R+JKWEhrhYXoUEIsIew== [node02][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-3/keyring [node02][WARNIN] added entity osd.3 auth(key=AQCg5qxfQuI4KRAAAL0R+JKWEhrhYXoUEIsIew==) [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/keyring [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/ [node02][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 3 --monmap /var/lib/ceph/osd/ceph-3/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-3/ --osd-uuid 381b6597-3d41-41bd-8304-cc485c348b52 --setuser ceph --setgroup ceph [node02][WARNIN] stderr: 2020-11-12 15:39:14.975 7fb218602a80 -1 bluestore(/var/lib/ceph/osd/ceph-3/) _read_fsid unparsable uuid [node02][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdb [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 [node02][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb/osd-block-381b6597-3d41-41bd-8304-cc485c348b52 --path /var/lib/ceph/osd/ceph-3 --no-mon-config [node02][WARNIN] Running command: /bin/ln -snf /dev/ceph-91956a7c-5375-4dea-81c0-5cb9e2f514fb/osd-block-381b6597-3d41-41bd-8304-cc485c348b52 /var/lib/ceph/osd/ceph-3/block [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-3/block [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-2 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 [node02][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-3-381b6597-3d41-41bd-8304-cc485c348b52 [node02][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-3-381b6597-3d41-41bd-8304-cc485c348b52.service to /usr/lib/systemd/system/ceph-volume@.service. [node02][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@3 [node02][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@3.service to /usr/lib/systemd/system/ceph-osd@.service. [node02][WARNIN] Running command: /bin/systemctl start ceph-osd@3 [node02][WARNIN] --> ceph-volume lvm activate successful for osd ID: 3 [node02][WARNIN] --> ceph-volume lvm create successful for: /dev/sdb [node02][INFO ] checking OSD status... [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node02 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 41m) mgr: no daemons active osd: 4 osds: 4 up (since 8s), 4 in (since 8s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdc [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node02 --data /dev/sdc [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7ff0f7ec6f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node02 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7ff0f87462a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdc [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdc [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdc [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 [node02][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0 /dev/sdc [node02][WARNIN] stdout: Physical volume "/dev/sdc" successfully created. [node02][WARNIN] stdout: Volume group "ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0" successfully created [node02][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0 [node02][WARNIN] stdout: Logical volume "osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203" created. [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-4 [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0/osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3 [node02][WARNIN] Running command: /bin/ln -s /dev/ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0/osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 /var/lib/ceph/osd/ceph-4/block [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-4/activate.monmap [node02][WARNIN] stderr: 2020-11-12 15:39:57.360 7fb72422a700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node02][WARNIN] 2020-11-12 15:39:57.360 7fb72422a700 -1 AuthRegistry(0x7fb71c065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node02][WARNIN] stderr: got monmap epoch 2 [node02][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-4/keyring --create-keyring --name osd.4 --add-key AQDM5qxf6FjpGBAArvaVdfRxM7iMqMB9OI10Dw== [node02][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-4/keyring [node02][WARNIN] added entity osd.4 auth(key=AQDM5qxf6FjpGBAArvaVdfRxM7iMqMB9OI10Dw==) [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4/keyring [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4/ [node02][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 4 --monmap /var/lib/ceph/osd/ceph-4/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-4/ --osd-uuid b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 --setuser ceph --setgroup ceph [node02][WARNIN] stderr: 2020-11-12 15:39:57.764 7f3698667a80 -1 bluestore(/var/lib/ceph/osd/ceph-4/) _read_fsid unparsable uuid [node02][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdc [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 [node02][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0/osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 --path /var/lib/ceph/osd/ceph-4 --no-mon-config [node02][WARNIN] Running command: /bin/ln -snf /dev/ceph-9aeac6d5-82eb-425c-9b8f-44d075b34bc0/osd-block-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 /var/lib/ceph/osd/ceph-4/block [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-4/block [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 [node02][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-4-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203 [node02][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-4-b8e2c607-aed1-45a0-a93e-8c6a2ec8e203.service to /usr/lib/systemd/system/ceph-volume@.service. [node02][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@4 [node02][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@4.service to /usr/lib/systemd/system/ceph-osd@.service. [node02][WARNIN] Running command: /bin/systemctl start ceph-osd@4 [node02][WARNIN] --> ceph-volume lvm activate successful for osd ID: 4 [node02][WARNIN] --> ceph-volume lvm create successful for: /dev/sdc [node02][INFO ] checking OSD status... [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node02 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 42m) mgr: no daemons active osd: 5 osds: 5 up (since 5s), 5 in (since 5s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node02 --data /dev/sdd [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node02 --data /dev/sdd [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fc1aef61f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node02 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7fc1af7e12a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdd [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdd [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [node02][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdd [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new c8049801-9529-406e-a8bf-35664a49d762 [node02][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40 /dev/sdd [node02][WARNIN] stdout: Physical volume "/dev/sdd" successfully created. [node02][WARNIN] stdout: Volume group "ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40" successfully created [node02][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-c8049801-9529-406e-a8bf-35664a49d762 ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40 [node02][WARNIN] stdout: Logical volume "osd-block-c8049801-9529-406e-a8bf-35664a49d762" created. [node02][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node02][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-5 [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40/osd-block-c8049801-9529-406e-a8bf-35664a49d762 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-4 [node02][WARNIN] Running command: /bin/ln -s /dev/ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40/osd-block-c8049801-9529-406e-a8bf-35664a49d762 /var/lib/ceph/osd/ceph-5/block [node02][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-5/activate.monmap [node02][WARNIN] stderr: 2020-11-12 15:40:35.710 7f34b5f69700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node02][WARNIN] 2020-11-12 15:40:35.710 7f34b5f69700 -1 AuthRegistry(0x7f34b0065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node02][WARNIN] stderr: got monmap epoch 2 [node02][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-5/keyring --create-keyring --name osd.5 --add-key AQDy5qxfT8bsKhAA7mF48viHKgy1j21rpIuJhQ== [node02][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-5/keyring [node02][WARNIN] added entity osd.5 auth(key=AQDy5qxfT8bsKhAA7mF48viHKgy1j21rpIuJhQ==) [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5/keyring [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5/ [node02][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 5 --monmap /var/lib/ceph/osd/ceph-5/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-5/ --osd-uuid c8049801-9529-406e-a8bf-35664a49d762 --setuser ceph --setgroup ceph [node02][WARNIN] stderr: 2020-11-12 15:40:37.286 7f5f98187a80 -1 bluestore(/var/lib/ceph/osd/ceph-5/) _read_fsid unparsable uuid [node02][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdd [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 [node02][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40/osd-block-c8049801-9529-406e-a8bf-35664a49d762 --path /var/lib/ceph/osd/ceph-5 --no-mon-config [node02][WARNIN] Running command: /bin/ln -snf /dev/ceph-db29c417-b9ce-4b5c-bfe3-ee5cab54da40/osd-block-c8049801-9529-406e-a8bf-35664a49d762 /var/lib/ceph/osd/ceph-5/block [node02][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-5/block [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-4 [node02][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 [node02][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-5-c8049801-9529-406e-a8bf-35664a49d762 [node02][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-5-c8049801-9529-406e-a8bf-35664a49d762.service to /usr/lib/systemd/system/ceph-volume@.service. [node02][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@5 [node02][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@5.service to /usr/lib/systemd/system/ceph-osd@.service. [node02][WARNIN] Running command: /bin/systemctl start ceph-osd@5 [node02][WARNIN] --> ceph-volume lvm activate successful for osd ID: 5 [node02][WARNIN] --> ceph-volume lvm create successful for: /dev/sdd [node02][INFO ] checking OSD status... [node02][DEBUG ] find the location of an executable [node02][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node02 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 43m) mgr: no daemons active osd: 6 osds: 6 up (since 6s), 6 in (since 6s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdb [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node03 --data /dev/sdb [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fb54f86ef80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node03 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7fb5500ee2a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdb [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][WARNIN] osd keyring does not exist yet, creating one [node03][DEBUG ] create a keyring file [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 4ce83292-8b3f-45c5-bb41-ddb3ef789a5c [node03][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-bd30de39-0962-4012-b54e-c353bf3f8330 /dev/sdb [node03][WARNIN] stdout: Physical volume "/dev/sdb" successfully created. [node03][WARNIN] stdout: Volume group "ceph-bd30de39-0962-4012-b54e-c353bf3f8330" successfully created [node03][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c ceph-bd30de39-0962-4012-b54e-c353bf3f8330 [node03][WARNIN] stdout: Logical volume "osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c" created. [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-6 [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-bd30de39-0962-4012-b54e-c353bf3f8330/osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-2 [node03][WARNIN] Running command: /bin/ln -s /dev/ceph-bd30de39-0962-4012-b54e-c353bf3f8330/osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c /var/lib/ceph/osd/ceph-6/block [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-6/activate.monmap [node03][WARNIN] stderr: 2020-11-12 15:41:19.876 7f2a5ab24700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node03][WARNIN] 2020-11-12 15:41:19.876 7f2a5ab24700 -1 AuthRegistry(0x7f2a54065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node03][WARNIN] stderr: got monmap epoch 2 [node03][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-6/keyring --create-keyring --name osd.6 --add-key AQAf56xfAwafBhAANdJIGFiX7+tbdQttMqLobg== [node03][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-6/keyring [node03][WARNIN] added entity osd.6 auth(key=AQAf56xfAwafBhAANdJIGFiX7+tbdQttMqLobg==) [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6/keyring [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6/ [node03][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 6 --monmap /var/lib/ceph/osd/ceph-6/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-6/ --osd-uuid 4ce83292-8b3f-45c5-bb41-ddb3ef789a5c --setuser ceph --setgroup ceph [node03][WARNIN] stderr: 2020-11-12 15:41:20.528 7f608fe1fa80 -1 bluestore(/var/lib/ceph/osd/ceph-6/) _read_fsid unparsable uuid [node03][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdb [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 [node03][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-bd30de39-0962-4012-b54e-c353bf3f8330/osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c --path /var/lib/ceph/osd/ceph-6 --no-mon-config [node03][WARNIN] Running command: /bin/ln -snf /dev/ceph-bd30de39-0962-4012-b54e-c353bf3f8330/osd-block-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c /var/lib/ceph/osd/ceph-6/block [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-6/block [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-2 [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 [node03][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-6-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c [node03][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-6-4ce83292-8b3f-45c5-bb41-ddb3ef789a5c.service to /usr/lib/systemd/system/ceph-volume@.service. [node03][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@6 [node03][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@6.service to /usr/lib/systemd/system/ceph-osd@.service. [node03][WARNIN] Running command: /bin/systemctl start ceph-osd@6 [node03][WARNIN] --> ceph-volume lvm activate successful for osd ID: 6 [node03][WARNIN] --> ceph-volume lvm create successful for: /dev/sdb [node03][INFO ] checking OSD status... [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node03 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 43m) mgr: no daemons active osd: 7 osds: 7 up (since 6s), 7 in (since 6s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdc [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node03 --data /dev/sdc [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fec94eaef80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node03 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7fec9572e2a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdc [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdc [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdc [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 714d5d28-07a9-41f4-b507-cc68d5515c9f [node03][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc /dev/sdc [node03][WARNIN] stdout: Physical volume "/dev/sdc" successfully created. [node03][WARNIN] stdout: Volume group "ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc" successfully created [node03][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc [node03][WARNIN] stdout: Logical volume "osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f" created. [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-7 [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc/osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3 [node03][WARNIN] Running command: /bin/ln -s /dev/ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc/osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f /var/lib/ceph/osd/ceph-7/block [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-7/activate.monmap [node03][WARNIN] stderr: 2020-11-12 15:42:04.999 7f0c45466700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node03][WARNIN] 2020-11-12 15:42:04.999 7f0c45466700 -1 AuthRegistry(0x7f0c40065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node03][WARNIN] stderr: got monmap epoch 2 [node03][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-7/keyring --create-keyring --name osd.7 --add-key AQBM56xfhXaRChAALfvZ8pWEKy/qEI/z2D1AnQ== [node03][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-7/keyring [node03][WARNIN] added entity osd.7 auth(key=AQBM56xfhXaRChAALfvZ8pWEKy/qEI/z2D1AnQ==) [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7/keyring [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7/ [node03][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 7 --monmap /var/lib/ceph/osd/ceph-7/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-7/ --osd-uuid 714d5d28-07a9-41f4-b507-cc68d5515c9f --setuser ceph --setgroup ceph [node03][WARNIN] stderr: 2020-11-12 15:42:05.342 7f96acbaca80 -1 bluestore(/var/lib/ceph/osd/ceph-7/) _read_fsid unparsable uuid [node03][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdc [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 [node03][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc/osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f --path /var/lib/ceph/osd/ceph-7 --no-mon-config [node03][WARNIN] Running command: /bin/ln -snf /dev/ceph-4e04f1f1-9018-40d4-ac70-e82c2d5aa9cc/osd-block-714d5d28-07a9-41f4-b507-cc68d5515c9f /var/lib/ceph/osd/ceph-7/block [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-7/block [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-3 [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 [node03][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-7-714d5d28-07a9-41f4-b507-cc68d5515c9f [node03][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-7-714d5d28-07a9-41f4-b507-cc68d5515c9f.service to /usr/lib/systemd/system/ceph-volume@.service. [node03][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@7 [node03][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@7.service to /usr/lib/systemd/system/ceph-osd@.service. [node03][WARNIN] Running command: /bin/systemctl start ceph-osd@7 [node03][WARNIN] --> ceph-volume lvm activate successful for osd ID: 7 [node03][WARNIN] --> ceph-volume lvm create successful for: /dev/sdc [node03][INFO ] checking OSD status... [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node03 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 44m) mgr: no daemons active osd: 8 osds: 8 up (since 5s), 8 in (since 5s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy osd create node03 --data /dev/sdd [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy osd create node03 --data /dev/sdd [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] bluestore : None [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f0618744f80> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] fs_type : xfs [ceph_deploy.cli][INFO ] block_wal : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] journal : None [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] host : node03 [ceph_deploy.cli][INFO ] filestore : None [ceph_deploy.cli][INFO ] func : <function osd at 0x7f0618fc42a8> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] zap_disk : False [ceph_deploy.cli][INFO ] data : /dev/sdd [ceph_deploy.cli][INFO ] block_db : None [ceph_deploy.cli][INFO ] dmcrypt : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] debug : False [ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdd [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [node03][DEBUG ] find the location of an executable [ceph_deploy.osd][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.osd][DEBUG ] Deploying osd to node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdd [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 5913569b-0783-4a5a-9157-bbf936bf9a88 [node03][WARNIN] Running command: /usr/sbin/vgcreate --force --yes ceph-b661755e-39af-425f-83c6-a908fb0ad2f9 /dev/sdd [node03][WARNIN] stdout: Physical volume "/dev/sdd" successfully created. [node03][WARNIN] stdout: Volume group "ceph-b661755e-39af-425f-83c6-a908fb0ad2f9" successfully created [node03][WARNIN] Running command: /usr/sbin/lvcreate --yes -l 2559 -n osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88 ceph-b661755e-39af-425f-83c6-a908fb0ad2f9 [node03][WARNIN] stdout: Logical volume "osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88" created. [node03][WARNIN] Running command: /bin/ceph-authtool --gen-print-key [node03][WARNIN] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-8 [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /dev/ceph-b661755e-39af-425f-83c6-a908fb0ad2f9/osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88 [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-4 [node03][WARNIN] Running command: /bin/ln -s /dev/ceph-b661755e-39af-425f-83c6-a908fb0ad2f9/osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88 /var/lib/ceph/osd/ceph-8/block [node03][WARNIN] Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-8/activate.monmap [node03][WARNIN] stderr: 2020-11-12 15:42:29.428 7f8722fc3700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory [node03][WARNIN] 2020-11-12 15:42:29.428 7f8722fc3700 -1 AuthRegistry(0x7f871c065a98) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx [node03][WARNIN] stderr: got monmap epoch 2 [node03][WARNIN] Running command: /bin/ceph-authtool /var/lib/ceph/osd/ceph-8/keyring --create-keyring --name osd.8 --add-key AQBk56xfIFKiIhAA3KEZc2bNrEHUH8cdb/0ZcQ== [node03][WARNIN] stdout: creating /var/lib/ceph/osd/ceph-8/keyring [node03][WARNIN] added entity osd.8 auth(key=AQBk56xfIFKiIhAA3KEZc2bNrEHUH8cdb/0ZcQ==) [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-8/keyring [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-8/ [node03][WARNIN] Running command: /bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 8 --monmap /var/lib/ceph/osd/ceph-8/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-8/ --osd-uuid 5913569b-0783-4a5a-9157-bbf936bf9a88 --setuser ceph --setgroup ceph [node03][WARNIN] stderr: 2020-11-12 15:42:30.155 7fdc18603a80 -1 bluestore(/var/lib/ceph/osd/ceph-8/) _read_fsid unparsable uuid [node03][WARNIN] --> ceph-volume lvm prepare successful for: /dev/sdd [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-8 [node03][WARNIN] Running command: /bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-b661755e-39af-425f-83c6-a908fb0ad2f9/osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88 --path /var/lib/ceph/osd/ceph-8 --no-mon-config [node03][WARNIN] Running command: /bin/ln -snf /dev/ceph-b661755e-39af-425f-83c6-a908fb0ad2f9/osd-block-5913569b-0783-4a5a-9157-bbf936bf9a88 /var/lib/ceph/osd/ceph-8/block [node03][WARNIN] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-8/block [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /dev/dm-4 [node03][WARNIN] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-8 [node03][WARNIN] Running command: /bin/systemctl enable ceph-volume@lvm-8-5913569b-0783-4a5a-9157-bbf936bf9a88 [node03][WARNIN] stderr: Created symlink from /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-8-5913569b-0783-4a5a-9157-bbf936bf9a88.service to /usr/lib/systemd/system/ceph-volume@.service. [node03][WARNIN] Running command: /bin/systemctl enable --runtime ceph-osd@8 [node03][WARNIN] stderr: Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@8.service to /usr/lib/systemd/system/ceph-osd@.service. [node03][WARNIN] Running command: /bin/systemctl start ceph-osd@8 [node03][WARNIN] --> ceph-volume lvm activate successful for osd ID: 8 [node03][WARNIN] --> ceph-volume lvm create successful for: /dev/sdd [node03][INFO ] checking OSD status... [node03][DEBUG ] find the location of an executable [node03][INFO ] Running command: /bin/ceph --cluster=ceph osd stat --format=json [ceph_deploy.osd][DEBUG ] Host node03 is now ready for osd use. [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_WARN no active mgr services: mon: 3 daemons, quorum node01,node02,node03 (age 44m) mgr: no daemons active osd: 9 osds: 9 up (since 5s), 9 in (since 5s) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: [root@node01 cluster]#
[root@node01 cluster]# ceph-deploy mgr create node01 node02 node03
[root@node01 cluster]# ceph-deploy rgw create node01 node02 node03
[root@node01 cluster]# netstat -tnlp|grep -aiE 6789
[root@node02 ~]# netstat -tnlp|grep -aiE 6789
[root@node03 ~]# netstat -tnlp|grep -aiE 6789
[root@node01 cluster]# ceph-deploy mgr create node01 node02 node03 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mgr create node01 node02 node03 [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] mgr : [('node01', 'node01'), ('node02', 'node02'), ('node03', 'node03')] [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f167b3d5200> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] func : <function mgr at 0x7f167bc42aa0> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.mgr][DEBUG ] Deploying mgr, cluster ceph hosts node01:node01 node02:node02 node03:node03 [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [ceph_deploy.mgr][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.mgr][DEBUG ] remote host will use systemd [ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][WARNIN] mgr keyring does not exist yet, creating one [node01][DEBUG ] create a keyring file [node01][DEBUG ] create path recursively if it doesn't exist [node01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.node01 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-node01/keyring [node01][INFO ] Running command: systemctl enable ceph-mgr@node01 [node01][WARNIN] Created symlink from /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@node01.service to /usr/lib/systemd/system/ceph-mgr@.service. [node01][INFO ] Running command: systemctl start ceph-mgr@node01 [node01][INFO ] Running command: systemctl enable ceph.target [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [ceph_deploy.mgr][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.mgr][DEBUG ] remote host will use systemd [ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][WARNIN] mgr keyring does not exist yet, creating one [node02][DEBUG ] create a keyring file [node02][DEBUG ] create path recursively if it doesn't exist [node02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.node02 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-node02/keyring [node02][INFO ] Running command: systemctl enable ceph-mgr@node02 [node02][WARNIN] Created symlink from /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@node02.service to /usr/lib/systemd/system/ceph-mgr@.service. [node02][INFO ] Running command: systemctl start ceph-mgr@node02 [node02][INFO ] Running command: systemctl enable ceph.target [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [ceph_deploy.mgr][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.mgr][DEBUG ] remote host will use systemd [ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][WARNIN] mgr keyring does not exist yet, creating one [node03][DEBUG ] create a keyring file [node03][DEBUG ] create path recursively if it doesn't exist [node03][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.node03 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-node03/keyring [node03][INFO ] Running command: systemctl enable ceph-mgr@node03 [node03][WARNIN] Created symlink from /etc/systemd/system/ceph-mgr.target.wants/ceph-mgr@node03.service to /usr/lib/systemd/system/ceph-mgr@.service. [node03][INFO ] Running command: systemctl start ceph-mgr@node03 [node03][INFO ] Running command: systemctl enable ceph.target [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 47m) mgr: node01(active, since 9s), standbys: node03, node02 osd: 9 osds: 9 up (since 3m), 9 in (since 3m) data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 9.0 GiB used, 81 GiB / 90 GiB avail pgs: [root@node01 cluster]# [root@node01 cluster]# ceph-deploy rgw create node01 node02 node03 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy rgw create node01 node02 node03 [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] rgw : [('node01', 'rgw.node01'), ('node02', 'rgw.node02'), ('node03', 'rgw.node03')] [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] subcommand : create [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fa344a4a5f0> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] func : <function rgw at 0x7fa3455169b0> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts node01:rgw.node01 node02:rgw.node02 node03:rgw.node03 [node01][DEBUG ] connected to host: node01 [node01][DEBUG ] detect platform information from remote host [node01][DEBUG ] detect machine type [ceph_deploy.rgw][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.rgw][DEBUG ] remote host will use systemd [ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to node01 [node01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node01][WARNIN] rgw keyring does not exist yet, creating one [node01][DEBUG ] create a keyring file [node01][DEBUG ] create path recursively if it doesn't exist [node01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.node01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.node01/keyring [node01][INFO ] Running command: systemctl enable ceph-radosgw@rgw.node01 [node01][WARNIN] Created symlink from /etc/systemd/system/ceph-radosgw.target.wants/ceph-radosgw@rgw.node01.service to /usr/lib/systemd/system/ceph-radosgw@.service. [node01][INFO ] Running command: systemctl start ceph-radosgw@rgw.node01 [node01][INFO ] Running command: systemctl enable ceph.target [ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host node01 and default port 7480 [node02][DEBUG ] connected to host: node02 [node02][DEBUG ] detect platform information from remote host [node02][DEBUG ] detect machine type [ceph_deploy.rgw][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.rgw][DEBUG ] remote host will use systemd [ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to node02 [node02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node02][WARNIN] rgw keyring does not exist yet, creating one [node02][DEBUG ] create a keyring file [node02][DEBUG ] create path recursively if it doesn't exist [node02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.node02 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.node02/keyring [node02][INFO ] Running command: systemctl enable ceph-radosgw@rgw.node02 [node02][WARNIN] Created symlink from /etc/systemd/system/ceph-radosgw.target.wants/ceph-radosgw@rgw.node02.service to /usr/lib/systemd/system/ceph-radosgw@.service. [node02][INFO ] Running command: systemctl start ceph-radosgw@rgw.node02 [node02][INFO ] Running command: systemctl enable ceph.target [ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host node02 and default port 7480 [node03][DEBUG ] connected to host: node03 [node03][DEBUG ] detect platform information from remote host [node03][DEBUG ] detect machine type [ceph_deploy.rgw][INFO ] Distro info: CentOS Linux 7.6.1810 Core [ceph_deploy.rgw][DEBUG ] remote host will use systemd [ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to node03 [node03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [node03][WARNIN] rgw keyring does not exist yet, creating one [node03][DEBUG ] create a keyring file [node03][DEBUG ] create path recursively if it doesn't exist [node03][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.node03 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.node03/keyring [node03][INFO ] Running command: systemctl enable ceph-radosgw@rgw.node03 [node03][WARNIN] Created symlink from /etc/systemd/system/ceph-radosgw.target.wants/ceph-radosgw@rgw.node03.service to /usr/lib/systemd/system/ceph-radosgw@.service. [node03][INFO ] Running command: systemctl start ceph-radosgw@rgw.node03 [node03][INFO ] Running command: systemctl enable ceph.target [ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host node03 and default port 7480 [root@node01 cluster]# [root@node01 cluster]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 49m) mgr: node01(active, since 110s), standbys: node03, node02 osd: 9 osds: 9 up (since 4m), 9 in (since 4m) rgw: 3 daemons active (node01, node02, node03) task status: data: pools: 4 pools, 128 pgs objects: 187 objects, 1.2 KiB usage: 9.1 GiB used, 81 GiB / 90 GiB avail pgs: 128 active+clean io: client: 73 KiB/s rd, 0 B/s wr, 80 op/s rd, 53 op/s wr [root@node01 cluster]# [root@node01 cluster]# netstat -tnlp|grep -aiE 6789 tcp 0 0 192.168.1.181:6789 0.0.0.0:* LISTEN 9811/ceph-mon [root@node01 cluster]# [root@node02 ~]# netstat -tnlp|grep -aiE 6789 tcp 0 0 192.168.1.182:6789 0.0.0.0:* LISTEN 9596/ceph-mon [root@node02 ~]# [root@node03 ~]# netstat -tnlp|grep -aiE 6789 tcp 0 0 192.168.1.183:6789 0.0.0.0:* LISTEN 9600/ceph-mon [root@node03 ~]#
[root@node01 ~]# ps -ef|grep ceph
[root@node01 ~]# netstat -tnlp
[root@node01 ~]# ceph -s
[root@node01 ~]# ps -ef|grep ceph root 9206 1 0 14:57 ? 00:00:00 /usr/bin/python2.7 /usr/bin/ceph-crash ceph 9811 1 0 14:57 ? 00:00:13 /usr/bin/ceph-mon -f --cluster ceph --id node01 --setuser ceph --setgroup ceph ceph 11775 1 0 15:34 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 0 --setuser ceph --setgroup ceph ceph 12731 1 0 15:37 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 1 --setuser ceph --setgroup ceph ceph 13232 1 0 15:38 ? 00:00:07 /usr/bin/ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup ceph ceph 13762 1 1 15:45 ? 00:00:17 /usr/bin/ceph-mgr -f --cluster ceph --id node01 --setuser ceph --setgroup ceph ceph 14008 1 0 15:46 ? 00:00:02 /usr/bin/radosgw -f --cluster ceph --name client.rgw.node01 --setuser ceph --setgroup ceph root 15116 10741 0 16:11 pts/0 00:00:00 grep --color=auto ceph [root@node01 ~]# [root@node01 ~]# netstat -tnlp Active Internet connections (only servers) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:6823 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 0.0.0.0:6824 0.0.0.0:* LISTEN 13762/ceph-mgr tcp 0 0 0.0.0.0:6825 0.0.0.0:* LISTEN 13762/ceph-mgr tcp 0 0 192.168.1.181:2379 0.0.0.0:* LISTEN 9817/etcd tcp 0 0 127.0.0.1:2379 0.0.0.0:* LISTEN 9817/etcd tcp 0 0 192.168.1.181:2380 0.0.0.0:* LISTEN 9817/etcd tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN 1/systemd tcp 0 0 0.0.0.0:6800 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6000 0.0.0.0:* LISTEN 10120/X tcp 0 0 0.0.0.0:6801 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6802 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6803 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6804 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6805 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:6806 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 9812/sshd tcp 0 0 0.0.0.0:6807 0.0.0.0:* LISTEN 11775/ceph-osd tcp 0 0 127.0.0.1:631 0.0.0.0:* LISTEN 9815/cupsd tcp 0 0 0.0.0.0:7480 0.0.0.0:* LISTEN 14008/radosgw tcp 0 0 0.0.0.0:6808 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6809 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN 10059/master tcp 0 0 0.0.0.0:6810 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6811 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6812 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6813 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6814 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6815 0.0.0.0:* LISTEN 12731/ceph-osd tcp 0 0 0.0.0.0:6816 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 0.0.0.0:6817 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 0.0.0.0:6818 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 0.0.0.0:6819 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 0.0.0.0:6820 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 192.168.1.181:3300 0.0.0.0:* LISTEN 9811/ceph-mon tcp 0 0 0.0.0.0:6821 0.0.0.0:* LISTEN 13232/ceph-osd tcp 0 0 192.168.1.181:6789 0.0.0.0:* LISTEN 9811/ceph-mon tcp 0 0 0.0.0.0:6822 0.0.0.0:* LISTEN 13232/ceph-osd tcp6 0 0 :::6443 :::* LISTEN 10467/kube-apiserve tcp6 0 0 :::10251 :::* LISTEN 9186/kube-scheduler tcp6 0 0 :::10252 :::* LISTEN 9250/kube-controlle tcp6 0 0 :::111 :::* LISTEN 1/systemd tcp6 0 0 :::8080 :::* LISTEN 10467/kube-apiserve tcp6 0 0 :::6000 :::* LISTEN 10120/X tcp6 0 0 :::22 :::* LISTEN 9812/sshd tcp6 0 0 ::1:631 :::* LISTEN 9815/cupsd tcp6 0 0 :::7480 :::* LISTEN 14008/radosgw tcp6 0 0 ::1:25 :::* LISTEN 10059/master [root@node01 ~]# [root@node01 ~]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 73m) mgr: node01(active, since 25m), standbys: node03, node02 osd: 9 osds: 9 up (since 28m), 9 in (since 28m) rgw: 3 daemons active (node01, node02, node03) task status: data: pools: 4 pools, 128 pgs objects: 187 objects, 1.2 KiB usage: 9.1 GiB used, 81 GiB / 90 GiB avail pgs: 128 active+clean [root@node01 ~]#
[root@node02 ~]# ps -ef|grep ceph
[root@node02 ~]# netstat -tnlp
[root@node02 ~]# ceph -s
[root@node02 ~]# ps -ef|grep ceph root 9142 1 0 14:57 ? 00:00:00 /usr/bin/python2.7 /usr/bin/ceph-crash ceph 9596 1 0 14:57 ? 00:00:09 /usr/bin/ceph-mon -f --cluster ceph --id node02 --setuser ceph --setgroup ceph ceph 48042 1 0 15:39 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 3 --setuser ceph --setgroup ceph ceph 48997 1 0 15:39 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 4 --setuser ceph --setgroup ceph ceph 49927 1 0 15:40 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 5 --setuser ceph --setgroup ceph ceph 53291 1 0 15:45 ? 00:00:04 /usr/bin/ceph-mgr -f --cluster ceph --id node02 --setuser ceph --setgroup ceph ceph 54454 1 0 15:47 ? 00:00:03 /usr/bin/radosgw -f --cluster ceph --name client.rgw.node02 --setuser ceph --setgroup ceph root 70838 21557 0 16:11 pts/0 00:00:00 grep --color=auto ceph [root@node02 ~]# [root@node02 ~]# netstat -tnlp Active Internet connections (only servers) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:6823 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 127.0.0.1:10248 0.0.0.0:* LISTEN 10235/kubelet tcp 0 0 127.0.0.1:10249 0.0.0.0:* LISTEN 9605/kube-proxy tcp 0 0 192.168.1.182:2379 0.0.0.0:* LISTEN 9603/etcd tcp 0 0 127.0.0.1:2379 0.0.0.0:* LISTEN 9603/etcd tcp 0 0 192.168.1.182:2380 0.0.0.0:* LISTEN 9603/etcd tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN 1/systemd tcp 0 0 0.0.0.0:6800 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6000 0.0.0.0:* LISTEN 9918/X tcp 0 0 0.0.0.0:6801 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6802 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6803 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6804 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6805 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:6806 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 9594/sshd tcp 0 0 0.0.0.0:6807 0.0.0.0:* LISTEN 48042/ceph-osd tcp 0 0 127.0.0.1:631 0.0.0.0:* LISTEN 9600/cupsd tcp 0 0 0.0.0.0:7480 0.0.0.0:* LISTEN 54454/radosgw tcp 0 0 0.0.0.0:6808 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6809 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN 10038/master tcp 0 0 0.0.0.0:6810 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6811 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6812 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6813 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6814 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6815 0.0.0.0:* LISTEN 48997/ceph-osd tcp 0 0 0.0.0.0:6816 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 0.0.0.0:6817 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 0.0.0.0:6818 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 0.0.0.0:6819 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 0.0.0.0:6820 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 192.168.1.182:3300 0.0.0.0:* LISTEN 9596/ceph-mon tcp 0 0 0.0.0.0:6821 0.0.0.0:* LISTEN 49927/ceph-osd tcp 0 0 192.168.1.182:6789 0.0.0.0:* LISTEN 9596/ceph-mon tcp 0 0 0.0.0.0:6822 0.0.0.0:* LISTEN 49927/ceph-osd tcp6 0 0 :::10250 :::* LISTEN 10235/kubelet tcp6 0 0 :::10255 :::* LISTEN 10235/kubelet tcp6 0 0 :::111 :::* LISTEN 1/systemd tcp6 0 0 :::6000 :::* LISTEN 9918/X tcp6 0 0 :::22 :::* LISTEN 9594/sshd tcp6 0 0 ::1:631 :::* LISTEN 9600/cupsd tcp6 0 0 :::7480 :::* LISTEN 54454/radosgw tcp6 0 0 ::1:25 :::* LISTEN 10038/master tcp6 0 0 :::4194 :::* LISTEN 10235/kubelet [root@node02 ~]# [root@node02 ~]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 73m) mgr: node01(active, since 25m), standbys: node03, node02 osd: 9 osds: 9 up (since 28m), 9 in (since 28m) rgw: 3 daemons active (node01, node02, node03) task status: data: pools: 4 pools, 128 pgs objects: 187 objects, 1.2 KiB usage: 9.1 GiB used, 81 GiB / 90 GiB avail pgs: 128 active+clean [root@node02 ~]#
[root@node03 ~]# ps -ef|grep ceph
[root@node03 ~]# netstat -tnlp
[root@node03 ~]# ceph -s
[root@node03 ~]# ps -ef|grep ceph root 9222 1 0 14:57 ? 00:00:00 /usr/bin/python2.7 /usr/bin/ceph-crash ceph 9600 1 0 14:57 ? 00:00:08 /usr/bin/ceph-mon -f --cluster ceph --id node03 --setuser ceph --setgroup ceph ceph 41148 1 0 15:41 ? 00:00:09 /usr/bin/ceph-osd -f --cluster ceph --id 6 --setuser ceph --setgroup ceph ceph 42129 1 0 15:42 ? 00:00:08 /usr/bin/ceph-osd -f --cluster ceph --id 7 --setuser ceph --setgroup ceph ceph 42893 1 0 15:42 ? 00:00:07 /usr/bin/ceph-osd -f --cluster ceph --id 8 --setuser ceph --setgroup ceph ceph 45082 1 0 15:45 ? 00:00:02 /usr/bin/ceph-mgr -f --cluster ceph --id node03 --setuser ceph --setgroup ceph ceph 46277 1 0 15:47 ? 00:00:03 /usr/bin/radosgw -f --cluster ceph --name client.rgw.node03 --setuser ceph --setgroup ceph root 62692 11703 0 16:11 pts/0 00:00:00 grep --color=auto ceph [root@node03 ~]# [root@node03 ~]# netstat -tnlp Active Internet connections (only servers) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:6823 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 127.0.0.1:10248 0.0.0.0:* LISTEN 10326/kubelet tcp 0 0 127.0.0.1:10249 0.0.0.0:* LISTEN 9591/kube-proxy tcp 0 0 192.168.1.183:2379 0.0.0.0:* LISTEN 62633/etcd tcp 0 0 127.0.0.1:2379 0.0.0.0:* LISTEN 62633/etcd tcp 0 0 127.0.0.1:2380 0.0.0.0:* LISTEN 62633/etcd tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN 1/systemd tcp 0 0 0.0.0.0:6800 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6000 0.0.0.0:* LISTEN 9735/X tcp 0 0 0.0.0.0:6801 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6802 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6803 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6804 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6805 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:6806 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 9587/sshd tcp 0 0 0.0.0.0:6807 0.0.0.0:* LISTEN 41148/ceph-osd tcp 0 0 127.0.0.1:631 0.0.0.0:* LISTEN 9584/cupsd tcp 0 0 0.0.0.0:7480 0.0.0.0:* LISTEN 46277/radosgw tcp 0 0 0.0.0.0:6808 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6809 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 127.0.0.1:25 0.0.0.0:* LISTEN 9757/master tcp 0 0 0.0.0.0:6810 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6811 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6812 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6813 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6814 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6815 0.0.0.0:* LISTEN 42129/ceph-osd tcp 0 0 0.0.0.0:6816 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 0.0.0.0:6817 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 0.0.0.0:6818 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 0.0.0.0:6819 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 0.0.0.0:6820 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 192.168.1.183:3300 0.0.0.0:* LISTEN 9600/ceph-mon tcp 0 0 0.0.0.0:6821 0.0.0.0:* LISTEN 42893/ceph-osd tcp 0 0 192.168.1.183:6789 0.0.0.0:* LISTEN 9600/ceph-mon tcp 0 0 0.0.0.0:6822 0.0.0.0:* LISTEN 42893/ceph-osd tcp6 0 0 :::10250 :::* LISTEN 10326/kubelet tcp6 0 0 :::10255 :::* LISTEN 10326/kubelet tcp6 0 0 :::111 :::* LISTEN 1/systemd tcp6 0 0 :::6000 :::* LISTEN 9735/X tcp6 0 0 :::22 :::* LISTEN 9587/sshd tcp6 0 0 ::1:631 :::* LISTEN 9584/cupsd tcp6 0 0 :::7480 :::* LISTEN 46277/radosgw tcp6 0 0 ::1:25 :::* LISTEN 9757/master tcp6 0 0 :::4194 :::* LISTEN 10326/kubelet [root@node03 ~]# [root@node03 ~]# ceph -s cluster: id: c90f3c76-20f0-4091-a66b-74ec0e6f4ec8 health: HEALTH_OK services: mon: 3 daemons, quorum node01,node02,node03 (age 73m) mgr: node01(active, since 25m), standbys: node03, node02 osd: 9 osds: 9 up (since 28m), 9 in (since 28m) rgw: 3 daemons active (node01, node02, node03) task status: data: pools: 4 pools, 128 pgs objects: 187 objects, 1.2 KiB usage: 9.1 GiB used, 81 GiB / 90 GiB avail pgs: 128 active+clean [root@node03 ~]#