.. _add_ceph_mds: ================== Ceph集群添加MDS ================== 现在部署 ``ceph-mds`` 服务,对外提供POSIX兼容元数据 - 在 ``z-b-data-1`` 上执行:: cluster=ceph id=z-b-data-1 sudo mkdir -p /var/lib/ceph/mds/${cluster}-${id} - 创建keyring:: sudo ceph-authtool --create-keyring /var/lib/ceph/mds/${cluster}-${id}/keyring --gen-key -n mds.${id} sudo chown -R ceph:ceph /var/lib/ceph/mds/${cluster}-${id} - 导入keyring和设置caps:: sudo ceph auth add mds.${id} osd "allow rwx" mds "allow *" mon "allow profile mds" -i /var/lib/ceph/mds/${cluster}-${id}/keyring - 启动服务:: sudo systemctl start ceph-mds@${id} sudo systemctl status ceph-mds@${id} 可以看到:: ● ceph-mds@z-b-data-1.service - Ceph metadata server daemon Loaded: loaded (/lib/systemd/system/ceph-mds@.service; disabled; vendor preset: enabled) Active: active (running) since Sun 2021-12-05 17:42:46 CST; 5s ago Main PID: 4700 (ceph-mds) Tasks: 15 Memory: 10.4M CGroup: /system.slice/system-ceph\x2dmds.slice/ceph-mds@z-b-data-1.service └─4700 /usr/bin/ceph-mds -f --cluster ceph --id z-b-data-1 --setuser ceph --setgroup ceph Dec 05 17:42:46 z-b-data-1 systemd[1]: Started Ceph metadata server daemon. Dec 05 17:42:46 z-b-data-1 ceph-mds[4700]: starting mds.z-b-data-1 at - 检查集群状态:: sudo ceph -s 显示:: cluster: id: 0e6c8b6f-0d32-4cdb-a45d-85f8c7997c17 health: HEALTH_OK services: mon: 3 daemons, quorum z-b-data-1,z-b-data-2,z-b-data-3 (age 19h) mgr: z-b-data-1(active, since 2d) mds: 1 up:standby osd: 3 osds: 3 up (since 18h), 3 in (since 18h) data: pools: 1 pools, 1 pgs objects: 3 objects, 0 B usage: 3.0 GiB used, 1.4 TiB / 1.4 TiB avail pgs: 1 active+clean 添加另外2个MDS ==================== - 修改每个服务器上 ``/etc/ceph/ceph.conf`` 配置,添加:: [mds.z-b-data-1] host = 192.168.6.204 [mds.z-b-data-2] host = 192.168.6.205 [mds.z-b-data-3] host = 192.168.6.206 然后重启每个服务器上 ``ceph-mon`` :: sudo systemctl restart ceph-mon@`hostname` - 在 ``z-b-data-2`` 和 ``z-b-data-3`` 上执行以下命令将 ``z-b-data-1`` 主机上 keyring 复制过来(这里举例是 ``z-b-data-2`` ):: cluster=ceph id=z-b-data-2 sudo mkdir /var/lib/ceph/mds/${cluster}-${id} sudo ceph-authtool --create-keyring /var/lib/ceph/mds/${cluster}-${id}/keyring --gen-key -n mds.${id} sudo chown -R ceph:ceph /var/lib/ceph/mds/${cluster}-${id} - 导入keyring和设置caps:: sudo ceph auth add mds.${id} osd "allow rwx" mds "allow *" mon "allow profile mds" -i /var/lib/ceph/mds/${cluster}-${id}/keyring - 启动服务器:: sudo systemctl start ceph-mds@${id} sudo systemctl enable ceph-mds@${id} 同样在 ``z-b-data-3`` 上完成上述操作 MDS检查 ========== 最后完成,执行状态检查:: sudo ceph -s 可以看到状态:: cluster: id: 0e6c8b6f-0d32-4cdb-a45d-85f8c7997c17 health: HEALTH_OK services: mon: 3 daemons, quorum z-b-data-1,z-b-data-2,z-b-data-3 (age 21m) mgr: z-b-data-1(active, since 29m) mds: 3 up:standby osd: 3 osds: 3 up (since 21m), 3 in (since 23h) data: pools: 1 pools, 1 pgs objects: 3 objects, 0 B usage: 3.0 GiB used, 1.4 TiB / 1.4 TiB avail pgs: 1 active+clean 上述状态中 ``services`` :: mds: 3 up:standby 根据 `Ceph文档MDS STATES `_ - 检查文件系统:: sudo ceph fs dump 显示:: dumped fsmap epoch 11 e11 enable_multiple, ever_enabled_multiple: 0,0 compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2} legacy client fscid: -1 No filesystems configured Standby daemons: [mds.z-b-data-1{-1:94320} state up:standby seq 1 addr [v2:192.168.6.204:6810/3515377620,v1:192.168.6.204:6811/3515377620]] [mds.z-b-data-2{-1:104110} state up:standby seq 1 addr [v2:192.168.6.205:6808/1674656387,v1:192.168.6.205:6809/1674656387]] [mds.z-b-data-3{-1:114122} state up:standby seq 2 addr [v2:192.168.6.206:6808/3637086204,v1:192.168.6.206:6809/3637086204]] 参考 ========= - `CEPH by hand `_