天天看點

CEPH 添加OSD操作

[talen@ceph_admin ~]$ ceph-deploy osd prepare ceph_monitor:/osd

[ceph_deploy.conf][DEBUG ] found configuration file at: /home/talen/.cephdeploy.conf

[ceph_deploy.cli][INFO  ] Invoked (1.5.28): /bin/ceph-deploy osd prepare ceph_monitor:/osd

[ceph_deploy.cli][INFO  ] ceph-deploy options:

[ceph_deploy.cli][INFO  ]  username                      : None

[ceph_deploy.cli][INFO  ]  disk                          : [('ceph_monitor', '/osd', None)]

[ceph_deploy.cli][INFO  ]  dmcrypt                       : False

[ceph_deploy.cli][INFO  ]  verbose                       : False

[ceph_deploy.cli][INFO  ]  overwrite_conf                : False

[ceph_deploy.cli][INFO  ]  subcommand                    : prepare

[ceph_deploy.cli][INFO  ]  dmcrypt_key_dir               : /etc/ceph/dmcrypt-keys

[ceph_deploy.cli][INFO  ]  quiet                         : False

[ceph_deploy.cli][INFO  ]  cd_conf                       :

[ceph_deploy.cli][INFO  ]  cluster                       : ceph

[ceph_deploy.cli][INFO  ]  fs_type                       : xfs

[ceph_deploy.cli][INFO  ]  func                          :

[ceph_deploy.cli][INFO  ]  ceph_conf                     : None

[ceph_deploy.cli][INFO  ]  default_release               : False

[ceph_deploy.cli][INFO  ]  zap_disk                      : False

[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks ceph_monitor:/osd:

[ceph_monitor][DEBUG ] connection detected need for sudo

[ceph_monitor][DEBUG ] connected to host: ceph_monitor

[ceph_monitor][DEBUG ] detect platform information from remote host

[ceph_monitor][DEBUG ] detect machine type

[ceph_monitor][DEBUG ] find the location of an executable

[ceph_deploy.osd][INFO  ] Distro info: CentOS Linux 7.1.1503 Core

[ceph_deploy.osd][DEBUG ] Deploying osd to ceph_monitor

[ceph_monitor][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf

[ceph_monitor][INFO  ] Running command: sudo udevadm trigger --subsystem-match=block --action=add

[ceph_deploy.osd][DEBUG ] Preparing host ceph_monitor disk /osd journal None activate False

[ceph_monitor][INFO  ] Running command: sudo ceph-disk -v prepare --cluster ceph --fs-type xfs -- /osd

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=fsid

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_mkfs_options_xfs

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_fs_mkfs_options_xfs

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_mount_options_xfs

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_fs_mount_options_xfs

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=osd_journal_size

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_cryptsetup_parameters

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_key_size

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_type

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Preparing osd data dir /osd

[ceph_monitor][INFO  ] checking OSD status...

[ceph_monitor][INFO  ] Running command: sudo ceph --cluster=ceph osd stat --format=json

[ceph_monitor][WARNIN] there are 5 OSDs down

[ceph_monitor][WARNIN] there are 5 OSDs out

[ceph_deploy.osd][DEBUG ] Host ceph_monitor is now ready for osd use.

[talen@ceph_admin ~]$

[talen@ceph_admin ~]$ ceph-deploy osd activate ceph_monitor:/osd

[ceph_deploy.cli][INFO  ] Invoked (1.5.28): /bin/ceph-deploy osd activate ceph_monitor:/osd

[ceph_deploy.cli][INFO  ]  subcommand                    : activate

[ceph_deploy.osd][DEBUG ] Activating cluster ceph disks ceph_monitor:/osd:

[ceph_deploy.osd][DEBUG ] activating host ceph_monitor disk /osd

[ceph_deploy.osd][DEBUG ] will use init type: sysvinit

[ceph_monitor][INFO  ] Running command: sudo ceph-disk -v activate --mark-init sysvinit --mount /osd

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Cluster uuid is 58514e13-d332-4a7e-9760-e3fccb9e2c76

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Cluster name is ceph

[ceph_monitor][WARNIN] DEBUG:ceph-disk:OSD uuid is 86491f86-a204-4cee-acb5-9d7a7c26f784

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Allocating OSD id...

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd create --concise 86491f86-a204-4cee-acb5-9d7a7c26f784

[ceph_monitor][WARNIN] DEBUG:ceph-disk:OSD id is 7

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Initializing OSD...

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /osd/activate.monmap

[ceph_monitor][WARNIN] got monmap epoch 1

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 7 --monmap /osd/activate.monmap --osd-data /osd --osd-journal /osd/journal --osd-uuid 86491f86-a204-4cee-acb5-9d7a7c26f784 --keyring /osd/keyring

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.300887 7f3da9997880 -1 journal FileJournal::_open: disabling aio for non-block journal.  Use journal_force_aio to force use of aio anyway

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.667066 7f3da9997880 -1 journal FileJournal::_open: disabling aio for non-block journal.  Use journal_force_aio to force use of aio anyway

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.669354 7f3da9997880 -1 filestore(/osd) could not find 23c2fcde/osd_superblock/0//-1 in index: (2) No such file or directory

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.917080 7f3da9997880 -1 created object store /osd journal /osd/journal for osd.7 fsid f35a65ad-1a6a-4e8d-8f7e-cb5f113c0a02

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.917200 7f3da9997880 -1 auth: error reading file: /osd/keyring: can't open /osd/keyring: (2) No such file or directory

[ceph_monitor][WARNIN] 2015-09-14 15:22:49.917469 7f3da9997880 -1 created new key in keyring /osd/keyring

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Marking with init system sysvinit

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Authorizing OSD key...

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring auth add osd.7 -i /osd/keyring osd allow * mon allow profile osd

[ceph_monitor][WARNIN] added key for osd.7

[ceph_monitor][WARNIN] DEBUG:ceph-disk:ceph osd.7 data dir is ready at /osd

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Creating symlink /var/lib/ceph/osd/ceph-7 -> /osd

[ceph_monitor][WARNIN] DEBUG:ceph-disk:Starting ceph osd.7...

[ceph_monitor][WARNIN] INFO:ceph-disk:Running command: /usr/sbin/service ceph --cluster ceph start osd.7

[ceph_monitor][DEBUG ] === osd.7 ===

[ceph_monitor][WARNIN] create-or-move updating item name 'osd.7' weight 0.01 at location {host=ceph_monitor,root=default} to crush map

[ceph_monitor][DEBUG ] Starting Ceph osd.7 on ceph_monitor...

[ceph_monitor][WARNIN] Running as unit run-12576.service.

[ceph_monitor][INFO  ] Running command: sudo systemctl enable ceph

[ceph_monitor][WARNIN] ceph.service is not a native service, redirecting to /sbin/chkconfig.

[ceph_monitor][WARNIN] Executing /sbin/chkconfig ceph on

[ceph_monitor][WARNIN] The unit files have no [Install] section. They are not meant to be enabled

[ceph_monitor][WARNIN] using systemctl.

[ceph_monitor][WARNIN] Possible reasons for having this kind of units are:

[ceph_monitor][WARNIN] 1) A unit may be statically enabled by being symlinked from another unit's

[ceph_monitor][WARNIN]    .wants/ or .requires/ directory.

[ceph_monitor][WARNIN] 2) A unit's purpose may be to act as a helper for some other unit which has

[ceph_monitor][WARNIN]    a requirement dependency on it.

[ceph_monitor][WARNIN] 3) A unit may be started when needed via activation (socket, path, timer,

[ceph_monitor][WARNIN]    D-Bus, udev, scripted systemctl call, ...).

[talen@ceph_admin ~]$ ceph -w

    cluster f35a65ad-1a6a-4e8d-8f7e-cb5f113c0a02

     health HEALTH_WARN

            64 pgs stale

            64 pgs stuck stale

     monmap e1: 1 mons at {ceph_monitor=10.0.2.33:6789/0}

            election epoch 1, quorum 0 ceph_monitor

     osdmap e47: 7 osds: 2 up, 2 in

      pgmap v1677: 64 pgs, 1 pools, 0 bytes data, 0 objects

            10305 MB used, 6056 MB / 16362 MB avail

                  64 stale+active+clean

2015-09-14 15:11:44.150240 mon.0 [INF] pgmap v1677: 64 pgs: 64 stale+active+clean; 0 bytes data, 10305 MB used, 6056 MB / 16362 MB avail

2015-09-14 15:22:48.442026 mon.0 [INF] from='client.? 10.0.2.33:0/1012278' entity='client.bootstrap-osd' cmd=[{"prefix": "osd create", "uuid": "86491f86-a204-4cee-acb5-9d7a7c26f784"}]: dispatch

2015-09-14 15:22:48.552414 mon.0 [INF] from='client.? 10.0.2.33:0/1012278' entity='client.bootstrap-osd' cmd='[{"prefix": "osd create", "uuid": "86491f86-a204-4cee-acb5-9d7a7c26f784"}]': finished

2015-09-14 15:22:48.665616 mon.0 [INF] osdmap e48: 8 osds: 2 up, 2 in

2015-09-14 15:22:48.685799 mon.0 [INF] pgmap v1678: 64 pgs: 64 stale+active+clean; 0 bytes data, 10305 MB used, 6056 MB / 16362 MB avail

2015-09-14 15:22:50.271971 mon.0 [INF] from='client.? 10.0.2.33:0/1012361' entity='client.bootstrap-osd' cmd=[{"prefix": "auth add", "entity": "osd.7", "caps": ["osd", "allow *", "mon", "allow profile osd"]}]: dispatch

2015-09-14 15:22:50.310775 mon.0 [INF] from='client.? 10.0.2.33:0/1012361' entity='client.bootstrap-osd' cmd='[{"prefix": "auth add", "entity": "osd.7", "caps": ["osd", "allow *", "mon", "allow profile osd"]}]': finished

2015-09-14 15:22:51.148684 mon.0 [INF] from='client.? 10.0.2.33:0/1012532' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "args": ["host=ceph_monitor", "root=default"], "id": 7, "weight": 0.01}]: dispatch

2015-09-14 15:22:51.895119 mon.0 [INF] from='client.? 10.0.2.33:0/1012532' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "args": ["host=ceph_monitor", "root=default"], "id": 7, "weight": 0.01}]': finished

2015-09-14 15:22:51.936453 mon.0 [INF] osdmap e49: 8 osds: 2 up, 2 in

2015-09-14 15:22:52.015062 mon.0 [INF] pgmap v1679: 64 pgs: 64 stale+active+clean; 0 bytes data, 10305 MB used, 6056 MB / 16362 MB avail

2015-09-14 15:22:53.077570 mon.0 [INF] osd.7 10.0.2.33:6800/12581 boot

2015-09-14 15:22:53.178958 mon.0 [INF] osdmap e50: 8 osds: 3 up, 3 in

2015-09-14 15:22:53.197124 mon.0 [INF] pgmap v1680: 64 pgs: 64 stale+active+clean; 0 bytes data, 10305 MB used, 6056 MB / 16362 MB avail

2015-09-14 15:22:58.593083 mon.0 [INF] pgmap v1681: 64 pgs: 64 stale+active+clean; 0 bytes data, 15460 MB used, 9082 MB / 24543 MB avail

繼續閱讀