ceph命令拷屏
常用命令
ceph -w
ceph df
ceph features
ceph fs ls
ceph fs status
ceph fsid
ceph health
ceph -s
ceph status
ceph mgr module ls
ceph mgr module enable dashboard
ceph mgr services
ceph mon feature ls
ceph node ls
ceph osd crush rule ls
ceph osd crush rule dump
ceph osd df tree
ceph osd lspools
ceph osd perf
watch ceph osd perf
ceph osd pool get kycrbd all
ceph osd pool ls
ceph osd pool ls detail
ceph osd pool stats
ceph osd status
ceph osd tree
ceph osd utilization pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief [all|summary|sum|delta|pools|osds|pgs|pgs_brief...]}
ceph pg dump all
ceph pg dump summary
ceph pg dump sum
ceph pg dump delta
ceph pg dump pools
ceph pg dump osds
ceph pg dump pgs
ceph pg dump pgs_brief
ceph pg dump pgs
ceph pg ls
ceph pg ls-by-osd osd.0
ceph pg ls-by-pool kycfs_metadata
ceph pg ls-by-primary
ceph pg map 7.1e8
ceph report ceph time-sync-status
ceph version
ceph versions
root@cu-pve04:~# ceph fs get kycfs
Filesystem 'kycfs' ()
fs_name kycfs
epoch
flags c
created -- ::48.957941
modified -- ::33.599472
tableserver
root
session_timeout
session_autoclose
max_file_size
last_failure
last_failure_osd_epoch
compat compat={},rocompat={},incompat={=base v0.,=client writeable ranges,=default file layouts on dirs,=dir inode in separate object,=mds uses versioned encoding,=dirfrag is stored in omap,=no anchor table,=file layout v2}
max_mds
in
up {=}
failed
damaged
stopped
data_pools []
metadata_pool
inline_data disabled
balancer
standby_count_wanted
: 192.168.7.205:/ 'cu-pve05' mds.0.12 up:active seq (standby for rank - 'pve') root@cu-pve04:~# ceph fs ls
name: kycfs, metadata pool: kycfs_metadata, data pools: [kycfs_data ] root@cu-pve04:~# ceph fs status
kycfs - clients
=====
+------+--------+----------+---------------+-------+-------+
| Rank | State | MDS | Activity | dns | inos |
+------+--------+----------+---------------+-------+-------+
| | active | cu-pve05 | Reqs: /s | | |
+------+--------+----------+---------------+-------+-------+
+----------------+----------+-------+-------+
| Pool | type | used | avail |
+----------------+----------+-------+-------+
| kycfs_metadata | metadata | 89.7M | .3T |
| kycfs_data | data | .0G | .3T |
+----------------+----------+-------+-------+ +-------------+
| Standby MDS |
+-------------+
| cu-pve04 |
| cu-pve06 |
+-------------+
MDS version: ceph version 12.2. (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable) root@cu-pve04:~# ceph fsid
b5fd132b-9ff4-470a-9a14-172eb48dc973
root@cu-pve04:~# ceph health
HEALTH_OK
root@cu-pve04:~# ceph -s
cluster:
id: b5fd132b-9ff4-470a-9a14-172eb48dc973
health: HEALTH_OK services:
mon: daemons, quorum cu-pve04,cu-pve05,cu-pve06
mgr: cu-pve04(active), standbys: cu-pve05, cu-pve06
mds: kycfs-// up {=cu-pve05=up:active}, up:standby
osd: osds: up, in data:
pools: pools, pgs
objects: .35k objects, 176GiB
usage: 550GiB used, .9TiB / .4TiB avail
pgs: active+clean io:
client: 0B/s rd, .5KiB/s wr, 0op/s rd, 6op/s wr root@cu-pve04:~# ceph mgr module ls
{
"enabled_modules": [
"balancer",
"dashboard",
"restful",
"status"
],
"disabled_modules": [
"influx",
"localpool",
"prometheus",
"selftest",
"zabbix"
]
} root@cu-pve04:~# ceph mgr module enable dashboard root@cu-pve04:~# ceph mgr services
{
"dashboard": "http://cu-pve04.ka1che.com:7000/"
} root@cu-pve04:~# ceph -v
ceph version 12.2. (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable) root@cu-pve04:~# ceph mds versions
{
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
}
root@cu-pve04:~# ceph mgr versions
{
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
}
root@cu-pve04:~# ceph mon versions
{
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
}
root@cu-pve04:~# ceph osd versions
{
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
} root@cu-pve04:~# ceph mon feature ls
all features
supported: [kraken,luminous]
persistent: [kraken,luminous]
on current monmap (epoch )
persistent: [kraken,luminous]
required: [kraken,luminous] root@cu-pve04:~# ceph mds stat
kycfs-// up {=cu-pve05=up:active}, up:standby root@cu-pve04:~# ceph mon stat
e3: mons at {cu-pve04=192.168.7.204:/,cu-pve05=192.168.7.205:/,cu-pve06=192.168.7.206:/}, election epoch , leader cu-pve04, quorum ,, cu-pve04,cu-pve05,cu-pve06 root@cu-pve04:~# ceph osd stat
osds: up, in root@cu-pve04:~# ceph pg stat
pgs: active+clean; 176GiB data, 550GiB used, .9TiB / .4TiB avail; 673B/s rd, 197KiB/s wr, 23op/s root@cu-pve04:~# ceph node ls
{
"mon": {
"cu-pve04": [ ],
"cu-pve05": [ ],
"cu-pve06": [ ]
},
"osd": {
"cu-pve04": [
,
,
,
,
,
,
, ],
"cu-pve05": [
,
,
,
,
,
,
, ],
"cu-pve06": [
,
,
,
,
,
,
, ]
},
"mds": {
"cu-pve04": [
-
],
"cu-pve05": [ ],
"cu-pve06": [
-
]
}
} root@cu-pve04:~# ceph osd crush rule ls
replicated_rule
root@cu-pve04:~# ceph osd crush rule dump
[
{
"rule_id": ,
"rule_name": "replicated_rule",
"ruleset": ,
"type": ,
"min_size": ,
"max_size": ,
"steps": [
{
"op": "take",
"item": -,
"item_name": "default"
},
{
"op": "chooseleaf_firstn",
"num": ,
"type": "host"
},
{
"op": "emit"
}
]
}
] root@cu-pve04:~# ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME
- 52.39417 - .4TiB 550GiB .9TiB 1.03 1.00 - root default
- 17.46472 - .5TiB 183GiB .3TiB 1.03 1.00 - host cu-pve04
hdd 2.18309 1.00000 .18TiB .1GiB .16TiB 1.04 1.01 osd.
hdd 2.18309 1.00000 .18TiB .2GiB .16TiB 0.90 0.88 osd.
hdd 2.18309 1.00000 .18TiB .1GiB .16TiB 1.12 1.10 osd.
hdd 2.18309 1.00000 .18TiB .0GiB .16TiB 1.21 1.18 osd.
hdd 2.18309 1.00000 .18TiB .1GiB .16TiB 0.85 0.83 osd.
hdd 2.18309 1.00000 .18TiB .1GiB .16TiB 1.12 1.09 osd.
hdd 2.18309 1.00000 .18TiB .2GiB .16TiB 1.04 1.01 osd.
hdd 2.18309 1.00000 .18TiB .6GiB .16TiB 0.92 0.90 osd.
- 17.46472 - .5TiB 183GiB .3TiB 1.03 1.00 - host cu-pve05
hdd 2.18309 1.00000 .18TiB .0GiB .16TiB 1.21 1.18 osd.
hdd 2.18309 1.00000 .18TiB .4GiB .16TiB 1.09 1.07 osd.
hdd 2.18309 1.00000 .18TiB .4GiB .16TiB 1.09 1.06 osd.
hdd 2.18309 1.00000 .18TiB .2GiB .16TiB 0.99 0.97 osd.
hdd 2.18309 1.00000 .18TiB .9GiB .16TiB 1.02 1.00 osd.
hdd 2.18309 1.00000 .18TiB .2GiB .16TiB 1.00 0.97 osd.
hdd 2.18309 1.00000 .18TiB .3GiB .16TiB 0.91 0.89 osd.
hdd 2.18309 1.00000 .18TiB .9GiB .16TiB 0.89 0.87 osd.
- 17.46472 - .5TiB 183GiB .3TiB 1.03 1.00 - host cu-pve06
hdd 2.18309 1.00000 .18TiB .9GiB .16TiB 1.03 1.00 osd.
hdd 2.18309 1.00000 .18TiB .3GiB .16TiB 1.04 1.02 osd.
hdd 2.18309 1.00000 .18TiB .0GiB .16TiB 1.16 1.13 osd.
hdd 2.18309 1.00000 .18TiB .0GiB .16TiB 0.94 0.92 osd.
hdd 2.18309 1.00000 .18TiB .4GiB .16TiB 1.14 1.11 osd.
hdd 2.18309 1.00000 .18TiB .8GiB .16TiB 0.84 0.82 osd.
hdd 2.18309 1.00000 .18TiB .4GiB .16TiB 1.09 1.06 osd.
hdd 2.18309 1.00000 .18TiB .5GiB .16TiB 0.96 0.94 osd.
TOTAL .4TiB 550GiB .9TiB 1.03
MIN/MAX VAR: 0.82/1.18 STDDEV: 0.11 root@cu-pve04:~# ceph osd lspools
kycfs_data, kycfs_metadata, kycrbd, root@cu-pve04:~# ceph osd perf
osd commit_latency(ms) apply_latency(ms) root@cu-pve04:~# ceph osd pool get kycrbd all
size:
min_size:
crash_replay_interval:
pg_num:
pgp_num:
crush_rule: replicated_rule
hashpspool: true
nodelete: false
nopgchange: false
nosizechange: false
write_fadvise_dontneed: false
noscrub: false
nodeep-scrub: false
use_gmt_hitset:
auid:
fast_read: [root@ceph1 ceph]# ceph osd pool create cfs_data
pool 'cfs_data' created
[root@ceph1 ceph]# ceph osd pool create cfs_meta
pool 'cfs_meta' created
[root@ceph1 ceph]# ceph fs new cefs cfs_meta cfs_data
new fs with metadata pool and data pool root@cu-pve04:~# ceph osd pool ls
kycfs_data
kycfs_metadata
kycrbd root@cu-pve04:~# ceph osd pool ls detail
pool 'kycfs_data' replicated size min_size crush_rule object_hash rjenkins pg_num pgp_num last_change flags hashpspool stripe_width application cephfs
pool 'kycfs_metadata' replicated size min_size crush_rule object_hash rjenkins pg_num pgp_num last_change flags hashpspool stripe_width application cephfs
pool 'kycrbd' replicated size min_size crush_rule object_hash rjenkins pg_num pgp_num last_change flags hashpspool stripe_width application rbd
removed_snaps [~] root@cu-pve04:~# ceph osd pool stats
pool kycfs_data id
client io .42KiB/s wr, 0op/s rd, 0op/s wr pool kycfs_metadata id
client io .08KiB/s wr, 0op/s rd, 0op/s wr pool kycrbd id
client io 0B/s rd, 357KiB/s wr, 0op/s rd, 25op/s wr root@cu-pve04:~# ceph osd status
+----+----------+-------+-------+--------+---------+--------+---------+-----------+
| id | host | used | avail | wr ops | wr data | rd ops | rd data | state |
+----+----------+-------+-------+--------+---------+--------+---------+-----------+
| | cu-pve04 | .1G | 2212G | | | | | exists,up |
| | cu-pve04 | .1G | 2215G | | .8k | | | exists,up |
| | cu-pve04 | .1G | 2210G | | | | | exists,up |
| | cu-pve04 | .0G | 2208G | | .2k | | | exists,up |
| | cu-pve04 | .0G | 2216G | | | | | exists,up |
| | cu-pve04 | .0G | 2210G | | .5k | | | exists,up |
| | cu-pve04 | .2G | 2212G | | .0k | | | exists,up |
| | cu-pve04 | .5G | 2214G | | .0k | | | exists,up |
| | cu-pve05 | .0G | 2208G | | .2k | | | exists,up |
| | cu-pve05 | .4G | 2211G | | | | | exists,up |
| | cu-pve05 | .3G | 2211G | | .4k | | | exists,up |
| | cu-pve05 | .2G | 2213G | | .8k | | | exists,up |
| | cu-pve05 | .8G | 2212G | | | | | exists,up |
| | cu-pve05 | .2G | 2213G | | .1k | | | exists,up |
| | cu-pve05 | .3G | 2215G | | .8k | | | exists,up |
| | cu-pve05 | .8G | 2215G | | | | | exists,up |
| | cu-pve06 | .9G | 2212G | | .4k | | | exists,up |
| | cu-pve06 | .3G | 2212G | | .6k | | | exists,up |
| | cu-pve06 | .9G | 2209G | | | | | exists,up |
| | cu-pve06 | .0G | 2214G | | | | | exists,up |
| | cu-pve06 | .4G | 2210G | | .2k | | | exists,up |
| | cu-pve06 | .8G | 2216G | | | | | exists,up |
| | cu-pve06 | .3G | 2211G | | .9k | | | exists,up |
| | cu-pve06 | .4G | 2214G | | | | | exists,up |
+----+----------+-------+-------+--------+---------+--------+---------+-----------+ root@cu-pve04:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
- 52.39417 root default
- 17.46472 host cu-pve04
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
- 17.46472 host cu-pve05
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
- 17.46472 host cu-pve06
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000
hdd 2.18309 osd. up 1.00000 1.00000 root@cu-pve04:~# ceph osd utilization
avg
stddev 9.49561 (expected baseline 11.7473)
min osd. with pgs (0.875 * mean)
max osd. with pgs (1.13889 * mean) root@cu-pve04:~# ceph pg dump sum
dumped sum
version
stamp -- ::45.513442
last_osdmap_epoch
last_pg_scan
full_ratio
nearfull_ratio
PG_STAT OBJECTS MISSING_ON_PRIMARY DEGRADED MISPLACED UNFOUND BYTES LOG DISK_LOG
sum
OSD_STAT USED AVAIL TOTAL
sum 550GiB .9TiB .4TiB
root@cu-pve04:~# ceph pg dump pools
dumped pools
POOLID OBJECTS MISSING_ON_PRIMARY DEGRADED MISPLACED UNFOUND BYTES LOG DISK_LOG root@cu-pve04:~# ceph pg dump osds
dumped osds
OSD_STAT USED AVAIL TOTAL HB_PEERS PG_SUM PRIMARY_PG_SUM
.5GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,]
.4GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.4GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.0GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,]
.6GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,]
.2GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.1GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.1GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.2GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,]
.2GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.1GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.0GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.4GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.2GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.9GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.2GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.3GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.9GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,]
.9GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,,,]
.3GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.0GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.0GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.4GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
.8GiB .16TiB .18TiB [,,,,,,,,,,,,,,,,,]
sum 550GiB .9TiB .4TiB root@cu-pve04:~# ceph pg map 7.1e8
osdmap e190 pg 7.1e8 (7.1e8) -> up [,,] acting [,,] root@cu-pve04:~# ceph status
cluster:
id: b5fd132b-9ff4-470a-9a14-172eb48dc973
health: HEALTH_OK services:
mon: daemons, quorum cu-pve04,cu-pve05,cu-pve06
mgr: cu-pve04(active), standbys: cu-pve05, cu-pve06
mds: kycfs-// up {=cu-pve05=up:active}, up:standby
osd: osds: up, in data:
pools: pools, pgs
objects: .35k objects, 176GiB
usage: 550GiB used, .9TiB / .4TiB avail
pgs: active+clean io:
client: 0B/s rd, 290KiB/s wr, 0op/s rd, 15op/s wr root@cu-pve04:~# ceph time-sync-status
{
"time_skew_status": {
"cu-pve04": {
"skew": 0.000000,
"latency": 0.000000,
"health": "HEALTH_OK"
},
"cu-pve05": {
"skew": 0.002848,
"latency": 0.001070,
"health": "HEALTH_OK"
},
"cu-pve06": {
"skew": 0.002570,
"latency": 0.001064,
"health": "HEALTH_OK"
}
},
"timechecks": {
"epoch": ,
"round": ,
"round_status": "finished"
}
} root@cu-pve04:~# ceph versions
{
"mon": {
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
},
"mgr": {
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
},
"osd": {
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
},
"mds": {
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
},
"overall": {
"ceph version 12.2.12 (39cfebf25a7011204a9876d2950e4b28aba66d11) luminous (stable)":
}
}
========================================= [sceph@ceph1 ~]$ ceph-authtool ceph.mon.keyring -l
[mon.]
key = AQBYF5JcAAAAABAAZageA/U12ulwiTj1qy9jKw==
caps mon = "allow *"
[sceph@ceph1 ~]$ ceph-authtool ceph.client.admin.keyring -l
[client.admin]
key = AQBaPZNcCalvLRAAt4iyva3DHfb8NbOX4MxBAw==
caps mds = "allow *"
caps mgr = "allow *"
caps mon = "allow *"
caps osd = "allow *" =========================================
[sceph@ceph1 ~]$ sudo ceph auth ls
installed auth entries: mds.ceph1
key: AQBUmpRc/KdcGhAAx3uWwlKVGu296HWFL3YhCw==
caps: [mds] allow
caps: [mon] allow profile mds
caps: [osd] allow rwx
mds.ceph2
key: AQCelpRcyn1WJBAAeXJ2e2ykDEHq7BYEFD57Tw==
caps: [mds] allow
caps: [mon] allow profile mds
caps: [osd] allow rwx
osd.
key: AQDrWpNcAextBRAA7usr2GT7OiEmnH5+Ya7iGg==
caps: [mgr] allow profile osd
caps: [mon] allow profile osd
caps: [osd] allow *
osd.
key: AQBGXJNc2fVyGhAAvNLbJSssGM6W9Om9gvGH/Q==
caps: [mgr] allow profile osd
caps: [mon] allow profile osd
caps: [osd] allow *
osd.
key: AQBcXJNcqPGOJxAA+U57mkFuRrNUjzEaR6EjIA==
caps: [mgr] allow profile osd
caps: [mon] allow profile osd
caps: [osd] allow *
client.admin
key: AQBaPZNcCalvLRAAt4iyva3DHfb8NbOX4MxBAw==
caps: [mds] allow *
caps: [mgr] allow *
caps: [mon] allow *
caps: [osd] allow *
client.bootstrap-mds
key: AQBaPZNcqO1vLRAANqPF730wvwPJWBbCqeW12w==
caps: [mon] allow profile bootstrap-mds
client.bootstrap-mgr
key: AQBaPZNcCCBwLRAAMGaeplDux+rd0jbTQVLNVw==
caps: [mon] allow profile bootstrap-mgr
client.bootstrap-osd
key: AQBaPZNcVE5wLRAA61JRSlzl72n65Dp5ZLpa/A==
caps: [mon] allow profile bootstrap-osd
client.bootstrap-rbd
key: AQBaPZNcpn5wLRAAps+/Xoxs7JoPHqO19KKQOA==
caps: [mon] allow profile bootstrap-rbd
client.bootstrap-rgw
key: AQBaPZNcEqtwLRAA/aW2qqnW+1uC4HAj1deONg==
caps: [mon] allow profile bootstrap-rgw
client.rgw.ceph1
key: AQDCl5RcUlRJEBAA25xPrLTfwnAwD+uSzc2T4Q==
caps: [mon] allow rw
caps: [osd] allow rwx
mgr.ceph2
key: AQDeWJNcqqItORAAPwDv8I4BcudMqzuzZFaY6w==
caps: [mds] allow *
caps: [mon] allow profile mgr
caps: [osd] allow *
[sceph@ceph1 ~]$ ===========================================
admin socket root@cu-pve04:~# ceph daemon mon.cu-pve04 help
root@cu-pve04:~# ceph daemon mon.cu-pve04 sessions
[root@ceph1 ceph]# ceph daemon osd. config show [root@ceph1 rbdpool]# ceph daemon osd. help
{
"calc_objectstore_db_histogram": "Generate key value histogram of kvdb(rocksdb) which used by bluestore",
"compact": "Commpact object store's omap. WARNING: Compaction probably slows your requests",
"config diff": "dump diff of current config and default config",
"config diff get": "dump diff get <field>: dump diff of current and default config setting <field>",
"config get": "config get <field>: get the config value",
"config help": "get config setting schema and descriptions",
"config set": "config set <field> <val> [<val> ...]: set a config variable",
"config show": "dump current config settings",
"config unset": "config unset <field>: unset a config variable",
"dump_blacklist": "dump blacklisted clients and times",
"dump_blocked_ops": "show the blocked ops currently in flight",
"dump_historic_ops": "show recent ops",
"dump_historic_ops_by_duration": "show slowest recent ops, sorted by duration",
"dump_historic_slow_ops": "show slowest recent ops",
"dump_mempools": "get mempool stats",
"dump_objectstore_kv_stats": "print statistics of kvdb which used by bluestore",
"dump_op_pq_state": "dump op priority queue state",
"dump_ops_in_flight": "show the ops currently in flight",
"dump_pgstate_history": "show recent state history",
"dump_reservations": "show recovery reservations",
"dump_scrubs": "print scheduled scrubs",
"dump_watchers": "show clients which have active watches, and on which objects",
"flush_journal": "flush the journal to permanent store",
"flush_store_cache": "Flush bluestore internal cache",
"get_command_descriptions": "list available commands",
"get_heap_property": "get malloc extension heap property",
"get_latest_osdmap": "force osd to update the latest map from the mon",
"get_mapped_pools": "dump pools whose PG(s) are mapped to this OSD.",
"getomap": "output entire object map",
"git_version": "get git sha1",
"heap": "show heap usage info (available only if compiled with tcmalloc)",
"help": "list available commands",
"injectdataerr": "inject data error to an object",
"injectfull": "Inject a full disk (optional count times)",
"injectmdataerr": "inject metadata error to an object",
"list_devices": "list OSD devices.",
"log dump": "dump recent log entries to log file",
"log flush": "flush log entries to log file",
"log reopen": "reopen log file",
"objecter_requests": "show in-progress osd requests",
"ops": "show the ops currently in flight",
"perf dump": "dump perfcounters value",
"perf histogram dump": "dump perf histogram values",
"perf histogram schema": "dump perf histogram schema",
"perf reset": "perf reset <name>: perf reset all or one perfcounter name",
"perf schema": "dump perfcounters schema",
"rmomapkey": "remove omap key",
"set_heap_property": "update malloc extension heap property",
"set_recovery_delay": "Delay osd recovery by specified seconds",
"setomapheader": "set omap header",
"setomapval": "set omap key",
"smart": "probe OSD devices for SMART data.",
"status": "high-level status of OSD",
"trigger_deep_scrub": "Trigger a scheduled deep scrub ",
"trigger_scrub": "Trigger a scheduled scrub ",
"truncobj": "truncate object to length",
"version": "get ceph version"
}
[root@ceph1 rbdpool]# [root@ceph1 rbdpool]# ceph daemon mon.ceph1 sessions
[
"MonSession(mon.0 192.168.7.151:6789/0 is open allow *, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(osd.0 192.168.7.151:6800/1988823 is open allow profile osd, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(osd.1 192.168.7.152:6801/1821392 is open allow profile osd, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(mds.? 192.168.7.152:6805/1783208616 is open allow profile mds, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(mds.? 192.168.7.151:6804/3007499436 is open allow profile mds, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(client.? 192.168.7.151:0/2871664294 is open allow rw, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(osd.2 192.168.7.153:6800/6408 is open allow profile osd, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(unknown.0 192.168.7.161:0/2782938665 is open allow *, features 0x27018fb86aa42ada (jewel))",
"MonSession(mgr.4729 192.168.7.152:0/2358460 is open allow profile mgr, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(client.? 192.168.7.152:0/1860240871 is open allow profile mgr, features 0x3ffddff8ffacfffb (luminous))",
"MonSession(unknown.0 192.168.7.151:0/819943570 is open allow *, features 0x27018fb86aa42ada (jewel))"
]
ceph命令拷屏的更多相关文章
- sed命令拷屏
http://blog.sina.com.cn/s/blog_45497dfa0100w6r3.html sed样例较多,可以参考 http://blog.sina.com.cn/s/blog_6d ...
- awk命令拷屏
如果不指明采取什么动作,awk默认打印出所有浏览出的记录,与{print $}是一样的 模式和动作两者是可选的,如果没有模式,则action应用到全部记录,如果没有action,则输出匹配全部记录. ...
- rbd_rados命令拷屏
mimic或者luminous rbd_rados sudo mount -t ceph 192.168.7.151:6789:/ /mnt -o name=admin,secret=AQBaPZNc ...
- 使用ceph命令提示handle_connect_reply connect got BADAUTHORIZER
输入命令提示如下错误: [root@node1 ~]# rados -p testpool ls 2017-10-21 06:13:25.743045 7f8f89b6d700 0 -- 192.16 ...
- VI打开和编辑多个文件的命令 分屏操作
VI打开和编辑多个文件的命令 可分两种情况: 1.在同一窗口中打开多个文件: vi file1 file2 file3:n 切换到下一个文件 (n=next):N 切换到上一个文件 2.在不同窗口 ...
- Ceph 命令
引用自: https://www.cnblogs.com/schangech/p/8036161.html 一.集群 1.启动一个ceph 进程启动mon进程 service ceph start ...
- script命令录屏
关于linux上的操作,我们的确可以使用'history'命令来显示出来操作记录,但是有些时候,我们不仅仅需要知道做了什么,还需要知道操作的时候,产生了什么效果,这个时候‘history’命令就显示无 ...
- ceph命令
chen@admin-node:~$ ceph --help General usage: ============== usage: ceph [-h] [-c CEPHCONF] [-i INPU ...
- C# 图像处理:复制屏幕到内存中,拷屏操作
/// <summary> /// 复制屏幕到内存中 /// </summary> /// <returns>返回内存流</returns> publi ...
随机推荐
- VS中C语言scanf函数报错
在VS中创建C项目使用scanf方法时会如下报错 解决方案 方法1:文件中设置 在第一行设置代码 #define _CRT_SECURE_NO_WARNINGS 或者添加警告忽略 #pr ...
- C++内存修改器开源代码
我们玩单机游戏时,游戏难度可能过大, 或者游戏已经比较熟练,想要增加游戏的玩法,这时候可以使用修改器. 内存式游戏修改器主要对游戏内存修改 修改时有两种方式,一是定时对内存数值进行修改.实现类似锁定的 ...
- 神他么奇怪NoClassDefFoundError
javaweb 项目,部署在Tomcat服务器 十分奇葩的问题 之前一直好好的,今天突然就给报错.代码没动过.真是奇葩.创建User对象报NoClassDefFoundError, User类是一个普 ...
- inode的若干锚
/** * __insert_inode_hash - hash an inode * @inode: unhashed inode * @hashval: unsigned long value u ...
- 关于deepin下安装ssh以后root用户登陆报错的解决
最近刚刚接触到deepin,觉得,wow,除了mac,还有这么好看的非win系统,而且第测出那个Linux,宽容度很高,非常适合我这种比较喜欢折腾的人,于是下载了deepin15版本并将其当作虚拟机成 ...
- luoguP1313 计算系数 题解(NOIP2011)
P1313 计算系数 题目 #include<iostream> #include<cstdlib> #include<cstdio> #include<cm ...
- object of type 'Response' has no len()
看见没,这里括号弄错了! 网上解释是requests.get()得到的是一个response对象,无法用BeautifulSoup解析,如果要解析,解析对象应该是requests.get().cont ...
- 第一节 初识RabbitMQ
原文:第一节 初识RabbitMQ 版权声明:未经本人同意,不得转载该文章,谢谢 https://blog.csdn.net/phocus1/article/details/87280120 1.什么 ...
- ionic3中使用docker 完成build代码,更新过程记录。
1.若未安装cordova 需先安装cordova 包: npm install -g cordova 2.安装docker 可查看官方文档进行一步步的安装:https://docs.docker.c ...
- JS中的call()和apply()方法(转)
转自:http://uule.iteye.com/blog/1158829 JS中的call()和apply()方法 博客分类: JS 1.方法定义 call方法: 语法:call([thisOb ...