ceph-pve英語

adapted accordingly
並相應地調整node

silos
n. 筒倉;糧倉;貯倉(silo的複數)app

saturate
vt. 浸透,使溼透;使飽和,使充滿
While one HDD might not saturate a 1 Gb linkide

likelihood
n. 可能性,可能orm

aforementioned
adj. 上述的;前面說起的server

fail-safe
n. 自動防故障裝置three

colocated
駐紮在同一地點ip

budgeted
adj. 已安排預算的ci

devoted
adj. 獻身的;忠誠的rem

In general SSDs will provide more IOPs than spinning disks. This fact and the higher cost may make a class
based Section 4.2.9 separation of pools appealing. Another possibility to speedup OSDs is to use a faster
disk as journal or DB/WAL device, see creating Ceph OSDs Section 4.2.7. If a faster disk is used for multiple
OSDs, a proper balance between OSD and WAL / DB (or journal) disk must be selected, otherwise the faster
disk becomes the bottleneck for all linked OSDs.
Aside from the disk type, Ceph best performs with an even sized and distributed amount of disks per node.
For example, 4 x 500 GB disks with in each node is better than a mixed setup with a single 1 TB and three
250 GB disk.
One also need to balance OSD count and single OSD capacity. More capacity allows to increase storage
density, but it also means that a single OSD failure forces ceph to recover more data at once.get

OSDs can also be backed by a combination of devices, like a HDD for most data and an SSD (or partition of an SSD) for some metadata.

BlueStore allows its internal journal (write-ahead log) to be written to a separate, high-speed device (like an SSD, NVMe, or NVDIMM) to increased performance.

However, the most common practice is to partition the journal drive (often an SSD),
and mount it such that Ceph uses the entire partition for the journal.

Sizing for block.db should be as large as possible to avoid performance penalties otherwise.

When using a mixed spinning and solid drive setup it is important to make a large-enough block.db logical volume for Bluestore.

The Ceph objecter handles where to place the objects and the tiering agent determines when to flush objects from the cache to the backing storage tier.

VMIDs < 100 are reserved for internal purposes, and VMIDs need to be unique cluster wide.

block就是primary device
block.db or block.wal
A DB device (identified as block.db
A WAL device (identified as block.wal


兩種普通的使用場景
1BLOCK (DATA) ONLY
it makes sense to just deploy with block only and not try to separate block.db or block.wal.

2BLOCK AND BLOCK.DB


-----------------------------------------------
Bibliography

n. 參考書目;文獻目錄

The Proxmox VE management tool (pvesh) allows to directly invoke API function, without using the REST/HTTPS server.

# single time output
pve# ceph -s
# continuously output status changes (press CTRL+C to stop)
pve# ceph -w

-------------------------------------
A volume is identified by the <STORAGE_ID>, followed by a storage type dependent volume name, separated by colon. A valid <VOLUME_ID> looks like:

local:230/example-image.raw

pvesm path <VOLUME_ID>

root@cu-pve04:/var/lib/vz# pvesm path kycfs:iso/CentOS-7-x86_64-Minimal-1810.iso
/mnt/pve/kycfs/template/iso/CentOS-7-x86_64-Minimal-1810.iso

root@cu-pve04:/var/lib/vz# pvesm path kycrbd:vm-102-disk-0rbd:kycrbd/vm-102-disk-0:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/kycrbd.keyring

相關文章
相關標籤/搜索