[ceph-users] PGs get placed in the same datacenter (Trying to make a hybrid NVMe/HDD pool with 6 servers, 2 in each datacenter)
sinan at turka.nl
Sat Oct 7 11:41:34 PDT 2017
You are talking about the min_size, which should be 2 according to your text.
Please be aware, the min_size in your CRUSH is _not_ the replica size. The replica size is set with your pools.
> Op 7 okt. 2017 om 19:39 heeft Peter Linder <peter.linder at fiberdirekt.se> het volgende geschreven:
>> On 10/7/2017 7:36 PM, Дробышевский, Владимир wrote:
>> 2017-10-07 19:12 GMT+05:00 Peter Linder <peter.linder at fiberdirekt.se>:
>>> The idea is to select an nvme osd, and
>>> then select the rest from hdd osds in different datacenters (see crush
>>> map below for hierarchy).
>> It's a little bit aside of the question, but why do you want to mix SSDs and HDDs in the same pool? Do you have read-intensive workload and going to use primary-affinity to get all reads from nvme?
> Yes, this is pretty much the idea, getting the performance from NVMe reads, while still maintaining triple redundancy and a reasonable cost.
> ceph-users mailing list
> ceph-users at lists.ceph.com
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the ceph-users