[ceph-users] tunable question
lists at merit.unu.edu
Tue Oct 3 05:38:12 PDT 2017
What would make the decision easier: if we knew that we could easily
> "ceph osd crush tunables optimal"
once it has begun rebalancing data?
Meaning: if we notice that impact is too high, or it will take too long,
that we could simply again say
> "ceph osd crush tunables hammer"
and the cluster would calm down again?
On 2-10-2017 9:41, Manuel Lausch wrote:
> We have similar issues.
> After upgradeing from hammer to jewel the tunable "choose leave stabel"
> was introduces. If we activate it nearly all data will be moved. The
> cluster has 2400 OSD on 40 nodes over two datacenters and is filled with
> 2,5 PB Data.
> We tried to enable it but the backfillingtraffic is to high to be
> handled without impacting other services on the Network.
> Do someone know if it is neccessary to enable this tunable? And could
> it be a problem in the future if we want to upgrade to newer versions
> wihout it enabled?
> Manuel Lausch
More information about the ceph-users