[ceph-users] New OSD with weight 0, rebalance still happen...
gaio at sv.lnf.it
Fri Nov 23 02:12:18 PST 2018
Mandi! Paweł Sadowski
In chel di` si favelave...
> This is most probably due to big difference in weights between your hosts (the
> new one has 20x lower weight than the old ones) which in combination with straw
> algorithm is a 'known' issue.
Ok. I've reweighted back that disk to '1' and status goes back to
> You could try to increase choose_total_tries in
> your crush map from 50 to some bigger number. The best IMO would be to use
> straw2 (which will cause some rebalance) and then use 'ceph osd crush reweight'
> (instead of 'ceph osd reweight') with small steps to slowly rebalance data onto
> new OSDs.
For now i'm putting in the new disks with 'ceph osd reweight',
probably when i'm on 50% of new disks i'll start to use 'ceph osd crush reweight'
against the old one.
dott. Marco Gaiarin GNUPG Key ID: 240A3D66
Associazione ``La Nostra Famiglia'' http://www.lanostrafamiglia.it/
Polo FVG - Via della Bontà, 7 - 33078 - San Vito al Tagliamento (PN)
marco.gaiarin(at)lanostrafamiglia.it t +39-0434-842711 f +39-0434-842797
Dona il 5 PER MILLE a LA NOSTRA FAMIGLIA!
(cf 00307430132, categoria ONLUS oppure RICERCA SANITARIA)
More information about the ceph-users