[ceph-users] 13.2.4 odd memory leak?
paul.emmerich at croit.io
Mon Mar 4 07:09:38 PST 2019
Bloated to ~4 GB per OSD and you are on HDDs?
13.2.3 backported the cache auto-tuning which targets 4 GB memory
usage by default.
The bluestore_cache_* options are no longer needed. They are replaced
by osd_memory_target, defaulting to 4GB. BlueStore will expand
and contract its cache to attempt to stay within this
limit. Users upgrading should note this is a higher default
than the previous bluestore_cache_size of 1GB, so OSDs using
BlueStore will use more memory by default.
For more details, see the BlueStore docs.
Looking for help with your Ceph cluster? Contact us at https://croit.io
Tel: +49 89 1896585 90
On Mon, Mar 4, 2019 at 3:55 PM Steffen Winther Sørensen
<stefws at gmail.com> wrote:
> List Members,
> patched a centos 7 based cluster from 13.2.2 to 13.2.4 last monday, everything appeared working fine.
> Only this morning I found all OSDs in the cluster to be bloated in memory foot print, possible after weekend backup through MDS.
> Anyone else seeing possible memory leak in 13.2.4 OSD possible primarily when using MDS?
> ceph-users mailing list
> ceph-users at lists.ceph.com
More information about the ceph-users