[ceph-users] 13.2.4 odd memory leak?

Paul Emmerich paul.emmerich at croit.io
Mon Mar 4 07:09:38 PST 2019


Bloated to ~4 GB per OSD and you are on HDDs?

13.2.3 backported the cache auto-tuning which targets 4 GB memory
usage by default.

See https://ceph.com/releases/13-2-4-mimic-released/

The bluestore_cache_* options are no longer needed. They are replaced
by osd_memory_target, defaulting to 4GB. BlueStore will expand
and contract its cache to attempt to stay within this
limit. Users upgrading should note this is a higher default
than the previous bluestore_cache_size of 1GB, so OSDs using
BlueStore will use more memory by default.
For more details, see the BlueStore docs.

Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

On Mon, Mar 4, 2019 at 3:55 PM Steffen Winther Sørensen
<stefws at gmail.com> wrote:
>
> List Members,
>
> patched a centos 7  based cluster from 13.2.2 to 13.2.4 last monday, everything appeared working fine.
>
> Only this morning I found all OSDs in the cluster to be bloated in memory foot print, possible after weekend backup through MDS.
>
> Anyone else seeing possible memory leak in 13.2.4 OSD possible primarily when using MDS?
>
> TIA
>
> /Steffen
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


More information about the ceph-users mailing list