In the mists of time (Luminous) I experienced a cluster reacting rather badly to a user issuing ~5000 RBD snap trims at once. At the time I raised my local values of osd_snap_trim_cost and osd_snap_trim_sleep (now osd_snap_trim_sleep_???) way up to spread the impact. The subject cluster was roughly half Filestore and half BlueStore; the BlueStore OSDs handled it much better than did Filestore. > On Apr 25, 2025, at 2:11 PM, Eugen Block <eblock@xxxxxx> wrote: > > Interesting, I did something similar just a few weeks back, flattening some images (didn’t look at their sizes or overlaps of the snapshots). But I don’t see any spikes in memory usage at all. At that time the cluster was running latest Pacific. > > Zitat von Dominique Ramaekers <dominique.ramaekers@xxxxxxxxxx>: > >> Hi, >> >> Housekeeping... I was cleaning up my snapshots and was flattening clones... Suddenly I ran out of memory on my nodes! >> >> 4 node cluster with each 10 ssd osd's with total storage size 25TiB. Each node has about 45 GiB of free (available) memory in normal operation. >> >> After flattening several images and removing dozens of snapshots, the free memory probably was already lower than the usual 45GiB. I was flattening an image of 75GiB and 2 out of 4 nodes ran out of memory. One node even automatically killed processes randomly to free up memory... After regaining control over the system, I put that node in maintenance and rebooted this node. After reboot the free mem was like 70Gib. Over night all of the nodes were back on the usual 45GiB. >> >> Today I checked the free mem of each node: 45GiB free. So I flattened again an image of 75GiB. And yes, the free mem dropped from 45GiB to 5GiB really fast! >> >> Is there a way to avoid this behavior of the cluster? >> >> Greetings, >> >> Dominique. >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx