Re: How important is the "default" data pool being replicated for CephFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Mikael,

This might be a long shot, but I have to ask: have you checked the average
file size on the current CephFS filesystem? Apart from extreme cost
efficiency or a bad design, the EC choice on HDD could have had a
legitimate reason in the past. It's probably not enough to make it the
default data pool, but it might help explain the current design.
I don't know the exact amount of data or how many users this filesystem
has, but the safest long-term approach in your situation is probably to
create a new filesystem and then migrate the data over, as you imagined.

Best regards,
Frédéric.

--
Frédéric Nass
Ceph Ambassador France | Senior Ceph Engineer @ CLYSO
Try our Ceph Analyzer -- https://analyzer.clyso.com/
https://clyso.com | frederic.nass@xxxxxxxxx


Le mer. 10 sept. 2025 à 16:09, Mikael Öhman <micketeer@xxxxxxxxx> a écrit :

> The recommendations for cephfs is to make a replicated default data pool,
> and adding any EC data pools using layouts:
> https://docs.ceph.com/en/latest/cephfs/createfs/
>
> > If erasure-coded pools are planned for file system data, it is best to
> configure the default as a replicated pool to improve small-object write
> and read performance when updating backtraces.
>
> I have an cephfs that unfortunately wasn't set up like this: they just made
> an EC pool on the slow HDDs as the default, which sounds like the worst
> case scenario to me. I would like to add a NVME data pool to this ceph fs,
> but recommended gives me pause on if i should instead go through the hassle
> of creating a new cephfs and migrating all users.
>
> I've tried to run some mdtest with small 1k files to see if i could measure
> this difference, but speed is about the same in my relatively small tests
> so far. I'm also not sure what impact I should realistically expect here. I
> don't even know if creating files counts as "updating backtraces", so my
> testing might just be pointless.
>
> I guess my core question is; just how important is this suggestion to keep
> the default data pool on replicated NVME?
>
> Setup:
> 14 hosts x 42 HDD + 3 NVMEs for db/wal  2*2x25 GbitE bonds
> 12 hosts x 10 NVME. 2*2x100 GbitE bonds
>
> Old CephFS setup:
> - metadata: replicated NVME
> - data-pools: EC 10+2 on HDD  (i plan to add a EC NVME pool here via
> layouts)
>
> New CephFS setup as recommended:
> - metadata: replicated NVME
> - data-pools: replicated NVME (default), EC 8+2 on HDD via layout, EC 8+2
> on NVME via layout.
>
> Ceph 18.2.7
>
>
> Best regards, Mikael
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux