Re: [PATCH 1/6] NFSD: add the ability to enable use of RWF_DONTCACHE for all IO

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 6/17/25 1:22 PM, Mike Snitzer wrote:
> On Mon, Jun 16, 2025 at 12:10:38PM -0400, Mike Snitzer wrote:
>> On Mon, Jun 16, 2025 at 09:32:16AM -0400, Chuck Lever wrote:
>>> On 6/12/25 12:00 PM, Mike Snitzer wrote:
>>>> On Thu, Jun 12, 2025 at 09:21:35AM -0400, Chuck Lever wrote:
>>>>> On 6/11/25 3:18 PM, Mike Snitzer wrote:
>>>>>> On Wed, Jun 11, 2025 at 10:31:20AM -0400, Chuck Lever wrote:
>>>>>>> On 6/10/25 4:57 PM, Mike Snitzer wrote:
>>>>>>>> Add 'enable-dontcache' to NFSD's debugfs interface so that: Any data
>>>>>>>> read or written by NFSD will either not be cached (thanks to O_DIRECT)
>>>>>>>> or will be removed from the page cache upon completion (DONTCACHE).
>>>>>>>
>>>>>>> I thought we were going to do two switches: One for reads and one for
>>>>>>> writes? I could be misremembering.
>>>>>>
>>>>>> We did discuss the possibility of doing that.  Still can-do if that's
>>>>>> what you'd prefer.
>>>>>
>>>>> For our experimental interface, I think having read and write enablement
>>>>> as separate settings is wise, so please do that.
>>>>>
>>>>> One quibble, though: The name "enable_dontcache" might be directly
>>>>> meaningful to you, but I think others might find "enable_dont" to be
>>>>> oxymoronic. And, it ties the setting to a specific kernel technology:
>>>>> RWF_DONTCACHE.
>>>>>
>>>>> So: Can we call these settings "io_cache_read" and "io_cache_write" ?
>>>>>
>>>>> They could each carry multiple settings:
>>>>>
>>>>> 0: Use page cache
>>>>> 1: Use RWF_DONTCACHE
>>>>> 2: Use O_DIRECT
>>>>>
>>>>> You can choose to implement any or all of the above three mechanisms.
>>>>
>>>> I like it, will do for v2. But will have O_DIRECT=1 and RWF_DONTCACHE=2.
>>>
>>> For io_cache_read, either settings 1 and 2 need to set
>>> disable_splice_read, or the io_cache_read setting has to be considered
>>> by nfsd_read_splice_ok() when deciding to use nfsd_iter_read() or
>>> splice read.
>>
>> Yes, I understand.
>>  
>>> However, it would be slightly nicer if we could decide whether splice
>>> read can be removed /before/ this series is merged. Can you get NFSD
>>> tested with IOR with disable_splice_read both enabled and disabled (no
>>> direct I/O)? Then we can compare the results to ensure that there is no
>>> negative performance impact for removing the splice read code.
>>
>> I can ask if we have a small window of opportunity to get this tested,
>> will let you know if so.
>>
> 
> I was able to enlist the help of Keith (cc'd) to get some runs in to
> compare splice_read vs vectored read.  A picture is worth 1000 words:
> https://original.art/NFSD_splice_vs_buffered_read_IOR_EASY.jpg
> 
> Left side is with splice_read running IOR_EASY with 48, 64, 96 PPN
> (Processes Per Node on each client) respectively.  Then the same
> IOR_EASY workload progression for buffered IO on the right side.
> 
> 6x servers with 1TB memory and 48 cpus, each configured with 32 NFSD
> threads, with CPU pinning and 4M Read Ahead. 6x clients running IOR_EASY. 
> 
> This was Keith's take on splice_read's benefits:
> - Is overall faster than buffered at any PPN.
> - Is able to scale higher with PPN (whereas buffered is flat).
> - Safe to say splice_read allows NFSD to do more IO then standard
>   buffered.

I thank you and Keith for the data!


> (These results came _after_ I did the patch to remove all the
> splice_read related code from NFSD and SUNRPC.. while cathartic, alas
> it seems it isn't meant to be at this point.  I'll let you do the
> honors in the future if/when you deem splice_read worthy of removal.)

If we were to make all NFS READ operations use O_DIRECT, then of course
NFSD's splice read should be removed at that point.


-- 
Chuck Lever




[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [NTFS 3]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [NTFS 3]     [Samba]     [Device Mapper]     [CEPH Development]

  Powered by Linux