On Fri 05-09-25 07:19:05, Brian Foster wrote: > On Thu, Sep 04, 2025 at 05:14:21PM -0700, Joanne Koong wrote: > > On Thu, Sep 4, 2025 at 1:07 PM Darrick J. Wong <djwong@xxxxxxxxxx> wrote: > > > On Thu, Sep 04, 2025 at 07:47:11AM -0400, Brian Foster wrote: > > > > On Wed, Sep 03, 2025 at 05:35:51PM -0700, Joanne Koong wrote: > > > > > On Wed, Sep 3, 2025 at 11:44 AM Brian Foster <bfoster@xxxxxxxxxx> wrote: > > > > > > On Tue, Sep 02, 2025 at 04:46:04PM -0700, Darrick J. Wong wrote: > > > > > > > On Fri, Aug 29, 2025 at 04:39:42PM -0700, Joanne Koong wrote: > > > > > > > > Add granular dirty and writeback accounting for large folios. These > > > > > > > > stats are used by the mm layer for dirty balancing and throttling. > > > > > > > > Having granular dirty and writeback accounting helps prevent > > > > > > > > over-aggressive balancing and throttling. > > > > > > > > > > > > > > > > There are 4 places in iomap this commit affects: > > > > > > > > a) filemap dirtying, which now calls filemap_dirty_folio_pages() > > > > > > > > b) writeback_iter with setting the wbc->no_stats_accounting bit and > > > > > > > > calling clear_dirty_for_io_stats() > > > > > > > > c) starting writeback, which now calls __folio_start_writeback() > > > > > > > > d) ending writeback, which now calls folio_end_writeback_pages() > > > > > > > > > > > > > > > > This relies on using the ifs->state dirty bitmap to track dirty pages in > > > > > > > > the folio. As such, this can only be utilized on filesystems where the > > > > > > > > block size >= PAGE_SIZE. > > > > > > > > > > > > > > Er... is this statement correct? I thought that you wanted the granular > > > > > > > dirty page accounting when it's possible that individual sub-pages of a > > > > > > > folio could be dirty. > > > > > > > > > > > > > > If i_blocksize >= PAGE_SIZE, then we'll have set the min folio order and > > > > > > > there will be exactly one (large) folio for a single fsblock. Writeback > > > > > > > > > > Oh interesting, this is the part I'm confused about. With i_blocksize > > > > > >= PAGE_SIZE, isn't there still the situation where the folio itself > > > > > could be a lot larger, like 1MB? That's what I've been seeing on fuse > > > > > where "blocksize" == PAGE_SIZE == 4096. I see that xfs sets the min > > > > > folio order through mapping_set_folio_min_order() but I'm not seeing > > > > > how that ensures "there will be exactly one large folio for a single > > > > > fsblock"? My understanding is that that only ensures the folio is at > > > > > least the size of the fsblock but that the folio size can be larger > > > > > than that too. Am I understanding this incorrectly? > > > > > > > > > > > > must happen in units of fsblocks, so there's no point in doing the extra > > > > > > > accounting calculations if there's only one fsblock. > > > > > > > > > > > > > > Waitaminute, I think the logic to decide if you're going to use the > > > > > > > granular accounting is: > > > > > > > > > > > > > > (folio_size > PAGE_SIZE && folio_size > i_blocksize) > > > > > > > > > > > > > > > > > Yeah, you're right about this - I had used "ifs && i_blocksize >= > > > > > PAGE_SIZE" as the check, which translates to "i_blocks_per_folio > 1 > > > > > && i_block_size >= PAGE_SIZE", which in effect does the same thing as > > > > > what you wrote but has the additional (and now I'm realizing, > > > > > unnecessary) stipulation that block_size can't be less than PAGE_SIZE. > > > > > > > > > > > > Hrm? > > > > > > > > > > > > > > > > > > > I'm also a little confused why this needs to be restricted to blocksize > > > > > > gte PAGE_SIZE. The lower level helpers all seem to be managing block > > > > > > ranges, and then apparently just want to be able to use that directly as > > > > > > a page count (for accounting purposes). > > > > > > > > > > > > Is there any reason the lower level functions couldn't return block > > > > > > units, then the higher level code can use a blocks_per_page or some such > > > > > > to translate that to a base page count..? As Darrick points out I assume > > > > > > you'd want to shortcut the folio_nr_pages() == 1 case to use a min page > > > > > > count of 1, but otherwise ISTM that would allow this to work with > > > > > > configs like 64k pagesize and 4k blocks as well. Am I missing something? > > > > > > > > > > > > > > > > No, I don't think you're missing anything, it should have been done > > > > > like this in the first place. > > > > > > > > > > > > > Ok. Something that came to mind after thinking about this some more is > > > > whether there is risk for the accounting to get wonky.. For example, > > > > consider 4k blocks, 64k pages, and then a large folio on top of that. If > > > > a couple or so blocks are dirtied at one time, you'd presumably want to > > > > account that as the minimum of 1 dirty page. Then if a couple more > > > > blocks are dirtied in the same large folio, how do you determine whether > > > > those blocks are a newly dirtied page or part of the already accounted > > > > dirty page? I wonder if perhaps this is the value of the no sub-page > > > > sized blocks restriction, because you can imply that newly dirtied > > > > blocks means newly dirtied pages..? > > > > > > > > I suppose if that is an issue it might still be manageable. Perhaps we'd > > > > have to scan the bitmap in blks per page windows and use that to > > > > determine how many base pages are accounted for at any time. So for > > > > example, 3 dirty 4k blocks all within the same 64k page size window > > > > still accounts as 1 dirty page, vs. dirty blocks in multiple page size > > > > windows might mean multiple dirty pages, etc. That way writeback > > > > accounting remains consistent with dirty accounting. Hm? > > > > > > Yes, I think that's correct -- one has to track which basepages /were/ > > > dirty, and then which ones become dirty after updating the ifs dirty > > > bitmap. > > > > > > For example, if you have a 1k fsblock filesystem, 4k base pages, and a > > > 64k folio, you could write a single byte at offset 0, then come back and > > > write to a byte at offset 1024. The first write will result in a charge > > > of one basepage, but so will the second, I think. That results > > > incharges for two dirty pages, when you've really only dirtied a single > > > basepage. > > > > Does it matter though which blocks map to which pages? AFAIU, the > > "block size" is the granularity for disk io and is not really related > > to pages (eg for writing out to disk, only the block gets written, not > > the whole page). The stats (as i understand it) are used to throttle > > how much data gets written back to disk, and the primary thing it > > cares about is how many bytes that is, not how many pages, it's just > > that it's in PAGE_SIZE granularity because prior to iomap there was no > > dirty tracking of individual blocks within a page/folio; it seems like > > it suffices then to just keep track of total # of dirty blocks, > > multiply that by blocksize, and roundup divide that by PAGE_SIZE and > > pass that to the stats. > > > > I suppose it may not matter in terms of the purpose of the mechanism > itself. In fact if the whole thing could just be converted to track > bytes, at least internally, then maybe that would eliminate some of the > confusion in dealing with different granularity of units..? I have no > idea how practical or appropriate that is, though. :) > > The concern Darrick and I were discussing is more about maintaining > accounting consistency in the event that we do continue translating > blocks to pages and ultimately add support for the block size < page > size case. > > In that case the implication is that we'd still need to account > something when we dirty a single block out of a page (i.e. use > Darrick's example where we dirty a 1k fs block out of a 4k page). If we > round that partial page case up to 1 dirty page and repeat as each 1k > block is dirtied, then we have to make sure accounting remains > consistent in the case where we dirty account each sub-block of a page > through separate writes, but then clear dirty accounting for the entire > folio once at writeback time. > > But I suppose we are projecting the implementation a bit so it might not > be worth getting this far into the weeds until you determine what > direction you want to go with this and have more code to review. All in > all, I do agree with Jan's general concern that I'd rather not have to > deal with multiple variants of sub-page state tracking in iomap. It's > already a challenge to support multiple different filesystems. This does > seem like a useful enhancement to me however, so IMO it would be fine to > just try and make it more generic (short of something more generic on > the mm side or whatever) than it is currently. > > > But, as Jan pointed out to me in his comment, the stats are also used > > for monitoring the health of reclaim, so maybe it does matter then how > > the blocks translate to pages. > > > > Random thought, but would having an additional/optional stat to track > bytes (alongside the existing page granularity counts) help at all? For > example, if throttling could use optional byte granular dirty/writeback > counters when they are enabled instead of the traditional page granular, > would that solve your problem and be less disruptive to other things > that actually prefer the page count? FWIW my current thinking is that the best might be to do byte granularity tracking for wb_stat_ counters and leave current coarse-grained accounting for the zone / memcg stats. That way mm counters could be fully managed within mm code and iomap wouldn't have to care and writeback counters (which care about amount of IO, not amount of pinned memory) would be maintained by filesystems / iomap. We'd just need to come up with sensible rules where writeback counters should be updated when mm doesn't do it. Honza -- Jan Kara <jack@xxxxxxxx> SUSE Labs, CR