diff options
author | 2023-07-24 16:12:29 -0700 | |
---|---|---|
committer | 2023-07-24 16:12:29 -0700 | |
commit | d42bd17c6a20638ddf96862bfc0c47e481c28392 (patch) | |
tree | 97191965a56afcb9f875fcc8dfd44a1bec80136f /Documentation/filesystems | |
parent | Linux 6.5-rc3 (diff) | |
parent | iomap: Copy larger chunks from userspace (diff) | |
download | wireguard-linux-d42bd17c6a20638ddf96862bfc0c47e481c28392.tar.xz wireguard-linux-d42bd17c6a20638ddf96862bfc0c47e481c28392.zip |
Merge tag 'large-folio-writes' of git://git.infradead.org/users/willy/pagecache into iomap-6.6-merge
Create large folios in iomap buffered write path
Commit ebb7fb1557b1 limited the length of ioend chains to 4096 entries
to improve worst-case latency. Unfortunately, this had the effect of
limiting the performance of:
fio -name write-bandwidth -rw=write -bs=1024Ki -size=32Gi -runtime=30 \
-iodepth 1 -ioengine sync -zero_buffers=1 -direct=0 -end_fsync=1 \
-numjobs=4 -directory=/mnt/test
https://lore.kernel.org/linux-xfs/20230508172406.1CF3.409509F4@e16-tech.com/
The problem ends up being lock contention on the i_pages spinlock as we
clear the writeback bit on each folio (and propagate that up through
the tree). By using larger folios, we decrease the number of folios
to be processed by a factor of 256 for this benchmark, eliminating the
lock contention.
Creating large folios in the buffered write path is also the right
thing to do. It's a project that has been on the back burner for years,
it just hasn't been important enough to do before now.
* tag 'large-folio-writes' of git://git.infradead.org/users/willy/pagecache:
iomap: Copy larger chunks from userspace
iomap: Create large folios in the buffered write path
filemap: Allow __filemap_get_folio to allocate large folios
filemap: Add fgf_t typedef
iomap: Remove unnecessary test from iomap_release_folio()
doc: Correct the description of ->release_folio
iomap: Remove large folio handling in iomap_invalidate_folio()
iov_iter: Add copy_folio_from_iter_atomic()
iov_iter: Handle compound highmem pages in copy_page_from_iter_atomic()
iov_iter: Map the page later in copy_page_from_iter_atomic()
[djwong: yay amortizations!]
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
Diffstat (limited to 'Documentation/filesystems')
-rw-r--r-- | Documentation/filesystems/locking.rst | 15 |
1 files changed, 11 insertions, 4 deletions
diff --git a/Documentation/filesystems/locking.rst b/Documentation/filesystems/locking.rst index ed148919e11a..ca3d24becbf5 100644 --- a/Documentation/filesystems/locking.rst +++ b/Documentation/filesystems/locking.rst @@ -374,10 +374,17 @@ invalidate_lock before invalidating page cache in truncate / hole punch path (and thus calling into ->invalidate_folio) to block races between page cache invalidation and page cache filling functions (fault, read, ...). -->release_folio() is called when the kernel is about to try to drop the -buffers from the folio in preparation for freeing it. It returns false to -indicate that the buffers are (or may be) freeable. If ->release_folio is -NULL, the kernel assumes that the fs has no private interest in the buffers. +->release_folio() is called when the MM wants to make a change to the +folio that would invalidate the filesystem's private data. For example, +it may be about to be removed from the address_space or split. The folio +is locked and not under writeback. It may be dirty. The gfp parameter +is not usually used for allocation, but rather to indicate what the +filesystem may do to attempt to free the private data. The filesystem may +return false to indicate that the folio's private data cannot be freed. +If it returns true, it should have already removed the private data from +the folio. If a filesystem does not provide a ->release_folio method, +the pagecache will assume that private data is buffer_heads and call +try_to_free_buffers(). ->free_folio() is called when the kernel has dropped the folio from the page cache. |