aboutsummaryrefslogtreecommitdiffstatshomepage
path: root/include/linux/writeback.h
diff options
context:
space:
mode:
authorMiaohe Lin <linmiaohe@huawei.com>2022-09-09 10:57:11 +0800
committerAndrew Morton <akpm@linux-foundation.org>2022-10-03 14:03:08 -0700
commit13cc378403a83e70430ae9bad53fd65199f21fe1 (patch)
tree8726dcee7d4b9e798e9d694bbd49db813236d04f /include/linux/writeback.h
parentmm: use nth_page instead of mem_map_offset mem_map_next (diff)
downloadwireguard-linux-13cc378403a83e70430ae9bad53fd65199f21fe1.tar.xz
wireguard-linux-13cc378403a83e70430ae9bad53fd65199f21fe1.zip
writeback: remove unused macro DIRTY_FULL_SCOPE
It's introduced but never used. Remove it. Link: https://lkml.kernel.org/r/20220909025711.32012-1-linmiaohe@huawei.com Signed-off-by: Miaohe Lin <linmiaohe@huawei.com> Reviewed-by: Jan Kara <jack@suse.cz> Acked-by: Jens Axboe <axboe@kernel.dk> Cc: Bart Van Assche <bvanassche@acm.org> Cc: David Howells <dhowells@redhat.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: NeilBrown <neilb@suse.de> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: zhanglianjie <zhanglianjie@uniontech.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Diffstat (limited to 'include/linux/writeback.h')
-rw-r--r--include/linux/writeback.h8
1 files changed, 0 insertions, 8 deletions
diff --git a/include/linux/writeback.h b/include/linux/writeback.h
index 3f045f6d6c4f..06f9291b6fd5 100644
--- a/include/linux/writeback.h
+++ b/include/linux/writeback.h
@@ -17,20 +17,12 @@ struct bio;
DECLARE_PER_CPU(int, dirty_throttle_leaks);
/*
- * The 1/4 region under the global dirty thresh is for smooth dirty throttling:
- *
- * (thresh - thresh/DIRTY_FULL_SCOPE, thresh)
- *
- * Further beyond, all dirtier tasks will enter a loop waiting (possibly long
- * time) for the dirty pages to drop, unless written enough pages.
- *
* The global dirty threshold is normally equal to the global dirty limit,
* except when the system suddenly allocates a lot of anonymous memory and
* knocks down the global dirty threshold quickly, in which case the global
* dirty limit will follow down slowly to prevent livelocking all dirtier tasks.
*/
#define DIRTY_SCOPE 8
-#define DIRTY_FULL_SCOPE (DIRTY_SCOPE / 2)
struct backing_dev_info;