Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752116AbdI0UP0 (ORCPT ); Wed, 27 Sep 2017 16:15:26 -0400 Received: from mail-wm0-f53.google.com ([74.125.82.53]:46071 "EHLO mail-wm0-f53.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751974AbdI0UO0 (ORCPT ); Wed, 27 Sep 2017 16:14:26 -0400 X-Google-Smtp-Source: AOwi7QDo2HrPtbbkarKs37RzVXithY2jLwZmEJGXKgxNG+wBW8vWp8JgYRio0DVfoxqkzoyYMwUbNg== From: Jens Axboe To: linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org Cc: hannes@cmpxchg.org, jack@suse.cz, torvalds@linux-foundation.org, Jens Axboe Subject: [PATCH 09/12] writeback: move nr_pages == 0 logic to one location Date: Wed, 27 Sep 2017 14:13:56 -0600 Message-Id: <1506543239-31470-10-git-send-email-axboe@kernel.dk> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1506543239-31470-1-git-send-email-axboe@kernel.dk> References: <1506543239-31470-1-git-send-email-axboe@kernel.dk> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 3765 Lines: 119 Now that we have no external callers of wb_start_writeback(), we can shuffle the passing in of 'nr_pages'. Everybody passes in 0 at this point, so just kill the argument and move the dirty count retrieval to that function. Acked-by: Johannes Weiner Tested-by: Chris Mason Reviewed-by: Jan Kara Reviewed-by: Christoph Hellwig Signed-off-by: Jens Axboe --- fs/fs-writeback.c | 42 ++++++++++++++++++------------------------ 1 file changed, 18 insertions(+), 24 deletions(-) diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c index 571a5702b568..d9be3dc34302 100644 --- a/fs/fs-writeback.c +++ b/fs/fs-writeback.c @@ -933,8 +933,19 @@ static void bdi_split_work_to_wbs(struct backing_dev_info *bdi, #endif /* CONFIG_CGROUP_WRITEBACK */ -static void wb_start_writeback(struct bdi_writeback *wb, long nr_pages, - bool range_cyclic, enum wb_reason reason) +/* + * Add in the number of potentially dirty inodes, because each inode + * write can dirty pagecache in the underlying blockdev. + */ +static unsigned long get_nr_dirty_pages(void) +{ + return global_node_page_state(NR_FILE_DIRTY) + + global_node_page_state(NR_UNSTABLE_NFS) + + get_nr_dirty_inodes(); +} + +static void wb_start_writeback(struct bdi_writeback *wb, bool range_cyclic, + enum wb_reason reason) { struct wb_writeback_work *work; @@ -954,7 +965,7 @@ static void wb_start_writeback(struct bdi_writeback *wb, long nr_pages, } work->sync_mode = WB_SYNC_NONE; - work->nr_pages = nr_pages; + work->nr_pages = wb_split_bdi_pages(wb, get_nr_dirty_pages()); work->range_cyclic = range_cyclic; work->reason = reason; work->auto_free = 1; @@ -1814,17 +1825,6 @@ static struct wb_writeback_work *get_next_work_item(struct bdi_writeback *wb) return work; } -/* - * Add in the number of potentially dirty inodes, because each inode - * write can dirty pagecache in the underlying blockdev. - */ -static unsigned long get_nr_dirty_pages(void) -{ - return global_node_page_state(NR_FILE_DIRTY) + - global_node_page_state(NR_UNSTABLE_NFS) + - get_nr_dirty_inodes(); -} - static long wb_check_background_flush(struct bdi_writeback *wb) { if (wb_over_bg_thresh(wb)) { @@ -1951,7 +1951,7 @@ void wb_workfn(struct work_struct *work) * write back the whole world. */ static void __wakeup_flusher_threads_bdi(struct backing_dev_info *bdi, - long nr_pages, enum wb_reason reason) + enum wb_reason reason) { struct bdi_writeback *wb; @@ -1959,17 +1959,14 @@ static void __wakeup_flusher_threads_bdi(struct backing_dev_info *bdi, return; list_for_each_entry_rcu(wb, &bdi->wb_list, bdi_node) - wb_start_writeback(wb, wb_split_bdi_pages(wb, nr_pages), - true, reason); + wb_start_writeback(wb, true, reason); } void wakeup_flusher_threads_bdi(struct backing_dev_info *bdi, enum wb_reason reason) { - long nr_pages = get_nr_dirty_pages(); - rcu_read_lock(); - __wakeup_flusher_threads_bdi(bdi, nr_pages, reason); + __wakeup_flusher_threads_bdi(bdi, reason); rcu_read_unlock(); } @@ -1979,7 +1976,6 @@ void wakeup_flusher_threads_bdi(struct backing_dev_info *bdi, void wakeup_flusher_threads(enum wb_reason reason) { struct backing_dev_info *bdi; - long nr_pages; /* * If we are expecting writeback progress we must submit plugged IO. @@ -1987,11 +1983,9 @@ void wakeup_flusher_threads(enum wb_reason reason) if (blk_needs_flush_plug(current)) blk_schedule_flush_plug(current); - nr_pages = get_nr_dirty_pages(); - rcu_read_lock(); list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) - __wakeup_flusher_threads_bdi(bdi, nr_pages, reason); + __wakeup_flusher_threads_bdi(bdi, reason); rcu_read_unlock(); } -- 2.7.4