From: Dave Chinner <[email protected]>
This function is now entirely done by kswapd, so we don't need the
worker thread to do async reclaim anymore.
Signed-off-by: Dave Chinner <[email protected]>
Reviewed-by: Christoph Hellwig <[email protected]>
---
fs/xfs/xfs_icache.c | 44 --------------------------------------------
fs/xfs/xfs_icache.h | 2 --
fs/xfs/xfs_mount.c | 2 --
fs/xfs/xfs_mount.h | 2 --
fs/xfs/xfs_super.c | 11 +----------
5 files changed, 1 insertion(+), 60 deletions(-)
diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c
index 189cf423fe8f..7e175304e146 100644
--- a/fs/xfs/xfs_icache.c
+++ b/fs/xfs/xfs_icache.c
@@ -138,44 +138,6 @@ xfs_inode_free(
__xfs_inode_free(ip);
}
-/*
- * Queue a new inode reclaim pass if there are reclaimable inodes and there
- * isn't a reclaim pass already in progress. By default it runs every 5s based
- * on the xfs periodic sync default of 30s. Perhaps this should have it's own
- * tunable, but that can be done if this method proves to be ineffective or too
- * aggressive.
- */
-static void
-xfs_reclaim_work_queue(
- struct xfs_mount *mp)
-{
-
- rcu_read_lock();
- if (radix_tree_tagged(&mp->m_perag_tree, XFS_ICI_RECLAIM_TAG)) {
- queue_delayed_work(mp->m_reclaim_workqueue, &mp->m_reclaim_work,
- msecs_to_jiffies(xfs_syncd_centisecs / 6 * 10));
- }
- rcu_read_unlock();
-}
-
-/*
- * This is a fast pass over the inode cache to try to get reclaim moving on as
- * many inodes as possible in a short period of time. It kicks itself every few
- * seconds, as well as being kicked by the inode cache shrinker when memory
- * goes low. It scans as quickly as possible avoiding locked inodes or those
- * already being flushed, and once done schedules a future pass.
- */
-void
-xfs_reclaim_worker(
- struct work_struct *work)
-{
- struct xfs_mount *mp = container_of(to_delayed_work(work),
- struct xfs_mount, m_reclaim_work);
-
- xfs_reclaim_inodes(mp, SYNC_TRYLOCK);
- xfs_reclaim_work_queue(mp);
-}
-
static void
xfs_perag_set_reclaim_tag(
struct xfs_perag *pag)
@@ -192,9 +154,6 @@ xfs_perag_set_reclaim_tag(
XFS_ICI_RECLAIM_TAG);
spin_unlock(&mp->m_perag_lock);
- /* schedule periodic background inode reclaim */
- xfs_reclaim_work_queue(mp);
-
trace_xfs_perag_set_reclaim(mp, pag->pag_agno, -1, _RET_IP_);
}
@@ -1393,9 +1352,6 @@ xfs_reclaim_inodes_nr(
{
int sync_mode = SYNC_TRYLOCK;
- /* kick background reclaimer */
- xfs_reclaim_work_queue(mp);
-
/*
* For kswapd, we kick background inode writeback. For direct
* reclaim, we issue and wait on inode writeback to throttle
diff --git a/fs/xfs/xfs_icache.h b/fs/xfs/xfs_icache.h
index 48f1fd2bb6ad..4c0d8920cc54 100644
--- a/fs/xfs/xfs_icache.h
+++ b/fs/xfs/xfs_icache.h
@@ -49,8 +49,6 @@ int xfs_iget(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t ino,
struct xfs_inode * xfs_inode_alloc(struct xfs_mount *mp, xfs_ino_t ino);
void xfs_inode_free(struct xfs_inode *ip);
-void xfs_reclaim_worker(struct work_struct *work);
-
int xfs_reclaim_inodes(struct xfs_mount *mp, int mode);
int xfs_reclaim_inodes_count(struct xfs_mount *mp);
long xfs_reclaim_inodes_nr(struct xfs_mount *mp, int nr_to_scan);
diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c
index 3e8eedf01eb2..8f76c2add18b 100644
--- a/fs/xfs/xfs_mount.c
+++ b/fs/xfs/xfs_mount.c
@@ -952,7 +952,6 @@ xfs_mountfs(
* qm_unmount_quotas and therefore rely on qm_unmount to release the
* quota inodes.
*/
- cancel_delayed_work_sync(&mp->m_reclaim_work);
xfs_reclaim_inodes(mp, SYNC_WAIT);
xfs_health_unmount(mp);
out_log_dealloc:
@@ -1035,7 +1034,6 @@ xfs_unmountfs(
* reclaim just to be sure. We can stop background inode reclaim
* here as well if it is still running.
*/
- cancel_delayed_work_sync(&mp->m_reclaim_work);
xfs_reclaim_inodes(mp, SYNC_WAIT);
xfs_health_unmount(mp);
diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h
index a46cb3fd24b1..8c6885d3b085 100644
--- a/fs/xfs/xfs_mount.h
+++ b/fs/xfs/xfs_mount.h
@@ -163,7 +163,6 @@ typedef struct xfs_mount {
uint m_chsize; /* size of next field */
atomic_t m_active_trans; /* number trans frozen */
struct xfs_mru_cache *m_filestream; /* per-mount filestream data */
- struct delayed_work m_reclaim_work; /* background inode reclaim */
struct delayed_work m_eofblocks_work; /* background eof blocks
trimming */
struct delayed_work m_cowblocks_work; /* background cow blocks
@@ -180,7 +179,6 @@ typedef struct xfs_mount {
struct workqueue_struct *m_buf_workqueue;
struct workqueue_struct *m_unwritten_workqueue;
struct workqueue_struct *m_cil_workqueue;
- struct workqueue_struct *m_reclaim_workqueue;
struct workqueue_struct *m_eofblocks_workqueue;
struct workqueue_struct *m_sync_workqueue;
diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c
index ebe2ccd36127..a4fe679207ef 100644
--- a/fs/xfs/xfs_super.c
+++ b/fs/xfs/xfs_super.c
@@ -794,15 +794,10 @@ xfs_init_mount_workqueues(
if (!mp->m_cil_workqueue)
goto out_destroy_unwritten;
- mp->m_reclaim_workqueue = alloc_workqueue("xfs-reclaim/%s",
- WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
- if (!mp->m_reclaim_workqueue)
- goto out_destroy_cil;
-
mp->m_eofblocks_workqueue = alloc_workqueue("xfs-eofblocks/%s",
WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
if (!mp->m_eofblocks_workqueue)
- goto out_destroy_reclaim;
+ goto out_destroy_cil;
mp->m_sync_workqueue = alloc_workqueue("xfs-sync/%s", WQ_FREEZABLE, 0,
mp->m_fsname);
@@ -813,8 +808,6 @@ xfs_init_mount_workqueues(
out_destroy_eofb:
destroy_workqueue(mp->m_eofblocks_workqueue);
-out_destroy_reclaim:
- destroy_workqueue(mp->m_reclaim_workqueue);
out_destroy_cil:
destroy_workqueue(mp->m_cil_workqueue);
out_destroy_unwritten:
@@ -831,7 +824,6 @@ xfs_destroy_mount_workqueues(
{
destroy_workqueue(mp->m_sync_workqueue);
destroy_workqueue(mp->m_eofblocks_workqueue);
- destroy_workqueue(mp->m_reclaim_workqueue);
destroy_workqueue(mp->m_cil_workqueue);
destroy_workqueue(mp->m_unwritten_workqueue);
destroy_workqueue(mp->m_buf_workqueue);
@@ -1520,7 +1512,6 @@ xfs_mount_alloc(
spin_lock_init(&mp->m_perag_lock);
mutex_init(&mp->m_growlock);
atomic_set(&mp->m_active_trans, 0);
- INIT_DELAYED_WORK(&mp->m_reclaim_work, xfs_reclaim_worker);
INIT_DELAYED_WORK(&mp->m_eofblocks_work, xfs_eofblocks_worker);
INIT_DELAYED_WORK(&mp->m_cowblocks_work, xfs_cowblocks_worker);
mp->m_kobj.kobject.kset = xfs_kset;
--
2.24.0.rc0
On Fri, Nov 01, 2019 at 10:46:10AM +1100, Dave Chinner wrote:
> From: Dave Chinner <[email protected]>
>
> This function is now entirely done by kswapd, so we don't need the
> worker thread to do async reclaim anymore.
>
> Signed-off-by: Dave Chinner <[email protected]>
> Reviewed-by: Christoph Hellwig <[email protected]>
> ---
Reviewed-by: Brian Foster <[email protected]>
> fs/xfs/xfs_icache.c | 44 --------------------------------------------
> fs/xfs/xfs_icache.h | 2 --
> fs/xfs/xfs_mount.c | 2 --
> fs/xfs/xfs_mount.h | 2 --
> fs/xfs/xfs_super.c | 11 +----------
> 5 files changed, 1 insertion(+), 60 deletions(-)
>
> diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c
> index 189cf423fe8f..7e175304e146 100644
> --- a/fs/xfs/xfs_icache.c
> +++ b/fs/xfs/xfs_icache.c
> @@ -138,44 +138,6 @@ xfs_inode_free(
> __xfs_inode_free(ip);
> }
>
> -/*
> - * Queue a new inode reclaim pass if there are reclaimable inodes and there
> - * isn't a reclaim pass already in progress. By default it runs every 5s based
> - * on the xfs periodic sync default of 30s. Perhaps this should have it's own
> - * tunable, but that can be done if this method proves to be ineffective or too
> - * aggressive.
> - */
> -static void
> -xfs_reclaim_work_queue(
> - struct xfs_mount *mp)
> -{
> -
> - rcu_read_lock();
> - if (radix_tree_tagged(&mp->m_perag_tree, XFS_ICI_RECLAIM_TAG)) {
> - queue_delayed_work(mp->m_reclaim_workqueue, &mp->m_reclaim_work,
> - msecs_to_jiffies(xfs_syncd_centisecs / 6 * 10));
> - }
> - rcu_read_unlock();
> -}
> -
> -/*
> - * This is a fast pass over the inode cache to try to get reclaim moving on as
> - * many inodes as possible in a short period of time. It kicks itself every few
> - * seconds, as well as being kicked by the inode cache shrinker when memory
> - * goes low. It scans as quickly as possible avoiding locked inodes or those
> - * already being flushed, and once done schedules a future pass.
> - */
> -void
> -xfs_reclaim_worker(
> - struct work_struct *work)
> -{
> - struct xfs_mount *mp = container_of(to_delayed_work(work),
> - struct xfs_mount, m_reclaim_work);
> -
> - xfs_reclaim_inodes(mp, SYNC_TRYLOCK);
> - xfs_reclaim_work_queue(mp);
> -}
> -
> static void
> xfs_perag_set_reclaim_tag(
> struct xfs_perag *pag)
> @@ -192,9 +154,6 @@ xfs_perag_set_reclaim_tag(
> XFS_ICI_RECLAIM_TAG);
> spin_unlock(&mp->m_perag_lock);
>
> - /* schedule periodic background inode reclaim */
> - xfs_reclaim_work_queue(mp);
> -
> trace_xfs_perag_set_reclaim(mp, pag->pag_agno, -1, _RET_IP_);
> }
>
> @@ -1393,9 +1352,6 @@ xfs_reclaim_inodes_nr(
> {
> int sync_mode = SYNC_TRYLOCK;
>
> - /* kick background reclaimer */
> - xfs_reclaim_work_queue(mp);
> -
> /*
> * For kswapd, we kick background inode writeback. For direct
> * reclaim, we issue and wait on inode writeback to throttle
> diff --git a/fs/xfs/xfs_icache.h b/fs/xfs/xfs_icache.h
> index 48f1fd2bb6ad..4c0d8920cc54 100644
> --- a/fs/xfs/xfs_icache.h
> +++ b/fs/xfs/xfs_icache.h
> @@ -49,8 +49,6 @@ int xfs_iget(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t ino,
> struct xfs_inode * xfs_inode_alloc(struct xfs_mount *mp, xfs_ino_t ino);
> void xfs_inode_free(struct xfs_inode *ip);
>
> -void xfs_reclaim_worker(struct work_struct *work);
> -
> int xfs_reclaim_inodes(struct xfs_mount *mp, int mode);
> int xfs_reclaim_inodes_count(struct xfs_mount *mp);
> long xfs_reclaim_inodes_nr(struct xfs_mount *mp, int nr_to_scan);
> diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c
> index 3e8eedf01eb2..8f76c2add18b 100644
> --- a/fs/xfs/xfs_mount.c
> +++ b/fs/xfs/xfs_mount.c
> @@ -952,7 +952,6 @@ xfs_mountfs(
> * qm_unmount_quotas and therefore rely on qm_unmount to release the
> * quota inodes.
> */
> - cancel_delayed_work_sync(&mp->m_reclaim_work);
> xfs_reclaim_inodes(mp, SYNC_WAIT);
> xfs_health_unmount(mp);
> out_log_dealloc:
> @@ -1035,7 +1034,6 @@ xfs_unmountfs(
> * reclaim just to be sure. We can stop background inode reclaim
> * here as well if it is still running.
> */
> - cancel_delayed_work_sync(&mp->m_reclaim_work);
> xfs_reclaim_inodes(mp, SYNC_WAIT);
> xfs_health_unmount(mp);
>
> diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h
> index a46cb3fd24b1..8c6885d3b085 100644
> --- a/fs/xfs/xfs_mount.h
> +++ b/fs/xfs/xfs_mount.h
> @@ -163,7 +163,6 @@ typedef struct xfs_mount {
> uint m_chsize; /* size of next field */
> atomic_t m_active_trans; /* number trans frozen */
> struct xfs_mru_cache *m_filestream; /* per-mount filestream data */
> - struct delayed_work m_reclaim_work; /* background inode reclaim */
> struct delayed_work m_eofblocks_work; /* background eof blocks
> trimming */
> struct delayed_work m_cowblocks_work; /* background cow blocks
> @@ -180,7 +179,6 @@ typedef struct xfs_mount {
> struct workqueue_struct *m_buf_workqueue;
> struct workqueue_struct *m_unwritten_workqueue;
> struct workqueue_struct *m_cil_workqueue;
> - struct workqueue_struct *m_reclaim_workqueue;
> struct workqueue_struct *m_eofblocks_workqueue;
> struct workqueue_struct *m_sync_workqueue;
>
> diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c
> index ebe2ccd36127..a4fe679207ef 100644
> --- a/fs/xfs/xfs_super.c
> +++ b/fs/xfs/xfs_super.c
> @@ -794,15 +794,10 @@ xfs_init_mount_workqueues(
> if (!mp->m_cil_workqueue)
> goto out_destroy_unwritten;
>
> - mp->m_reclaim_workqueue = alloc_workqueue("xfs-reclaim/%s",
> - WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
> - if (!mp->m_reclaim_workqueue)
> - goto out_destroy_cil;
> -
> mp->m_eofblocks_workqueue = alloc_workqueue("xfs-eofblocks/%s",
> WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
> if (!mp->m_eofblocks_workqueue)
> - goto out_destroy_reclaim;
> + goto out_destroy_cil;
>
> mp->m_sync_workqueue = alloc_workqueue("xfs-sync/%s", WQ_FREEZABLE, 0,
> mp->m_fsname);
> @@ -813,8 +808,6 @@ xfs_init_mount_workqueues(
>
> out_destroy_eofb:
> destroy_workqueue(mp->m_eofblocks_workqueue);
> -out_destroy_reclaim:
> - destroy_workqueue(mp->m_reclaim_workqueue);
> out_destroy_cil:
> destroy_workqueue(mp->m_cil_workqueue);
> out_destroy_unwritten:
> @@ -831,7 +824,6 @@ xfs_destroy_mount_workqueues(
> {
> destroy_workqueue(mp->m_sync_workqueue);
> destroy_workqueue(mp->m_eofblocks_workqueue);
> - destroy_workqueue(mp->m_reclaim_workqueue);
> destroy_workqueue(mp->m_cil_workqueue);
> destroy_workqueue(mp->m_unwritten_workqueue);
> destroy_workqueue(mp->m_buf_workqueue);
> @@ -1520,7 +1512,6 @@ xfs_mount_alloc(
> spin_lock_init(&mp->m_perag_lock);
> mutex_init(&mp->m_growlock);
> atomic_set(&mp->m_active_trans, 0);
> - INIT_DELAYED_WORK(&mp->m_reclaim_work, xfs_reclaim_worker);
> INIT_DELAYED_WORK(&mp->m_eofblocks_work, xfs_eofblocks_worker);
> INIT_DELAYED_WORK(&mp->m_cowblocks_work, xfs_cowblocks_worker);
> mp->m_kobj.kobject.kset = xfs_kset;
> --
> 2.24.0.rc0
>