Commit 3e5a428b authored by Darrick J. Wong's avatar Darrick J. Wong

xfs: poll waiting for quotacheck

Create a pwork destroy function that uses polling instead of
uninterruptible sleep to wait for work items to finish so that we can
touch the softlockup watchdog.  IOWs, gross hack.
Signed-off-by: default avatarDarrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: default avatarBrian Foster <bfoster@redhat.com>
parent 40786717
...@@ -588,6 +588,7 @@ xfs_iwalk_threaded( ...@@ -588,6 +588,7 @@ xfs_iwalk_threaded(
xfs_ino_t startino, xfs_ino_t startino,
xfs_iwalk_fn iwalk_fn, xfs_iwalk_fn iwalk_fn,
unsigned int inode_records, unsigned int inode_records,
bool polled,
void *data) void *data)
{ {
struct xfs_pwork_ctl pctl; struct xfs_pwork_ctl pctl;
...@@ -619,6 +620,8 @@ xfs_iwalk_threaded( ...@@ -619,6 +620,8 @@ xfs_iwalk_threaded(
startino = XFS_AGINO_TO_INO(mp, agno + 1, 0); startino = XFS_AGINO_TO_INO(mp, agno + 1, 0);
} }
if (polled)
xfs_pwork_poll(&pctl);
return xfs_pwork_destroy(&pctl); return xfs_pwork_destroy(&pctl);
} }
......
...@@ -16,7 +16,8 @@ typedef int (*xfs_iwalk_fn)(struct xfs_mount *mp, struct xfs_trans *tp, ...@@ -16,7 +16,8 @@ typedef int (*xfs_iwalk_fn)(struct xfs_mount *mp, struct xfs_trans *tp,
int xfs_iwalk(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t startino, int xfs_iwalk(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t startino,
xfs_iwalk_fn iwalk_fn, unsigned int inode_records, void *data); xfs_iwalk_fn iwalk_fn, unsigned int inode_records, void *data);
int xfs_iwalk_threaded(struct xfs_mount *mp, xfs_ino_t startino, int xfs_iwalk_threaded(struct xfs_mount *mp, xfs_ino_t startino,
xfs_iwalk_fn iwalk_fn, unsigned int inode_records, void *data); xfs_iwalk_fn iwalk_fn, unsigned int inode_records, bool poll,
void *data);
/* Walk all inode btree records in the filesystem starting from @startino. */ /* Walk all inode btree records in the filesystem starting from @startino. */
typedef int (*xfs_inobt_walk_fn)(struct xfs_mount *mp, struct xfs_trans *tp, typedef int (*xfs_inobt_walk_fn)(struct xfs_mount *mp, struct xfs_trans *tp,
......
...@@ -13,6 +13,7 @@ ...@@ -13,6 +13,7 @@
#include "xfs_trace.h" #include "xfs_trace.h"
#include "xfs_sysctl.h" #include "xfs_sysctl.h"
#include "xfs_pwork.h" #include "xfs_pwork.h"
#include <linux/nmi.h>
/* /*
* Parallel Work Queue * Parallel Work Queue
...@@ -46,6 +47,8 @@ xfs_pwork_work( ...@@ -46,6 +47,8 @@ xfs_pwork_work(
error = pctl->work_fn(pctl->mp, pwork); error = pctl->work_fn(pctl->mp, pwork);
if (error && !pctl->error) if (error && !pctl->error)
pctl->error = error; pctl->error = error;
if (atomic_dec_and_test(&pctl->nr_work))
wake_up(&pctl->poll_wait);
} }
/* /*
...@@ -74,6 +77,8 @@ xfs_pwork_init( ...@@ -74,6 +77,8 @@ xfs_pwork_init(
pctl->work_fn = work_fn; pctl->work_fn = work_fn;
pctl->error = 0; pctl->error = 0;
pctl->mp = mp; pctl->mp = mp;
atomic_set(&pctl->nr_work, 0);
init_waitqueue_head(&pctl->poll_wait);
return 0; return 0;
} }
...@@ -86,6 +91,7 @@ xfs_pwork_queue( ...@@ -86,6 +91,7 @@ xfs_pwork_queue(
{ {
INIT_WORK(&pwork->work, xfs_pwork_work); INIT_WORK(&pwork->work, xfs_pwork_work);
pwork->pctl = pctl; pwork->pctl = pctl;
atomic_inc(&pctl->nr_work);
queue_work(pctl->wq, &pwork->work); queue_work(pctl->wq, &pwork->work);
} }
...@@ -99,6 +105,19 @@ xfs_pwork_destroy( ...@@ -99,6 +105,19 @@ xfs_pwork_destroy(
return pctl->error; return pctl->error;
} }
/*
* Wait for the work to finish by polling completion status and touch the soft
* lockup watchdog. This is for callers such as mount which hold locks.
*/
void
xfs_pwork_poll(
struct xfs_pwork_ctl *pctl)
{
while (wait_event_timeout(pctl->poll_wait,
atomic_read(&pctl->nr_work) == 0, HZ) == 0)
touch_softlockup_watchdog();
}
/* /*
* Return the amount of parallelism that the data device can handle, or 0 for * Return the amount of parallelism that the data device can handle, or 0 for
* no limit. * no limit.
......
...@@ -18,6 +18,8 @@ struct xfs_pwork_ctl { ...@@ -18,6 +18,8 @@ struct xfs_pwork_ctl {
struct workqueue_struct *wq; struct workqueue_struct *wq;
struct xfs_mount *mp; struct xfs_mount *mp;
xfs_pwork_work_fn work_fn; xfs_pwork_work_fn work_fn;
struct wait_queue_head poll_wait;
atomic_t nr_work;
int error; int error;
}; };
...@@ -53,6 +55,7 @@ int xfs_pwork_init(struct xfs_mount *mp, struct xfs_pwork_ctl *pctl, ...@@ -53,6 +55,7 @@ int xfs_pwork_init(struct xfs_mount *mp, struct xfs_pwork_ctl *pctl,
unsigned int nr_threads); unsigned int nr_threads);
void xfs_pwork_queue(struct xfs_pwork_ctl *pctl, struct xfs_pwork *pwork); void xfs_pwork_queue(struct xfs_pwork_ctl *pctl, struct xfs_pwork *pwork);
int xfs_pwork_destroy(struct xfs_pwork_ctl *pctl); int xfs_pwork_destroy(struct xfs_pwork_ctl *pctl);
void xfs_pwork_poll(struct xfs_pwork_ctl *pctl);
unsigned int xfs_pwork_guess_datadev_parallelism(struct xfs_mount *mp); unsigned int xfs_pwork_guess_datadev_parallelism(struct xfs_mount *mp);
#endif /* __XFS_PWORK_H__ */ #endif /* __XFS_PWORK_H__ */
...@@ -1300,7 +1300,7 @@ xfs_qm_quotacheck( ...@@ -1300,7 +1300,7 @@ xfs_qm_quotacheck(
flags |= XFS_PQUOTA_CHKD; flags |= XFS_PQUOTA_CHKD;
} }
error = xfs_iwalk_threaded(mp, 0, xfs_qm_dqusage_adjust, 0, NULL); error = xfs_iwalk_threaded(mp, 0, xfs_qm_dqusage_adjust, 0, true, NULL);
if (error) if (error)
goto error_return; goto error_return;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment