Commit b44166cd authored by Bob Liu's avatar Bob Liu Committed by David Vrabel

xen/grant: introduce func gnttab_unmap_refs_sync()

There are several place using gnttab async unmap and wait for
completion, so move the common code to a function
gnttab_unmap_refs_sync().
Signed-off-by: default avatarBob Liu <bob.liu@oracle.com>
Acked-by: default avatarRoger Pau Monné <roger.pau@citrix.com>
Acked-by: default avatarKonrad Rzeszutek Wilk <konrad.wilk@oracle.com>
Signed-off-by: default avatarDavid Vrabel <david.vrabel@citrix.com>
parent 325d73bf
...@@ -265,17 +265,6 @@ static void put_persistent_gnt(struct xen_blkif *blkif, ...@@ -265,17 +265,6 @@ static void put_persistent_gnt(struct xen_blkif *blkif,
atomic_dec(&blkif->persistent_gnt_in_use); atomic_dec(&blkif->persistent_gnt_in_use);
} }
static void free_persistent_gnts_unmap_callback(int result,
struct gntab_unmap_queue_data *data)
{
struct completion *c = data->data;
/* BUG_ON used to reproduce existing behaviour,
but is this the best way to deal with this? */
BUG_ON(result);
complete(c);
}
static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root, static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root,
unsigned int num) unsigned int num)
{ {
...@@ -285,12 +274,7 @@ static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root, ...@@ -285,12 +274,7 @@ static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root,
struct rb_node *n; struct rb_node *n;
int segs_to_unmap = 0; int segs_to_unmap = 0;
struct gntab_unmap_queue_data unmap_data; struct gntab_unmap_queue_data unmap_data;
struct completion unmap_completion;
init_completion(&unmap_completion);
unmap_data.data = &unmap_completion;
unmap_data.done = &free_persistent_gnts_unmap_callback;
unmap_data.pages = pages; unmap_data.pages = pages;
unmap_data.unmap_ops = unmap; unmap_data.unmap_ops = unmap;
unmap_data.kunmap_ops = NULL; unmap_data.kunmap_ops = NULL;
...@@ -310,8 +294,7 @@ static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root, ...@@ -310,8 +294,7 @@ static void free_persistent_gnts(struct xen_blkif *blkif, struct rb_root *root,
!rb_next(&persistent_gnt->node)) { !rb_next(&persistent_gnt->node)) {
unmap_data.count = segs_to_unmap; unmap_data.count = segs_to_unmap;
gnttab_unmap_refs_async(&unmap_data); BUG_ON(gnttab_unmap_refs_sync(&unmap_data));
wait_for_completion(&unmap_completion);
put_free_pages(blkif, pages, segs_to_unmap); put_free_pages(blkif, pages, segs_to_unmap);
segs_to_unmap = 0; segs_to_unmap = 0;
...@@ -332,12 +315,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work) ...@@ -332,12 +315,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work)
int segs_to_unmap = 0; int segs_to_unmap = 0;
struct xen_blkif *blkif = container_of(work, typeof(*blkif), persistent_purge_work); struct xen_blkif *blkif = container_of(work, typeof(*blkif), persistent_purge_work);
struct gntab_unmap_queue_data unmap_data; struct gntab_unmap_queue_data unmap_data;
struct completion unmap_completion;
init_completion(&unmap_completion);
unmap_data.data = &unmap_completion;
unmap_data.done = &free_persistent_gnts_unmap_callback;
unmap_data.pages = pages; unmap_data.pages = pages;
unmap_data.unmap_ops = unmap; unmap_data.unmap_ops = unmap;
unmap_data.kunmap_ops = NULL; unmap_data.kunmap_ops = NULL;
...@@ -357,9 +335,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work) ...@@ -357,9 +335,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work)
if (++segs_to_unmap == BLKIF_MAX_SEGMENTS_PER_REQUEST) { if (++segs_to_unmap == BLKIF_MAX_SEGMENTS_PER_REQUEST) {
unmap_data.count = segs_to_unmap; unmap_data.count = segs_to_unmap;
gnttab_unmap_refs_async(&unmap_data); BUG_ON(gnttab_unmap_refs_sync(&unmap_data));
wait_for_completion(&unmap_completion);
put_free_pages(blkif, pages, segs_to_unmap); put_free_pages(blkif, pages, segs_to_unmap);
segs_to_unmap = 0; segs_to_unmap = 0;
} }
...@@ -367,8 +343,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work) ...@@ -367,8 +343,7 @@ void xen_blkbk_unmap_purged_grants(struct work_struct *work)
} }
if (segs_to_unmap > 0) { if (segs_to_unmap > 0) {
unmap_data.count = segs_to_unmap; unmap_data.count = segs_to_unmap;
gnttab_unmap_refs_async(&unmap_data); BUG_ON(gnttab_unmap_refs_sync(&unmap_data));
wait_for_completion(&unmap_completion);
put_free_pages(blkif, pages, segs_to_unmap); put_free_pages(blkif, pages, segs_to_unmap);
} }
} }
......
...@@ -327,30 +327,10 @@ static int map_grant_pages(struct grant_map *map) ...@@ -327,30 +327,10 @@ static int map_grant_pages(struct grant_map *map)
return err; return err;
} }
struct unmap_grant_pages_callback_data
{
struct completion completion;
int result;
};
static void unmap_grant_callback(int result,
struct gntab_unmap_queue_data *data)
{
struct unmap_grant_pages_callback_data* d = data->data;
d->result = result;
complete(&d->completion);
}
static int __unmap_grant_pages(struct grant_map *map, int offset, int pages) static int __unmap_grant_pages(struct grant_map *map, int offset, int pages)
{ {
int i, err = 0; int i, err = 0;
struct gntab_unmap_queue_data unmap_data; struct gntab_unmap_queue_data unmap_data;
struct unmap_grant_pages_callback_data data;
init_completion(&data.completion);
unmap_data.data = &data;
unmap_data.done= &unmap_grant_callback;
if (map->notify.flags & UNMAP_NOTIFY_CLEAR_BYTE) { if (map->notify.flags & UNMAP_NOTIFY_CLEAR_BYTE) {
int pgno = (map->notify.addr >> PAGE_SHIFT); int pgno = (map->notify.addr >> PAGE_SHIFT);
...@@ -367,11 +347,9 @@ static int __unmap_grant_pages(struct grant_map *map, int offset, int pages) ...@@ -367,11 +347,9 @@ static int __unmap_grant_pages(struct grant_map *map, int offset, int pages)
unmap_data.pages = map->pages + offset; unmap_data.pages = map->pages + offset;
unmap_data.count = pages; unmap_data.count = pages;
gnttab_unmap_refs_async(&unmap_data); err = gnttab_unmap_refs_sync(&unmap_data);
if (err)
wait_for_completion(&data.completion); return err;
if (data.result)
return data.result;
for (i = 0; i < pages; i++) { for (i = 0; i < pages; i++) {
if (map->unmap_ops[offset+i].status) if (map->unmap_ops[offset+i].status)
......
...@@ -123,6 +123,11 @@ struct gnttab_ops { ...@@ -123,6 +123,11 @@ struct gnttab_ops {
int (*query_foreign_access)(grant_ref_t ref); int (*query_foreign_access)(grant_ref_t ref);
}; };
struct unmap_refs_callback_data {
struct completion completion;
int result;
};
static struct gnttab_ops *gnttab_interface; static struct gnttab_ops *gnttab_interface;
static int grant_table_version; static int grant_table_version;
...@@ -863,6 +868,29 @@ void gnttab_unmap_refs_async(struct gntab_unmap_queue_data* item) ...@@ -863,6 +868,29 @@ void gnttab_unmap_refs_async(struct gntab_unmap_queue_data* item)
} }
EXPORT_SYMBOL_GPL(gnttab_unmap_refs_async); EXPORT_SYMBOL_GPL(gnttab_unmap_refs_async);
static void unmap_refs_callback(int result,
struct gntab_unmap_queue_data *data)
{
struct unmap_refs_callback_data *d = data->data;
d->result = result;
complete(&d->completion);
}
int gnttab_unmap_refs_sync(struct gntab_unmap_queue_data *item)
{
struct unmap_refs_callback_data data;
init_completion(&data.completion);
item->data = &data;
item->done = &unmap_refs_callback;
gnttab_unmap_refs_async(item);
wait_for_completion(&data.completion);
return data.result;
}
EXPORT_SYMBOL_GPL(gnttab_unmap_refs_sync);
static int gnttab_map_frames_v1(xen_pfn_t *frames, unsigned int nr_gframes) static int gnttab_map_frames_v1(xen_pfn_t *frames, unsigned int nr_gframes)
{ {
int rc; int rc;
......
...@@ -191,6 +191,7 @@ int gnttab_unmap_refs(struct gnttab_unmap_grant_ref *unmap_ops, ...@@ -191,6 +191,7 @@ int gnttab_unmap_refs(struct gnttab_unmap_grant_ref *unmap_ops,
struct gnttab_unmap_grant_ref *kunmap_ops, struct gnttab_unmap_grant_ref *kunmap_ops,
struct page **pages, unsigned int count); struct page **pages, unsigned int count);
void gnttab_unmap_refs_async(struct gntab_unmap_queue_data* item); void gnttab_unmap_refs_async(struct gntab_unmap_queue_data* item);
int gnttab_unmap_refs_sync(struct gntab_unmap_queue_data *item);
/* Perform a batch of grant map/copy operations. Retry every batch slot /* Perform a batch of grant map/copy operations. Retry every batch slot
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment