Commit 56594ae1 authored by Parav Pandit's avatar Parav Pandit Committed by Jason Gunthorpe

RDMA/core: Annotate destroy of mutex to ensure that it is released as unlocked

While compiled with CONFIG_DEBUG_MUTEXES, the kernel ensures that mutex is
not held during destroy. Hence add mutex_destroy() for mutexes used in
RDMA modules.

Link: https://lore.kernel.org/r/20190723065733.4899-2-leon@kernel.orgSigned-off-by: default avatarParav Pandit <parav@mellanox.com>
Signed-off-by: default avatarLeon Romanovsky <leonro@mellanox.com>
Reviewed-by: default avatarJason Gunthorpe <jgg@mellanox.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@mellanox.com>
parent a511f822
...@@ -810,6 +810,7 @@ static void release_gid_table(struct ib_device *device, ...@@ -810,6 +810,7 @@ static void release_gid_table(struct ib_device *device,
if (leak) if (leak)
return; return;
mutex_destroy(&table->lock);
kfree(table->data_vec); kfree(table->data_vec);
kfree(table); kfree(table);
} }
......
...@@ -342,12 +342,18 @@ static struct configfs_subsystem cma_subsys = { ...@@ -342,12 +342,18 @@ static struct configfs_subsystem cma_subsys = {
int __init cma_configfs_init(void) int __init cma_configfs_init(void)
{ {
int ret;
config_group_init(&cma_subsys.su_group); config_group_init(&cma_subsys.su_group);
mutex_init(&cma_subsys.su_mutex); mutex_init(&cma_subsys.su_mutex);
return configfs_register_subsystem(&cma_subsys); ret = configfs_register_subsystem(&cma_subsys);
if (ret)
mutex_destroy(&cma_subsys.su_mutex);
return ret;
} }
void __exit cma_configfs_exit(void) void __exit cma_configfs_exit(void)
{ {
configfs_unregister_subsystem(&cma_subsys); configfs_unregister_subsystem(&cma_subsys);
mutex_destroy(&cma_subsys.su_mutex);
} }
...@@ -592,7 +592,7 @@ int rdma_counter_get_mode(struct ib_device *dev, u8 port, ...@@ -592,7 +592,7 @@ int rdma_counter_get_mode(struct ib_device *dev, u8 port,
void rdma_counter_init(struct ib_device *dev) void rdma_counter_init(struct ib_device *dev)
{ {
struct rdma_port_counter *port_counter; struct rdma_port_counter *port_counter;
u32 port; u32 port, i;
if (!dev->ops.alloc_hw_stats || !dev->port_data) if (!dev->ops.alloc_hw_stats || !dev->port_data)
return; return;
...@@ -610,13 +610,12 @@ void rdma_counter_init(struct ib_device *dev) ...@@ -610,13 +610,12 @@ void rdma_counter_init(struct ib_device *dev)
return; return;
fail: fail:
rdma_for_each_port(dev, port) { for (i = port; i >= rdma_start_port(dev); i--) {
port_counter = &dev->port_data[port].port_counter; port_counter = &dev->port_data[port].port_counter;
kfree(port_counter->hstats); kfree(port_counter->hstats);
port_counter->hstats = NULL; port_counter->hstats = NULL;
mutex_destroy(&port_counter->lock);
} }
return;
} }
void rdma_counter_release(struct ib_device *dev) void rdma_counter_release(struct ib_device *dev)
...@@ -630,5 +629,6 @@ void rdma_counter_release(struct ib_device *dev) ...@@ -630,5 +629,6 @@ void rdma_counter_release(struct ib_device *dev)
rdma_for_each_port(dev, port) { rdma_for_each_port(dev, port) {
port_counter = &dev->port_data[port].port_counter; port_counter = &dev->port_data[port].port_counter;
kfree(port_counter->hstats); kfree(port_counter->hstats);
mutex_destroy(&port_counter->lock);
} }
} }
...@@ -508,6 +508,9 @@ static void ib_device_release(struct device *device) ...@@ -508,6 +508,9 @@ static void ib_device_release(struct device *device)
rcu_head); rcu_head);
} }
mutex_destroy(&dev->unregistration_lock);
mutex_destroy(&dev->compat_devs_mutex);
xa_destroy(&dev->compat_devs); xa_destroy(&dev->compat_devs);
xa_destroy(&dev->client_data); xa_destroy(&dev->client_data);
kfree_rcu(dev, rcu_head); kfree_rcu(dev, rcu_head);
......
...@@ -1038,7 +1038,7 @@ static int ib_umad_close(struct inode *inode, struct file *filp) ...@@ -1038,7 +1038,7 @@ static int ib_umad_close(struct inode *inode, struct file *filp)
ib_unregister_mad_agent(file->agent[i]); ib_unregister_mad_agent(file->agent[i]);
mutex_unlock(&file->port->file_mutex); mutex_unlock(&file->port->file_mutex);
mutex_destroy(&file->mutex);
kfree(file); kfree(file);
return 0; return 0;
} }
......
...@@ -120,6 +120,8 @@ static void ib_uverbs_release_dev(struct device *device) ...@@ -120,6 +120,8 @@ static void ib_uverbs_release_dev(struct device *device)
uverbs_destroy_api(dev->uapi); uverbs_destroy_api(dev->uapi);
cleanup_srcu_struct(&dev->disassociate_srcu); cleanup_srcu_struct(&dev->disassociate_srcu);
mutex_destroy(&dev->lists_mutex);
mutex_destroy(&dev->xrcd_tree_mutex);
kfree(dev); kfree(dev);
} }
...@@ -212,6 +214,8 @@ void ib_uverbs_release_file(struct kref *ref) ...@@ -212,6 +214,8 @@ void ib_uverbs_release_file(struct kref *ref)
if (file->disassociate_page) if (file->disassociate_page)
__free_pages(file->disassociate_page, 0); __free_pages(file->disassociate_page, 0);
mutex_destroy(&file->umap_lock);
mutex_destroy(&file->ucontext_lock);
kfree(file); kfree(file);
} }
......
...@@ -2259,6 +2259,7 @@ int ib_dealloc_xrcd(struct ib_xrcd *xrcd, struct ib_udata *udata) ...@@ -2259,6 +2259,7 @@ int ib_dealloc_xrcd(struct ib_xrcd *xrcd, struct ib_udata *udata)
if (ret) if (ret)
return ret; return ret;
} }
mutex_destroy(&xrcd->tgt_qp_mutex);
return xrcd->device->ops.dealloc_xrcd(xrcd, udata); return xrcd->device->ops.dealloc_xrcd(xrcd, udata);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment