Commit 451bc8eb authored by Christian König's avatar Christian König Committed by Alex Deucher

drm/amdgpu: fix PRT teardown on VM fini v3

v2: new approach fixing this by registering a fence callback for
    all users of the VM on teardown
v3: agd: rebase
Signed-off-by: default avatarChristian König <christian.koenig@amd.com>
Reviewed-by: default avatarNicolai Hähnle <nicolai.haehnle@amd.com>
Signed-off-by: default avatarAlex Deucher <alexander.deucher@amd.com>
parent 0b15f2fc
...@@ -1188,22 +1188,31 @@ static void amdgpu_vm_update_prt_state(struct amdgpu_device *adev) ...@@ -1188,22 +1188,31 @@ static void amdgpu_vm_update_prt_state(struct amdgpu_device *adev)
bool enable; bool enable;
spin_lock_irqsave(&adev->vm_manager.prt_lock, flags); spin_lock_irqsave(&adev->vm_manager.prt_lock, flags);
enable = !!atomic_read(&adev->vm_manager.num_prt_mappings); enable = !!atomic_read(&adev->vm_manager.num_prt_users);
adev->gart.gart_funcs->set_prt(adev, enable); adev->gart.gart_funcs->set_prt(adev, enable);
spin_unlock_irqrestore(&adev->vm_manager.prt_lock, flags); spin_unlock_irqrestore(&adev->vm_manager.prt_lock, flags);
} }
/**
* amdgpu_vm_prt_put - add a PRT user
*/
static void amdgpu_vm_prt_get(struct amdgpu_device *adev)
{
if (atomic_inc_return(&adev->vm_manager.num_prt_users) == 1)
amdgpu_vm_update_prt_state(adev);
}
/** /**
* amdgpu_vm_prt_put - drop a PRT user * amdgpu_vm_prt_put - drop a PRT user
*/ */
static void amdgpu_vm_prt_put(struct amdgpu_device *adev) static void amdgpu_vm_prt_put(struct amdgpu_device *adev)
{ {
if (atomic_dec_return(&adev->vm_manager.num_prt_mappings) == 0) if (atomic_dec_return(&adev->vm_manager.num_prt_users) == 0)
amdgpu_vm_update_prt_state(adev); amdgpu_vm_update_prt_state(adev);
} }
/** /**
* amdgpu_vm_prt - callback for updating the PRT status * amdgpu_vm_prt_cb - callback for updating the PRT status
*/ */
static void amdgpu_vm_prt_cb(struct dma_fence *fence, struct dma_fence_cb *_cb) static void amdgpu_vm_prt_cb(struct dma_fence *fence, struct dma_fence_cb *_cb)
{ {
...@@ -1214,21 +1223,11 @@ static void amdgpu_vm_prt_cb(struct dma_fence *fence, struct dma_fence_cb *_cb) ...@@ -1214,21 +1223,11 @@ static void amdgpu_vm_prt_cb(struct dma_fence *fence, struct dma_fence_cb *_cb)
} }
/** /**
* amdgpu_vm_free_mapping - free a mapping * amdgpu_vm_add_prt_cb - add callback for updating the PRT status
*
* @adev: amdgpu_device pointer
* @vm: requested vm
* @mapping: mapping to be freed
* @fence: fence of the unmap operation
*
* Free a mapping and make sure we decrease the PRT usage count if applicable.
*/ */
static void amdgpu_vm_free_mapping(struct amdgpu_device *adev, static void amdgpu_vm_add_prt_cb(struct amdgpu_device *adev,
struct amdgpu_vm *vm,
struct amdgpu_bo_va_mapping *mapping,
struct dma_fence *fence) struct dma_fence *fence)
{ {
if (mapping->flags & AMDGPU_PTE_PRT) {
struct amdgpu_prt_cb *cb = kmalloc(sizeof(struct amdgpu_prt_cb), struct amdgpu_prt_cb *cb = kmalloc(sizeof(struct amdgpu_prt_cb),
GFP_KERNEL); GFP_KERNEL);
...@@ -1244,10 +1243,66 @@ static void amdgpu_vm_free_mapping(struct amdgpu_device *adev, ...@@ -1244,10 +1243,66 @@ static void amdgpu_vm_free_mapping(struct amdgpu_device *adev,
amdgpu_vm_prt_cb)) amdgpu_vm_prt_cb))
amdgpu_vm_prt_cb(fence, &cb->cb); amdgpu_vm_prt_cb(fence, &cb->cb);
} }
} }
/**
* amdgpu_vm_free_mapping - free a mapping
*
* @adev: amdgpu_device pointer
* @vm: requested vm
* @mapping: mapping to be freed
* @fence: fence of the unmap operation
*
* Free a mapping and make sure we decrease the PRT usage count if applicable.
*/
static void amdgpu_vm_free_mapping(struct amdgpu_device *adev,
struct amdgpu_vm *vm,
struct amdgpu_bo_va_mapping *mapping,
struct dma_fence *fence)
{
if (mapping->flags & AMDGPU_PTE_PRT)
amdgpu_vm_add_prt_cb(adev, fence);
kfree(mapping); kfree(mapping);
} }
/**
* amdgpu_vm_prt_fini - finish all prt mappings
*
* @adev: amdgpu_device pointer
* @vm: requested vm
*
* Register a cleanup callback to disable PRT support after VM dies.
*/
static void amdgpu_vm_prt_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
{
struct reservation_object *resv = vm->page_directory->tbo.resv;
struct dma_fence *excl, **shared;
unsigned i, shared_count;
int r;
r = reservation_object_get_fences_rcu(resv, &excl,
&shared_count, &shared);
if (r) {
/* Not enough memory to grab the fence list, as last resort
* block for all the fences to complete.
*/
reservation_object_wait_timeout_rcu(resv, true, false,
MAX_SCHEDULE_TIMEOUT);
return;
}
/* Add a callback for each fence in the reservation object */
amdgpu_vm_prt_get(adev);
amdgpu_vm_add_prt_cb(adev, excl);
for (i = 0; i < shared_count; ++i) {
amdgpu_vm_prt_get(adev);
amdgpu_vm_add_prt_cb(adev, shared[i]);
}
kfree(shared);
}
/** /**
* amdgpu_vm_clear_freed - clear freed BOs in the PT * amdgpu_vm_clear_freed - clear freed BOs in the PT
* *
...@@ -1395,8 +1450,7 @@ int amdgpu_vm_bo_map(struct amdgpu_device *adev, ...@@ -1395,8 +1450,7 @@ int amdgpu_vm_bo_map(struct amdgpu_device *adev,
if (!adev->gart.gart_funcs->set_prt) if (!adev->gart.gart_funcs->set_prt)
return -EINVAL; return -EINVAL;
if (atomic_inc_return(&adev->vm_manager.num_prt_mappings) == 1) amdgpu_vm_prt_get(adev);
amdgpu_vm_update_prt_state(adev);
} }
/* make sure object fit at this offset */ /* make sure object fit at this offset */
...@@ -1699,6 +1753,7 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm) ...@@ -1699,6 +1753,7 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm)
void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
{ {
struct amdgpu_bo_va_mapping *mapping, *tmp; struct amdgpu_bo_va_mapping *mapping, *tmp;
bool prt_fini_called = false;
int i; int i;
amd_sched_entity_fini(vm->entity.sched, &vm->entity); amd_sched_entity_fini(vm->entity.sched, &vm->entity);
...@@ -1712,13 +1767,14 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) ...@@ -1712,13 +1767,14 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
kfree(mapping); kfree(mapping);
} }
list_for_each_entry_safe(mapping, tmp, &vm->freed, list) { list_for_each_entry_safe(mapping, tmp, &vm->freed, list) {
if (mapping->flags & AMDGPU_PTE_PRT) if (mapping->flags & AMDGPU_PTE_PRT && !prt_fini_called) {
continue; amdgpu_vm_prt_fini(adev, vm);
prt_fini_called = true;
}
list_del(&mapping->list); list_del(&mapping->list);
kfree(mapping); amdgpu_vm_free_mapping(adev, vm, mapping, NULL);
} }
amdgpu_vm_clear_freed(adev, vm);
for (i = 0; i < amdgpu_vm_num_pdes(adev); i++) { for (i = 0; i < amdgpu_vm_num_pdes(adev); i++) {
struct amdgpu_bo *pt = vm->page_tables[i].bo; struct amdgpu_bo *pt = vm->page_tables[i].bo;
...@@ -1765,7 +1821,7 @@ void amdgpu_vm_manager_init(struct amdgpu_device *adev) ...@@ -1765,7 +1821,7 @@ void amdgpu_vm_manager_init(struct amdgpu_device *adev)
atomic_set(&adev->vm_manager.vm_pte_next_ring, 0); atomic_set(&adev->vm_manager.vm_pte_next_ring, 0);
atomic64_set(&adev->vm_manager.client_counter, 0); atomic64_set(&adev->vm_manager.client_counter, 0);
spin_lock_init(&adev->vm_manager.prt_lock); spin_lock_init(&adev->vm_manager.prt_lock);
atomic_set(&adev->vm_manager.num_prt_mappings, 0); atomic_set(&adev->vm_manager.num_prt_users, 0);
} }
/** /**
......
...@@ -164,7 +164,7 @@ struct amdgpu_vm_manager { ...@@ -164,7 +164,7 @@ struct amdgpu_vm_manager {
/* partial resident texture handling */ /* partial resident texture handling */
spinlock_t prt_lock; spinlock_t prt_lock;
atomic_t num_prt_mappings; atomic_t num_prt_users;
}; };
void amdgpu_vm_manager_init(struct amdgpu_device *adev); void amdgpu_vm_manager_init(struct amdgpu_device *adev);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment