Commit af4c0f65 authored by Christian König's avatar Christian König Committed by Alex Deucher

drm/amdgpu: rework VM state machine lock handling v2

Only the moved state needs a separate spin lock protection. All other
states are protected by reserving the VM anyway.

v2: fix some more incorrect cases
Signed-off-by: default avatarChristian König <christian.koenig@amd.com>
Reviewed-by: default avatarChunming Zhou <david1.zhou@amd.com>
Signed-off-by: default avatarAlex Deucher <alexander.deucher@amd.com>
parent 22cc6c5e
...@@ -119,9 +119,7 @@ static void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base, ...@@ -119,9 +119,7 @@ static void amdgpu_vm_bo_base_init(struct amdgpu_vm_bo_base *base,
* is currently evicted. add the bo to the evicted list to make sure it * is currently evicted. add the bo to the evicted list to make sure it
* is validated on next vm use to avoid fault. * is validated on next vm use to avoid fault.
* */ * */
spin_lock(&vm->status_lock);
list_move_tail(&base->vm_status, &vm->evicted); list_move_tail(&base->vm_status, &vm->evicted);
spin_unlock(&vm->status_lock);
} }
/** /**
...@@ -228,7 +226,6 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, ...@@ -228,7 +226,6 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
struct ttm_bo_global *glob = adev->mman.bdev.glob; struct ttm_bo_global *glob = adev->mman.bdev.glob;
int r; int r;
spin_lock(&vm->status_lock);
while (!list_empty(&vm->evicted)) { while (!list_empty(&vm->evicted)) {
struct amdgpu_vm_bo_base *bo_base; struct amdgpu_vm_bo_base *bo_base;
struct amdgpu_bo *bo; struct amdgpu_bo *bo;
...@@ -236,10 +233,8 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, ...@@ -236,10 +233,8 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
bo_base = list_first_entry(&vm->evicted, bo_base = list_first_entry(&vm->evicted,
struct amdgpu_vm_bo_base, struct amdgpu_vm_bo_base,
vm_status); vm_status);
spin_unlock(&vm->status_lock);
bo = bo_base->bo; bo = bo_base->bo;
BUG_ON(!bo);
if (bo->parent) { if (bo->parent) {
r = validate(param, bo); r = validate(param, bo);
if (r) if (r)
...@@ -259,13 +254,14 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, ...@@ -259,13 +254,14 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
return r; return r;
} }
spin_lock(&vm->status_lock); if (bo->tbo.type != ttm_bo_type_kernel) {
if (bo->tbo.type != ttm_bo_type_kernel) spin_lock(&vm->moved_lock);
list_move(&bo_base->vm_status, &vm->moved); list_move(&bo_base->vm_status, &vm->moved);
else spin_unlock(&vm->moved_lock);
} else {
list_move(&bo_base->vm_status, &vm->relocated); list_move(&bo_base->vm_status, &vm->relocated);
} }
spin_unlock(&vm->status_lock); }
return 0; return 0;
} }
...@@ -279,13 +275,7 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm, ...@@ -279,13 +275,7 @@ int amdgpu_vm_validate_pt_bos(struct amdgpu_device *adev, struct amdgpu_vm *vm,
*/ */
bool amdgpu_vm_ready(struct amdgpu_vm *vm) bool amdgpu_vm_ready(struct amdgpu_vm *vm)
{ {
bool ready; return list_empty(&vm->evicted);
spin_lock(&vm->status_lock);
ready = list_empty(&vm->evicted);
spin_unlock(&vm->status_lock);
return ready;
} }
/** /**
...@@ -477,9 +467,7 @@ static int amdgpu_vm_alloc_levels(struct amdgpu_device *adev, ...@@ -477,9 +467,7 @@ static int amdgpu_vm_alloc_levels(struct amdgpu_device *adev,
pt->parent = amdgpu_bo_ref(parent->base.bo); pt->parent = amdgpu_bo_ref(parent->base.bo);
amdgpu_vm_bo_base_init(&entry->base, vm, pt); amdgpu_vm_bo_base_init(&entry->base, vm, pt);
spin_lock(&vm->status_lock);
list_move(&entry->base.vm_status, &vm->relocated); list_move(&entry->base.vm_status, &vm->relocated);
spin_unlock(&vm->status_lock);
} }
if (level < AMDGPU_VM_PTB) { if (level < AMDGPU_VM_PTB) {
...@@ -926,10 +914,8 @@ static void amdgpu_vm_invalidate_level(struct amdgpu_device *adev, ...@@ -926,10 +914,8 @@ static void amdgpu_vm_invalidate_level(struct amdgpu_device *adev,
if (!entry->base.bo) if (!entry->base.bo)
continue; continue;
spin_lock(&vm->status_lock);
if (list_empty(&entry->base.vm_status)) if (list_empty(&entry->base.vm_status))
list_add(&entry->base.vm_status, &vm->relocated); list_add(&entry->base.vm_status, &vm->relocated);
spin_unlock(&vm->status_lock);
amdgpu_vm_invalidate_level(adev, vm, entry, level + 1); amdgpu_vm_invalidate_level(adev, vm, entry, level + 1);
} }
} }
...@@ -974,7 +960,6 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev, ...@@ -974,7 +960,6 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
params.func = amdgpu_vm_do_set_ptes; params.func = amdgpu_vm_do_set_ptes;
} }
spin_lock(&vm->status_lock);
while (!list_empty(&vm->relocated)) { while (!list_empty(&vm->relocated)) {
struct amdgpu_vm_bo_base *bo_base, *parent; struct amdgpu_vm_bo_base *bo_base, *parent;
struct amdgpu_vm_pt *pt, *entry; struct amdgpu_vm_pt *pt, *entry;
...@@ -984,13 +969,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev, ...@@ -984,13 +969,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
struct amdgpu_vm_bo_base, struct amdgpu_vm_bo_base,
vm_status); vm_status);
list_del_init(&bo_base->vm_status); list_del_init(&bo_base->vm_status);
spin_unlock(&vm->status_lock);
bo = bo_base->bo->parent; bo = bo_base->bo->parent;
if (!bo) { if (!bo)
spin_lock(&vm->status_lock);
continue; continue;
}
parent = list_first_entry(&bo->va, struct amdgpu_vm_bo_base, parent = list_first_entry(&bo->va, struct amdgpu_vm_bo_base,
bo_list); bo_list);
...@@ -999,12 +981,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev, ...@@ -999,12 +981,10 @@ int amdgpu_vm_update_directories(struct amdgpu_device *adev,
amdgpu_vm_update_pde(&params, vm, pt, entry); amdgpu_vm_update_pde(&params, vm, pt, entry);
spin_lock(&vm->status_lock);
if (!vm->use_cpu_for_update && if (!vm->use_cpu_for_update &&
(ndw - params.ib->length_dw) < 32) (ndw - params.ib->length_dw) < 32)
break; break;
} }
spin_unlock(&vm->status_lock);
if (vm->use_cpu_for_update) { if (vm->use_cpu_for_update) {
/* Flush HDP */ /* Flush HDP */
...@@ -1107,9 +1087,7 @@ static void amdgpu_vm_handle_huge_pages(struct amdgpu_pte_update_params *p, ...@@ -1107,9 +1087,7 @@ static void amdgpu_vm_handle_huge_pages(struct amdgpu_pte_update_params *p,
if (entry->huge) { if (entry->huge) {
/* Add the entry to the relocated list to update it. */ /* Add the entry to the relocated list to update it. */
entry->huge = false; entry->huge = false;
spin_lock(&p->vm->status_lock);
list_move(&entry->base.vm_status, &p->vm->relocated); list_move(&entry->base.vm_status, &p->vm->relocated);
spin_unlock(&p->vm->status_lock);
} }
return; return;
} }
...@@ -1588,8 +1566,9 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, ...@@ -1588,8 +1566,9 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
amdgpu_asic_flush_hdp(adev, NULL); amdgpu_asic_flush_hdp(adev, NULL);
} }
spin_lock(&vm->status_lock); spin_lock(&vm->moved_lock);
list_del_init(&bo_va->base.vm_status); list_del_init(&bo_va->base.vm_status);
spin_unlock(&vm->moved_lock);
/* If the BO is not in its preferred location add it back to /* If the BO is not in its preferred location add it back to
* the evicted list so that it gets validated again on the * the evicted list so that it gets validated again on the
...@@ -1599,7 +1578,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, ...@@ -1599,7 +1578,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev,
!(bo->preferred_domains & !(bo->preferred_domains &
amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type))) amdgpu_mem_type_to_domain(bo->tbo.mem.mem_type)))
list_add_tail(&bo_va->base.vm_status, &vm->evicted); list_add_tail(&bo_va->base.vm_status, &vm->evicted);
spin_unlock(&vm->status_lock);
list_splice_init(&bo_va->invalids, &bo_va->valids); list_splice_init(&bo_va->invalids, &bo_va->valids);
bo_va->cleared = clear; bo_va->cleared = clear;
...@@ -1811,14 +1789,14 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev, ...@@ -1811,14 +1789,14 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
bool clear; bool clear;
int r = 0; int r = 0;
spin_lock(&vm->status_lock); spin_lock(&vm->moved_lock);
while (!list_empty(&vm->moved)) { while (!list_empty(&vm->moved)) {
struct amdgpu_bo_va *bo_va; struct amdgpu_bo_va *bo_va;
struct reservation_object *resv; struct reservation_object *resv;
bo_va = list_first_entry(&vm->moved, bo_va = list_first_entry(&vm->moved,
struct amdgpu_bo_va, base.vm_status); struct amdgpu_bo_va, base.vm_status);
spin_unlock(&vm->status_lock); spin_unlock(&vm->moved_lock);
resv = bo_va->base.bo->tbo.resv; resv = bo_va->base.bo->tbo.resv;
...@@ -1839,9 +1817,9 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev, ...@@ -1839,9 +1817,9 @@ int amdgpu_vm_handle_moved(struct amdgpu_device *adev,
if (!clear && resv != vm->root.base.bo->tbo.resv) if (!clear && resv != vm->root.base.bo->tbo.resv)
reservation_object_unlock(resv); reservation_object_unlock(resv);
spin_lock(&vm->status_lock); spin_lock(&vm->moved_lock);
} }
spin_unlock(&vm->status_lock); spin_unlock(&vm->moved_lock);
return r; return r;
} }
...@@ -1903,10 +1881,10 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev, ...@@ -1903,10 +1881,10 @@ static void amdgpu_vm_bo_insert_map(struct amdgpu_device *adev,
amdgpu_vm_prt_get(adev); amdgpu_vm_prt_get(adev);
if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv) { if (bo && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
spin_lock(&vm->status_lock); spin_lock(&vm->moved_lock);
if (list_empty(&bo_va->base.vm_status)) if (list_empty(&bo_va->base.vm_status))
list_add(&bo_va->base.vm_status, &vm->moved); list_add(&bo_va->base.vm_status, &vm->moved);
spin_unlock(&vm->status_lock); spin_unlock(&vm->moved_lock);
} }
trace_amdgpu_vm_bo_map(bo_va, mapping); trace_amdgpu_vm_bo_map(bo_va, mapping);
} }
...@@ -2216,9 +2194,9 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev, ...@@ -2216,9 +2194,9 @@ void amdgpu_vm_bo_rmv(struct amdgpu_device *adev,
list_del(&bo_va->base.bo_list); list_del(&bo_va->base.bo_list);
spin_lock(&vm->status_lock); spin_lock(&vm->moved_lock);
list_del(&bo_va->base.vm_status); list_del(&bo_va->base.vm_status);
spin_unlock(&vm->status_lock); spin_unlock(&vm->moved_lock);
list_for_each_entry_safe(mapping, next, &bo_va->valids, list) { list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
list_del(&mapping->list); list_del(&mapping->list);
...@@ -2261,28 +2239,24 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev, ...@@ -2261,28 +2239,24 @@ void amdgpu_vm_bo_invalidate(struct amdgpu_device *adev,
bo_base->moved = true; bo_base->moved = true;
if (evicted && bo->tbo.resv == vm->root.base.bo->tbo.resv) { if (evicted && bo->tbo.resv == vm->root.base.bo->tbo.resv) {
spin_lock(&bo_base->vm->status_lock);
if (bo->tbo.type == ttm_bo_type_kernel) if (bo->tbo.type == ttm_bo_type_kernel)
list_move(&bo_base->vm_status, &vm->evicted); list_move(&bo_base->vm_status, &vm->evicted);
else else
list_move_tail(&bo_base->vm_status, list_move_tail(&bo_base->vm_status,
&vm->evicted); &vm->evicted);
spin_unlock(&bo_base->vm->status_lock);
continue; continue;
} }
if (bo->tbo.type == ttm_bo_type_kernel) { if (bo->tbo.type == ttm_bo_type_kernel) {
spin_lock(&bo_base->vm->status_lock);
if (list_empty(&bo_base->vm_status)) if (list_empty(&bo_base->vm_status))
list_add(&bo_base->vm_status, &vm->relocated); list_add(&bo_base->vm_status, &vm->relocated);
spin_unlock(&bo_base->vm->status_lock);
continue; continue;
} }
spin_lock(&bo_base->vm->status_lock); spin_lock(&bo_base->vm->moved_lock);
if (list_empty(&bo_base->vm_status)) if (list_empty(&bo_base->vm_status))
list_add(&bo_base->vm_status, &vm->moved); list_add(&bo_base->vm_status, &vm->moved);
spin_unlock(&bo_base->vm->status_lock); spin_unlock(&bo_base->vm->moved_lock);
} }
} }
...@@ -2391,9 +2365,9 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm, ...@@ -2391,9 +2365,9 @@ int amdgpu_vm_init(struct amdgpu_device *adev, struct amdgpu_vm *vm,
vm->va = RB_ROOT_CACHED; vm->va = RB_ROOT_CACHED;
for (i = 0; i < AMDGPU_MAX_VMHUBS; i++) for (i = 0; i < AMDGPU_MAX_VMHUBS; i++)
vm->reserved_vmid[i] = NULL; vm->reserved_vmid[i] = NULL;
spin_lock_init(&vm->status_lock);
INIT_LIST_HEAD(&vm->evicted); INIT_LIST_HEAD(&vm->evicted);
INIT_LIST_HEAD(&vm->relocated); INIT_LIST_HEAD(&vm->relocated);
spin_lock_init(&vm->moved_lock);
INIT_LIST_HEAD(&vm->moved); INIT_LIST_HEAD(&vm->moved);
INIT_LIST_HEAD(&vm->freed); INIT_LIST_HEAD(&vm->freed);
......
...@@ -168,9 +168,6 @@ struct amdgpu_vm { ...@@ -168,9 +168,6 @@ struct amdgpu_vm {
/* tree of virtual addresses mapped */ /* tree of virtual addresses mapped */
struct rb_root_cached va; struct rb_root_cached va;
/* protecting invalidated */
spinlock_t status_lock;
/* BOs who needs a validation */ /* BOs who needs a validation */
struct list_head evicted; struct list_head evicted;
...@@ -179,6 +176,7 @@ struct amdgpu_vm { ...@@ -179,6 +176,7 @@ struct amdgpu_vm {
/* BOs moved, but not yet updated in the PT */ /* BOs moved, but not yet updated in the PT */
struct list_head moved; struct list_head moved;
spinlock_t moved_lock;
/* BO mappings freed, but not yet updated in the PT */ /* BO mappings freed, but not yet updated in the PT */
struct list_head freed; struct list_head freed;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment