提交 ede0dd86 编写于 作者: F Felix Kuehling 提交者: Oded Gabbay

drm/amdgpu: Add kfd2kgd interface to acquire an existing VM

This allows acquiring an existing VM from a render node FD to use it
for a compute process.

Such VMs get destroyed when the original file descriptor is released.
Added a callback from amdgpu_vm_fini to handle KFD VM destruction
correctly in this case.

v2:
* Removed vm->vm_context check in amdgpu_amdkfd_gpuvm_destroy_cb,
  check vm->process_info earlier instead
Signed-off-by: NFelix Kuehling <Felix.Kuehling@amd.com>
Acked-by: NChristian König <christian.koenig@amd.com>
Signed-off-by: NOded Gabbay <oded.gabbay@gmail.com>
上级 b236fa1d
...@@ -145,6 +145,12 @@ uint64_t amdgpu_amdkfd_get_vram_usage(struct kgd_dev *kgd); ...@@ -145,6 +145,12 @@ uint64_t amdgpu_amdkfd_get_vram_usage(struct kgd_dev *kgd);
int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm, int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm,
void **process_info, void **process_info,
struct dma_fence **ef); struct dma_fence **ef);
int amdgpu_amdkfd_gpuvm_acquire_process_vm(struct kgd_dev *kgd,
struct file *filp,
void **vm, void **process_info,
struct dma_fence **ef);
void amdgpu_amdkfd_gpuvm_destroy_cb(struct amdgpu_device *adev,
struct amdgpu_vm *vm);
void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm); void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm);
uint32_t amdgpu_amdkfd_gpuvm_get_process_page_dir(void *vm); uint32_t amdgpu_amdkfd_gpuvm_get_process_page_dir(void *vm);
int amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu( int amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu(
......
...@@ -205,6 +205,7 @@ static const struct kfd2kgd_calls kfd2kgd = { ...@@ -205,6 +205,7 @@ static const struct kfd2kgd_calls kfd2kgd = {
.get_cu_info = get_cu_info, .get_cu_info = get_cu_info,
.get_vram_usage = amdgpu_amdkfd_get_vram_usage, .get_vram_usage = amdgpu_amdkfd_get_vram_usage,
.create_process_vm = amdgpu_amdkfd_gpuvm_create_process_vm, .create_process_vm = amdgpu_amdkfd_gpuvm_create_process_vm,
.acquire_process_vm = amdgpu_amdkfd_gpuvm_acquire_process_vm,
.destroy_process_vm = amdgpu_amdkfd_gpuvm_destroy_process_vm, .destroy_process_vm = amdgpu_amdkfd_gpuvm_destroy_process_vm,
.get_process_page_dir = amdgpu_amdkfd_gpuvm_get_process_page_dir, .get_process_page_dir = amdgpu_amdkfd_gpuvm_get_process_page_dir,
.set_vm_context_page_table_base = set_vm_context_page_table_base, .set_vm_context_page_table_base = set_vm_context_page_table_base,
......
...@@ -165,6 +165,7 @@ static const struct kfd2kgd_calls kfd2kgd = { ...@@ -165,6 +165,7 @@ static const struct kfd2kgd_calls kfd2kgd = {
.get_cu_info = get_cu_info, .get_cu_info = get_cu_info,
.get_vram_usage = amdgpu_amdkfd_get_vram_usage, .get_vram_usage = amdgpu_amdkfd_get_vram_usage,
.create_process_vm = amdgpu_amdkfd_gpuvm_create_process_vm, .create_process_vm = amdgpu_amdkfd_gpuvm_create_process_vm,
.acquire_process_vm = amdgpu_amdkfd_gpuvm_acquire_process_vm,
.destroy_process_vm = amdgpu_amdkfd_gpuvm_destroy_process_vm, .destroy_process_vm = amdgpu_amdkfd_gpuvm_destroy_process_vm,
.get_process_page_dir = amdgpu_amdkfd_gpuvm_get_process_page_dir, .get_process_page_dir = amdgpu_amdkfd_gpuvm_get_process_page_dir,
.set_vm_context_page_table_base = set_vm_context_page_table_base, .set_vm_context_page_table_base = set_vm_context_page_table_base,
......
...@@ -806,32 +806,16 @@ static int process_update_pds(struct amdkfd_process_info *process_info, ...@@ -806,32 +806,16 @@ static int process_update_pds(struct amdkfd_process_info *process_info,
return 0; return 0;
} }
int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm, static int init_kfd_vm(struct amdgpu_vm *vm, void **process_info,
void **process_info,
struct dma_fence **ef) struct dma_fence **ef)
{ {
int ret;
struct amdgpu_vm *new_vm;
struct amdkfd_process_info *info = NULL; struct amdkfd_process_info *info = NULL;
struct amdgpu_device *adev = get_amdgpu_device(kgd); int ret;
new_vm = kzalloc(sizeof(*new_vm), GFP_KERNEL);
if (!new_vm)
return -ENOMEM;
/* Initialize the VM context, allocate the page directory and zero it */
ret = amdgpu_vm_init(adev, new_vm, AMDGPU_VM_CONTEXT_COMPUTE, 0);
if (ret) {
pr_err("Failed init vm ret %d\n", ret);
goto vm_init_fail;
}
if (!*process_info) { if (!*process_info) {
info = kzalloc(sizeof(*info), GFP_KERNEL); info = kzalloc(sizeof(*info), GFP_KERNEL);
if (!info) { if (!info)
ret = -ENOMEM; return -ENOMEM;
goto alloc_process_info_fail;
}
mutex_init(&info->lock); mutex_init(&info->lock);
INIT_LIST_HEAD(&info->vm_list_head); INIT_LIST_HEAD(&info->vm_list_head);
...@@ -842,6 +826,7 @@ int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm, ...@@ -842,6 +826,7 @@ int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm,
current->mm); current->mm);
if (!info->eviction_fence) { if (!info->eviction_fence) {
pr_err("Failed to create eviction fence\n"); pr_err("Failed to create eviction fence\n");
ret = -ENOMEM;
goto create_evict_fence_fail; goto create_evict_fence_fail;
} }
...@@ -849,77 +834,137 @@ int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm, ...@@ -849,77 +834,137 @@ int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm,
*ef = dma_fence_get(&info->eviction_fence->base); *ef = dma_fence_get(&info->eviction_fence->base);
} }
new_vm->process_info = *process_info; vm->process_info = *process_info;
/* Validate page directory and attach eviction fence */ /* Validate page directory and attach eviction fence */
ret = amdgpu_bo_reserve(new_vm->root.base.bo, true); ret = amdgpu_bo_reserve(vm->root.base.bo, true);
if (ret) if (ret)
goto reserve_pd_fail; goto reserve_pd_fail;
ret = vm_validate_pt_pd_bos(new_vm); ret = vm_validate_pt_pd_bos(vm);
if (ret) { if (ret) {
pr_err("validate_pt_pd_bos() failed\n"); pr_err("validate_pt_pd_bos() failed\n");
goto validate_pd_fail; goto validate_pd_fail;
} }
ret = ttm_bo_wait(&new_vm->root.base.bo->tbo, false, false); ret = ttm_bo_wait(&vm->root.base.bo->tbo, false, false);
if (ret) if (ret)
goto wait_pd_fail; goto wait_pd_fail;
amdgpu_bo_fence(new_vm->root.base.bo, amdgpu_bo_fence(vm->root.base.bo,
&new_vm->process_info->eviction_fence->base, true); &vm->process_info->eviction_fence->base, true);
amdgpu_bo_unreserve(new_vm->root.base.bo); amdgpu_bo_unreserve(vm->root.base.bo);
/* Update process info */ /* Update process info */
mutex_lock(&new_vm->process_info->lock); mutex_lock(&vm->process_info->lock);
list_add_tail(&new_vm->vm_list_node, list_add_tail(&vm->vm_list_node,
&(new_vm->process_info->vm_list_head)); &(vm->process_info->vm_list_head));
new_vm->process_info->n_vms++; vm->process_info->n_vms++;
mutex_unlock(&new_vm->process_info->lock); mutex_unlock(&vm->process_info->lock);
*vm = (void *) new_vm;
pr_debug("Created process vm %p\n", *vm); return 0;
return ret;
wait_pd_fail: wait_pd_fail:
validate_pd_fail: validate_pd_fail:
amdgpu_bo_unreserve(new_vm->root.base.bo); amdgpu_bo_unreserve(vm->root.base.bo);
reserve_pd_fail: reserve_pd_fail:
vm->process_info = NULL;
if (info) {
/* Two fence references: one in info and one in *ef */
dma_fence_put(&info->eviction_fence->base);
dma_fence_put(*ef);
*ef = NULL;
*process_info = NULL;
create_evict_fence_fail: create_evict_fence_fail:
mutex_destroy(&info->lock); mutex_destroy(&info->lock);
kfree(info); kfree(info);
alloc_process_info_fail: }
return ret;
}
int amdgpu_amdkfd_gpuvm_create_process_vm(struct kgd_dev *kgd, void **vm,
void **process_info,
struct dma_fence **ef)
{
struct amdgpu_device *adev = get_amdgpu_device(kgd);
struct amdgpu_vm *new_vm;
int ret;
new_vm = kzalloc(sizeof(*new_vm), GFP_KERNEL);
if (!new_vm)
return -ENOMEM;
/* Initialize AMDGPU part of the VM */
ret = amdgpu_vm_init(adev, new_vm, AMDGPU_VM_CONTEXT_COMPUTE, 0);
if (ret) {
pr_err("Failed init vm ret %d\n", ret);
goto amdgpu_vm_init_fail;
}
/* Initialize KFD part of the VM and process info */
ret = init_kfd_vm(new_vm, process_info, ef);
if (ret)
goto init_kfd_vm_fail;
*vm = (void *) new_vm;
return 0;
init_kfd_vm_fail:
amdgpu_vm_fini(adev, new_vm); amdgpu_vm_fini(adev, new_vm);
vm_init_fail: amdgpu_vm_init_fail:
kfree(new_vm); kfree(new_vm);
return ret; return ret;
} }
void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm) int amdgpu_amdkfd_gpuvm_acquire_process_vm(struct kgd_dev *kgd,
struct file *filp,
void **vm, void **process_info,
struct dma_fence **ef)
{ {
struct amdgpu_device *adev = get_amdgpu_device(kgd); struct amdgpu_device *adev = get_amdgpu_device(kgd);
struct amdgpu_vm *avm = (struct amdgpu_vm *)vm; struct drm_file *drm_priv = filp->private_data;
struct amdgpu_bo *pd; struct amdgpu_fpriv *drv_priv = drm_priv->driver_priv;
struct amdkfd_process_info *process_info; struct amdgpu_vm *avm = &drv_priv->vm;
int ret;
if (WARN_ON(!kgd || !vm)) /* Already a compute VM? */
if (avm->process_info)
return -EINVAL;
/* Convert VM into a compute VM */
ret = amdgpu_vm_make_compute(adev, avm);
if (ret)
return ret;
/* Initialize KFD part of the VM and process info */
ret = init_kfd_vm(avm, process_info, ef);
if (ret)
return ret;
*vm = (void *)avm;
return 0;
}
void amdgpu_amdkfd_gpuvm_destroy_cb(struct amdgpu_device *adev,
struct amdgpu_vm *vm)
{
struct amdkfd_process_info *process_info = vm->process_info;
struct amdgpu_bo *pd = vm->root.base.bo;
if (!process_info)
return; return;
pr_debug("Destroying process vm %p\n", vm);
/* Release eviction fence from PD */ /* Release eviction fence from PD */
pd = avm->root.base.bo;
amdgpu_bo_reserve(pd, false); amdgpu_bo_reserve(pd, false);
amdgpu_bo_fence(pd, NULL, false); amdgpu_bo_fence(pd, NULL, false);
amdgpu_bo_unreserve(pd); amdgpu_bo_unreserve(pd);
process_info = avm->process_info; /* Update process info */
mutex_lock(&process_info->lock); mutex_lock(&process_info->lock);
process_info->n_vms--; process_info->n_vms--;
list_del(&avm->vm_list_node); list_del(&vm->vm_list_node);
mutex_unlock(&process_info->lock); mutex_unlock(&process_info->lock);
/* Release per-process resources */ /* Release per-process resources when last compute VM is destroyed */
if (!process_info->n_vms) { if (!process_info->n_vms) {
WARN_ON(!list_empty(&process_info->kfd_bo_list)); WARN_ON(!list_empty(&process_info->kfd_bo_list));
...@@ -927,6 +972,17 @@ void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm) ...@@ -927,6 +972,17 @@ void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm)
mutex_destroy(&process_info->lock); mutex_destroy(&process_info->lock);
kfree(process_info); kfree(process_info);
} }
}
void amdgpu_amdkfd_gpuvm_destroy_process_vm(struct kgd_dev *kgd, void *vm)
{
struct amdgpu_device *adev = get_amdgpu_device(kgd);
struct amdgpu_vm *avm = (struct amdgpu_vm *)vm;
if (WARN_ON(!kgd || !vm))
return;
pr_debug("Destroying process vm %p\n", vm);
/* Release the VM context */ /* Release the VM context */
amdgpu_vm_fini(adev, avm); amdgpu_vm_fini(adev, avm);
......
...@@ -32,6 +32,7 @@ ...@@ -32,6 +32,7 @@
#include <drm/amdgpu_drm.h> #include <drm/amdgpu_drm.h>
#include "amdgpu.h" #include "amdgpu.h"
#include "amdgpu_trace.h" #include "amdgpu_trace.h"
#include "amdgpu_amdkfd.h"
/* /*
* GPUVM * GPUVM
...@@ -2575,6 +2576,8 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm) ...@@ -2575,6 +2576,8 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
u64 fault; u64 fault;
int i, r; int i, r;
amdgpu_amdkfd_gpuvm_destroy_cb(adev, vm);
/* Clear pending page faults from IH when the VM is destroyed */ /* Clear pending page faults from IH when the VM is destroyed */
while (kfifo_get(&vm->faults, &fault)) while (kfifo_get(&vm->faults, &fault))
amdgpu_ih_clear_fault(adev, fault); amdgpu_ih_clear_fault(adev, fault);
......
...@@ -336,6 +336,8 @@ struct kfd2kgd_calls { ...@@ -336,6 +336,8 @@ struct kfd2kgd_calls {
int (*create_process_vm)(struct kgd_dev *kgd, void **vm, int (*create_process_vm)(struct kgd_dev *kgd, void **vm,
void **process_info, struct dma_fence **ef); void **process_info, struct dma_fence **ef);
int (*acquire_process_vm)(struct kgd_dev *kgd, struct file *filp,
void **vm, void **process_info, struct dma_fence **ef);
void (*destroy_process_vm)(struct kgd_dev *kgd, void *vm); void (*destroy_process_vm)(struct kgd_dev *kgd, void *vm);
uint32_t (*get_process_page_dir)(void *vm); uint32_t (*get_process_page_dir)(void *vm);
void (*set_vm_context_page_table_base)(struct kgd_dev *kgd, void (*set_vm_context_page_table_base)(struct kgd_dev *kgd,
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册