Commit 014f831a authored by Danilo Krummrich's avatar Danilo Krummrich

drm/nouveau: use GPUVM common infrastructure

GPUVM provides common infrastructure to track external and evicted GEM
objects as well as locking and validation helpers.

Especially external and evicted object tracking is a huge improvement
compared to the current brute force approach of iterating all mappings
in order to lock and validate the GPUVM's GEM objects. Hence, make us of
it.
Signed-off-by: default avatarDanilo Krummrich <dakr@redhat.com>
Reviewed-by: default avatarDave Airlie <airlied@redhat.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20231113221202.7203-1-dakr@redhat.com
parent 6c18005d
...@@ -1061,17 +1061,18 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, ...@@ -1061,17 +1061,18 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict,
{ {
struct nouveau_drm *drm = nouveau_bdev(bo->bdev); struct nouveau_drm *drm = nouveau_bdev(bo->bdev);
struct nouveau_bo *nvbo = nouveau_bo(bo); struct nouveau_bo *nvbo = nouveau_bo(bo);
struct drm_gem_object *obj = &bo->base;
struct ttm_resource *old_reg = bo->resource; struct ttm_resource *old_reg = bo->resource;
struct nouveau_drm_tile *new_tile = NULL; struct nouveau_drm_tile *new_tile = NULL;
int ret = 0; int ret = 0;
if (new_reg->mem_type == TTM_PL_TT) { if (new_reg->mem_type == TTM_PL_TT) {
ret = nouveau_ttm_tt_bind(bo->bdev, bo->ttm, new_reg); ret = nouveau_ttm_tt_bind(bo->bdev, bo->ttm, new_reg);
if (ret) if (ret)
return ret; return ret;
} }
drm_gpuvm_bo_gem_evict(obj, evict);
nouveau_bo_move_ntfy(bo, new_reg); nouveau_bo_move_ntfy(bo, new_reg);
ret = ttm_bo_wait_ctx(bo, ctx); ret = ttm_bo_wait_ctx(bo, ctx);
if (ret) if (ret)
...@@ -1136,6 +1137,7 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, ...@@ -1136,6 +1137,7 @@ nouveau_bo_move(struct ttm_buffer_object *bo, bool evict,
out_ntfy: out_ntfy:
if (ret) { if (ret) {
nouveau_bo_move_ntfy(bo, bo->resource); nouveau_bo_move_ntfy(bo, bo->resource);
drm_gpuvm_bo_gem_evict(obj, !evict);
} }
return ret; return ret;
} }
......
// SPDX-License-Identifier: MIT // SPDX-License-Identifier: MIT
#include <drm/drm_exec.h>
#include "nouveau_drv.h" #include "nouveau_drv.h"
#include "nouveau_gem.h" #include "nouveau_gem.h"
#include "nouveau_mem.h" #include "nouveau_mem.h"
...@@ -86,14 +84,12 @@ ...@@ -86,14 +84,12 @@
*/ */
static int static int
nouveau_exec_job_submit(struct nouveau_job *job) nouveau_exec_job_submit(struct nouveau_job *job,
struct drm_gpuvm_exec *vme)
{ {
struct nouveau_exec_job *exec_job = to_nouveau_exec_job(job); struct nouveau_exec_job *exec_job = to_nouveau_exec_job(job);
struct nouveau_cli *cli = job->cli; struct nouveau_cli *cli = job->cli;
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli); struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(cli);
struct drm_exec *exec = &job->exec;
struct drm_gem_object *obj;
unsigned long index;
int ret; int ret;
/* Create a new fence, but do not emit yet. */ /* Create a new fence, but do not emit yet. */
...@@ -102,52 +98,29 @@ nouveau_exec_job_submit(struct nouveau_job *job) ...@@ -102,52 +98,29 @@ nouveau_exec_job_submit(struct nouveau_job *job)
return ret; return ret;
nouveau_uvmm_lock(uvmm); nouveau_uvmm_lock(uvmm);
drm_exec_init(exec, DRM_EXEC_INTERRUPTIBLE_WAIT | ret = drm_gpuvm_exec_lock(vme);
DRM_EXEC_IGNORE_DUPLICATES); if (ret) {
drm_exec_until_all_locked(exec) { nouveau_uvmm_unlock(uvmm);
struct drm_gpuva *va; return ret;
drm_gpuvm_for_each_va(va, &uvmm->base) {
if (unlikely(va == &uvmm->base.kernel_alloc_node))
continue;
ret = drm_exec_prepare_obj(exec, va->gem.obj, 1);
drm_exec_retry_on_contention(exec);
if (ret)
goto err_uvmm_unlock;
}
} }
nouveau_uvmm_unlock(uvmm); nouveau_uvmm_unlock(uvmm);
drm_exec_for_each_locked_object(exec, index, obj) { ret = drm_gpuvm_exec_validate(vme);
struct nouveau_bo *nvbo = nouveau_gem_object(obj); if (ret) {
drm_gpuvm_exec_unlock(vme);
ret = nouveau_bo_validate(nvbo, true, false); return ret;
if (ret)
goto err_exec_fini;
} }
return 0; return 0;
err_uvmm_unlock:
nouveau_uvmm_unlock(uvmm);
err_exec_fini:
drm_exec_fini(exec);
return ret;
} }
static void static void
nouveau_exec_job_armed_submit(struct nouveau_job *job) nouveau_exec_job_armed_submit(struct nouveau_job *job,
struct drm_gpuvm_exec *vme)
{ {
struct drm_exec *exec = &job->exec; drm_gpuvm_exec_resv_add_fence(vme, job->done_fence,
struct drm_gem_object *obj; job->resv_usage, job->resv_usage);
unsigned long index; drm_gpuvm_exec_unlock(vme);
drm_exec_for_each_locked_object(exec, index, obj)
dma_resv_add_fence(obj->resv, job->done_fence, job->resv_usage);
drm_exec_fini(exec);
} }
static struct dma_fence * static struct dma_fence *
......
...@@ -3,16 +3,12 @@ ...@@ -3,16 +3,12 @@
#ifndef __NOUVEAU_EXEC_H__ #ifndef __NOUVEAU_EXEC_H__
#define __NOUVEAU_EXEC_H__ #define __NOUVEAU_EXEC_H__
#include <drm/drm_exec.h>
#include "nouveau_drv.h" #include "nouveau_drv.h"
#include "nouveau_sched.h" #include "nouveau_sched.h"
struct nouveau_exec_job_args { struct nouveau_exec_job_args {
struct drm_file *file_priv; struct drm_file *file_priv;
struct nouveau_sched_entity *sched_entity; struct nouveau_sched_entity *sched_entity;
struct drm_exec exec;
struct nouveau_channel *chan; struct nouveau_channel *chan;
struct { struct {
......
...@@ -263,6 +263,11 @@ nouveau_job_submit(struct nouveau_job *job) ...@@ -263,6 +263,11 @@ nouveau_job_submit(struct nouveau_job *job)
{ {
struct nouveau_sched_entity *entity = to_nouveau_sched_entity(job->base.entity); struct nouveau_sched_entity *entity = to_nouveau_sched_entity(job->base.entity);
struct dma_fence *done_fence = NULL; struct dma_fence *done_fence = NULL;
struct drm_gpuvm_exec vm_exec = {
.vm = &nouveau_cli_uvmm(job->cli)->base,
.flags = DRM_EXEC_IGNORE_DUPLICATES,
.num_fences = 1,
};
int ret; int ret;
ret = nouveau_job_add_deps(job); ret = nouveau_job_add_deps(job);
...@@ -282,7 +287,7 @@ nouveau_job_submit(struct nouveau_job *job) ...@@ -282,7 +287,7 @@ nouveau_job_submit(struct nouveau_job *job)
* successfully. * successfully.
*/ */
if (job->ops->submit) { if (job->ops->submit) {
ret = job->ops->submit(job); ret = job->ops->submit(job, &vm_exec);
if (ret) if (ret)
goto err_cleanup; goto err_cleanup;
} }
...@@ -315,7 +320,7 @@ nouveau_job_submit(struct nouveau_job *job) ...@@ -315,7 +320,7 @@ nouveau_job_submit(struct nouveau_job *job)
set_bit(DRM_SCHED_FENCE_DONT_PIPELINE, &job->done_fence->flags); set_bit(DRM_SCHED_FENCE_DONT_PIPELINE, &job->done_fence->flags);
if (job->ops->armed_submit) if (job->ops->armed_submit)
job->ops->armed_submit(job); job->ops->armed_submit(job, &vm_exec);
nouveau_job_fence_attach(job); nouveau_job_fence_attach(job);
......
...@@ -5,7 +5,7 @@ ...@@ -5,7 +5,7 @@
#include <linux/types.h> #include <linux/types.h>
#include <drm/drm_exec.h> #include <drm/drm_gpuvm.h>
#include <drm/gpu_scheduler.h> #include <drm/gpu_scheduler.h>
#include "nouveau_drv.h" #include "nouveau_drv.h"
...@@ -54,7 +54,6 @@ struct nouveau_job { ...@@ -54,7 +54,6 @@ struct nouveau_job {
struct drm_file *file_priv; struct drm_file *file_priv;
struct nouveau_cli *cli; struct nouveau_cli *cli;
struct drm_exec exec;
enum dma_resv_usage resv_usage; enum dma_resv_usage resv_usage;
struct dma_fence *done_fence; struct dma_fence *done_fence;
...@@ -76,8 +75,8 @@ struct nouveau_job { ...@@ -76,8 +75,8 @@ struct nouveau_job {
/* If .submit() returns without any error, it is guaranteed that /* If .submit() returns without any error, it is guaranteed that
* armed_submit() is called. * armed_submit() is called.
*/ */
int (*submit)(struct nouveau_job *); int (*submit)(struct nouveau_job *, struct drm_gpuvm_exec *);
void (*armed_submit)(struct nouveau_job *); void (*armed_submit)(struct nouveau_job *, struct drm_gpuvm_exec *);
struct dma_fence *(*run)(struct nouveau_job *); struct dma_fence *(*run)(struct nouveau_job *);
void (*free)(struct nouveau_job *); void (*free)(struct nouveau_job *);
enum drm_gpu_sched_stat (*timeout)(struct nouveau_job *); enum drm_gpu_sched_stat (*timeout)(struct nouveau_job *);
......
...@@ -438,8 +438,9 @@ nouveau_uvma_region_complete(struct nouveau_uvma_region *reg) ...@@ -438,8 +438,9 @@ nouveau_uvma_region_complete(struct nouveau_uvma_region *reg)
static void static void
op_map_prepare_unwind(struct nouveau_uvma *uvma) op_map_prepare_unwind(struct nouveau_uvma *uvma)
{ {
struct drm_gpuva *va = &uvma->va;
nouveau_uvma_gem_put(uvma); nouveau_uvma_gem_put(uvma);
drm_gpuva_remove(&uvma->va); drm_gpuva_remove(va);
nouveau_uvma_free(uvma); nouveau_uvma_free(uvma);
} }
...@@ -468,6 +469,7 @@ nouveau_uvmm_sm_prepare_unwind(struct nouveau_uvmm *uvmm, ...@@ -468,6 +469,7 @@ nouveau_uvmm_sm_prepare_unwind(struct nouveau_uvmm *uvmm,
break; break;
case DRM_GPUVA_OP_REMAP: { case DRM_GPUVA_OP_REMAP: {
struct drm_gpuva_op_remap *r = &op->remap; struct drm_gpuva_op_remap *r = &op->remap;
struct drm_gpuva *va = r->unmap->va;
if (r->next) if (r->next)
op_map_prepare_unwind(new->next); op_map_prepare_unwind(new->next);
...@@ -475,7 +477,7 @@ nouveau_uvmm_sm_prepare_unwind(struct nouveau_uvmm *uvmm, ...@@ -475,7 +477,7 @@ nouveau_uvmm_sm_prepare_unwind(struct nouveau_uvmm *uvmm,
if (r->prev) if (r->prev)
op_map_prepare_unwind(new->prev); op_map_prepare_unwind(new->prev);
op_unmap_prepare_unwind(r->unmap->va); op_unmap_prepare_unwind(va);
break; break;
} }
case DRM_GPUVA_OP_UNMAP: case DRM_GPUVA_OP_UNMAP:
...@@ -634,6 +636,7 @@ nouveau_uvmm_sm_prepare(struct nouveau_uvmm *uvmm, ...@@ -634,6 +636,7 @@ nouveau_uvmm_sm_prepare(struct nouveau_uvmm *uvmm,
goto unwind; goto unwind;
} }
} }
break; break;
} }
case DRM_GPUVA_OP_REMAP: { case DRM_GPUVA_OP_REMAP: {
...@@ -1135,12 +1138,53 @@ bind_link_gpuvas(struct bind_job_op *bop) ...@@ -1135,12 +1138,53 @@ bind_link_gpuvas(struct bind_job_op *bop)
} }
static int static int
nouveau_uvmm_bind_job_submit(struct nouveau_job *job) bind_lock_validate(struct nouveau_job *job, struct drm_exec *exec,
unsigned int num_fences)
{
struct nouveau_uvmm_bind_job *bind_job = to_uvmm_bind_job(job);
struct bind_job_op *op;
int ret;
list_for_each_op(op, &bind_job->ops) {
struct drm_gpuva_op *va_op;
if (!op->ops)
continue;
drm_gpuva_for_each_op(va_op, op->ops) {
struct drm_gem_object *obj = op_gem_obj(va_op);
if (unlikely(!obj))
continue;
ret = drm_exec_prepare_obj(exec, obj, num_fences);
if (ret)
return ret;
/* Don't validate GEMs backing mappings we're about to
* unmap, it's not worth the effort.
*/
if (va_op->op == DRM_GPUVA_OP_UNMAP)
continue;
ret = nouveau_bo_validate(nouveau_gem_object(obj),
true, false);
if (ret)
return ret;
}
}
return 0;
}
static int
nouveau_uvmm_bind_job_submit(struct nouveau_job *job,
struct drm_gpuvm_exec *vme)
{ {
struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli); struct nouveau_uvmm *uvmm = nouveau_cli_uvmm(job->cli);
struct nouveau_uvmm_bind_job *bind_job = to_uvmm_bind_job(job); struct nouveau_uvmm_bind_job *bind_job = to_uvmm_bind_job(job);
struct nouveau_sched_entity *entity = job->entity; struct nouveau_sched_entity *entity = job->entity;
struct drm_exec *exec = &job->exec; struct drm_exec *exec = &vme->exec;
struct bind_job_op *op; struct bind_job_op *op;
int ret; int ret;
...@@ -1157,6 +1201,8 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ...@@ -1157,6 +1201,8 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job)
dma_resv_unlock(obj->resv); dma_resv_unlock(obj->resv);
if (IS_ERR(op->vm_bo)) if (IS_ERR(op->vm_bo))
return PTR_ERR(op->vm_bo); return PTR_ERR(op->vm_bo);
drm_gpuvm_bo_extobj_add(op->vm_bo);
} }
ret = bind_validate_op(job, op); ret = bind_validate_op(job, op);
...@@ -1179,6 +1225,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ...@@ -1179,6 +1225,7 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job)
* unwind all GPU VA space changes on failure. * unwind all GPU VA space changes on failure.
*/ */
nouveau_uvmm_lock(uvmm); nouveau_uvmm_lock(uvmm);
list_for_each_op(op, &bind_job->ops) { list_for_each_op(op, &bind_job->ops) {
switch (op->op) { switch (op->op) {
case OP_MAP_SPARSE: case OP_MAP_SPARSE:
...@@ -1290,57 +1337,15 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ...@@ -1290,57 +1337,15 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job)
} }
} }
drm_exec_init(exec, DRM_EXEC_INTERRUPTIBLE_WAIT | drm_exec_init(exec, vme->flags);
DRM_EXEC_IGNORE_DUPLICATES);
drm_exec_until_all_locked(exec) { drm_exec_until_all_locked(exec) {
list_for_each_op(op, &bind_job->ops) { ret = bind_lock_validate(job, exec, vme->num_fences);
struct drm_gpuva_op *va_op;
if (IS_ERR_OR_NULL(op->ops))
continue;
drm_gpuva_for_each_op(va_op, op->ops) {
struct drm_gem_object *obj = op_gem_obj(va_op);
if (unlikely(!obj))
continue;
ret = drm_exec_prepare_obj(exec, obj, 1);
drm_exec_retry_on_contention(exec); drm_exec_retry_on_contention(exec);
if (ret) { if (ret) {
op = list_last_op(&bind_job->ops); op = list_last_op(&bind_job->ops);
goto unwind; goto unwind;
} }
} }
}
}
list_for_each_op(op, &bind_job->ops) {
struct drm_gpuva_op *va_op;
if (IS_ERR_OR_NULL(op->ops))
continue;
drm_gpuva_for_each_op(va_op, op->ops) {
struct drm_gem_object *obj = op_gem_obj(va_op);
if (unlikely(!obj))
continue;
/* Don't validate GEMs backing mappings we're about to
* unmap, it's not worth the effort.
*/
if (unlikely(va_op->op == DRM_GPUVA_OP_UNMAP))
continue;
ret = nouveau_bo_validate(nouveau_gem_object(obj),
true, false);
if (ret) {
op = list_last_op(&bind_job->ops);
goto unwind;
}
}
}
/* Link and unlink GPUVAs while holding the dma_resv lock. /* Link and unlink GPUVAs while holding the dma_resv lock.
* *
...@@ -1413,21 +1418,17 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job) ...@@ -1413,21 +1418,17 @@ nouveau_uvmm_bind_job_submit(struct nouveau_job *job)
} }
nouveau_uvmm_unlock(uvmm); nouveau_uvmm_unlock(uvmm);
drm_exec_fini(exec); drm_gpuvm_exec_unlock(vme);
return ret; return ret;
} }
static void static void
nouveau_uvmm_bind_job_armed_submit(struct nouveau_job *job) nouveau_uvmm_bind_job_armed_submit(struct nouveau_job *job,
struct drm_gpuvm_exec *vme)
{ {
struct drm_exec *exec = &job->exec; drm_gpuvm_exec_resv_add_fence(vme, job->done_fence,
struct drm_gem_object *obj; job->resv_usage, job->resv_usage);
unsigned long index; drm_gpuvm_exec_unlock(vme);
drm_exec_for_each_locked_object(exec, index, obj)
dma_resv_add_fence(obj->resv, job->done_fence, job->resv_usage);
drm_exec_fini(exec);
} }
static struct dma_fence * static struct dma_fence *
...@@ -1815,8 +1816,17 @@ nouveau_uvmm_free(struct drm_gpuvm *gpuvm) ...@@ -1815,8 +1816,17 @@ nouveau_uvmm_free(struct drm_gpuvm *gpuvm)
kfree(uvmm); kfree(uvmm);
} }
static int
nouveau_uvmm_bo_validate(struct drm_gpuvm_bo *vm_bo, struct drm_exec *exec)
{
struct nouveau_bo *nvbo = nouveau_gem_object(vm_bo->obj);
return nouveau_bo_validate(nvbo, true, false);
}
static const struct drm_gpuvm_ops gpuvm_ops = { static const struct drm_gpuvm_ops gpuvm_ops = {
.vm_free = nouveau_uvmm_free, .vm_free = nouveau_uvmm_free,
.vm_bo_validate = nouveau_uvmm_bo_validate,
}; };
int int
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment