Commit 347987a2 authored by Christian König's avatar Christian König

drm/ttm: rename and cleanup ttm_bo_init

Rename ttm_bo_init to ttm_bo_init_validate since that better matches
what the function is actually doing.

Remove the unused size parameter, move the function's kerneldoc to the
implementation and cleanup the whole error handling.
Signed-off-by: default avatarChristian König <christian.koenig@amd.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20220707102453.3633-2-christian.koenig@amd.comReviewed-by: default avatarMichael J. Ruhl <michael.j.ruhl@intel.com>
parent 20529e26
...@@ -591,7 +591,7 @@ int amdgpu_bo_create(struct amdgpu_device *adev, ...@@ -591,7 +591,7 @@ int amdgpu_bo_create(struct amdgpu_device *adev,
if (!bp->destroy) if (!bp->destroy)
bp->destroy = &amdgpu_bo_destroy; bp->destroy = &amdgpu_bo_destroy;
r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, size, bp->type, r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, bp->type,
&bo->placement, page_align, &ctx, NULL, &bo->placement, page_align, &ctx, NULL,
bp->resv, bp->destroy); bp->resv, bp->destroy);
if (unlikely(r != 0)) if (unlikely(r != 0))
......
...@@ -226,9 +226,9 @@ struct drm_gem_vram_object *drm_gem_vram_create(struct drm_device *dev, ...@@ -226,9 +226,9 @@ struct drm_gem_vram_object *drm_gem_vram_create(struct drm_device *dev,
* A failing ttm_bo_init will call ttm_buffer_object_destroy * A failing ttm_bo_init will call ttm_buffer_object_destroy
* to release gbo->bo.base and kfree gbo. * to release gbo->bo.base and kfree gbo.
*/ */
ret = ttm_bo_init(bdev, &gbo->bo, size, ttm_bo_type_device, ret = ttm_bo_init_validate(bdev, &gbo->bo, ttm_bo_type_device,
&gbo->placement, pg_align, false, NULL, NULL, &gbo->placement, pg_align, false, NULL, NULL,
ttm_buffer_object_destroy); ttm_buffer_object_destroy);
if (ret) if (ret)
return ERR_PTR(ret); return ERR_PTR(ret);
......
...@@ -1229,9 +1229,8 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem, ...@@ -1229,9 +1229,8 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem,
* Similarly, in delayed_destroy, we can't call ttm_bo_put() * Similarly, in delayed_destroy, we can't call ttm_bo_put()
* until successful initialization. * until successful initialization.
*/ */
ret = ttm_bo_init_reserved(&i915->bdev, i915_gem_to_ttm(obj), size, ret = ttm_bo_init_reserved(&i915->bdev, i915_gem_to_ttm(obj), bo_type,
bo_type, &i915_sys_placement, &i915_sys_placement, page_size >> PAGE_SHIFT,
page_size >> PAGE_SHIFT,
&ctx, NULL, NULL, i915_ttm_bo_destroy); &ctx, NULL, NULL, i915_ttm_bo_destroy);
if (ret) if (ret)
return i915_ttm_err_to_gem(ret); return i915_ttm_err_to_gem(ret);
......
...@@ -309,9 +309,9 @@ nouveau_bo_init(struct nouveau_bo *nvbo, u64 size, int align, u32 domain, ...@@ -309,9 +309,9 @@ nouveau_bo_init(struct nouveau_bo *nvbo, u64 size, int align, u32 domain,
nouveau_bo_placement_set(nvbo, domain, 0); nouveau_bo_placement_set(nvbo, domain, 0);
INIT_LIST_HEAD(&nvbo->io_reserve_lru); INIT_LIST_HEAD(&nvbo->io_reserve_lru);
ret = ttm_bo_init(nvbo->bo.bdev, &nvbo->bo, size, type, ret = ttm_bo_init_validate(nvbo->bo.bdev, &nvbo->bo, type,
&nvbo->placement, align >> PAGE_SHIFT, false, sg, &nvbo->placement, align >> PAGE_SHIFT, false,
robj, nouveau_bo_del_ttm); sg, robj, nouveau_bo_del_ttm);
if (ret) { if (ret) {
/* ttm will call nouveau_bo_del_ttm if it fails.. */ /* ttm will call nouveau_bo_del_ttm if it fails.. */
return ret; return ret;
......
...@@ -141,7 +141,7 @@ int qxl_bo_create(struct qxl_device *qdev, unsigned long size, ...@@ -141,7 +141,7 @@ int qxl_bo_create(struct qxl_device *qdev, unsigned long size,
qxl_ttm_placement_from_domain(bo, domain); qxl_ttm_placement_from_domain(bo, domain);
bo->tbo.priority = priority; bo->tbo.priority = priority;
r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, size, type, r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, type,
&bo->placement, 0, &ctx, NULL, NULL, &bo->placement, 0, &ctx, NULL, NULL,
&qxl_ttm_bo_destroy); &qxl_ttm_bo_destroy);
if (unlikely(r != 0)) { if (unlikely(r != 0)) {
......
...@@ -202,9 +202,9 @@ int radeon_bo_create(struct radeon_device *rdev, ...@@ -202,9 +202,9 @@ int radeon_bo_create(struct radeon_device *rdev,
radeon_ttm_placement_from_domain(bo, domain); radeon_ttm_placement_from_domain(bo, domain);
/* Kernel allocation are uninterruptible */ /* Kernel allocation are uninterruptible */
down_read(&rdev->pm.mclk_lock); down_read(&rdev->pm.mclk_lock);
r = ttm_bo_init(&rdev->mman.bdev, &bo->tbo, size, type, r = ttm_bo_init_validate(&rdev->mman.bdev, &bo->tbo, type,
&bo->placement, page_align, !kernel, sg, resv, &bo->placement, page_align, !kernel, sg, resv,
&radeon_ttm_bo_destroy); &radeon_ttm_bo_destroy);
up_read(&rdev->pm.mclk_lock); up_read(&rdev->pm.mclk_lock);
if (unlikely(r != 0)) { if (unlikely(r != 0)) {
return r; return r;
......
...@@ -915,36 +915,61 @@ int ttm_bo_validate(struct ttm_buffer_object *bo, ...@@ -915,36 +915,61 @@ int ttm_bo_validate(struct ttm_buffer_object *bo,
} }
EXPORT_SYMBOL(ttm_bo_validate); EXPORT_SYMBOL(ttm_bo_validate);
int ttm_bo_init_reserved(struct ttm_device *bdev, /**
struct ttm_buffer_object *bo, * ttm_bo_init_reserved
size_t size, *
enum ttm_bo_type type, * @bdev: Pointer to a ttm_device struct.
struct ttm_placement *placement, * @bo: Pointer to a ttm_buffer_object to be initialized.
uint32_t page_alignment, * @type: Requested type of buffer object.
struct ttm_operation_ctx *ctx, * @placement: Initial placement for buffer object.
struct sg_table *sg, * @alignment: Data alignment in pages.
struct dma_resv *resv, * @ctx: TTM operation context for memory allocation.
* @sg: Scatter-gather table.
* @resv: Pointer to a dma_resv, or NULL to let ttm allocate one.
* @destroy: Destroy function. Use NULL for kfree().
*
* This function initializes a pre-allocated struct ttm_buffer_object.
* As this object may be part of a larger structure, this function,
* together with the @destroy function, enables driver-specific objects
* derived from a ttm_buffer_object.
*
* On successful return, the caller owns an object kref to @bo. The kref and
* list_kref are usually set to 1, but note that in some situations, other
* tasks may already be holding references to @bo as well.
* Furthermore, if resv == NULL, the buffer's reservation lock will be held,
* and it is the caller's responsibility to call ttm_bo_unreserve.
*
* If a failure occurs, the function will call the @destroy function. Thus,
* after a failure, dereferencing @bo is illegal and will likely cause memory
* corruption.
*
* Returns
* -ENOMEM: Out of memory.
* -EINVAL: Invalid placement flags.
* -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources.
*/
int ttm_bo_init_reserved(struct ttm_device *bdev, struct ttm_buffer_object *bo,
enum ttm_bo_type type, struct ttm_placement *placement,
uint32_t alignment, struct ttm_operation_ctx *ctx,
struct sg_table *sg, struct dma_resv *resv,
void (*destroy) (struct ttm_buffer_object *)) void (*destroy) (struct ttm_buffer_object *))
{ {
static const struct ttm_place sys_mem = { .mem_type = TTM_PL_SYSTEM }; static const struct ttm_place sys_mem = { .mem_type = TTM_PL_SYSTEM };
bool locked;
int ret; int ret;
bo->destroy = destroy;
kref_init(&bo->kref); kref_init(&bo->kref);
INIT_LIST_HEAD(&bo->ddestroy); INIT_LIST_HEAD(&bo->ddestroy);
bo->bdev = bdev; bo->bdev = bdev;
bo->type = type; bo->type = type;
bo->page_alignment = page_alignment; bo->page_alignment = alignment;
bo->destroy = destroy;
bo->pin_count = 0; bo->pin_count = 0;
bo->sg = sg; bo->sg = sg;
bo->bulk_move = NULL; bo->bulk_move = NULL;
if (resv) { if (resv)
bo->base.resv = resv; bo->base.resv = resv;
dma_resv_assert_held(bo->base.resv); else
} else {
bo->base.resv = &bo->base._resv; bo->base.resv = &bo->base._resv;
}
atomic_inc(&ttm_glob.bo_count); atomic_inc(&ttm_glob.bo_count);
ret = ttm_resource_alloc(bo, &sys_mem, &bo->resource); ret = ttm_resource_alloc(bo, &sys_mem, &bo->resource);
...@@ -957,50 +982,84 @@ int ttm_bo_init_reserved(struct ttm_device *bdev, ...@@ -957,50 +982,84 @@ int ttm_bo_init_reserved(struct ttm_device *bdev,
* For ttm_bo_type_device buffers, allocate * For ttm_bo_type_device buffers, allocate
* address space from the device. * address space from the device.
*/ */
if (bo->type == ttm_bo_type_device || if (bo->type == ttm_bo_type_device || bo->type == ttm_bo_type_sg) {
bo->type == ttm_bo_type_sg)
ret = drm_vma_offset_add(bdev->vma_manager, &bo->base.vma_node, ret = drm_vma_offset_add(bdev->vma_manager, &bo->base.vma_node,
bo->resource->num_pages); PFN_UP(bo->base.size));
if (ret)
goto err_put;
}
/* passed reservation objects should already be locked, /* passed reservation objects should already be locked,
* since otherwise lockdep will be angered in radeon. * since otherwise lockdep will be angered in radeon.
*/ */
if (!resv) { if (!resv)
locked = dma_resv_trylock(bo->base.resv); WARN_ON(!dma_resv_trylock(bo->base.resv));
WARN_ON(!locked); else
} dma_resv_assert_held(resv);
if (likely(!ret)) ret = ttm_bo_validate(bo, placement, ctx);
ret = ttm_bo_validate(bo, placement, ctx); if (unlikely(ret))
goto err_unlock;
if (unlikely(ret)) { return 0;
if (!resv)
ttm_bo_unreserve(bo);
ttm_bo_put(bo); err_unlock:
return ret; if (!resv)
} dma_resv_unlock(bo->base.resv);
err_put:
ttm_bo_put(bo);
return ret; return ret;
} }
EXPORT_SYMBOL(ttm_bo_init_reserved); EXPORT_SYMBOL(ttm_bo_init_reserved);
int ttm_bo_init(struct ttm_device *bdev, /**
struct ttm_buffer_object *bo, * ttm_bo_init_validate
size_t size, *
enum ttm_bo_type type, * @bdev: Pointer to a ttm_device struct.
struct ttm_placement *placement, * @bo: Pointer to a ttm_buffer_object to be initialized.
uint32_t page_alignment, * @type: Requested type of buffer object.
bool interruptible, * @placement: Initial placement for buffer object.
struct sg_table *sg, * @alignment: Data alignment in pages.
struct dma_resv *resv, * @interruptible: If needing to sleep to wait for GPU resources,
void (*destroy) (struct ttm_buffer_object *)) * sleep interruptible.
* pinned in physical memory. If this behaviour is not desired, this member
* holds a pointer to a persistent shmem object. Typically, this would
* point to the shmem object backing a GEM object if TTM is used to back a
* GEM user interface.
* @sg: Scatter-gather table.
* @resv: Pointer to a dma_resv, or NULL to let ttm allocate one.
* @destroy: Destroy function. Use NULL for kfree().
*
* This function initializes a pre-allocated struct ttm_buffer_object.
* As this object may be part of a larger structure, this function,
* together with the @destroy function,
* enables driver-specific objects derived from a ttm_buffer_object.
*
* On successful return, the caller owns an object kref to @bo. The kref and
* list_kref are usually set to 1, but note that in some situations, other
* tasks may already be holding references to @bo as well.
*
* If a failure occurs, the function will call the @destroy function, Thus,
* after a failure, dereferencing @bo is illegal and will likely cause memory
* corruption.
*
* Returns
* -ENOMEM: Out of memory.
* -EINVAL: Invalid placement flags.
* -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources.
*/
int ttm_bo_init_validate(struct ttm_device *bdev, struct ttm_buffer_object *bo,
enum ttm_bo_type type, struct ttm_placement *placement,
uint32_t alignment, bool interruptible,
struct sg_table *sg, struct dma_resv *resv,
void (*destroy) (struct ttm_buffer_object *))
{ {
struct ttm_operation_ctx ctx = { interruptible, false }; struct ttm_operation_ctx ctx = { interruptible, false };
int ret; int ret;
ret = ttm_bo_init_reserved(bdev, bo, size, type, placement, ret = ttm_bo_init_reserved(bdev, bo, type, placement, alignment, &ctx,
page_alignment, &ctx, sg, resv, destroy); sg, resv, destroy);
if (ret) if (ret)
return ret; return ret;
...@@ -1009,7 +1068,7 @@ int ttm_bo_init(struct ttm_device *bdev, ...@@ -1009,7 +1068,7 @@ int ttm_bo_init(struct ttm_device *bdev,
return 0; return 0;
} }
EXPORT_SYMBOL(ttm_bo_init); EXPORT_SYMBOL(ttm_bo_init_validate);
/* /*
* buffer object vm functions. * buffer object vm functions.
......
...@@ -429,9 +429,9 @@ int vmw_bo_create_kernel(struct vmw_private *dev_priv, unsigned long size, ...@@ -429,9 +429,9 @@ int vmw_bo_create_kernel(struct vmw_private *dev_priv, unsigned long size,
drm_gem_private_object_init(vdev, &bo->base, size); drm_gem_private_object_init(vdev, &bo->base, size);
ret = ttm_bo_init_reserved(&dev_priv->bdev, bo, size, ret = ttm_bo_init_reserved(&dev_priv->bdev, bo, ttm_bo_type_kernel,
ttm_bo_type_kernel, placement, 0, placement, 0, &ctx, NULL, NULL,
&ctx, NULL, NULL, vmw_bo_default_destroy); vmw_bo_default_destroy);
if (unlikely(ret)) if (unlikely(ret))
goto error_free; goto error_free;
...@@ -512,10 +512,8 @@ int vmw_bo_init(struct vmw_private *dev_priv, ...@@ -512,10 +512,8 @@ int vmw_bo_init(struct vmw_private *dev_priv,
size = ALIGN(size, PAGE_SIZE); size = ALIGN(size, PAGE_SIZE);
drm_gem_private_object_init(vdev, &vmw_bo->base.base, size); drm_gem_private_object_init(vdev, &vmw_bo->base.base, size);
ret = ttm_bo_init_reserved(bdev, &vmw_bo->base, size, ret = ttm_bo_init_reserved(bdev, &vmw_bo->base, ttm_bo_type_device,
ttm_bo_type_device, placement, 0, &ctx, NULL, NULL, bo_free);
placement,
0, &ctx, NULL, NULL, bo_free);
if (unlikely(ret)) { if (unlikely(ret)) {
return ret; return ret;
} }
......
...@@ -317,93 +317,16 @@ void ttm_bo_unlock_delayed_workqueue(struct ttm_device *bdev, int resched); ...@@ -317,93 +317,16 @@ void ttm_bo_unlock_delayed_workqueue(struct ttm_device *bdev, int resched);
bool ttm_bo_eviction_valuable(struct ttm_buffer_object *bo, bool ttm_bo_eviction_valuable(struct ttm_buffer_object *bo,
const struct ttm_place *place); const struct ttm_place *place);
/** int ttm_bo_init_reserved(struct ttm_device *bdev, struct ttm_buffer_object *bo,
* ttm_bo_init_reserved enum ttm_bo_type type, struct ttm_placement *placement,
* uint32_t alignment, struct ttm_operation_ctx *ctx,
* @bdev: Pointer to a ttm_device struct. struct sg_table *sg, struct dma_resv *resv,
* @bo: Pointer to a ttm_buffer_object to be initialized. void (*destroy) (struct ttm_buffer_object *));
* @size: Requested size of buffer object. int ttm_bo_init_validate(struct ttm_device *bdev, struct ttm_buffer_object *bo,
* @type: Requested type of buffer object. enum ttm_bo_type type, struct ttm_placement *placement,
* @placement: Initial placement for buffer object. uint32_t alignment, bool interruptible,
* @page_alignment: Data alignment in pages.
* @ctx: TTM operation context for memory allocation.
* @sg: Scatter-gather table.
* @resv: Pointer to a dma_resv, or NULL to let ttm allocate one.
* @destroy: Destroy function. Use NULL for kfree().
*
* This function initializes a pre-allocated struct ttm_buffer_object.
* As this object may be part of a larger structure, this function,
* together with the @destroy function,
* enables driver-specific objects derived from a ttm_buffer_object.
*
* On successful return, the caller owns an object kref to @bo. The kref and
* list_kref are usually set to 1, but note that in some situations, other
* tasks may already be holding references to @bo as well.
* Furthermore, if resv == NULL, the buffer's reservation lock will be held,
* and it is the caller's responsibility to call ttm_bo_unreserve.
*
* If a failure occurs, the function will call the @destroy function, or
* kfree() if @destroy is NULL. Thus, after a failure, dereferencing @bo is
* illegal and will likely cause memory corruption.
*
* Returns
* -ENOMEM: Out of memory.
* -EINVAL: Invalid placement flags.
* -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources.
*/
int ttm_bo_init_reserved(struct ttm_device *bdev,
struct ttm_buffer_object *bo,
size_t size, enum ttm_bo_type type,
struct ttm_placement *placement,
uint32_t page_alignment,
struct ttm_operation_ctx *ctx,
struct sg_table *sg, struct dma_resv *resv, struct sg_table *sg, struct dma_resv *resv,
void (*destroy) (struct ttm_buffer_object *)); void (*destroy) (struct ttm_buffer_object *));
/**
* ttm_bo_init
*
* @bdev: Pointer to a ttm_device struct.
* @bo: Pointer to a ttm_buffer_object to be initialized.
* @size: Requested size of buffer object.
* @type: Requested type of buffer object.
* @placement: Initial placement for buffer object.
* @page_alignment: Data alignment in pages.
* @interruptible: If needing to sleep to wait for GPU resources,
* sleep interruptible.
* pinned in physical memory. If this behaviour is not desired, this member
* holds a pointer to a persistent shmem object. Typically, this would
* point to the shmem object backing a GEM object if TTM is used to back a
* GEM user interface.
* @sg: Scatter-gather table.
* @resv: Pointer to a dma_resv, or NULL to let ttm allocate one.
* @destroy: Destroy function. Use NULL for kfree().
*
* This function initializes a pre-allocated struct ttm_buffer_object.
* As this object may be part of a larger structure, this function,
* together with the @destroy function,
* enables driver-specific objects derived from a ttm_buffer_object.
*
* On successful return, the caller owns an object kref to @bo. The kref and
* list_kref are usually set to 1, but note that in some situations, other
* tasks may already be holding references to @bo as well.
*
* If a failure occurs, the function will call the @destroy function, or
* kfree() if @destroy is NULL. Thus, after a failure, dereferencing @bo is
* illegal and will likely cause memory corruption.
*
* Returns
* -ENOMEM: Out of memory.
* -EINVAL: Invalid placement flags.
* -ERESTARTSYS: Interrupted by signal while sleeping waiting for resources.
*/
int ttm_bo_init(struct ttm_device *bdev, struct ttm_buffer_object *bo,
size_t size, enum ttm_bo_type type,
struct ttm_placement *placement,
uint32_t page_alignment, bool interrubtible,
struct sg_table *sg, struct dma_resv *resv,
void (*destroy) (struct ttm_buffer_object *));
/** /**
* ttm_kmap_obj_virtual * ttm_kmap_obj_virtual
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment