Commit 4ce9891e authored by Chunming Zhou's avatar Chunming Zhou Committed by Alex Deucher

drm/amdgpu: improve sa_bo->fence by kernel fence

Signed-off-by: default avatarChunming Zhou <david1.zhou@amd.com>
Reviewed-by: default avatarChristian K?nig <christian.koenig@amd.com>
parent 1aa4051b
...@@ -441,7 +441,7 @@ int amdgpu_fence_wait_empty(struct amdgpu_ring *ring); ...@@ -441,7 +441,7 @@ int amdgpu_fence_wait_empty(struct amdgpu_ring *ring);
unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring); unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring);
signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev, signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev,
struct amdgpu_fence **array, struct fence **array,
uint32_t count, uint32_t count,
bool wait_all, bool wait_all,
bool intr, bool intr,
...@@ -654,7 +654,7 @@ struct amdgpu_sa_bo { ...@@ -654,7 +654,7 @@ struct amdgpu_sa_bo {
struct amdgpu_sa_manager *manager; struct amdgpu_sa_manager *manager;
unsigned soffset; unsigned soffset;
unsigned eoffset; unsigned eoffset;
struct amdgpu_fence *fence; struct fence *fence;
}; };
/* /*
...@@ -696,7 +696,7 @@ bool amdgpu_semaphore_emit_wait(struct amdgpu_ring *ring, ...@@ -696,7 +696,7 @@ bool amdgpu_semaphore_emit_wait(struct amdgpu_ring *ring,
struct amdgpu_semaphore *semaphore); struct amdgpu_semaphore *semaphore);
void amdgpu_semaphore_free(struct amdgpu_device *adev, void amdgpu_semaphore_free(struct amdgpu_device *adev,
struct amdgpu_semaphore **semaphore, struct amdgpu_semaphore **semaphore,
struct amdgpu_fence *fence); struct fence *fence);
/* /*
* Synchronization * Synchronization
...@@ -717,7 +717,7 @@ int amdgpu_sync_resv(struct amdgpu_device *adev, ...@@ -717,7 +717,7 @@ int amdgpu_sync_resv(struct amdgpu_device *adev,
int amdgpu_sync_rings(struct amdgpu_sync *sync, int amdgpu_sync_rings(struct amdgpu_sync *sync,
struct amdgpu_ring *ring); struct amdgpu_ring *ring);
void amdgpu_sync_free(struct amdgpu_device *adev, struct amdgpu_sync *sync, void amdgpu_sync_free(struct amdgpu_device *adev, struct amdgpu_sync *sync,
struct amdgpu_fence *fence); struct fence *fence);
/* /*
* GART structures, functions & helpers * GART structures, functions & helpers
......
...@@ -836,30 +836,30 @@ static inline bool amdgpu_test_signaled(struct amdgpu_fence *fence) ...@@ -836,30 +836,30 @@ static inline bool amdgpu_test_signaled(struct amdgpu_fence *fence)
return test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags); return test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags);
} }
static bool amdgpu_test_signaled_any(struct amdgpu_fence **fences, uint32_t count) static bool amdgpu_test_signaled_any(struct fence **fences, uint32_t count)
{ {
int idx; int idx;
struct amdgpu_fence *fence; struct fence *fence;
for (idx = 0; idx < count; ++idx) { for (idx = 0; idx < count; ++idx) {
fence = fences[idx]; fence = fences[idx];
if (fence) { if (fence) {
if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags)) if (test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
return true; return true;
} }
} }
return false; return false;
} }
static bool amdgpu_test_signaled_all(struct amdgpu_fence **fences, uint32_t count) static bool amdgpu_test_signaled_all(struct fence **fences, uint32_t count)
{ {
int idx; int idx;
struct amdgpu_fence *fence; struct fence *fence;
for (idx = 0; idx < count; ++idx) { for (idx = 0; idx < count; ++idx) {
fence = fences[idx]; fence = fences[idx];
if (fence) { if (fence) {
if (!test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->base.flags)) if (!test_bit(FENCE_FLAG_SIGNALED_BIT, &fence->flags))
return false; return false;
} }
} }
...@@ -885,7 +885,7 @@ static signed long amdgpu_fence_default_wait(struct fence *f, bool intr, ...@@ -885,7 +885,7 @@ static signed long amdgpu_fence_default_wait(struct fence *f, bool intr,
struct amdgpu_fence *fence = to_amdgpu_fence(f); struct amdgpu_fence *fence = to_amdgpu_fence(f);
struct amdgpu_device *adev = fence->ring->adev; struct amdgpu_device *adev = fence->ring->adev;
return amdgpu_fence_wait_multiple(adev, &fence, 1, false, intr, t); return amdgpu_fence_wait_multiple(adev, &f, 1, false, intr, t);
} }
/** /**
...@@ -902,7 +902,7 @@ static signed long amdgpu_fence_default_wait(struct fence *f, bool intr, ...@@ -902,7 +902,7 @@ static signed long amdgpu_fence_default_wait(struct fence *f, bool intr,
* If wait_all is false, it will return when any fence is signaled or timeout. * If wait_all is false, it will return when any fence is signaled or timeout.
*/ */
signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev, signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev,
struct amdgpu_fence **array, struct fence **array,
uint32_t count, uint32_t count,
bool wait_all, bool wait_all,
bool intr, bool intr,
...@@ -910,7 +910,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev, ...@@ -910,7 +910,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev,
{ {
long idx = 0; long idx = 0;
struct amdgpu_wait_cb *cb; struct amdgpu_wait_cb *cb;
struct amdgpu_fence *fence; struct fence *fence;
BUG_ON(!array); BUG_ON(!array);
...@@ -924,7 +924,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev, ...@@ -924,7 +924,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev,
fence = array[idx]; fence = array[idx];
if (fence) { if (fence) {
cb[idx].task = current; cb[idx].task = current;
if (fence_add_callback(&fence->base, if (fence_add_callback(fence,
&cb[idx].base, amdgpu_fence_wait_cb)) { &cb[idx].base, amdgpu_fence_wait_cb)) {
/* The fence is already signaled */ /* The fence is already signaled */
if (wait_all) if (wait_all)
...@@ -967,7 +967,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev, ...@@ -967,7 +967,7 @@ signed long amdgpu_fence_wait_multiple(struct amdgpu_device *adev,
for (idx = 0; idx < count; ++idx) { for (idx = 0; idx < count; ++idx) {
fence = array[idx]; fence = array[idx];
if (fence) if (fence)
fence_remove_callback(&fence->base, &cb[idx].base); fence_remove_callback(fence, &cb[idx].base);
} }
err_free_cb: err_free_cb:
......
...@@ -93,8 +93,8 @@ int amdgpu_ib_get(struct amdgpu_ring *ring, struct amdgpu_vm *vm, ...@@ -93,8 +93,8 @@ int amdgpu_ib_get(struct amdgpu_ring *ring, struct amdgpu_vm *vm,
*/ */
void amdgpu_ib_free(struct amdgpu_device *adev, struct amdgpu_ib *ib) void amdgpu_ib_free(struct amdgpu_device *adev, struct amdgpu_ib *ib)
{ {
amdgpu_sync_free(adev, &ib->sync, ib->fence); amdgpu_sync_free(adev, &ib->sync, &ib->fence->base);
amdgpu_sa_bo_free(adev, &ib->sa_bo, ib->fence); amdgpu_sa_bo_free(adev, &ib->sa_bo, &ib->fence->base);
amdgpu_fence_unref(&ib->fence); amdgpu_fence_unref(&ib->fence);
} }
......
...@@ -193,7 +193,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev, ...@@ -193,7 +193,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev,
unsigned size, unsigned align); unsigned size, unsigned align);
void amdgpu_sa_bo_free(struct amdgpu_device *adev, void amdgpu_sa_bo_free(struct amdgpu_device *adev,
struct amdgpu_sa_bo **sa_bo, struct amdgpu_sa_bo **sa_bo,
struct amdgpu_fence *fence); struct fence *fence);
#if defined(CONFIG_DEBUG_FS) #if defined(CONFIG_DEBUG_FS)
void amdgpu_sa_bo_dump_debug_info(struct amdgpu_sa_manager *sa_manager, void amdgpu_sa_bo_dump_debug_info(struct amdgpu_sa_manager *sa_manager,
struct seq_file *m); struct seq_file *m);
......
...@@ -139,6 +139,20 @@ int amdgpu_sa_bo_manager_suspend(struct amdgpu_device *adev, ...@@ -139,6 +139,20 @@ int amdgpu_sa_bo_manager_suspend(struct amdgpu_device *adev,
return r; return r;
} }
static uint32_t amdgpu_sa_get_ring_from_fence(struct fence *f)
{
struct amdgpu_fence *a_fence;
struct amd_sched_fence *s_fence;
s_fence = to_amd_sched_fence(f);
if (s_fence)
return s_fence->entity->scheduler->ring_id;
a_fence = to_amdgpu_fence(f);
if (a_fence)
return a_fence->ring->idx;
return 0;
}
static void amdgpu_sa_bo_remove_locked(struct amdgpu_sa_bo *sa_bo) static void amdgpu_sa_bo_remove_locked(struct amdgpu_sa_bo *sa_bo)
{ {
struct amdgpu_sa_manager *sa_manager = sa_bo->manager; struct amdgpu_sa_manager *sa_manager = sa_bo->manager;
...@@ -147,7 +161,7 @@ static void amdgpu_sa_bo_remove_locked(struct amdgpu_sa_bo *sa_bo) ...@@ -147,7 +161,7 @@ static void amdgpu_sa_bo_remove_locked(struct amdgpu_sa_bo *sa_bo)
} }
list_del_init(&sa_bo->olist); list_del_init(&sa_bo->olist);
list_del_init(&sa_bo->flist); list_del_init(&sa_bo->flist);
amdgpu_fence_unref(&sa_bo->fence); fence_put(sa_bo->fence);
kfree(sa_bo); kfree(sa_bo);
} }
...@@ -161,7 +175,7 @@ static void amdgpu_sa_bo_try_free(struct amdgpu_sa_manager *sa_manager) ...@@ -161,7 +175,7 @@ static void amdgpu_sa_bo_try_free(struct amdgpu_sa_manager *sa_manager)
sa_bo = list_entry(sa_manager->hole->next, struct amdgpu_sa_bo, olist); sa_bo = list_entry(sa_manager->hole->next, struct amdgpu_sa_bo, olist);
list_for_each_entry_safe_from(sa_bo, tmp, &sa_manager->olist, olist) { list_for_each_entry_safe_from(sa_bo, tmp, &sa_manager->olist, olist) {
if (sa_bo->fence == NULL || if (sa_bo->fence == NULL ||
!fence_is_signaled(&sa_bo->fence->base)) { !fence_is_signaled(sa_bo->fence)) {
return; return;
} }
amdgpu_sa_bo_remove_locked(sa_bo); amdgpu_sa_bo_remove_locked(sa_bo);
...@@ -246,7 +260,7 @@ static bool amdgpu_sa_event(struct amdgpu_sa_manager *sa_manager, ...@@ -246,7 +260,7 @@ static bool amdgpu_sa_event(struct amdgpu_sa_manager *sa_manager,
} }
static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager, static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager,
struct amdgpu_fence **fences, struct fence **fences,
unsigned *tries) unsigned *tries)
{ {
struct amdgpu_sa_bo *best_bo = NULL; struct amdgpu_sa_bo *best_bo = NULL;
...@@ -275,7 +289,7 @@ static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager, ...@@ -275,7 +289,7 @@ static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager,
sa_bo = list_first_entry(&sa_manager->flist[i], sa_bo = list_first_entry(&sa_manager->flist[i],
struct amdgpu_sa_bo, flist); struct amdgpu_sa_bo, flist);
if (!fence_is_signaled(&sa_bo->fence->base)) { if (!fence_is_signaled(sa_bo->fence)) {
fences[i] = sa_bo->fence; fences[i] = sa_bo->fence;
continue; continue;
} }
...@@ -299,7 +313,8 @@ static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager, ...@@ -299,7 +313,8 @@ static bool amdgpu_sa_bo_next_hole(struct amdgpu_sa_manager *sa_manager,
} }
if (best_bo) { if (best_bo) {
++tries[best_bo->fence->ring->idx]; uint32_t idx = amdgpu_sa_get_ring_from_fence(best_bo->fence);
++tries[idx];
sa_manager->hole = best_bo->olist.prev; sa_manager->hole = best_bo->olist.prev;
/* we knew that this one is signaled, /* we knew that this one is signaled,
...@@ -315,7 +330,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev, ...@@ -315,7 +330,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev,
struct amdgpu_sa_bo **sa_bo, struct amdgpu_sa_bo **sa_bo,
unsigned size, unsigned align) unsigned size, unsigned align)
{ {
struct amdgpu_fence *fences[AMDGPU_MAX_RINGS]; struct fence *fences[AMDGPU_MAX_RINGS];
unsigned tries[AMDGPU_MAX_RINGS]; unsigned tries[AMDGPU_MAX_RINGS];
int i, r; int i, r;
signed long t; signed long t;
...@@ -373,7 +388,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev, ...@@ -373,7 +388,7 @@ int amdgpu_sa_bo_new(struct amdgpu_device *adev,
} }
void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo, void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo,
struct amdgpu_fence *fence) struct fence *fence)
{ {
struct amdgpu_sa_manager *sa_manager; struct amdgpu_sa_manager *sa_manager;
...@@ -383,10 +398,11 @@ void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo, ...@@ -383,10 +398,11 @@ void amdgpu_sa_bo_free(struct amdgpu_device *adev, struct amdgpu_sa_bo **sa_bo,
sa_manager = (*sa_bo)->manager; sa_manager = (*sa_bo)->manager;
spin_lock(&sa_manager->wq.lock); spin_lock(&sa_manager->wq.lock);
if (fence && !fence_is_signaled(&fence->base)) { if (fence && !fence_is_signaled(fence)) {
(*sa_bo)->fence = amdgpu_fence_ref(fence); uint32_t idx;
list_add_tail(&(*sa_bo)->flist, (*sa_bo)->fence = fence_get(fence);
&sa_manager->flist[fence->ring->idx]); idx = amdgpu_sa_get_ring_from_fence(fence);
list_add_tail(&(*sa_bo)->flist, &sa_manager->flist[idx]);
} else { } else {
amdgpu_sa_bo_remove_locked(*sa_bo); amdgpu_sa_bo_remove_locked(*sa_bo);
} }
...@@ -413,8 +429,16 @@ void amdgpu_sa_bo_dump_debug_info(struct amdgpu_sa_manager *sa_manager, ...@@ -413,8 +429,16 @@ void amdgpu_sa_bo_dump_debug_info(struct amdgpu_sa_manager *sa_manager,
seq_printf(m, "[0x%010llx 0x%010llx] size %8lld", seq_printf(m, "[0x%010llx 0x%010llx] size %8lld",
soffset, eoffset, eoffset - soffset); soffset, eoffset, eoffset - soffset);
if (i->fence) { if (i->fence) {
seq_printf(m, " protected by 0x%016llx on ring %d", struct amdgpu_fence *a_fence = to_amdgpu_fence(i->fence);
i->fence->seq, i->fence->ring->idx); struct amd_sched_fence *s_fence = to_amd_sched_fence(i->fence);
if (a_fence)
seq_printf(m, " protected by 0x%016llx on ring %d",
a_fence->seq, a_fence->ring->idx);
if (s_fence)
seq_printf(m, " protected by 0x%016llx on ring %d",
s_fence->v_seq,
s_fence->entity->scheduler->ring_id);
} }
seq_printf(m, "\n"); seq_printf(m, "\n");
} }
......
...@@ -87,7 +87,7 @@ bool amdgpu_semaphore_emit_wait(struct amdgpu_ring *ring, ...@@ -87,7 +87,7 @@ bool amdgpu_semaphore_emit_wait(struct amdgpu_ring *ring,
void amdgpu_semaphore_free(struct amdgpu_device *adev, void amdgpu_semaphore_free(struct amdgpu_device *adev,
struct amdgpu_semaphore **semaphore, struct amdgpu_semaphore **semaphore,
struct amdgpu_fence *fence) struct fence *fence)
{ {
if (semaphore == NULL || *semaphore == NULL) { if (semaphore == NULL || *semaphore == NULL) {
return; return;
......
...@@ -234,7 +234,7 @@ int amdgpu_sync_rings(struct amdgpu_sync *sync, ...@@ -234,7 +234,7 @@ int amdgpu_sync_rings(struct amdgpu_sync *sync,
*/ */
void amdgpu_sync_free(struct amdgpu_device *adev, void amdgpu_sync_free(struct amdgpu_device *adev,
struct amdgpu_sync *sync, struct amdgpu_sync *sync,
struct amdgpu_fence *fence) struct fence *fence)
{ {
unsigned i; unsigned i;
......
...@@ -1042,7 +1042,7 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring, ...@@ -1042,7 +1042,7 @@ int amdgpu_copy_buffer(struct amdgpu_ring *ring,
} }
amdgpu_ring_unlock_commit(ring); amdgpu_ring_unlock_commit(ring);
amdgpu_sync_free(adev, &sync, *fence); amdgpu_sync_free(adev, &sync, &(*fence)->base);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment