Commit b81e4d9b authored by Chris Wilson's avatar Chris Wilson

drm/i915/gt: Track engine round-trip times

Knowing the round trip time of an engine is useful for tracking the
health of the system as well as providing a metric for the baseline
responsiveness of the engine. We can use the latter metric for
automatically tuning our waits in selftests and when idling so we don't
confuse a slower system with a dead one.

Upon idling the engine, we send one last pulse to switch the context
away from precious user state to the volatile kernel context. We know
the engine is idle at this point, and the pulse is non-preemptible, so
this provides us with a good measurement of the round trip time. It also
provides us with faster engine parking for ringbuffer submission, which
is a welcome bonus (e.g. softer-rc6).
Signed-off-by: default avatarChris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Cc: Stuart Summers <stuart.summers@intel.com>
Reviewed-by: default avatarJoonas Lahtinen <joonas.lahtinen@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20191219105043.4169050-1-chris@chris-wilson.co.uk
Link: https://patchwork.freedesktop.org/patch/msgid/20191219124353.8607-2-chris@chris-wilson.co.uk
parent dc93c9b6
...@@ -334,6 +334,7 @@ static int intel_engine_setup(struct intel_gt *gt, enum intel_engine_id id) ...@@ -334,6 +334,7 @@ static int intel_engine_setup(struct intel_gt *gt, enum intel_engine_id id)
/* Nothing to do here, execute in order of dependencies */ /* Nothing to do here, execute in order of dependencies */
engine->schedule = NULL; engine->schedule = NULL;
ewma__engine_latency_init(&engine->latency);
seqlock_init(&engine->stats.lock); seqlock_init(&engine->stats.lock);
ATOMIC_INIT_NOTIFIER_HEAD(&engine->context_status_notifier); ATOMIC_INIT_NOTIFIER_HEAD(&engine->context_status_notifier);
...@@ -1481,6 +1482,8 @@ void intel_engine_dump(struct intel_engine_cs *engine, ...@@ -1481,6 +1482,8 @@ void intel_engine_dump(struct intel_engine_cs *engine,
drm_printf(m, "\tAwake? %d\n", atomic_read(&engine->wakeref.count)); drm_printf(m, "\tAwake? %d\n", atomic_read(&engine->wakeref.count));
drm_printf(m, "\tBarriers?: %s\n", drm_printf(m, "\tBarriers?: %s\n",
yesno(!llist_empty(&engine->barrier_tasks))); yesno(!llist_empty(&engine->barrier_tasks)));
drm_printf(m, "\tLatency: %luus\n",
ewma__engine_latency_read(&engine->latency));
rcu_read_lock(); rcu_read_lock();
rq = READ_ONCE(engine->heartbeat.systole); rq = READ_ONCE(engine->heartbeat.systole);
......
...@@ -73,6 +73,15 @@ static inline void __timeline_mark_unlock(struct intel_context *ce, ...@@ -73,6 +73,15 @@ static inline void __timeline_mark_unlock(struct intel_context *ce,
#endif /* !IS_ENABLED(CONFIG_LOCKDEP) */ #endif /* !IS_ENABLED(CONFIG_LOCKDEP) */
static void duration(struct dma_fence *fence, struct dma_fence_cb *cb)
{
struct i915_request *rq = to_request(fence);
ewma__engine_latency_add(&rq->engine->latency,
ktime_us_delta(rq->fence.timestamp,
rq->duration.emitted));
}
static void static void
__queue_and_release_pm(struct i915_request *rq, __queue_and_release_pm(struct i915_request *rq,
struct intel_timeline *tl, struct intel_timeline *tl,
...@@ -163,7 +172,18 @@ static bool switch_to_kernel_context(struct intel_engine_cs *engine) ...@@ -163,7 +172,18 @@ static bool switch_to_kernel_context(struct intel_engine_cs *engine)
/* Install ourselves as a preemption barrier */ /* Install ourselves as a preemption barrier */
rq->sched.attr.priority = I915_PRIORITY_BARRIER; rq->sched.attr.priority = I915_PRIORITY_BARRIER;
__i915_request_commit(rq); if (likely(!__i915_request_commit(rq))) { /* engine should be idle! */
/*
* Use an interrupt for precise measurement of duration,
* otherwise we rely on someone else retiring all the requests
* which may delay the signaling (i.e. we will likely wait
* until the background request retirement running every
* second or two).
*/
BUILD_BUG_ON(sizeof(rq->duration) > sizeof(rq->submitq));
dma_fence_add_callback(&rq->fence, &rq->duration.cb, duration);
rq->duration.emitted = ktime_get();
}
/* Expose ourselves to the world */ /* Expose ourselves to the world */
__queue_and_release_pm(rq, ce->timeline, engine); __queue_and_release_pm(rq, ce->timeline, engine);
......
...@@ -7,6 +7,7 @@ ...@@ -7,6 +7,7 @@
#ifndef __INTEL_ENGINE_TYPES__ #ifndef __INTEL_ENGINE_TYPES__
#define __INTEL_ENGINE_TYPES__ #define __INTEL_ENGINE_TYPES__
#include <linux/average.h>
#include <linux/hashtable.h> #include <linux/hashtable.h>
#include <linux/irq_work.h> #include <linux/irq_work.h>
#include <linux/kref.h> #include <linux/kref.h>
...@@ -119,6 +120,9 @@ enum intel_engine_id { ...@@ -119,6 +120,9 @@ enum intel_engine_id {
#define INVALID_ENGINE ((enum intel_engine_id)-1) #define INVALID_ENGINE ((enum intel_engine_id)-1)
}; };
/* A simple estimator for the round-trip latency of an engine */
DECLARE_EWMA(_engine_latency, 6, 4)
struct st_preempt_hang { struct st_preempt_hang {
struct completion completion; struct completion completion;
unsigned int count; unsigned int count;
...@@ -316,6 +320,13 @@ struct intel_engine_cs { ...@@ -316,6 +320,13 @@ struct intel_engine_cs {
struct intel_timeline *timeline; struct intel_timeline *timeline;
} legacy; } legacy;
/*
* We track the average duration of the idle pulse on parking the
* engine to keep an estimate of the how the fast the engine is
* under ideal conditions.
*/
struct ewma__engine_latency latency;
/* Rather than have every client wait upon all user interrupts, /* Rather than have every client wait upon all user interrupts,
* with the herd waking after every interrupt and each doing the * with the herd waking after every interrupt and each doing the
* heavyweight seqno dance, we delegate the task (of being the * heavyweight seqno dance, we delegate the task (of being the
......
...@@ -150,6 +150,10 @@ struct i915_request { ...@@ -150,6 +150,10 @@ struct i915_request {
union { union {
wait_queue_entry_t submitq; wait_queue_entry_t submitq;
struct i915_sw_dma_fence_cb dmaq; struct i915_sw_dma_fence_cb dmaq;
struct i915_request_duration_cb {
struct dma_fence_cb cb;
ktime_t emitted;
} duration;
}; };
struct list_head execute_cb; struct list_head execute_cb;
struct i915_sw_fence semaphore; struct i915_sw_fence semaphore;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment