Commit f6e8aa38 authored by Chris Wilson's avatar Chris Wilson

drm/i915: Report the number of closed vma held by each context in debugfs

Include the total size of closed vma when reporting the per_ctx_stats of
debugfs/i915_gem_objects.

Whilst adjusting the context tracking, note that we can simply use our
list of contexts in i915->contexts rather than circumlocute via
dev->filelist and the per-file context idr, with the result that we can
show objects allocated to different vm (i.e. contexts within a file).

We change the output to show every context of each client, with its own
unique set of objects (for full-ppgtt machines, i.e. gen7+, for older
hardware all objects are in the global gtt and so can not be associated
with a single context). That should result in no loss of information,
and for gen7+, no duplication of active objects.
Signed-off-by: default avatarChris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: default avatarMika Kuoppala <mika.kuoppala@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190107115509.12523-2-chris@chris-wilson.co.uk
parent e4fc69f2
...@@ -297,11 +297,12 @@ static int i915_gem_stolen_list_info(struct seq_file *m, void *data) ...@@ -297,11 +297,12 @@ static int i915_gem_stolen_list_info(struct seq_file *m, void *data)
} }
struct file_stats { struct file_stats {
struct drm_i915_file_private *file_priv; struct i915_address_space *vm;
unsigned long count; unsigned long count;
u64 total, unbound; u64 total, unbound;
u64 global, shared; u64 global, shared;
u64 active, inactive; u64 active, inactive;
u64 closed;
}; };
static int per_file_stats(int id, void *ptr, void *data) static int per_file_stats(int id, void *ptr, void *data)
...@@ -326,9 +327,7 @@ static int per_file_stats(int id, void *ptr, void *data) ...@@ -326,9 +327,7 @@ static int per_file_stats(int id, void *ptr, void *data)
if (i915_vma_is_ggtt(vma)) { if (i915_vma_is_ggtt(vma)) {
stats->global += vma->node.size; stats->global += vma->node.size;
} else { } else {
struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vma->vm); if (vma->vm != stats->vm)
if (ppgtt->vm.file != stats->file_priv)
continue; continue;
} }
...@@ -336,6 +335,9 @@ static int per_file_stats(int id, void *ptr, void *data) ...@@ -336,6 +335,9 @@ static int per_file_stats(int id, void *ptr, void *data)
stats->active += vma->node.size; stats->active += vma->node.size;
else else
stats->inactive += vma->node.size; stats->inactive += vma->node.size;
if (i915_vma_is_closed(vma))
stats->closed += vma->node.size;
} }
return 0; return 0;
...@@ -343,7 +345,7 @@ static int per_file_stats(int id, void *ptr, void *data) ...@@ -343,7 +345,7 @@ static int per_file_stats(int id, void *ptr, void *data)
#define print_file_stats(m, name, stats) do { \ #define print_file_stats(m, name, stats) do { \
if (stats.count) \ if (stats.count) \
seq_printf(m, "%s: %lu objects, %llu bytes (%llu active, %llu inactive, %llu global, %llu shared, %llu unbound)\n", \ seq_printf(m, "%s: %lu objects, %llu bytes (%llu active, %llu inactive, %llu global, %llu shared, %llu unbound, %llu closed)\n", \
name, \ name, \
stats.count, \ stats.count, \
stats.total, \ stats.total, \
...@@ -351,20 +353,19 @@ static int per_file_stats(int id, void *ptr, void *data) ...@@ -351,20 +353,19 @@ static int per_file_stats(int id, void *ptr, void *data)
stats.inactive, \ stats.inactive, \
stats.global, \ stats.global, \
stats.shared, \ stats.shared, \
stats.unbound); \ stats.unbound, \
stats.closed); \
} while (0) } while (0)
static void print_batch_pool_stats(struct seq_file *m, static void print_batch_pool_stats(struct seq_file *m,
struct drm_i915_private *dev_priv) struct drm_i915_private *dev_priv)
{ {
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
struct file_stats stats;
struct intel_engine_cs *engine; struct intel_engine_cs *engine;
struct file_stats stats = {};
enum intel_engine_id id; enum intel_engine_id id;
int j; int j;
memset(&stats, 0, sizeof(stats));
for_each_engine(engine, dev_priv, id) { for_each_engine(engine, dev_priv, id) {
for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) { for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) {
list_for_each_entry(obj, list_for_each_entry(obj,
...@@ -377,44 +378,47 @@ static void print_batch_pool_stats(struct seq_file *m, ...@@ -377,44 +378,47 @@ static void print_batch_pool_stats(struct seq_file *m,
print_file_stats(m, "[k]batch pool", stats); print_file_stats(m, "[k]batch pool", stats);
} }
static int per_file_ctx_stats(int idx, void *ptr, void *data) static void print_context_stats(struct seq_file *m,
struct drm_i915_private *i915)
{ {
struct i915_gem_context *ctx = ptr; struct file_stats kstats = {};
struct i915_gem_context *ctx;
list_for_each_entry(ctx, &i915->contexts.list, link) {
struct intel_engine_cs *engine; struct intel_engine_cs *engine;
enum intel_engine_id id; enum intel_engine_id id;
for_each_engine(engine, ctx->i915, id) { for_each_engine(engine, i915, id) {
struct intel_context *ce = to_intel_context(ctx, engine); struct intel_context *ce = to_intel_context(ctx, engine);
if (ce->state) if (ce->state)
per_file_stats(0, ce->state->obj, data); per_file_stats(0, ce->state->obj, &kstats);
if (ce->ring) if (ce->ring)
per_file_stats(0, ce->ring->vma->obj, data); per_file_stats(0, ce->ring->vma->obj, &kstats);
} }
return 0; if (!IS_ERR_OR_NULL(ctx->file_priv)) {
} struct file_stats stats = { .vm = &ctx->ppgtt->vm, };
struct drm_file *file = ctx->file_priv->file;
static void print_context_stats(struct seq_file *m, struct task_struct *task;
struct drm_i915_private *dev_priv) char name[80];
{
struct drm_device *dev = &dev_priv->drm;
struct file_stats stats;
struct drm_file *file;
memset(&stats, 0, sizeof(stats)); spin_lock(&file->table_lock);
idr_for_each(&file->object_idr, per_file_stats, &stats);
spin_unlock(&file->table_lock);
mutex_lock(&dev->struct_mutex); rcu_read_lock();
if (dev_priv->kernel_context) task = pid_task(ctx->pid ?: file->pid, PIDTYPE_PID);
per_file_ctx_stats(0, dev_priv->kernel_context, &stats); snprintf(name, sizeof(name), "%s/%d",
task ? task->comm : "<unknown>",
ctx->user_handle);
rcu_read_unlock();
list_for_each_entry(file, &dev->filelist, lhead) { print_file_stats(m, name, stats);
struct drm_i915_file_private *fpriv = file->driver_priv; }
idr_for_each(&fpriv->context_idr, per_file_ctx_stats, &stats);
} }
mutex_unlock(&dev->struct_mutex);
print_file_stats(m, "[k]contexts", stats); print_file_stats(m, "[k]contexts", kstats);
} }
static int i915_gem_object_info(struct seq_file *m, void *data) static int i915_gem_object_info(struct seq_file *m, void *data)
...@@ -426,14 +430,9 @@ static int i915_gem_object_info(struct seq_file *m, void *data) ...@@ -426,14 +430,9 @@ static int i915_gem_object_info(struct seq_file *m, void *data)
u64 size, mapped_size, purgeable_size, dpy_size, huge_size; u64 size, mapped_size, purgeable_size, dpy_size, huge_size;
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
unsigned int page_sizes = 0; unsigned int page_sizes = 0;
struct drm_file *file;
char buf[80]; char buf[80];
int ret; int ret;
ret = mutex_lock_interruptible(&dev->struct_mutex);
if (ret)
return ret;
seq_printf(m, "%u objects, %llu bytes\n", seq_printf(m, "%u objects, %llu bytes\n",
dev_priv->mm.object_count, dev_priv->mm.object_count,
dev_priv->mm.object_memory); dev_priv->mm.object_memory);
...@@ -514,43 +513,14 @@ static int i915_gem_object_info(struct seq_file *m, void *data) ...@@ -514,43 +513,14 @@ static int i915_gem_object_info(struct seq_file *m, void *data)
buf, sizeof(buf))); buf, sizeof(buf)));
seq_putc(m, '\n'); seq_putc(m, '\n');
print_batch_pool_stats(m, dev_priv);
mutex_unlock(&dev->struct_mutex);
mutex_lock(&dev->filelist_mutex); ret = mutex_lock_interruptible(&dev->struct_mutex);
print_context_stats(m, dev_priv); if (ret)
list_for_each_entry_reverse(file, &dev->filelist, lhead) { return ret;
struct file_stats stats;
struct drm_i915_file_private *file_priv = file->driver_priv;
struct i915_request *request;
struct task_struct *task;
mutex_lock(&dev->struct_mutex);
memset(&stats, 0, sizeof(stats));
stats.file_priv = file->driver_priv;
spin_lock(&file->table_lock);
idr_for_each(&file->object_idr, per_file_stats, &stats);
spin_unlock(&file->table_lock);
/*
* Although we have a valid reference on file->pid, that does
* not guarantee that the task_struct who called get_pid() is
* still alive (e.g. get_pid(current) => fork() => exit()).
* Therefore, we need to protect this ->comm access using RCU.
*/
request = list_first_entry_or_null(&file_priv->mm.request_list,
struct i915_request,
client_link);
rcu_read_lock();
task = pid_task(request && request->gem_context->pid ?
request->gem_context->pid : file->pid,
PIDTYPE_PID);
print_file_stats(m, task ? task->comm : "<unknown>", stats);
rcu_read_unlock();
print_batch_pool_stats(m, dev_priv);
print_context_stats(m, dev_priv);
mutex_unlock(&dev->struct_mutex); mutex_unlock(&dev->struct_mutex);
}
mutex_unlock(&dev->filelist_mutex);
return 0; return 0;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment