summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorBen Widawsky <benjamin.widawsky@intel.com>2014-04-03 09:28:21 -0700
committerBen Widawsky <benjamin.widawsky@intel.com>2014-06-20 17:05:01 -0700
commitcc307bbeca7b7645c6528469e8b7ac9e1784abc5 (patch)
treefe028616674166085cbe16b5498813ca12079908
parent543b8de52b354e7e423e637d32f9af702e951791 (diff)
drm/i915/bdw: Add dynamic page trace events
This works the same as GEN6. I was disappointed that I need to pass vm around now, but it's not so much uglier than the drm_device, and having the vm in trace events is hugely important. v2: Consolidate pagetable/pagedirectory events Signed-off-by: Ben Widawsky <ben@bwidawsk.net>
-rw-r--r--drivers/gpu/drm/i915/i915_gem_gtt.c41
-rw-r--r--drivers/gpu/drm/i915/i915_trace.h16
2 files changed, 47 insertions, 10 deletions
diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c
index 65900926de3f..7e52696c7101 100644
--- a/drivers/gpu/drm/i915/i915_gem_gtt.c
+++ b/drivers/gpu/drm/i915/i915_gem_gtt.c
@@ -565,19 +565,24 @@ static void __gen8_do_map_pt(gen8_ppgtt_pde_t *pde,
/* It's likely we'll map more than one pagetable at a time. This function will
* save us unnecessary kmap calls, but do no more functionally than multiple
* calls to map_pt. */
-static void gen8_map_pagetable_range(struct i915_pagedir *pd,
+static void gen8_map_pagetable_range(struct i915_address_space *vm,
+ struct i915_pagedir *pd,
uint64_t start,
- uint64_t length,
- struct drm_device *dev)
+ uint64_t length)
{
gen8_ppgtt_pde_t *pagedir = kmap_atomic(pd->page);
struct i915_pagetab *pt;
uint64_t temp, pde;
- gen8_for_each_pde(pt, pd, start, length, temp, pde)
- __gen8_do_map_pt(pagedir + pde, pt, dev);
+ gen8_for_each_pde(pt, pd, start, length, temp, pde) {
+ __gen8_do_map_pt(pagedir + pde, pt, vm->dev);
+ trace_i915_pagetable_map(vm, pde, pt,
+ gen8_pte_index(start),
+ gen8_pte_count(start, length),
+ GEN8_PTES_PER_PT);
+ }
- if (!HAS_LLC(dev))
+ if (!HAS_LLC(vm->dev))
drm_clflush_virt_range(pagedir, PAGE_SIZE);
kunmap_atomic(pagedir);
@@ -641,11 +646,20 @@ static void gen8_teardown_va_range(struct i915_address_space *vm,
"PDE %d not reserved, but is allocated (%p)",
pde, vm);
+ trace_i915_pagetable_unmap(vm, pde, pt,
+ gen8_pte_index(pd_start),
+ gen8_pte_count(pd_start, pd_len),
+ GEN8_PTES_PER_PT);
+
bitmap_clear(pt->used_ptes,
gen8_pte_index(pd_start),
gen8_pte_count(pd_start, pd_len));
if (bitmap_empty(pt->used_ptes, GEN8_PTES_PER_PT)) {
+ trace_i915_pagetable_destroy(vm,
+ pde,
+ pd_start & GENMASK_ULL(64, GEN8_PDE_SHIFT),
+ GEN8_PDE_SHIFT);
free_pt_single(pt, vm->dev);
/* This may be nixed later. Optimize? */
gen8_unmap_pagetable(ppgtt, pd, pde);
@@ -657,6 +671,9 @@ static void gen8_teardown_va_range(struct i915_address_space *vm,
if (bitmap_empty(pd->used_pdes, I915_PDES_PER_PD)) {
free_pd_single(pd, vm->dev);
ppgtt->pdp.pagedirs[pdpe] = NULL;
+ trace_i915_pagedirectory_destroy(vm, pdpe,
+ start & GENMASK_ULL(64, GEN8_PDPE_SHIFT),
+ GEN8_PDPE_SHIFT);
WARN_ON(!test_and_clear_bit(pdpe, ppgtt->pdp.used_pdpes));
}
}
@@ -705,6 +722,7 @@ static int gen8_ppgtt_alloc_pagetabs(struct i915_hw_ppgtt *ppgtt,
uint64_t length,
unsigned long *new_pts)
{
+ struct drm_device *dev = ppgtt->base.dev;
struct i915_pagetab *unused;
uint64_t temp;
uint32_t pde;
@@ -713,19 +731,20 @@ static int gen8_ppgtt_alloc_pagetabs(struct i915_hw_ppgtt *ppgtt,
if (unused)
continue;
- pd->page_tables[pde] = alloc_pt_single(ppgtt->base.dev);
+ pd->page_tables[pde] = alloc_pt_single(dev);
if (IS_ERR(pd->page_tables[pde]))
goto unwind_out;
set_bit(pde, new_pts);
+ trace_i915_pagetable_alloc(&ppgtt->base, pde, start, GEN8_PDE_SHIFT);
}
return 0;
unwind_out:
for_each_set_bit(pde, new_pts, I915_PDES_PER_PD)
- free_pt_single(pd->page_tables[pde], ppgtt->base.dev);
+ free_pt_single(pd->page_tables[pde], dev);
return -ENOMEM;
}
@@ -779,6 +798,8 @@ static int gen8_ppgtt_alloc_pagedirs(struct i915_hw_ppgtt *ppgtt,
pdp->pagedirs[pdpe] = pd;
set_bit(pdpe, new_pds);
+ trace_i915_pagedirectory_alloc(&ppgtt->base, pdpe, start,
+ GEN8_PDPE_SHIFT);
}
return 0;
@@ -909,7 +930,7 @@ static int gen8_alloc_va_range(struct i915_address_space *vm,
set_bit(pdpe, ppgtt->pdp.used_pdpes);
- gen8_map_pagetable_range(pd, start, length, ppgtt->base.dev);
+ gen8_map_pagetable_range(vm, pd, start, length);
}
free_gen8_temp_bitmaps(new_page_dirs, new_page_tables);
@@ -971,7 +992,7 @@ static int gen8_aliasing_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
}
gen8_for_each_pdpe(pd, &ppgtt->pdp, start, size, temp, pdpe)
- gen8_map_pagetable_range(pd, start, size, ppgtt->base.dev);
+ gen8_map_pagetable_range(&ppgtt->base, pd, start, size);
ppgtt->base.allocate_va_range = NULL;
ppgtt->base.teardown_va_range = NULL;
diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h
index 729d1198b1bd..7f54cd0f6c30 100644
--- a/drivers/gpu/drm/i915/i915_trace.h
+++ b/drivers/gpu/drm/i915/i915_trace.h
@@ -217,6 +217,22 @@ DEFINE_EVENT(i915_pagetable, i915_pagetable_destroy,
TP_ARGS(vm, pde, start, pde_shift)
);
+DEFINE_EVENT_PRINT(i915_pagetable, i915_pagedirectory_alloc,
+ TP_PROTO(struct i915_address_space *vm, u32 pdpe, u64 start, u64 pdpe_shift),
+ TP_ARGS(vm, pdpe, start, pdpe_shift),
+
+ TP_printk("vm=%p, pdpe=%d (0x%llx-0x%llx)",
+ __entry->vm, __entry->pde, __entry->start, __entry->end)
+);
+
+DEFINE_EVENT_PRINT(i915_pagetable, i915_pagedirectory_destroy,
+ TP_PROTO(struct i915_address_space *vm, u32 pdpe, u64 start, u64 pdpe_shift),
+ TP_ARGS(vm, pdpe, start, pdpe_shift),
+
+ TP_printk("vm=%p, pdpe=%d (0x%llx-0x%llx)",
+ __entry->vm, __entry->pde, __entry->start, __entry->end)
+);
+
/* Avoid extra math because we only support two sizes. The format is defined by
* bitmap_scnprintf. Each 32 bits is 8 HEX digits followed by comma */
#define TRACE_PT_SIZE(bits) \