Commit c5d092a4 authored by Chris Wilson's avatar Chris Wilson

drm/i915: Remove bitmap tracking for used-pml4

We only operate on known extents (both for alloc/clear) and so we can use
both the knowledge of the bind/unbind range along with the knowledge of
the existing pagetable to avoid having to allocate temporary and
auxiliary bitmaps.
Signed-off-by: default avatarChris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: default avatarMatthew Auld <matthew.auld@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/20170215084357.19977-15-chris@chris-wilson.co.uk
parent e2b763ca
...@@ -806,7 +806,6 @@ static void gen8_ppgtt_clear_4lvl(struct i915_address_space *vm, ...@@ -806,7 +806,6 @@ static void gen8_ppgtt_clear_4lvl(struct i915_address_space *vm,
continue; continue;
gen8_ppgtt_set_pml4e(pml4, vm->scratch_pdp, pml4e); gen8_ppgtt_set_pml4e(pml4, vm->scratch_pdp, pml4e);
__clear_bit(pml4e, pml4->used_pml4es);
free_pdp(vm, pdp); free_pdp(vm, pdp);
} }
...@@ -1027,8 +1026,8 @@ static void gen8_ppgtt_cleanup_4lvl(struct i915_hw_ppgtt *ppgtt) ...@@ -1027,8 +1026,8 @@ static void gen8_ppgtt_cleanup_4lvl(struct i915_hw_ppgtt *ppgtt)
{ {
int i; int i;
for_each_set_bit(i, ppgtt->pml4.used_pml4es, GEN8_PML4ES_PER_PML4) { for (i = 0; i < GEN8_PML4ES_PER_PML4; i++) {
if (WARN_ON(!ppgtt->pml4.pdps[i])) if (ppgtt->pml4.pdps[i] == ppgtt->base.scratch_pdp)
continue; continue;
gen8_ppgtt_cleanup_3lvl(&ppgtt->base, ppgtt->pml4.pdps[i]); gen8_ppgtt_cleanup_3lvl(&ppgtt->base, ppgtt->pml4.pdps[i]);
...@@ -1082,53 +1081,9 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm, ...@@ -1082,53 +1081,9 @@ static int gen8_ppgtt_alloc_pd(struct i915_address_space *vm,
return -ENOMEM; return -ENOMEM;
} }
/** static int gen8_ppgtt_alloc_pdp(struct i915_address_space *vm,
* gen8_ppgtt_alloc_page_dirpointers() - Allocate pdps for VA range. struct i915_page_directory_pointer *pdp,
* @vm: Master vm structure. u64 start, u64 length)
* @pml4: Page map level 4 for this address range.
* @start: Starting virtual address to begin allocations.
* @length: Size of the allocations.
* @new_pdps: Bitmap set by function with new allocations. Likely used by the
* caller to free on error.
*
* Allocate the required number of page directory pointers. Extremely similar to
* gen8_ppgtt_alloc_page_directories() and gen8_ppgtt_alloc_pd().
* The main difference is here we are limited by the pml4 boundary (instead of
* the page directory pointer).
*
* Return: 0 if success; negative error code otherwise.
*/
static int
gen8_ppgtt_alloc_page_dirpointers(struct i915_address_space *vm,
struct i915_pml4 *pml4,
uint64_t start,
uint64_t length)
{
struct i915_page_directory_pointer *pdp;
uint32_t pml4e;
gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) {
if (!test_bit(pml4e, pml4->used_pml4es)) {
pdp = alloc_pdp(vm);
if (IS_ERR(pdp))
return PTR_ERR(pdp);
gen8_initialize_pdp(vm, pdp);
pml4->pdps[pml4e] = pdp;
trace_i915_page_directory_pointer_entry_alloc(vm,
pml4e,
start,
GEN8_PML4E_SHIFT);
}
}
return 0;
}
static int gen8_alloc_va_range_3lvl(struct i915_address_space *vm,
struct i915_page_directory_pointer *pdp,
uint64_t start,
uint64_t length)
{ {
struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vm);
struct i915_page_directory *pd; struct i915_page_directory *pd;
...@@ -1164,58 +1119,46 @@ static int gen8_alloc_va_range_3lvl(struct i915_address_space *vm, ...@@ -1164,58 +1119,46 @@ static int gen8_alloc_va_range_3lvl(struct i915_address_space *vm,
return -ENOMEM; return -ENOMEM;
} }
static int gen8_alloc_va_range_4lvl(struct i915_address_space *vm, static int gen8_ppgtt_alloc_3lvl(struct i915_address_space *vm,
struct i915_pml4 *pml4, u64 start, u64 length)
uint64_t start,
uint64_t length)
{ {
DECLARE_BITMAP(new_pdps, GEN8_PML4ES_PER_PML4); return gen8_ppgtt_alloc_pdp(vm,
struct i915_page_directory_pointer *pdp; &i915_vm_to_ppgtt(vm)->pdp, start, length);
uint64_t pml4e; }
int ret = 0;
/* Do the pml4 allocations first, so we don't need to track the newly
* allocated tables below the pdp */
bitmap_zero(new_pdps, GEN8_PML4ES_PER_PML4);
/* The pagedirectory and pagetable allocations are done in the shared 3 static int gen8_ppgtt_alloc_4lvl(struct i915_address_space *vm,
* and 4 level code. Just allocate the pdps. u64 start, u64 length)
*/ {
ret = gen8_ppgtt_alloc_page_dirpointers(vm, pml4, start, length); struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vm);
if (ret) struct i915_pml4 *pml4 = &ppgtt->pml4;
return ret; struct i915_page_directory_pointer *pdp;
u64 from = start;
u32 pml4e;
int ret;
gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) {
WARN_ON(!pdp); if (pml4->pdps[pml4e] == vm->scratch_pdp) {
pdp = alloc_pdp(vm);
if (IS_ERR(pdp))
goto unwind;
ret = gen8_alloc_va_range_3lvl(vm, pdp, start, length); gen8_initialize_pdp(vm, pdp);
if (ret) gen8_ppgtt_set_pml4e(pml4, pdp, pml4e);
goto err_out; }
gen8_ppgtt_set_pml4e(pml4, pdp, pml4e); ret = gen8_ppgtt_alloc_pdp(vm, pdp, start, length);
if (unlikely(ret)) {
gen8_ppgtt_set_pml4e(pml4, vm->scratch_pdp, pml4e);
free_pdp(vm, pdp);
goto unwind;
}
} }
bitmap_or(pml4->used_pml4es, new_pdps, pml4->used_pml4es,
GEN8_PML4ES_PER_PML4);
return 0; return 0;
err_out: unwind:
for_each_set_bit(pml4e, new_pdps, GEN8_PML4ES_PER_PML4) gen8_ppgtt_clear_4lvl(vm, from, start - from);
gen8_ppgtt_cleanup_3lvl(vm, pml4->pdps[pml4e]); return -ENOMEM;
return ret;
}
static int gen8_alloc_va_range(struct i915_address_space *vm,
uint64_t start, uint64_t length)
{
struct i915_hw_ppgtt *ppgtt = i915_vm_to_ppgtt(vm);
if (USES_FULL_48BIT_PPGTT(vm->i915))
return gen8_alloc_va_range_4lvl(vm, &ppgtt->pml4, start, length);
else
return gen8_alloc_va_range_3lvl(vm, &ppgtt->pdp, start, length);
} }
static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt, static void gen8_dump_pdp(struct i915_hw_ppgtt *ppgtt,
...@@ -1289,7 +1232,7 @@ static void gen8_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m) ...@@ -1289,7 +1232,7 @@ static void gen8_dump_ppgtt(struct i915_hw_ppgtt *ppgtt, struct seq_file *m)
struct i915_page_directory_pointer *pdp; struct i915_page_directory_pointer *pdp;
gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) { gen8_for_each_pml4e(pdp, pml4, start, length, pml4e) {
if (!test_bit(pml4e, pml4->used_pml4es)) if (pml4->pdps[pml4e] == ppgtt->base.scratch_pdp)
continue; continue;
seq_printf(m, " PML4E #%llu\n", pml4e); seq_printf(m, " PML4E #%llu\n", pml4e);
...@@ -1348,7 +1291,6 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt) ...@@ -1348,7 +1291,6 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
ppgtt->base.start = 0; ppgtt->base.start = 0;
ppgtt->base.cleanup = gen8_ppgtt_cleanup; ppgtt->base.cleanup = gen8_ppgtt_cleanup;
ppgtt->base.allocate_va_range = gen8_alloc_va_range;
ppgtt->base.unbind_vma = ppgtt_unbind_vma; ppgtt->base.unbind_vma = ppgtt_unbind_vma;
ppgtt->base.bind_vma = ppgtt_bind_vma; ppgtt->base.bind_vma = ppgtt_bind_vma;
ppgtt->debug_dump = gen8_dump_ppgtt; ppgtt->debug_dump = gen8_dump_ppgtt;
...@@ -1369,6 +1311,7 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt) ...@@ -1369,6 +1311,7 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
ppgtt->base.total = 1ULL << 48; ppgtt->base.total = 1ULL << 48;
ppgtt->switch_mm = gen8_48b_mm_switch; ppgtt->switch_mm = gen8_48b_mm_switch;
ppgtt->base.allocate_va_range = gen8_ppgtt_alloc_4lvl;
ppgtt->base.insert_entries = gen8_ppgtt_insert_4lvl; ppgtt->base.insert_entries = gen8_ppgtt_insert_4lvl;
ppgtt->base.clear_range = gen8_ppgtt_clear_4lvl; ppgtt->base.clear_range = gen8_ppgtt_clear_4lvl;
} else { } else {
...@@ -1390,6 +1333,7 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt) ...@@ -1390,6 +1333,7 @@ static int gen8_ppgtt_init(struct i915_hw_ppgtt *ppgtt)
} }
} }
ppgtt->base.allocate_va_range = gen8_ppgtt_alloc_3lvl;
ppgtt->base.insert_entries = gen8_ppgtt_insert_3lvl; ppgtt->base.insert_entries = gen8_ppgtt_insert_3lvl;
ppgtt->base.clear_range = gen8_ppgtt_clear_3lvl; ppgtt->base.clear_range = gen8_ppgtt_clear_3lvl;
} }
......
...@@ -238,8 +238,6 @@ struct i915_page_directory_pointer { ...@@ -238,8 +238,6 @@ struct i915_page_directory_pointer {
struct i915_pml4 { struct i915_pml4 {
struct i915_page_dma base; struct i915_page_dma base;
DECLARE_BITMAP(used_pml4es, GEN8_PML4ES_PER_PML4);
struct i915_page_directory_pointer *pdps[GEN8_PML4ES_PER_PML4]; struct i915_page_directory_pointer *pdps[GEN8_PML4ES_PER_PML4];
}; };
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment