Commit 20fef4ef authored by Jason Gunthorpe's avatar Jason Gunthorpe

nouveau: use mmu_interval_notifier instead of hmm_mirror

Remove the hmm_mirror object and use the mmu_interval_notifier API instead
for the range, and use the normal mmu_notifier API for the general
invalidation callback.

While here re-organize the pagefault path so the locking pattern is clear.

nouveau is the only driver that uses a temporary range object and instead
forwards nearly every invalidation range directly to the HW. While this is
not how the mmu_interval_notifier was intended to be used, the overheads on
the pagefaulting path are similar to the existing hmm_mirror version.
Particularly since the interval tree will be small.

Link: https://lore.kernel.org/r/20191112202231.3856-10-jgg@ziepe.caTested-by: default avatarRalph Campbell <rcampbell@nvidia.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@mellanox.com>
parent c625c274
...@@ -96,8 +96,6 @@ struct nouveau_svmm { ...@@ -96,8 +96,6 @@ struct nouveau_svmm {
} unmanaged; } unmanaged;
struct mutex mutex; struct mutex mutex;
struct hmm_mirror mirror;
}; };
#define SVMM_DBG(s,f,a...) \ #define SVMM_DBG(s,f,a...) \
...@@ -293,23 +291,11 @@ static const struct mmu_notifier_ops nouveau_mn_ops = { ...@@ -293,23 +291,11 @@ static const struct mmu_notifier_ops nouveau_mn_ops = {
.free_notifier = nouveau_svmm_free_notifier, .free_notifier = nouveau_svmm_free_notifier,
}; };
static int
nouveau_svmm_sync_cpu_device_pagetables(struct hmm_mirror *mirror,
const struct mmu_notifier_range *update)
{
return 0;
}
static const struct hmm_mirror_ops nouveau_svmm = {
.sync_cpu_device_pagetables = nouveau_svmm_sync_cpu_device_pagetables,
};
void void
nouveau_svmm_fini(struct nouveau_svmm **psvmm) nouveau_svmm_fini(struct nouveau_svmm **psvmm)
{ {
struct nouveau_svmm *svmm = *psvmm; struct nouveau_svmm *svmm = *psvmm;
if (svmm) { if (svmm) {
hmm_mirror_unregister(&svmm->mirror);
mutex_lock(&svmm->mutex); mutex_lock(&svmm->mutex);
svmm->vmm = NULL; svmm->vmm = NULL;
mutex_unlock(&svmm->mutex); mutex_unlock(&svmm->mutex);
...@@ -357,15 +343,10 @@ nouveau_svmm_init(struct drm_device *dev, void *data, ...@@ -357,15 +343,10 @@ nouveau_svmm_init(struct drm_device *dev, void *data,
goto out_free; goto out_free;
down_write(&current->mm->mmap_sem); down_write(&current->mm->mmap_sem);
svmm->mirror.ops = &nouveau_svmm;
ret = hmm_mirror_register(&svmm->mirror, current->mm);
if (ret)
goto out_mm_unlock;
svmm->notifier.ops = &nouveau_mn_ops; svmm->notifier.ops = &nouveau_mn_ops;
ret = __mmu_notifier_register(&svmm->notifier, current->mm); ret = __mmu_notifier_register(&svmm->notifier, current->mm);
if (ret) if (ret)
goto out_hmm_unregister; goto out_mm_unlock;
/* Note, ownership of svmm transfers to mmu_notifier */ /* Note, ownership of svmm transfers to mmu_notifier */
cli->svm.svmm = svmm; cli->svm.svmm = svmm;
...@@ -374,8 +355,6 @@ nouveau_svmm_init(struct drm_device *dev, void *data, ...@@ -374,8 +355,6 @@ nouveau_svmm_init(struct drm_device *dev, void *data,
mutex_unlock(&cli->mutex); mutex_unlock(&cli->mutex);
return 0; return 0;
out_hmm_unregister:
hmm_mirror_unregister(&svmm->mirror);
out_mm_unlock: out_mm_unlock:
up_write(&current->mm->mmap_sem); up_write(&current->mm->mmap_sem);
out_free: out_free:
...@@ -503,43 +482,90 @@ nouveau_svm_fault_cache(struct nouveau_svm *svm, ...@@ -503,43 +482,90 @@ nouveau_svm_fault_cache(struct nouveau_svm *svm,
fault->inst, fault->addr, fault->access); fault->inst, fault->addr, fault->access);
} }
static inline bool struct svm_notifier {
nouveau_range_done(struct hmm_range *range) struct mmu_interval_notifier notifier;
struct nouveau_svmm *svmm;
};
static bool nouveau_svm_range_invalidate(struct mmu_interval_notifier *mni,
const struct mmu_notifier_range *range,
unsigned long cur_seq)
{ {
bool ret = hmm_range_valid(range); struct svm_notifier *sn =
container_of(mni, struct svm_notifier, notifier);
hmm_range_unregister(range); /*
return ret; * serializes the update to mni->invalidate_seq done by caller and
* prevents invalidation of the PTE from progressing while HW is being
* programmed. This is very hacky and only works because the normal
* notifier that does invalidation is always called after the range
* notifier.
*/
if (mmu_notifier_range_blockable(range))
mutex_lock(&sn->svmm->mutex);
else if (!mutex_trylock(&sn->svmm->mutex))
return false;
mmu_interval_set_seq(mni, cur_seq);
mutex_unlock(&sn->svmm->mutex);
return true;
} }
static int static const struct mmu_interval_notifier_ops nouveau_svm_mni_ops = {
nouveau_range_fault(struct nouveau_svmm *svmm, struct hmm_range *range) .invalidate = nouveau_svm_range_invalidate,
};
static int nouveau_range_fault(struct nouveau_svmm *svmm,
struct nouveau_drm *drm, void *data, u32 size,
u64 *pfns, struct svm_notifier *notifier)
{ {
unsigned long timeout =
jiffies + msecs_to_jiffies(HMM_RANGE_DEFAULT_TIMEOUT);
/* Have HMM fault pages within the fault window to the GPU. */
struct hmm_range range = {
.notifier = &notifier->notifier,
.start = notifier->notifier.interval_tree.start,
.end = notifier->notifier.interval_tree.last + 1,
.pfns = pfns,
.flags = nouveau_svm_pfn_flags,
.values = nouveau_svm_pfn_values,
.pfn_shift = NVIF_VMM_PFNMAP_V0_ADDR_SHIFT,
};
struct mm_struct *mm = notifier->notifier.mm;
long ret; long ret;
range->default_flags = 0; while (true) {
range->pfn_flags_mask = -1UL; if (time_after(jiffies, timeout))
return -EBUSY;
ret = hmm_range_register(range, &svmm->mirror); range.notifier_seq = mmu_interval_read_begin(range.notifier);
if (ret) { range.default_flags = 0;
up_read(&svmm->notifier.mm->mmap_sem); range.pfn_flags_mask = -1UL;
return (int)ret; down_read(&mm->mmap_sem);
} ret = hmm_range_fault(&range, 0);
up_read(&mm->mmap_sem);
if (ret <= 0) {
if (ret == 0 || ret == -EBUSY)
continue;
return ret;
}
if (!hmm_range_wait_until_valid(range, HMM_RANGE_DEFAULT_TIMEOUT)) { mutex_lock(&svmm->mutex);
up_read(&svmm->notifier.mm->mmap_sem); if (mmu_interval_read_retry(range.notifier,
return -EBUSY; range.notifier_seq)) {
mutex_unlock(&svmm->mutex);
continue;
}
break;
} }
ret = hmm_range_fault(range, 0); nouveau_dmem_convert_pfn(drm, &range);
if (ret <= 0) {
if (ret == 0) svmm->vmm->vmm.object.client->super = true;
ret = -EBUSY; ret = nvif_object_ioctl(&svmm->vmm->vmm.object, data, size, NULL);
up_read(&svmm->notifier.mm->mmap_sem); svmm->vmm->vmm.object.client->super = false;
hmm_range_unregister(range); mutex_unlock(&svmm->mutex);
return ret;
} return ret;
return 0;
} }
static int static int
...@@ -559,7 +585,6 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -559,7 +585,6 @@ nouveau_svm_fault(struct nvif_notify *notify)
} i; } i;
u64 phys[16]; u64 phys[16];
} args; } args;
struct hmm_range range;
struct vm_area_struct *vma; struct vm_area_struct *vma;
u64 inst, start, limit; u64 inst, start, limit;
int fi, fn, pi, fill; int fi, fn, pi, fill;
...@@ -615,6 +640,7 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -615,6 +640,7 @@ nouveau_svm_fault(struct nvif_notify *notify)
args.i.p.version = 0; args.i.p.version = 0;
for (fi = 0; fn = fi + 1, fi < buffer->fault_nr; fi = fn) { for (fi = 0; fn = fi + 1, fi < buffer->fault_nr; fi = fn) {
struct svm_notifier notifier;
struct mm_struct *mm; struct mm_struct *mm;
/* Cancel any faults from non-SVM channels. */ /* Cancel any faults from non-SVM channels. */
...@@ -623,7 +649,6 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -623,7 +649,6 @@ nouveau_svm_fault(struct nvif_notify *notify)
continue; continue;
} }
SVMM_DBG(svmm, "addr %016llx", buffer->fault[fi]->addr); SVMM_DBG(svmm, "addr %016llx", buffer->fault[fi]->addr);
mm = svmm->notifier.mm;
/* We try and group handling of faults within a small /* We try and group handling of faults within a small
* window into a single update. * window into a single update.
...@@ -637,6 +662,12 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -637,6 +662,12 @@ nouveau_svm_fault(struct nvif_notify *notify)
start = max_t(u64, start, svmm->unmanaged.limit); start = max_t(u64, start, svmm->unmanaged.limit);
SVMM_DBG(svmm, "wndw %016llx-%016llx", start, limit); SVMM_DBG(svmm, "wndw %016llx-%016llx", start, limit);
mm = svmm->notifier.mm;
if (!mmget_not_zero(mm)) {
nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]);
continue;
}
/* Intersect fault window with the CPU VMA, cancelling /* Intersect fault window with the CPU VMA, cancelling
* the fault if the address is invalid. * the fault if the address is invalid.
*/ */
...@@ -645,16 +676,18 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -645,16 +676,18 @@ nouveau_svm_fault(struct nvif_notify *notify)
if (!vma) { if (!vma) {
SVMM_ERR(svmm, "wndw %016llx-%016llx", start, limit); SVMM_ERR(svmm, "wndw %016llx-%016llx", start, limit);
up_read(&mm->mmap_sem); up_read(&mm->mmap_sem);
mmput(mm);
nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]); nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]);
continue; continue;
} }
start = max_t(u64, start, vma->vm_start); start = max_t(u64, start, vma->vm_start);
limit = min_t(u64, limit, vma->vm_end); limit = min_t(u64, limit, vma->vm_end);
up_read(&mm->mmap_sem);
SVMM_DBG(svmm, "wndw %016llx-%016llx", start, limit); SVMM_DBG(svmm, "wndw %016llx-%016llx", start, limit);
if (buffer->fault[fi]->addr != start) { if (buffer->fault[fi]->addr != start) {
SVMM_ERR(svmm, "addr %016llx", buffer->fault[fi]->addr); SVMM_ERR(svmm, "addr %016llx", buffer->fault[fi]->addr);
up_read(&mm->mmap_sem); mmput(mm);
nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]); nouveau_svm_fault_cancel_fault(svm, buffer->fault[fi]);
continue; continue;
} }
...@@ -710,33 +743,19 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -710,33 +743,19 @@ nouveau_svm_fault(struct nvif_notify *notify)
args.i.p.addr, args.i.p.addr,
args.i.p.addr + args.i.p.size, fn - fi); args.i.p.addr + args.i.p.size, fn - fi);
/* Have HMM fault pages within the fault window to the GPU. */ notifier.svmm = svmm;
range.start = args.i.p.addr; ret = mmu_interval_notifier_insert(&notifier.notifier,
range.end = args.i.p.addr + args.i.p.size; svmm->notifier.mm,
range.pfns = args.phys; args.i.p.addr, args.i.p.size,
range.flags = nouveau_svm_pfn_flags; &nouveau_svm_mni_ops);
range.values = nouveau_svm_pfn_values; if (!ret) {
range.pfn_shift = NVIF_VMM_PFNMAP_V0_ADDR_SHIFT; ret = nouveau_range_fault(
again: svmm, svm->drm, &args,
ret = nouveau_range_fault(svmm, &range); sizeof(args.i) + pi * sizeof(args.phys[0]),
if (ret == 0) { args.phys, &notifier);
mutex_lock(&svmm->mutex); mmu_interval_notifier_remove(&notifier.notifier);
if (!nouveau_range_done(&range)) {
mutex_unlock(&svmm->mutex);
goto again;
}
nouveau_dmem_convert_pfn(svm->drm, &range);
svmm->vmm->vmm.object.client->super = true;
ret = nvif_object_ioctl(&svmm->vmm->vmm.object,
&args, sizeof(args.i) +
pi * sizeof(args.phys[0]),
NULL);
svmm->vmm->vmm.object.client->super = false;
mutex_unlock(&svmm->mutex);
up_read(&mm->mmap_sem);
} }
mmput(mm);
/* Cancel any faults in the window whose pages didn't manage /* Cancel any faults in the window whose pages didn't manage
* to keep their valid bit, or stay writeable when required. * to keep their valid bit, or stay writeable when required.
...@@ -745,10 +764,10 @@ nouveau_svm_fault(struct nvif_notify *notify) ...@@ -745,10 +764,10 @@ nouveau_svm_fault(struct nvif_notify *notify)
*/ */
while (fi < fn) { while (fi < fn) {
struct nouveau_svm_fault *fault = buffer->fault[fi++]; struct nouveau_svm_fault *fault = buffer->fault[fi++];
pi = (fault->addr - range.start) >> PAGE_SHIFT; pi = (fault->addr - args.i.p.addr) >> PAGE_SHIFT;
if (ret || if (ret ||
!(range.pfns[pi] & NVIF_VMM_PFNMAP_V0_V) || !(args.phys[pi] & NVIF_VMM_PFNMAP_V0_V) ||
(!(range.pfns[pi] & NVIF_VMM_PFNMAP_V0_W) && (!(args.phys[pi] & NVIF_VMM_PFNMAP_V0_W) &&
fault->access != 0 && fault->access != 3)) { fault->access != 0 && fault->access != 3)) {
nouveau_svm_fault_cancel_fault(svm, fault); nouveau_svm_fault_cancel_fault(svm, fault);
continue; continue;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment