Commit 20dfa5b7 authored by Barry Song's avatar Barry Song Committed by Andrew Morton

mm: set pte writable while pte_soft_dirty() is true in do_swap_page()

This patch leverages the new pte_needs_soft_dirty_wp() helper to optimize
a scenario where softdirty is enabled, but the softdirty flag has already
been set in do_swap_page().  In this situation, we can use pte_mkwrite
instead of applying write-protection since we don't depend on write
faults.

Link: https://lkml.kernel.org/r/20240607211358.4660-3-21cnbao@gmail.comSigned-off-by: default avatarBarry Song <v-songbaohua@oppo.com>
Suggested-by: default avatarDavid Hildenbrand <david@redhat.com>
Acked-by: default avatarDavid Hildenbrand <david@redhat.com>
Cc: Chris Li <chrisl@kernel.org>
Cc: Kairui Song <kasong@tencent.com>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Ryan Roberts <ryan.roberts@arm.com>
Cc: Suren Baghdasaryan <surenb@google.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent f38ee285
...@@ -4324,7 +4324,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf) ...@@ -4324,7 +4324,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf)
if (!folio_test_ksm(folio) && if (!folio_test_ksm(folio) &&
(exclusive || folio_ref_count(folio) == 1)) { (exclusive || folio_ref_count(folio) == 1)) {
if ((vma->vm_flags & VM_WRITE) && !userfaultfd_pte_wp(vma, pte) && if ((vma->vm_flags & VM_WRITE) && !userfaultfd_pte_wp(vma, pte) &&
!vma_soft_dirty_enabled(vma)) { !pte_needs_soft_dirty_wp(vma, pte)) {
pte = pte_mkwrite(pte, vma); pte = pte_mkwrite(pte, vma);
if (vmf->flags & FAULT_FLAG_WRITE) { if (vmf->flags & FAULT_FLAG_WRITE) {
pte = pte_mkdirty(pte); pte = pte_mkdirty(pte);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment