Commit 8b5989f3 authored by Matthew Wilcox (Oracle)'s avatar Matthew Wilcox (Oracle) Committed by Andrew Morton

arm: implement the new page table range API

Add set_ptes(), update_mmu_cache_range(), flush_dcache_folio() and
flush_icache_pages().  Change the PG_dcache_clear flag from being per-page
to per-folio which makes __dma_page_dev_to_cpu() a bit more exciting. 
Also add flush_cache_pages(), even though this isn't used by generic code
(yet?)

[m.szyprowski@samsung.com: fix potential endless loop in __dma_page_dev_to_cpu()]
  Link: https://lkml.kernel.org/r/20230809172737.3574190-1-m.szyprowski@samsung.com
[willy@infradead.org: fix folio conversion in __dma_page_dev_to_cpu()]
  Link: https://lkml.kernel.org/r/20230823191852.1556561-1-willy@infradead.org
Link: https://lkml.kernel.org/r/20230802151406.3735276-10-willy@infradead.orgSigned-off-by: default avatarMatthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: default avatarMarek Szyprowski <m.szyprowski@samsung.com>
Acked-by: default avatarMike Rapoport (IBM) <rppt@kernel.org>
Reviewed-by: default avatarRussell King (Oracle) <rmk+kernel@armlinux.org.uk>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent ac4cfacc
...@@ -231,14 +231,15 @@ vivt_flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned ...@@ -231,14 +231,15 @@ vivt_flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned
vma->vm_flags); vma->vm_flags);
} }
static inline void static inline void vivt_flush_cache_pages(struct vm_area_struct *vma,
vivt_flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) unsigned long user_addr, unsigned long pfn, unsigned int nr)
{ {
struct mm_struct *mm = vma->vm_mm; struct mm_struct *mm = vma->vm_mm;
if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) { if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) {
unsigned long addr = user_addr & PAGE_MASK; unsigned long addr = user_addr & PAGE_MASK;
__cpuc_flush_user_range(addr, addr + PAGE_SIZE, vma->vm_flags); __cpuc_flush_user_range(addr, addr + nr * PAGE_SIZE,
vma->vm_flags);
} }
} }
...@@ -247,15 +248,17 @@ vivt_flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsig ...@@ -247,15 +248,17 @@ vivt_flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsig
vivt_flush_cache_mm(mm) vivt_flush_cache_mm(mm)
#define flush_cache_range(vma,start,end) \ #define flush_cache_range(vma,start,end) \
vivt_flush_cache_range(vma,start,end) vivt_flush_cache_range(vma,start,end)
#define flush_cache_page(vma,addr,pfn) \ #define flush_cache_pages(vma, addr, pfn, nr) \
vivt_flush_cache_page(vma,addr,pfn) vivt_flush_cache_pages(vma, addr, pfn, nr)
#else #else
extern void flush_cache_mm(struct mm_struct *mm); void flush_cache_mm(struct mm_struct *mm);
extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
extern void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn); void flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr,
unsigned long pfn, unsigned int nr);
#endif #endif
#define flush_cache_dup_mm(mm) flush_cache_mm(mm) #define flush_cache_dup_mm(mm) flush_cache_mm(mm)
#define flush_cache_page(vma, addr, pfn) flush_cache_pages(vma, addr, pfn, 1)
/* /*
* flush_icache_user_range is used when we want to ensure that the * flush_icache_user_range is used when we want to ensure that the
...@@ -289,7 +292,9 @@ extern void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr ...@@ -289,7 +292,9 @@ extern void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr
* See update_mmu_cache for the user space part. * See update_mmu_cache for the user space part.
*/ */
#define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1 #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1
extern void flush_dcache_page(struct page *); void flush_dcache_page(struct page *);
void flush_dcache_folio(struct folio *folio);
#define flush_dcache_folio flush_dcache_folio
#define ARCH_IMPLEMENTS_FLUSH_KERNEL_VMAP_RANGE 1 #define ARCH_IMPLEMENTS_FLUSH_KERNEL_VMAP_RANGE 1
static inline void flush_kernel_vmap_range(void *addr, int size) static inline void flush_kernel_vmap_range(void *addr, int size)
...@@ -321,6 +326,7 @@ static inline void flush_anon_page(struct vm_area_struct *vma, ...@@ -321,6 +326,7 @@ static inline void flush_anon_page(struct vm_area_struct *vma,
* duplicate cache flushing elsewhere performed by flush_dcache_page(). * duplicate cache flushing elsewhere performed by flush_dcache_page().
*/ */
#define flush_icache_page(vma,page) do { } while (0) #define flush_icache_page(vma,page) do { } while (0)
#define flush_icache_pages(vma, page, nr) do { } while (0)
/* /*
* flush_cache_vmap() is used when creating mappings (eg, via vmap, * flush_cache_vmap() is used when creating mappings (eg, via vmap,
......
...@@ -207,8 +207,9 @@ static inline void __sync_icache_dcache(pte_t pteval) ...@@ -207,8 +207,9 @@ static inline void __sync_icache_dcache(pte_t pteval)
extern void __sync_icache_dcache(pte_t pteval); extern void __sync_icache_dcache(pte_t pteval);
#endif #endif
void set_pte_at(struct mm_struct *mm, unsigned long addr, void set_ptes(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, pte_t pteval); pte_t *ptep, pte_t pteval, unsigned int nr);
#define set_ptes set_ptes
static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
{ {
......
...@@ -619,18 +619,22 @@ extern void flush_bp_all(void); ...@@ -619,18 +619,22 @@ extern void flush_bp_all(void);
* If PG_dcache_clean is not set for the page, we need to ensure that any * If PG_dcache_clean is not set for the page, we need to ensure that any
* cache entries for the kernels virtual memory range are written * cache entries for the kernels virtual memory range are written
* back to the page. On ARMv6 and later, the cache coherency is handled via * back to the page. On ARMv6 and later, the cache coherency is handled via
* the set_pte_at() function. * the set_ptes() function.
*/ */
#if __LINUX_ARM_ARCH__ < 6 #if __LINUX_ARM_ARCH__ < 6
extern void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma,
pte_t *ptep); unsigned long addr, pte_t *ptep, unsigned int nr);
#else #else
static inline void update_mmu_cache(struct vm_area_struct *vma, static inline void update_mmu_cache_range(struct vm_fault *vmf,
unsigned long addr, pte_t *ptep) struct vm_area_struct *vma, unsigned long addr, pte_t *ptep,
unsigned int nr)
{ {
} }
#endif #endif
#define update_mmu_cache(vma, addr, ptep) \
update_mmu_cache_range(NULL, vma, addr, ptep, 1)
#define update_mmu_cache_pmd(vma, address, pmd) do { } while (0) #define update_mmu_cache_pmd(vma, address, pmd) do { } while (0)
#endif #endif
......
...@@ -64,10 +64,11 @@ static void mc_copy_user_page(void *from, void *to) ...@@ -64,10 +64,11 @@ static void mc_copy_user_page(void *from, void *to)
void v4_mc_copy_user_highpage(struct page *to, struct page *from, void v4_mc_copy_user_highpage(struct page *to, struct page *from,
unsigned long vaddr, struct vm_area_struct *vma) unsigned long vaddr, struct vm_area_struct *vma)
{ {
struct folio *src = page_folio(from);
void *kto = kmap_atomic(to); void *kto = kmap_atomic(to);
if (!test_and_set_bit(PG_dcache_clean, &from->flags)) if (!test_and_set_bit(PG_dcache_clean, &src->flags))
__flush_dcache_page(page_mapping_file(from), from); __flush_dcache_folio(folio_flush_mapping(src), src);
raw_spin_lock(&minicache_lock); raw_spin_lock(&minicache_lock);
......
...@@ -69,11 +69,12 @@ static void discard_old_kernel_data(void *kto) ...@@ -69,11 +69,12 @@ static void discard_old_kernel_data(void *kto)
static void v6_copy_user_highpage_aliasing(struct page *to, static void v6_copy_user_highpage_aliasing(struct page *to,
struct page *from, unsigned long vaddr, struct vm_area_struct *vma) struct page *from, unsigned long vaddr, struct vm_area_struct *vma)
{ {
struct folio *src = page_folio(from);
unsigned int offset = CACHE_COLOUR(vaddr); unsigned int offset = CACHE_COLOUR(vaddr);
unsigned long kfrom, kto; unsigned long kfrom, kto;
if (!test_and_set_bit(PG_dcache_clean, &from->flags)) if (!test_and_set_bit(PG_dcache_clean, &src->flags))
__flush_dcache_page(page_mapping_file(from), from); __flush_dcache_folio(folio_flush_mapping(src), src);
/* FIXME: not highmem safe */ /* FIXME: not highmem safe */
discard_old_kernel_data(page_address(to)); discard_old_kernel_data(page_address(to));
......
...@@ -84,10 +84,11 @@ static void mc_copy_user_page(void *from, void *to) ...@@ -84,10 +84,11 @@ static void mc_copy_user_page(void *from, void *to)
void xscale_mc_copy_user_highpage(struct page *to, struct page *from, void xscale_mc_copy_user_highpage(struct page *to, struct page *from,
unsigned long vaddr, struct vm_area_struct *vma) unsigned long vaddr, struct vm_area_struct *vma)
{ {
struct folio *src = page_folio(from);
void *kto = kmap_atomic(to); void *kto = kmap_atomic(to);
if (!test_and_set_bit(PG_dcache_clean, &from->flags)) if (!test_and_set_bit(PG_dcache_clean, &src->flags))
__flush_dcache_page(page_mapping_file(from), from); __flush_dcache_folio(folio_flush_mapping(src), src);
raw_spin_lock(&minicache_lock); raw_spin_lock(&minicache_lock);
......
...@@ -709,19 +709,21 @@ static void __dma_page_dev_to_cpu(struct page *page, unsigned long off, ...@@ -709,19 +709,21 @@ static void __dma_page_dev_to_cpu(struct page *page, unsigned long off,
* Mark the D-cache clean for these pages to avoid extra flushing. * Mark the D-cache clean for these pages to avoid extra flushing.
*/ */
if (dir != DMA_TO_DEVICE && size >= PAGE_SIZE) { if (dir != DMA_TO_DEVICE && size >= PAGE_SIZE) {
unsigned long pfn; struct folio *folio = pfn_folio(paddr / PAGE_SIZE);
size_t left = size; size_t offset = offset_in_folio(folio, paddr);
pfn = page_to_pfn(page) + off / PAGE_SIZE; for (;;) {
off %= PAGE_SIZE; size_t sz = folio_size(folio) - offset;
if (off) {
pfn++; if (size < sz)
left -= PAGE_SIZE - off; break;
} if (!offset)
while (left >= PAGE_SIZE) { set_bit(PG_dcache_clean, &folio->flags);
page = pfn_to_page(pfn++); offset = 0;
set_bit(PG_dcache_clean, &page->flags); size -= sz;
left -= PAGE_SIZE; if (!size)
break;
folio = folio_next(folio);
} }
} }
} }
......
...@@ -180,12 +180,12 @@ make_coherent(struct address_space *mapping, struct vm_area_struct *vma, ...@@ -180,12 +180,12 @@ make_coherent(struct address_space *mapping, struct vm_area_struct *vma,
* *
* Note that the pte lock will be held. * Note that the pte lock will be held.
*/ */
void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma,
pte_t *ptep) unsigned long addr, pte_t *ptep, unsigned int nr)
{ {
unsigned long pfn = pte_pfn(*ptep); unsigned long pfn = pte_pfn(*ptep);
struct address_space *mapping; struct address_space *mapping;
struct page *page; struct folio *folio;
if (!pfn_valid(pfn)) if (!pfn_valid(pfn))
return; return;
...@@ -194,13 +194,13 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, ...@@ -194,13 +194,13 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr,
* The zero page is never written to, so never has any dirty * The zero page is never written to, so never has any dirty
* cache lines, and therefore never needs to be flushed. * cache lines, and therefore never needs to be flushed.
*/ */
page = pfn_to_page(pfn); if (is_zero_pfn(pfn))
if (page == ZERO_PAGE(0))
return; return;
mapping = page_mapping_file(page); folio = page_folio(pfn_to_page(pfn));
if (!test_and_set_bit(PG_dcache_clean, &page->flags)) mapping = folio_flush_mapping(folio);
__flush_dcache_page(mapping, page); if (!test_and_set_bit(PG_dcache_clean, &folio->flags))
__flush_dcache_folio(mapping, folio);
if (mapping) { if (mapping) {
if (cache_is_vivt()) if (cache_is_vivt())
make_coherent(mapping, vma, addr, ptep, pfn); make_coherent(mapping, vma, addr, ptep, pfn);
......
...@@ -95,10 +95,10 @@ void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned ...@@ -95,10 +95,10 @@ void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned
__flush_icache_all(); __flush_icache_all();
} }
void flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) void flush_cache_pages(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn, unsigned int nr)
{ {
if (cache_is_vivt()) { if (cache_is_vivt()) {
vivt_flush_cache_page(vma, user_addr, pfn); vivt_flush_cache_pages(vma, user_addr, pfn, nr);
return; return;
} }
...@@ -196,29 +196,31 @@ void copy_to_user_page(struct vm_area_struct *vma, struct page *page, ...@@ -196,29 +196,31 @@ void copy_to_user_page(struct vm_area_struct *vma, struct page *page,
#endif #endif
} }
void __flush_dcache_page(struct address_space *mapping, struct page *page) void __flush_dcache_folio(struct address_space *mapping, struct folio *folio)
{ {
/* /*
* Writeback any data associated with the kernel mapping of this * Writeback any data associated with the kernel mapping of this
* page. This ensures that data in the physical page is mutually * page. This ensures that data in the physical page is mutually
* coherent with the kernels mapping. * coherent with the kernels mapping.
*/ */
if (!PageHighMem(page)) { if (!folio_test_highmem(folio)) {
__cpuc_flush_dcache_area(page_address(page), page_size(page)); __cpuc_flush_dcache_area(folio_address(folio),
folio_size(folio));
} else { } else {
unsigned long i; unsigned long i;
if (cache_is_vipt_nonaliasing()) { if (cache_is_vipt_nonaliasing()) {
for (i = 0; i < compound_nr(page); i++) { for (i = 0; i < folio_nr_pages(folio); i++) {
void *addr = kmap_atomic(page + i); void *addr = kmap_local_folio(folio,
i * PAGE_SIZE);
__cpuc_flush_dcache_area(addr, PAGE_SIZE); __cpuc_flush_dcache_area(addr, PAGE_SIZE);
kunmap_atomic(addr); kunmap_local(addr);
} }
} else { } else {
for (i = 0; i < compound_nr(page); i++) { for (i = 0; i < folio_nr_pages(folio); i++) {
void *addr = kmap_high_get(page + i); void *addr = kmap_high_get(folio_page(folio, i));
if (addr) { if (addr) {
__cpuc_flush_dcache_area(addr, PAGE_SIZE); __cpuc_flush_dcache_area(addr, PAGE_SIZE);
kunmap_high(page + i); kunmap_high(folio_page(folio, i));
} }
} }
} }
...@@ -230,15 +232,14 @@ void __flush_dcache_page(struct address_space *mapping, struct page *page) ...@@ -230,15 +232,14 @@ void __flush_dcache_page(struct address_space *mapping, struct page *page)
* userspace colour, which is congruent with page->index. * userspace colour, which is congruent with page->index.
*/ */
if (mapping && cache_is_vipt_aliasing()) if (mapping && cache_is_vipt_aliasing())
flush_pfn_alias(page_to_pfn(page), flush_pfn_alias(folio_pfn(folio), folio_pos(folio));
page->index << PAGE_SHIFT);
} }
static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) static void __flush_dcache_aliases(struct address_space *mapping, struct folio *folio)
{ {
struct mm_struct *mm = current->active_mm; struct mm_struct *mm = current->active_mm;
struct vm_area_struct *mpnt; struct vm_area_struct *vma;
pgoff_t pgoff; pgoff_t pgoff, pgoff_end;
/* /*
* There are possible user space mappings of this page: * There are possible user space mappings of this page:
...@@ -246,21 +247,36 @@ static void __flush_dcache_aliases(struct address_space *mapping, struct page *p ...@@ -246,21 +247,36 @@ static void __flush_dcache_aliases(struct address_space *mapping, struct page *p
* data in the current VM view associated with this page. * data in the current VM view associated with this page.
* - aliasing VIPT: we only need to find one mapping of this page. * - aliasing VIPT: we only need to find one mapping of this page.
*/ */
pgoff = page->index; pgoff = folio->index;
pgoff_end = pgoff + folio_nr_pages(folio) - 1;
flush_dcache_mmap_lock(mapping); flush_dcache_mmap_lock(mapping);
vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { vma_interval_tree_foreach(vma, &mapping->i_mmap, pgoff, pgoff_end) {
unsigned long offset; unsigned long start, offset, pfn;
unsigned int nr;
/* /*
* If this VMA is not in our MM, we can ignore it. * If this VMA is not in our MM, we can ignore it.
*/ */
if (mpnt->vm_mm != mm) if (vma->vm_mm != mm)
continue; continue;
if (!(mpnt->vm_flags & VM_MAYSHARE)) if (!(vma->vm_flags & VM_MAYSHARE))
continue; continue;
offset = (pgoff - mpnt->vm_pgoff) << PAGE_SHIFT;
flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); start = vma->vm_start;
pfn = folio_pfn(folio);
nr = folio_nr_pages(folio);
offset = pgoff - vma->vm_pgoff;
if (offset > -nr) {
pfn -= offset;
nr += offset;
} else {
start += offset * PAGE_SIZE;
}
if (start + nr * PAGE_SIZE > vma->vm_end)
nr = (vma->vm_end - start) / PAGE_SIZE;
flush_cache_pages(vma, start, pfn, nr);
} }
flush_dcache_mmap_unlock(mapping); flush_dcache_mmap_unlock(mapping);
} }
...@@ -269,7 +285,7 @@ static void __flush_dcache_aliases(struct address_space *mapping, struct page *p ...@@ -269,7 +285,7 @@ static void __flush_dcache_aliases(struct address_space *mapping, struct page *p
void __sync_icache_dcache(pte_t pteval) void __sync_icache_dcache(pte_t pteval)
{ {
unsigned long pfn; unsigned long pfn;
struct page *page; struct folio *folio;
struct address_space *mapping; struct address_space *mapping;
if (cache_is_vipt_nonaliasing() && !pte_exec(pteval)) if (cache_is_vipt_nonaliasing() && !pte_exec(pteval))
...@@ -279,14 +295,14 @@ void __sync_icache_dcache(pte_t pteval) ...@@ -279,14 +295,14 @@ void __sync_icache_dcache(pte_t pteval)
if (!pfn_valid(pfn)) if (!pfn_valid(pfn))
return; return;
page = pfn_to_page(pfn); folio = page_folio(pfn_to_page(pfn));
if (cache_is_vipt_aliasing()) if (cache_is_vipt_aliasing())
mapping = page_mapping_file(page); mapping = folio_flush_mapping(folio);
else else
mapping = NULL; mapping = NULL;
if (!test_and_set_bit(PG_dcache_clean, &page->flags)) if (!test_and_set_bit(PG_dcache_clean, &folio->flags))
__flush_dcache_page(mapping, page); __flush_dcache_folio(mapping, folio);
if (pte_exec(pteval)) if (pte_exec(pteval))
__flush_icache_all(); __flush_icache_all();
...@@ -312,7 +328,7 @@ void __sync_icache_dcache(pte_t pteval) ...@@ -312,7 +328,7 @@ void __sync_icache_dcache(pte_t pteval)
* Note that we disable the lazy flush for SMP configurations where * Note that we disable the lazy flush for SMP configurations where
* the cache maintenance operations are not automatically broadcasted. * the cache maintenance operations are not automatically broadcasted.
*/ */
void flush_dcache_page(struct page *page) void flush_dcache_folio(struct folio *folio)
{ {
struct address_space *mapping; struct address_space *mapping;
...@@ -320,31 +336,36 @@ void flush_dcache_page(struct page *page) ...@@ -320,31 +336,36 @@ void flush_dcache_page(struct page *page)
* The zero page is never written to, so never has any dirty * The zero page is never written to, so never has any dirty
* cache lines, and therefore never needs to be flushed. * cache lines, and therefore never needs to be flushed.
*/ */
if (page == ZERO_PAGE(0)) if (is_zero_pfn(folio_pfn(folio)))
return; return;
if (!cache_ops_need_broadcast() && cache_is_vipt_nonaliasing()) { if (!cache_ops_need_broadcast() && cache_is_vipt_nonaliasing()) {
if (test_bit(PG_dcache_clean, &page->flags)) if (test_bit(PG_dcache_clean, &folio->flags))
clear_bit(PG_dcache_clean, &page->flags); clear_bit(PG_dcache_clean, &folio->flags);
return; return;
} }
mapping = page_mapping_file(page); mapping = folio_flush_mapping(folio);
if (!cache_ops_need_broadcast() && if (!cache_ops_need_broadcast() &&
mapping && !page_mapcount(page)) mapping && !folio_mapped(folio))
clear_bit(PG_dcache_clean, &page->flags); clear_bit(PG_dcache_clean, &folio->flags);
else { else {
__flush_dcache_page(mapping, page); __flush_dcache_folio(mapping, folio);
if (mapping && cache_is_vivt()) if (mapping && cache_is_vivt())
__flush_dcache_aliases(mapping, page); __flush_dcache_aliases(mapping, folio);
else if (mapping) else if (mapping)
__flush_icache_all(); __flush_icache_all();
set_bit(PG_dcache_clean, &page->flags); set_bit(PG_dcache_clean, &folio->flags);
} }
} }
EXPORT_SYMBOL(flush_dcache_page); EXPORT_SYMBOL(flush_dcache_folio);
void flush_dcache_page(struct page *page)
{
flush_dcache_folio(page_folio(page));
}
EXPORT_SYMBOL(flush_dcache_page);
/* /*
* Flush an anonymous page so that users of get_user_pages() * Flush an anonymous page so that users of get_user_pages()
* can safely access the data. The expected sequence is: * can safely access the data. The expected sequence is:
......
...@@ -45,7 +45,7 @@ struct mem_type { ...@@ -45,7 +45,7 @@ struct mem_type {
const struct mem_type *get_mem_type(unsigned int type); const struct mem_type *get_mem_type(unsigned int type);
extern void __flush_dcache_page(struct address_space *mapping, struct page *page); void __flush_dcache_folio(struct address_space *mapping, struct folio *folio);
/* /*
* ARM specific vm_struct->flags bits. * ARM specific vm_struct->flags bits.
......
...@@ -1789,7 +1789,7 @@ void __init paging_init(const struct machine_desc *mdesc) ...@@ -1789,7 +1789,7 @@ void __init paging_init(const struct machine_desc *mdesc)
bootmem_init(); bootmem_init();
empty_zero_page = virt_to_page(zero_page); empty_zero_page = virt_to_page(zero_page);
__flush_dcache_page(NULL, empty_zero_page); __flush_dcache_folio(NULL, page_folio(empty_zero_page));
} }
void __init early_mm_init(const struct machine_desc *mdesc) void __init early_mm_init(const struct machine_desc *mdesc)
...@@ -1798,8 +1798,8 @@ void __init early_mm_init(const struct machine_desc *mdesc) ...@@ -1798,8 +1798,8 @@ void __init early_mm_init(const struct machine_desc *mdesc)
early_paging_init(mdesc); early_paging_init(mdesc);
} }
void set_pte_at(struct mm_struct *mm, unsigned long addr, void set_ptes(struct mm_struct *mm, unsigned long addr,
pte_t *ptep, pte_t pteval) pte_t *ptep, pte_t pteval, unsigned int nr)
{ {
unsigned long ext = 0; unsigned long ext = 0;
...@@ -1809,5 +1809,11 @@ void set_pte_at(struct mm_struct *mm, unsigned long addr, ...@@ -1809,5 +1809,11 @@ void set_pte_at(struct mm_struct *mm, unsigned long addr,
ext |= PTE_EXT_NG; ext |= PTE_EXT_NG;
} }
for (;;) {
set_pte_ext(ptep, pteval, ext); set_pte_ext(ptep, pteval, ext);
if (--nr == 0)
break;
ptep++;
pte_val(pteval) += PAGE_SIZE;
}
} }
...@@ -180,6 +180,12 @@ void setup_mm_for_reboot(void) ...@@ -180,6 +180,12 @@ void setup_mm_for_reboot(void)
{ {
} }
void flush_dcache_folio(struct folio *folio)
{
__cpuc_flush_dcache_area(folio_address(folio), folio_size(folio));
}
EXPORT_SYMBOL(flush_dcache_folio);
void flush_dcache_page(struct page *page) void flush_dcache_page(struct page *page)
{ {
__cpuc_flush_dcache_area(page_address(page), PAGE_SIZE); __cpuc_flush_dcache_area(page_address(page), PAGE_SIZE);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment