Commit 2d4d7676 authored by Jason Gunthorpe's avatar Jason Gunthorpe Committed by Joerg Roedel

iommu/intel: Support the gfp argument to the map_pages op

Flow it down to alloc_pgtable_page() via pfn_to_dma_pte() and
__domain_mapping().
Reviewed-by: default avatarKevin Tian <kevin.tian@intel.com>
Reviewed-by: default avatarLu Baolu <baolu.lu@linux.intel.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@nvidia.com>
Link: https://lore.kernel.org/r/7-v3-76b587fe28df+6e3-iommu_map_gfp_jgg@nvidia.comSigned-off-by: default avatarJoerg Roedel <jroedel@suse.de>
parent 2552d3a2
...@@ -908,7 +908,8 @@ void dmar_fault_dump_ptes(struct intel_iommu *iommu, u16 source_id, ...@@ -908,7 +908,8 @@ void dmar_fault_dump_ptes(struct intel_iommu *iommu, u16 source_id,
#endif #endif
static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain, static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain,
unsigned long pfn, int *target_level) unsigned long pfn, int *target_level,
gfp_t gfp)
{ {
struct dma_pte *parent, *pte; struct dma_pte *parent, *pte;
int level = agaw_to_level(domain->agaw); int level = agaw_to_level(domain->agaw);
...@@ -935,7 +936,7 @@ static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain, ...@@ -935,7 +936,7 @@ static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain,
if (!dma_pte_present(pte)) { if (!dma_pte_present(pte)) {
uint64_t pteval; uint64_t pteval;
tmp_page = alloc_pgtable_page(domain->nid, GFP_ATOMIC); tmp_page = alloc_pgtable_page(domain->nid, gfp);
if (!tmp_page) if (!tmp_page)
return NULL; return NULL;
...@@ -2150,7 +2151,8 @@ static void switch_to_super_page(struct dmar_domain *domain, ...@@ -2150,7 +2151,8 @@ static void switch_to_super_page(struct dmar_domain *domain,
while (start_pfn <= end_pfn) { while (start_pfn <= end_pfn) {
if (!pte) if (!pte)
pte = pfn_to_dma_pte(domain, start_pfn, &level); pte = pfn_to_dma_pte(domain, start_pfn, &level,
GFP_ATOMIC);
if (dma_pte_present(pte)) { if (dma_pte_present(pte)) {
dma_pte_free_pagetable(domain, start_pfn, dma_pte_free_pagetable(domain, start_pfn,
...@@ -2172,7 +2174,8 @@ static void switch_to_super_page(struct dmar_domain *domain, ...@@ -2172,7 +2174,8 @@ static void switch_to_super_page(struct dmar_domain *domain,
static int static int
__domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn, __domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn,
unsigned long phys_pfn, unsigned long nr_pages, int prot) unsigned long phys_pfn, unsigned long nr_pages, int prot,
gfp_t gfp)
{ {
struct dma_pte *first_pte = NULL, *pte = NULL; struct dma_pte *first_pte = NULL, *pte = NULL;
unsigned int largepage_lvl = 0; unsigned int largepage_lvl = 0;
...@@ -2202,7 +2205,8 @@ __domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn, ...@@ -2202,7 +2205,8 @@ __domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn,
largepage_lvl = hardware_largepage_caps(domain, iov_pfn, largepage_lvl = hardware_largepage_caps(domain, iov_pfn,
phys_pfn, nr_pages); phys_pfn, nr_pages);
pte = pfn_to_dma_pte(domain, iov_pfn, &largepage_lvl); pte = pfn_to_dma_pte(domain, iov_pfn, &largepage_lvl,
gfp);
if (!pte) if (!pte)
return -ENOMEM; return -ENOMEM;
first_pte = pte; first_pte = pte;
...@@ -2368,7 +2372,7 @@ static int iommu_domain_identity_map(struct dmar_domain *domain, ...@@ -2368,7 +2372,7 @@ static int iommu_domain_identity_map(struct dmar_domain *domain,
return __domain_mapping(domain, first_vpfn, return __domain_mapping(domain, first_vpfn,
first_vpfn, last_vpfn - first_vpfn + 1, first_vpfn, last_vpfn - first_vpfn + 1,
DMA_PTE_READ|DMA_PTE_WRITE); DMA_PTE_READ|DMA_PTE_WRITE, GFP_ATOMIC);
} }
static int md_domain_init(struct dmar_domain *domain, int guest_width); static int md_domain_init(struct dmar_domain *domain, int guest_width);
...@@ -4298,7 +4302,7 @@ static int intel_iommu_map(struct iommu_domain *domain, ...@@ -4298,7 +4302,7 @@ static int intel_iommu_map(struct iommu_domain *domain,
the low bits of hpa would take us onto the next page */ the low bits of hpa would take us onto the next page */
size = aligned_nrpages(hpa, size); size = aligned_nrpages(hpa, size);
return __domain_mapping(dmar_domain, iova >> VTD_PAGE_SHIFT, return __domain_mapping(dmar_domain, iova >> VTD_PAGE_SHIFT,
hpa >> VTD_PAGE_SHIFT, size, prot); hpa >> VTD_PAGE_SHIFT, size, prot, gfp);
} }
static int intel_iommu_map_pages(struct iommu_domain *domain, static int intel_iommu_map_pages(struct iommu_domain *domain,
...@@ -4333,7 +4337,8 @@ static size_t intel_iommu_unmap(struct iommu_domain *domain, ...@@ -4333,7 +4337,8 @@ static size_t intel_iommu_unmap(struct iommu_domain *domain,
/* Cope with horrid API which requires us to unmap more than the /* Cope with horrid API which requires us to unmap more than the
size argument if it happens to be a large-page mapping. */ size argument if it happens to be a large-page mapping. */
BUG_ON(!pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level)); BUG_ON(!pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level,
GFP_ATOMIC));
if (size < VTD_PAGE_SIZE << level_to_offset_bits(level)) if (size < VTD_PAGE_SIZE << level_to_offset_bits(level))
size = VTD_PAGE_SIZE << level_to_offset_bits(level); size = VTD_PAGE_SIZE << level_to_offset_bits(level);
...@@ -4392,7 +4397,8 @@ static phys_addr_t intel_iommu_iova_to_phys(struct iommu_domain *domain, ...@@ -4392,7 +4397,8 @@ static phys_addr_t intel_iommu_iova_to_phys(struct iommu_domain *domain,
int level = 0; int level = 0;
u64 phys = 0; u64 phys = 0;
pte = pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level); pte = pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level,
GFP_ATOMIC);
if (pte && dma_pte_present(pte)) if (pte && dma_pte_present(pte))
phys = dma_pte_addr(pte) + phys = dma_pte_addr(pte) +
(iova & (BIT_MASK(level_to_offset_bits(level) + (iova & (BIT_MASK(level_to_offset_bits(level) +
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment