Commit 2fd8774c authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'stable/for-linus-4.10' of...

Merge branch 'stable/for-linus-4.10' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/swiotlb

Pull swiotlb fixes from Konrad Rzeszutek Wilk:
 "This has one fix to make i915 work when using Xen SWIOTLB, and a
  feature from Geert to aid in debugging of devices that can't do DMA
  outside the 32-bit address space.

  The feature from Geert is on top of v4.10 merge window commit
  (specifically you pulling my previous branch), as his changes were
  dependent on the Documentation/ movement patches.

  I figured it would just easier than me trying than to cherry-pick the
  Documentation patches to satisfy git.

  The patches have been soaking since 12/20, albeit I updated the last
  patch due to linux-next catching an compiler error and adding an
  Tested-and-Reported-by tag"

* 'stable/for-linus-4.10' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/swiotlb:
  swiotlb: Export swiotlb_max_segment to users
  swiotlb: Add swiotlb=noforce debug option
  swiotlb: Convert swiotlb_force from int to enum
  x86, swiotlb: Simplify pci_swiotlb_detect_override()
parents 65cdc405 7453c549
...@@ -3821,10 +3821,11 @@ ...@@ -3821,10 +3821,11 @@
it if 0 is given (See Documentation/cgroup-v1/memory.txt) it if 0 is given (See Documentation/cgroup-v1/memory.txt)
swiotlb= [ARM,IA-64,PPC,MIPS,X86] swiotlb= [ARM,IA-64,PPC,MIPS,X86]
Format: { <int> | force } Format: { <int> | force | noforce }
<int> -- Number of I/O TLB slabs <int> -- Number of I/O TLB slabs
force -- force using of bounce buffers even if they force -- force using of bounce buffers even if they
wouldn't be automatically used by the kernel wouldn't be automatically used by the kernel
noforce -- Never use bounce buffers (for debugging)
switches= [HW,M68k] switches= [HW,M68k]
......
...@@ -524,7 +524,8 @@ EXPORT_SYMBOL(dummy_dma_ops); ...@@ -524,7 +524,8 @@ EXPORT_SYMBOL(dummy_dma_ops);
static int __init arm64_dma_init(void) static int __init arm64_dma_init(void)
{ {
if (swiotlb_force || max_pfn > (arm64_dma_phys_limit >> PAGE_SHIFT)) if (swiotlb_force == SWIOTLB_FORCE ||
max_pfn > (arm64_dma_phys_limit >> PAGE_SHIFT))
swiotlb = 1; swiotlb = 1;
return atomic_pool_init(); return atomic_pool_init();
......
...@@ -401,7 +401,8 @@ static void __init free_unused_memmap(void) ...@@ -401,7 +401,8 @@ static void __init free_unused_memmap(void)
*/ */
void __init mem_init(void) void __init mem_init(void)
{ {
if (swiotlb_force || max_pfn > (arm64_dma_phys_limit >> PAGE_SHIFT)) if (swiotlb_force == SWIOTLB_FORCE ||
max_pfn > (arm64_dma_phys_limit >> PAGE_SHIFT))
swiotlb_init(1); swiotlb_init(1);
set_max_mapnr(pfn_to_page(max_pfn) - mem_map); set_max_mapnr(pfn_to_page(max_pfn) - mem_map);
......
...@@ -68,12 +68,10 @@ static struct dma_map_ops swiotlb_dma_ops = { ...@@ -68,12 +68,10 @@ static struct dma_map_ops swiotlb_dma_ops = {
*/ */
int __init pci_swiotlb_detect_override(void) int __init pci_swiotlb_detect_override(void)
{ {
int use_swiotlb = swiotlb | swiotlb_force; if (swiotlb_force == SWIOTLB_FORCE)
if (swiotlb_force)
swiotlb = 1; swiotlb = 1;
return use_swiotlb; return swiotlb;
} }
IOMMU_INIT_FINISH(pci_swiotlb_detect_override, IOMMU_INIT_FINISH(pci_swiotlb_detect_override,
pci_xen_swiotlb_detect, pci_xen_swiotlb_detect,
......
...@@ -48,7 +48,7 @@ int __init pci_xen_swiotlb_detect(void) ...@@ -48,7 +48,7 @@ int __init pci_xen_swiotlb_detect(void)
* activate this IOMMU. If running as PV privileged, activate it * activate this IOMMU. If running as PV privileged, activate it
* irregardless. * irregardless.
*/ */
if ((xen_initial_domain() || swiotlb || swiotlb_force)) if (xen_initial_domain() || swiotlb || swiotlb_force == SWIOTLB_FORCE)
xen_swiotlb = 1; xen_swiotlb = 1;
/* If we are running under Xen, we MUST disable the native SWIOTLB. /* If we are running under Xen, we MUST disable the native SWIOTLB.
......
...@@ -2306,15 +2306,6 @@ void __i915_gem_object_put_pages(struct drm_i915_gem_object *obj, ...@@ -2306,15 +2306,6 @@ void __i915_gem_object_put_pages(struct drm_i915_gem_object *obj,
mutex_unlock(&obj->mm.lock); mutex_unlock(&obj->mm.lock);
} }
static unsigned int swiotlb_max_size(void)
{
#if IS_ENABLED(CONFIG_SWIOTLB)
return rounddown(swiotlb_nr_tbl() << IO_TLB_SHIFT, PAGE_SIZE);
#else
return 0;
#endif
}
static void i915_sg_trim(struct sg_table *orig_st) static void i915_sg_trim(struct sg_table *orig_st)
{ {
struct sg_table new_st; struct sg_table new_st;
...@@ -2362,7 +2353,7 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj) ...@@ -2362,7 +2353,7 @@ i915_gem_object_get_pages_gtt(struct drm_i915_gem_object *obj)
GEM_BUG_ON(obj->base.read_domains & I915_GEM_GPU_DOMAINS); GEM_BUG_ON(obj->base.read_domains & I915_GEM_GPU_DOMAINS);
GEM_BUG_ON(obj->base.write_domain & I915_GEM_GPU_DOMAINS); GEM_BUG_ON(obj->base.write_domain & I915_GEM_GPU_DOMAINS);
max_segment = swiotlb_max_size(); max_segment = swiotlb_max_segment();
if (!max_segment) if (!max_segment)
max_segment = rounddown(UINT_MAX, PAGE_SIZE); max_segment = rounddown(UINT_MAX, PAGE_SIZE);
......
...@@ -275,6 +275,10 @@ int __ref xen_swiotlb_init(int verbose, bool early) ...@@ -275,6 +275,10 @@ int __ref xen_swiotlb_init(int verbose, bool early)
rc = 0; rc = 0;
} else } else
rc = swiotlb_late_init_with_tbl(xen_io_tlb_start, xen_io_tlb_nslabs); rc = swiotlb_late_init_with_tbl(xen_io_tlb_start, xen_io_tlb_nslabs);
if (!rc)
swiotlb_set_max_segment(PAGE_SIZE);
return rc; return rc;
error: error:
if (repeat--) { if (repeat--) {
...@@ -392,7 +396,7 @@ dma_addr_t xen_swiotlb_map_page(struct device *dev, struct page *page, ...@@ -392,7 +396,7 @@ dma_addr_t xen_swiotlb_map_page(struct device *dev, struct page *page,
if (dma_capable(dev, dev_addr, size) && if (dma_capable(dev, dev_addr, size) &&
!range_straddles_page_boundary(phys, size) && !range_straddles_page_boundary(phys, size) &&
!xen_arch_need_swiotlb(dev, phys, dev_addr) && !xen_arch_need_swiotlb(dev, phys, dev_addr) &&
!swiotlb_force) { (swiotlb_force != SWIOTLB_FORCE)) {
/* we are not interested in the dma_addr returned by /* we are not interested in the dma_addr returned by
* xen_dma_map_page, only in the potential cache flushes executed * xen_dma_map_page, only in the potential cache flushes executed
* by the function. */ * by the function. */
...@@ -552,7 +556,7 @@ xen_swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, ...@@ -552,7 +556,7 @@ xen_swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl,
phys_addr_t paddr = sg_phys(sg); phys_addr_t paddr = sg_phys(sg);
dma_addr_t dev_addr = xen_phys_to_bus(paddr); dma_addr_t dev_addr = xen_phys_to_bus(paddr);
if (swiotlb_force || if (swiotlb_force == SWIOTLB_FORCE ||
xen_arch_need_swiotlb(hwdev, paddr, dev_addr) || xen_arch_need_swiotlb(hwdev, paddr, dev_addr) ||
!dma_capable(hwdev, dev_addr, sg->length) || !dma_capable(hwdev, dev_addr, sg->length) ||
range_straddles_page_boundary(paddr, sg->length)) { range_straddles_page_boundary(paddr, sg->length)) {
......
...@@ -9,7 +9,13 @@ struct device; ...@@ -9,7 +9,13 @@ struct device;
struct page; struct page;
struct scatterlist; struct scatterlist;
extern int swiotlb_force; enum swiotlb_force {
SWIOTLB_NORMAL, /* Default - depending on HW DMA mask etc. */
SWIOTLB_FORCE, /* swiotlb=force */
SWIOTLB_NO_FORCE, /* swiotlb=noforce */
};
extern enum swiotlb_force swiotlb_force;
/* /*
* Maximum allowable number of contiguous slabs to map, * Maximum allowable number of contiguous slabs to map,
...@@ -108,11 +114,14 @@ swiotlb_dma_supported(struct device *hwdev, u64 mask); ...@@ -108,11 +114,14 @@ swiotlb_dma_supported(struct device *hwdev, u64 mask);
#ifdef CONFIG_SWIOTLB #ifdef CONFIG_SWIOTLB
extern void __init swiotlb_free(void); extern void __init swiotlb_free(void);
unsigned int swiotlb_max_segment(void);
#else #else
static inline void swiotlb_free(void) { } static inline void swiotlb_free(void) { }
static inline unsigned int swiotlb_max_segment(void) { return 0; }
#endif #endif
extern void swiotlb_print_info(void); extern void swiotlb_print_info(void);
extern int is_swiotlb_buffer(phys_addr_t paddr); extern int is_swiotlb_buffer(phys_addr_t paddr);
extern void swiotlb_set_max_segment(unsigned int);
#endif /* __LINUX_SWIOTLB_H */ #endif /* __LINUX_SWIOTLB_H */
...@@ -11,7 +11,7 @@ TRACE_EVENT(swiotlb_bounced, ...@@ -11,7 +11,7 @@ TRACE_EVENT(swiotlb_bounced,
TP_PROTO(struct device *dev, TP_PROTO(struct device *dev,
dma_addr_t dev_addr, dma_addr_t dev_addr,
size_t size, size_t size,
int swiotlb_force), enum swiotlb_force swiotlb_force),
TP_ARGS(dev, dev_addr, size, swiotlb_force), TP_ARGS(dev, dev_addr, size, swiotlb_force),
...@@ -20,7 +20,7 @@ TRACE_EVENT(swiotlb_bounced, ...@@ -20,7 +20,7 @@ TRACE_EVENT(swiotlb_bounced,
__field( u64, dma_mask ) __field( u64, dma_mask )
__field( dma_addr_t, dev_addr ) __field( dma_addr_t, dev_addr )
__field( size_t, size ) __field( size_t, size )
__field( int, swiotlb_force ) __field( enum swiotlb_force, swiotlb_force )
), ),
TP_fast_assign( TP_fast_assign(
...@@ -37,7 +37,10 @@ TRACE_EVENT(swiotlb_bounced, ...@@ -37,7 +37,10 @@ TRACE_EVENT(swiotlb_bounced,
__entry->dma_mask, __entry->dma_mask,
(unsigned long long)__entry->dev_addr, (unsigned long long)__entry->dev_addr,
__entry->size, __entry->size,
__entry->swiotlb_force ? "swiotlb_force" : "" ) __print_symbolic(__entry->swiotlb_force,
{ SWIOTLB_NORMAL, "NORMAL" },
{ SWIOTLB_FORCE, "FORCE" },
{ SWIOTLB_NO_FORCE, "NO_FORCE" }))
); );
#endif /* _TRACE_SWIOTLB_H */ #endif /* _TRACE_SWIOTLB_H */
......
...@@ -53,7 +53,7 @@ ...@@ -53,7 +53,7 @@
*/ */
#define IO_TLB_MIN_SLABS ((1<<20) >> IO_TLB_SHIFT) #define IO_TLB_MIN_SLABS ((1<<20) >> IO_TLB_SHIFT)
int swiotlb_force; enum swiotlb_force swiotlb_force;
/* /*
* Used to do a quick range check in swiotlb_tbl_unmap_single and * Used to do a quick range check in swiotlb_tbl_unmap_single and
...@@ -82,6 +82,12 @@ static phys_addr_t io_tlb_overflow_buffer; ...@@ -82,6 +82,12 @@ static phys_addr_t io_tlb_overflow_buffer;
static unsigned int *io_tlb_list; static unsigned int *io_tlb_list;
static unsigned int io_tlb_index; static unsigned int io_tlb_index;
/*
* Max segment that we can provide which (if pages are contingous) will
* not be bounced (unless SWIOTLB_FORCE is set).
*/
unsigned int max_segment;
/* /*
* We need to save away the original address corresponding to a mapped entry * We need to save away the original address corresponding to a mapped entry
* for the sync operations. * for the sync operations.
...@@ -106,8 +112,12 @@ setup_io_tlb_npages(char *str) ...@@ -106,8 +112,12 @@ setup_io_tlb_npages(char *str)
} }
if (*str == ',') if (*str == ',')
++str; ++str;
if (!strcmp(str, "force")) if (!strcmp(str, "force")) {
swiotlb_force = 1; swiotlb_force = SWIOTLB_FORCE;
} else if (!strcmp(str, "noforce")) {
swiotlb_force = SWIOTLB_NO_FORCE;
io_tlb_nslabs = 1;
}
return 0; return 0;
} }
...@@ -120,6 +130,20 @@ unsigned long swiotlb_nr_tbl(void) ...@@ -120,6 +130,20 @@ unsigned long swiotlb_nr_tbl(void)
} }
EXPORT_SYMBOL_GPL(swiotlb_nr_tbl); EXPORT_SYMBOL_GPL(swiotlb_nr_tbl);
unsigned int swiotlb_max_segment(void)
{
return max_segment;
}
EXPORT_SYMBOL_GPL(swiotlb_max_segment);
void swiotlb_set_max_segment(unsigned int val)
{
if (swiotlb_force == SWIOTLB_FORCE)
max_segment = 1;
else
max_segment = rounddown(val, PAGE_SIZE);
}
/* default to 64MB */ /* default to 64MB */
#define IO_TLB_DEFAULT_SIZE (64UL<<20) #define IO_TLB_DEFAULT_SIZE (64UL<<20)
unsigned long swiotlb_size_or_default(void) unsigned long swiotlb_size_or_default(void)
...@@ -201,6 +225,7 @@ int __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose) ...@@ -201,6 +225,7 @@ int __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose)
if (verbose) if (verbose)
swiotlb_print_info(); swiotlb_print_info();
swiotlb_set_max_segment(io_tlb_nslabs << IO_TLB_SHIFT);
return 0; return 0;
} }
...@@ -279,6 +304,7 @@ swiotlb_late_init_with_default_size(size_t default_size) ...@@ -279,6 +304,7 @@ swiotlb_late_init_with_default_size(size_t default_size)
rc = swiotlb_late_init_with_tbl(vstart, io_tlb_nslabs); rc = swiotlb_late_init_with_tbl(vstart, io_tlb_nslabs);
if (rc) if (rc)
free_pages((unsigned long)vstart, order); free_pages((unsigned long)vstart, order);
return rc; return rc;
} }
...@@ -333,6 +359,8 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -333,6 +359,8 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
late_alloc = 1; late_alloc = 1;
swiotlb_set_max_segment(io_tlb_nslabs << IO_TLB_SHIFT);
return 0; return 0;
cleanup4: cleanup4:
...@@ -347,6 +375,7 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -347,6 +375,7 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
io_tlb_end = 0; io_tlb_end = 0;
io_tlb_start = 0; io_tlb_start = 0;
io_tlb_nslabs = 0; io_tlb_nslabs = 0;
max_segment = 0;
return -ENOMEM; return -ENOMEM;
} }
...@@ -375,6 +404,7 @@ void __init swiotlb_free(void) ...@@ -375,6 +404,7 @@ void __init swiotlb_free(void)
PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT));
} }
io_tlb_nslabs = 0; io_tlb_nslabs = 0;
max_segment = 0;
} }
int is_swiotlb_buffer(phys_addr_t paddr) int is_swiotlb_buffer(phys_addr_t paddr)
...@@ -543,8 +573,15 @@ static phys_addr_t ...@@ -543,8 +573,15 @@ static phys_addr_t
map_single(struct device *hwdev, phys_addr_t phys, size_t size, map_single(struct device *hwdev, phys_addr_t phys, size_t size,
enum dma_data_direction dir, unsigned long attrs) enum dma_data_direction dir, unsigned long attrs)
{ {
dma_addr_t start_dma_addr = phys_to_dma(hwdev, io_tlb_start); dma_addr_t start_dma_addr;
if (swiotlb_force == SWIOTLB_NO_FORCE) {
dev_warn_ratelimited(hwdev, "Cannot do DMA to address %pa\n",
&phys);
return SWIOTLB_MAP_ERROR;
}
start_dma_addr = phys_to_dma(hwdev, io_tlb_start);
return swiotlb_tbl_map_single(hwdev, start_dma_addr, phys, size, return swiotlb_tbl_map_single(hwdev, start_dma_addr, phys, size,
dir, attrs); dir, attrs);
} }
...@@ -721,6 +758,9 @@ static void ...@@ -721,6 +758,9 @@ static void
swiotlb_full(struct device *dev, size_t size, enum dma_data_direction dir, swiotlb_full(struct device *dev, size_t size, enum dma_data_direction dir,
int do_panic) int do_panic)
{ {
if (swiotlb_force == SWIOTLB_NO_FORCE)
return;
/* /*
* Ran out of IOMMU space for this operation. This is very bad. * Ran out of IOMMU space for this operation. This is very bad.
* Unfortunately the drivers cannot handle this operation properly. * Unfortunately the drivers cannot handle this operation properly.
...@@ -763,7 +803,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, ...@@ -763,7 +803,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
* we can safely return the device addr and not worry about bounce * we can safely return the device addr and not worry about bounce
* buffering it. * buffering it.
*/ */
if (dma_capable(dev, dev_addr, size) && !swiotlb_force) if (dma_capable(dev, dev_addr, size) && swiotlb_force != SWIOTLB_FORCE)
return dev_addr; return dev_addr;
trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force); trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force);
...@@ -904,7 +944,7 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems, ...@@ -904,7 +944,7 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems,
phys_addr_t paddr = sg_phys(sg); phys_addr_t paddr = sg_phys(sg);
dma_addr_t dev_addr = phys_to_dma(hwdev, paddr); dma_addr_t dev_addr = phys_to_dma(hwdev, paddr);
if (swiotlb_force || if (swiotlb_force == SWIOTLB_FORCE ||
!dma_capable(hwdev, dev_addr, sg->length)) { !dma_capable(hwdev, dev_addr, sg->length)) {
phys_addr_t map = map_single(hwdev, sg_phys(sg), phys_addr_t map = map_single(hwdev, sg_phys(sg),
sg->length, dir, attrs); sg->length, dir, attrs);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment