Commit 7904ba8a authored by Peter Zijlstra's avatar Peter Zijlstra Committed by Thomas Gleixner

x86/mm/cpa: Optimize __cpa_flush_range()

If we IPI for WBINDV, then we might as well kill the entire TLB too.
But if we don't have to invalidate cache, there is no reason not to
use a range TLB flush.
Signed-off-by: default avatarPeter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
Reviewed-by: default avatarDave Hansen <dave.hansen@intel.com>
Cc: Bin Yang <bin.yang@intel.com>
Cc: Mark Gross <mark.gross@intel.com>
Link: https://lkml.kernel.org/r/20180919085948.195633798@infradead.org
parent 47e262ac
...@@ -291,7 +291,7 @@ static bool __cpa_flush_range(unsigned long start, int numpages, int cache) ...@@ -291,7 +291,7 @@ static bool __cpa_flush_range(unsigned long start, int numpages, int cache)
WARN_ON(PAGE_ALIGN(start) != start); WARN_ON(PAGE_ALIGN(start) != start);
if (!static_cpu_has(X86_FEATURE_CLFLUSH)) { if (cache && !static_cpu_has(X86_FEATURE_CLFLUSH)) {
cpa_flush_all(cache); cpa_flush_all(cache);
return true; return true;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment