Commit 18f4b843 authored by Tvrtko Ursulin's avatar Tvrtko Ursulin

drm/i915: Use atomic waits for short non-atomic ones

usleep_range is not recommended for waits shorten than 10us.

Make the wait_for_us use the atomic variant for such waits.

To do so we need to reimplement the _wait_for_atomic macro to
be safe with regards to preemption and interrupts.

v2: Reimplement _wait_for_atomic to be irq and preemption safe.
    (Chris Wilson and Imre Deak)

v3: Fixed in_atomic check due rebase error.
v4: Build bug on non-constant timeouts.
v5: Compile away cpu migration code in atomic paths.
Signed-off-by: default avatarTvrtko Ursulin <tvrtko.ursulin@intel.com>
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Imre Deak <imre.deak@intel.com>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: default avatarChris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1467114710-29989-1-git-send-email-tvrtko.ursulin@linux.intel.com
parent c68b0ab2
...@@ -69,39 +69,63 @@ ...@@ -69,39 +69,63 @@
}) })
#define wait_for(COND, MS) _wait_for((COND), (MS) * 1000, 1000) #define wait_for(COND, MS) _wait_for((COND), (MS) * 1000, 1000)
#define wait_for_us(COND, US) _wait_for((COND), (US), 1)
/* If CONFIG_PREEMPT_COUNT is disabled, in_atomic() always reports false. */ /* If CONFIG_PREEMPT_COUNT is disabled, in_atomic() always reports false. */
#if defined(CONFIG_DRM_I915_DEBUG) && defined(CONFIG_PREEMPT_COUNT) #if defined(CONFIG_DRM_I915_DEBUG) && defined(CONFIG_PREEMPT_COUNT)
# define _WAIT_FOR_ATOMIC_CHECK WARN_ON_ONCE(!in_atomic()) # define _WAIT_FOR_ATOMIC_CHECK(ATOMIC) WARN_ON_ONCE((ATOMIC) && !in_atomic())
#else #else
# define _WAIT_FOR_ATOMIC_CHECK do { } while (0) # define _WAIT_FOR_ATOMIC_CHECK(ATOMIC) do { } while (0)
#endif #endif
#define _wait_for_atomic(COND, US) ({ \ #define _wait_for_atomic(COND, US, ATOMIC) \
unsigned long end__; \ ({ \
int ret__ = 0; \ int cpu, ret, timeout = (US) * 1000; \
_WAIT_FOR_ATOMIC_CHECK; \ u64 base; \
_WAIT_FOR_ATOMIC_CHECK(ATOMIC); \
BUILD_BUG_ON((US) > 50000); \ BUILD_BUG_ON((US) > 50000); \
end__ = (local_clock() >> 10) + (US) + 1; \ if (!(ATOMIC)) { \
while (!(COND)) { \ preempt_disable(); \
if (time_after((unsigned long)(local_clock() >> 10), end__)) { \ cpu = smp_processor_id(); \
/* Unlike the regular wait_for(), this atomic variant \ } \
* cannot be preempted (and we'll just ignore the issue\ base = local_clock(); \
* of irq interruptions) and so we know that no time \ for (;;) { \
* has passed since the last check of COND and can \ u64 now = local_clock(); \
* immediately report the timeout. \ if (!(ATOMIC)) \
*/ \ preempt_enable(); \
ret__ = -ETIMEDOUT; \ if (COND) { \
ret = 0; \
break; \
} \
if (now - base >= timeout) { \
ret = -ETIMEDOUT; \
break; \ break; \
} \ } \
cpu_relax(); \ cpu_relax(); \
if (!(ATOMIC)) { \
preempt_disable(); \
if (unlikely(cpu != smp_processor_id())) { \
timeout -= now - base; \
cpu = smp_processor_id(); \
base = local_clock(); \
} \
} \
} \ } \
ret; \
})
#define wait_for_us(COND, US) \
({ \
int ret__; \
BUILD_BUG_ON(!__builtin_constant_p(US)); \
if ((US) > 10) \
ret__ = _wait_for((COND), (US), 10); \
else \
ret__ = _wait_for_atomic((COND), (US), 0); \
ret__; \ ret__; \
}) })
#define wait_for_atomic(COND, MS) _wait_for_atomic((COND), (MS) * 1000) #define wait_for_atomic(COND, MS) _wait_for_atomic((COND), (MS) * 1000, 1)
#define wait_for_atomic_us(COND, US) _wait_for_atomic((COND), (US)) #define wait_for_atomic_us(COND, US) _wait_for_atomic((COND), (US), 1)
#define KHz(x) (1000 * (x)) #define KHz(x) (1000 * (x))
#define MHz(x) KHz(1000 * (x)) #define MHz(x) KHz(1000 * (x))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment