Commit 6251f997 authored by Tejun Heo's avatar Tejun Heo

percpu_ref: minor code and comment updates

* Some comments became stale.  Updated.
* percpu_ref_tryget() unnecessarily initializes @ret.  Removed.
* A blank line removed from percpu_ref_kill_rcu().
* Explicit function name in a WARN format string replaced with __func__.
* WARN_ON() in percpu_ref_reinit() converted to WARN_ON_ONCE().
Signed-off-by: default avatarTejun Heo <tj@kernel.org>
Reviewed-by: default avatarKent Overstreet <kmo@daterainc.com>
parent a2237370
...@@ -115,8 +115,10 @@ static inline bool __pcpu_ref_alive(struct percpu_ref *ref, ...@@ -115,8 +115,10 @@ static inline bool __pcpu_ref_alive(struct percpu_ref *ref,
* percpu_ref_get - increment a percpu refcount * percpu_ref_get - increment a percpu refcount
* @ref: percpu_ref to get * @ref: percpu_ref to get
* *
* Analagous to atomic_inc(). * Analagous to atomic_long_inc().
*/ *
* This function is safe to call as long as @ref is between init and exit.
*/
static inline void percpu_ref_get(struct percpu_ref *ref) static inline void percpu_ref_get(struct percpu_ref *ref)
{ {
unsigned long __percpu *pcpu_count; unsigned long __percpu *pcpu_count;
...@@ -138,12 +140,12 @@ static inline void percpu_ref_get(struct percpu_ref *ref) ...@@ -138,12 +140,12 @@ static inline void percpu_ref_get(struct percpu_ref *ref)
* Increment a percpu refcount unless its count already reached zero. * Increment a percpu refcount unless its count already reached zero.
* Returns %true on success; %false on failure. * Returns %true on success; %false on failure.
* *
* The caller is responsible for ensuring that @ref stays accessible. * This function is safe to call as long as @ref is between init and exit.
*/ */
static inline bool percpu_ref_tryget(struct percpu_ref *ref) static inline bool percpu_ref_tryget(struct percpu_ref *ref)
{ {
unsigned long __percpu *pcpu_count; unsigned long __percpu *pcpu_count;
int ret = false; int ret;
rcu_read_lock_sched(); rcu_read_lock_sched();
...@@ -166,12 +168,13 @@ static inline bool percpu_ref_tryget(struct percpu_ref *ref) ...@@ -166,12 +168,13 @@ static inline bool percpu_ref_tryget(struct percpu_ref *ref)
* Increment a percpu refcount unless it has already been killed. Returns * Increment a percpu refcount unless it has already been killed. Returns
* %true on success; %false on failure. * %true on success; %false on failure.
* *
* Completion of percpu_ref_kill() in itself doesn't guarantee that tryget * Completion of percpu_ref_kill() in itself doesn't guarantee that this
* will fail. For such guarantee, percpu_ref_kill_and_confirm() should be * function will fail. For such guarantee, percpu_ref_kill_and_confirm()
* used. After the confirm_kill callback is invoked, it's guaranteed that * should be used. After the confirm_kill callback is invoked, it's
* no new reference will be given out by percpu_ref_tryget(). * guaranteed that no new reference will be given out by
* percpu_ref_tryget_live().
* *
* The caller is responsible for ensuring that @ref stays accessible. * This function is safe to call as long as @ref is between init and exit.
*/ */
static inline bool percpu_ref_tryget_live(struct percpu_ref *ref) static inline bool percpu_ref_tryget_live(struct percpu_ref *ref)
{ {
...@@ -196,6 +199,8 @@ static inline bool percpu_ref_tryget_live(struct percpu_ref *ref) ...@@ -196,6 +199,8 @@ static inline bool percpu_ref_tryget_live(struct percpu_ref *ref)
* *
* Decrement the refcount, and if 0, call the release function (which was passed * Decrement the refcount, and if 0, call the release function (which was passed
* to percpu_ref_init()) * to percpu_ref_init())
*
* This function is safe to call as long as @ref is between init and exit.
*/ */
static inline void percpu_ref_put(struct percpu_ref *ref) static inline void percpu_ref_put(struct percpu_ref *ref)
{ {
...@@ -216,6 +221,8 @@ static inline void percpu_ref_put(struct percpu_ref *ref) ...@@ -216,6 +221,8 @@ static inline void percpu_ref_put(struct percpu_ref *ref)
* @ref: percpu_ref to test * @ref: percpu_ref to test
* *
* Returns %true if @ref reached zero. * Returns %true if @ref reached zero.
*
* This function is safe to call as long as @ref is between init and exit.
*/ */
static inline bool percpu_ref_is_zero(struct percpu_ref *ref) static inline bool percpu_ref_is_zero(struct percpu_ref *ref)
{ {
......
...@@ -108,7 +108,6 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu) ...@@ -108,7 +108,6 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu)
* reaching 0 before we add the percpu counts. But doing it at the same * reaching 0 before we add the percpu counts. But doing it at the same
* time is equivalent and saves us atomic operations: * time is equivalent and saves us atomic operations:
*/ */
atomic_long_add((long)count - PCPU_COUNT_BIAS, &ref->count); atomic_long_add((long)count - PCPU_COUNT_BIAS, &ref->count);
WARN_ONCE(atomic_long_read(&ref->count) <= 0, WARN_ONCE(atomic_long_read(&ref->count) <= 0,
...@@ -120,8 +119,8 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu) ...@@ -120,8 +119,8 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu)
ref->confirm_kill(ref); ref->confirm_kill(ref);
/* /*
* Now we're in single atomic_t mode with a consistent refcount, so it's * Now we're in single atomic_long_t mode with a consistent
* safe to drop our initial ref: * refcount, so it's safe to drop our initial ref:
*/ */
percpu_ref_put(ref); percpu_ref_put(ref);
} }
...@@ -134,8 +133,8 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu) ...@@ -134,8 +133,8 @@ static void percpu_ref_kill_rcu(struct rcu_head *rcu)
* Equivalent to percpu_ref_kill() but also schedules kill confirmation if * Equivalent to percpu_ref_kill() but also schedules kill confirmation if
* @confirm_kill is not NULL. @confirm_kill, which may not block, will be * @confirm_kill is not NULL. @confirm_kill, which may not block, will be
* called after @ref is seen as dead from all CPUs - all further * called after @ref is seen as dead from all CPUs - all further
* invocations of percpu_ref_tryget() will fail. See percpu_ref_tryget() * invocations of percpu_ref_tryget_live() will fail. See
* for more details. * percpu_ref_tryget_live() for more details.
* *
* Due to the way percpu_ref is implemented, @confirm_kill will be called * Due to the way percpu_ref is implemented, @confirm_kill will be called
* after at least one full RCU grace period has passed but this is an * after at least one full RCU grace period has passed but this is an
...@@ -145,8 +144,7 @@ void percpu_ref_kill_and_confirm(struct percpu_ref *ref, ...@@ -145,8 +144,7 @@ void percpu_ref_kill_and_confirm(struct percpu_ref *ref,
percpu_ref_func_t *confirm_kill) percpu_ref_func_t *confirm_kill)
{ {
WARN_ONCE(ref->pcpu_count_ptr & PCPU_REF_DEAD, WARN_ONCE(ref->pcpu_count_ptr & PCPU_REF_DEAD,
"percpu_ref_kill() called more than once on %pf!", "%s called more than once on %pf!", __func__, ref->release);
ref->release);
ref->pcpu_count_ptr |= PCPU_REF_DEAD; ref->pcpu_count_ptr |= PCPU_REF_DEAD;
ref->confirm_kill = confirm_kill; ref->confirm_kill = confirm_kill;
...@@ -172,7 +170,7 @@ void percpu_ref_reinit(struct percpu_ref *ref) ...@@ -172,7 +170,7 @@ void percpu_ref_reinit(struct percpu_ref *ref)
int cpu; int cpu;
BUG_ON(!pcpu_count); BUG_ON(!pcpu_count);
WARN_ON(!percpu_ref_is_zero(ref)); WARN_ON_ONCE(!percpu_ref_is_zero(ref));
atomic_long_set(&ref->count, 1 + PCPU_COUNT_BIAS); atomic_long_set(&ref->count, 1 + PCPU_COUNT_BIAS);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment