Commit 7121dd91 authored by Frederic Weisbecker's avatar Frederic Weisbecker Committed by Neeraj Upadhyay

rcu/nocb: Introduce nocb mutex

The barrier_mutex is used currently to protect (de-)offloading
operations and prevent from nocb_lock locking imbalance in rcu_barrier()
and shrinker, and also from misordered RCU barrier invocation.

Now since RCU (de-)offloading is going to happen on offline CPUs, an RCU
barrier will have to be executed while transitionning from offloaded to
de-offloaded state. And this can't happen while holding the
barrier_mutex.

Introduce a NOCB mutex to protect (de-)offloading transitions. The
barrier_mutex is still held for now when necessary to avoid barrier
callbacks reordering and nocb_lock imbalance.
Signed-off-by: default avatarFrederic Weisbecker <frederic@kernel.org>
Signed-off-by: default avatarPaul E. McKenney <paulmck@kernel.org>
Reviewed-by: default avatarPaul E. McKenney <paulmck@kernel.org>
Signed-off-by: default avatarNeeraj Upadhyay <neeraj.upadhyay@kernel.org>
parent 7be88a85
...@@ -97,6 +97,9 @@ static struct rcu_state rcu_state = { ...@@ -97,6 +97,9 @@ static struct rcu_state rcu_state = {
.srs_cleanup_work = __WORK_INITIALIZER(rcu_state.srs_cleanup_work, .srs_cleanup_work = __WORK_INITIALIZER(rcu_state.srs_cleanup_work,
rcu_sr_normal_gp_cleanup_work), rcu_sr_normal_gp_cleanup_work),
.srs_cleanups_pending = ATOMIC_INIT(0), .srs_cleanups_pending = ATOMIC_INIT(0),
#ifdef CONFIG_RCU_NOCB_CPU
.nocb_mutex = __MUTEX_INITIALIZER(rcu_state.nocb_mutex),
#endif
}; };
/* Dump rcu_node combining tree at boot to verify correct setup. */ /* Dump rcu_node combining tree at boot to verify correct setup. */
......
...@@ -421,6 +421,7 @@ struct rcu_state { ...@@ -421,6 +421,7 @@ struct rcu_state {
atomic_t srs_cleanups_pending; /* srs inflight worker cleanups. */ atomic_t srs_cleanups_pending; /* srs inflight worker cleanups. */
#ifdef CONFIG_RCU_NOCB_CPU #ifdef CONFIG_RCU_NOCB_CPU
struct mutex nocb_mutex; /* Guards (de-)offloading */
int nocb_is_setup; /* nocb is setup from boot */ int nocb_is_setup; /* nocb is setup from boot */
#endif #endif
}; };
......
...@@ -1141,6 +1141,7 @@ int rcu_nocb_cpu_deoffload(int cpu) ...@@ -1141,6 +1141,7 @@ int rcu_nocb_cpu_deoffload(int cpu)
int ret = 0; int ret = 0;
cpus_read_lock(); cpus_read_lock();
mutex_lock(&rcu_state.nocb_mutex);
mutex_lock(&rcu_state.barrier_mutex); mutex_lock(&rcu_state.barrier_mutex);
if (rcu_rdp_is_offloaded(rdp)) { if (rcu_rdp_is_offloaded(rdp)) {
if (cpu_online(cpu)) { if (cpu_online(cpu)) {
...@@ -1153,6 +1154,7 @@ int rcu_nocb_cpu_deoffload(int cpu) ...@@ -1153,6 +1154,7 @@ int rcu_nocb_cpu_deoffload(int cpu)
} }
} }
mutex_unlock(&rcu_state.barrier_mutex); mutex_unlock(&rcu_state.barrier_mutex);
mutex_unlock(&rcu_state.nocb_mutex);
cpus_read_unlock(); cpus_read_unlock();
return ret; return ret;
...@@ -1228,6 +1230,7 @@ int rcu_nocb_cpu_offload(int cpu) ...@@ -1228,6 +1230,7 @@ int rcu_nocb_cpu_offload(int cpu)
int ret = 0; int ret = 0;
cpus_read_lock(); cpus_read_lock();
mutex_lock(&rcu_state.nocb_mutex);
mutex_lock(&rcu_state.barrier_mutex); mutex_lock(&rcu_state.barrier_mutex);
if (!rcu_rdp_is_offloaded(rdp)) { if (!rcu_rdp_is_offloaded(rdp)) {
if (cpu_online(cpu)) { if (cpu_online(cpu)) {
...@@ -1240,6 +1243,7 @@ int rcu_nocb_cpu_offload(int cpu) ...@@ -1240,6 +1243,7 @@ int rcu_nocb_cpu_offload(int cpu)
} }
} }
mutex_unlock(&rcu_state.barrier_mutex); mutex_unlock(&rcu_state.barrier_mutex);
mutex_unlock(&rcu_state.nocb_mutex);
cpus_read_unlock(); cpus_read_unlock();
return ret; return ret;
...@@ -1257,7 +1261,7 @@ lazy_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc) ...@@ -1257,7 +1261,7 @@ lazy_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
return 0; return 0;
/* Protect rcu_nocb_mask against concurrent (de-)offloading. */ /* Protect rcu_nocb_mask against concurrent (de-)offloading. */
if (!mutex_trylock(&rcu_state.barrier_mutex)) if (!mutex_trylock(&rcu_state.nocb_mutex))
return 0; return 0;
/* Snapshot count of all CPUs */ /* Snapshot count of all CPUs */
...@@ -1267,7 +1271,7 @@ lazy_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc) ...@@ -1267,7 +1271,7 @@ lazy_rcu_shrink_count(struct shrinker *shrink, struct shrink_control *sc)
count += READ_ONCE(rdp->lazy_len); count += READ_ONCE(rdp->lazy_len);
} }
mutex_unlock(&rcu_state.barrier_mutex); mutex_unlock(&rcu_state.nocb_mutex);
return count ? count : SHRINK_EMPTY; return count ? count : SHRINK_EMPTY;
} }
...@@ -1285,9 +1289,9 @@ lazy_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) ...@@ -1285,9 +1289,9 @@ lazy_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
* Protect against concurrent (de-)offloading. Otherwise nocb locking * Protect against concurrent (de-)offloading. Otherwise nocb locking
* may be ignored or imbalanced. * may be ignored or imbalanced.
*/ */
if (!mutex_trylock(&rcu_state.barrier_mutex)) { if (!mutex_trylock(&rcu_state.nocb_mutex)) {
/* /*
* But really don't insist if barrier_mutex is contended since we * But really don't insist if nocb_mutex is contended since we
* can't guarantee that it will never engage in a dependency * can't guarantee that it will never engage in a dependency
* chain involving memory allocation. The lock is seldom contended * chain involving memory allocation. The lock is seldom contended
* anyway. * anyway.
...@@ -1326,7 +1330,7 @@ lazy_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) ...@@ -1326,7 +1330,7 @@ lazy_rcu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc)
break; break;
} }
mutex_unlock(&rcu_state.barrier_mutex); mutex_unlock(&rcu_state.nocb_mutex);
return count ? count : SHRINK_STOP; return count ? count : SHRINK_STOP;
} }
...@@ -1473,15 +1477,15 @@ static void rcu_spawn_cpu_nocb_kthread(int cpu) ...@@ -1473,15 +1477,15 @@ static void rcu_spawn_cpu_nocb_kthread(int cpu)
* No need to protect against concurrent rcu_barrier() * No need to protect against concurrent rcu_barrier()
* because the number of callbacks should be 0 for a non-boot CPU, * because the number of callbacks should be 0 for a non-boot CPU,
* therefore rcu_barrier() shouldn't even try to grab the nocb_lock. * therefore rcu_barrier() shouldn't even try to grab the nocb_lock.
* But hold barrier_mutex to avoid nocb_lock imbalance from shrinker. * But hold nocb_mutex to avoid nocb_lock imbalance from shrinker.
*/ */
WARN_ON_ONCE(system_state > SYSTEM_BOOTING && rcu_segcblist_n_cbs(&rdp->cblist)); WARN_ON_ONCE(system_state > SYSTEM_BOOTING && rcu_segcblist_n_cbs(&rdp->cblist));
mutex_lock(&rcu_state.barrier_mutex); mutex_lock(&rcu_state.nocb_mutex);
if (rcu_rdp_is_offloaded(rdp)) { if (rcu_rdp_is_offloaded(rdp)) {
rcu_nocb_rdp_deoffload(rdp); rcu_nocb_rdp_deoffload(rdp);
cpumask_clear_cpu(cpu, rcu_nocb_mask); cpumask_clear_cpu(cpu, rcu_nocb_mask);
} }
mutex_unlock(&rcu_state.barrier_mutex); mutex_unlock(&rcu_state.nocb_mutex);
} }
/* How many CB CPU IDs per GP kthread? Default of -1 for sqrt(nr_cpu_ids). */ /* How many CB CPU IDs per GP kthread? Default of -1 for sqrt(nr_cpu_ids). */
......
...@@ -28,6 +28,7 @@ static bool rcu_rdp_is_offloaded(struct rcu_data *rdp) ...@@ -28,6 +28,7 @@ static bool rcu_rdp_is_offloaded(struct rcu_data *rdp)
!(lockdep_is_held(&rcu_state.barrier_mutex) || !(lockdep_is_held(&rcu_state.barrier_mutex) ||
(IS_ENABLED(CONFIG_HOTPLUG_CPU) && lockdep_is_cpus_held()) || (IS_ENABLED(CONFIG_HOTPLUG_CPU) && lockdep_is_cpus_held()) ||
lockdep_is_held(&rdp->nocb_lock) || lockdep_is_held(&rdp->nocb_lock) ||
lockdep_is_held(&rcu_state.nocb_mutex) ||
(!(IS_ENABLED(CONFIG_PREEMPT_COUNT) && preemptible()) && (!(IS_ENABLED(CONFIG_PREEMPT_COUNT) && preemptible()) &&
rdp == this_cpu_ptr(&rcu_data)) || rdp == this_cpu_ptr(&rcu_data)) ||
rcu_current_is_nocb_kthread(rdp)), rcu_current_is_nocb_kthread(rdp)),
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment