Commit 7d762e49 authored by Sebastian Andrzej Siewior's avatar Sebastian Andrzej Siewior Committed by Thomas Gleixner

perf/x86/amd/uncore: Prevent use after free

The resent conversion of the cpu hotplug support in the uncore driver
introduced a regression due to the way the callbacks are invoked at
initialization time.

The old code called the prepare/starting/online function on each online cpu
as a block. The new code registers the hotplug callbacks in the core for
each state. The core invokes the callbacks at each registration on all
online cpus.

The code implicitely relied on the prepare/starting/online callbacks being
called as combo on a particular cpu, which was not obvious and completely
undocumented.

The resulting subtle wreckage happens due to the way how the uncore code
manages shared data structures for cpus which share an uncore resource in
hardware. The sharing is determined in the cpu starting callback, but the
prepare callback allocates per cpu data for the upcoming cpu because
potential sharing is unknown at this point. If the starting callback finds
a online cpu which shares the hardware resource it takes a refcount on the
percpu data of that cpu and puts the own data structure into a
'free_at_online' pointer of that shared data structure. The online callback
frees that.

With the old model this worked because in a starting callback only one non
unused structure (the one of the starting cpu) was available. The new code
allocates the data structures for all cpus when the prepare callback is
registered.

Now the starting function iterates through all online cpus and looks for a
data structure (skipping its own) which has a matching hardware id. The id
member of the data structure is initialized to 0, but the hardware id can
be 0 as well. The resulting wreckage is:

  CPU0 finds a matching id on CPU1, takes a refcount on CPU1 data and puts
  its own data structure into CPU1s data structure to be freed.

  CPU1 skips CPU0 because the data structure is its allegedly unsued own.
  It finds a matching id on CPU2, takes a refcount on CPU1 data and puts
  its own data structure into CPU2s data structure to be freed.

  ....

Now the online callbacks are invoked.

  CPU0 has a pointer to CPU1s data and frees the original CPU0 data. So
  far so good.

  CPU1 has a pointer to CPU2s data and frees the original CPU1 data, which
  is still referenced by CPU0 ---> Booom

So there are two issues to be solved here:

1) The id field must be initialized at allocation time to a value which
   cannot be a valid hardware id, i.e. -1

   This prevents the above scenario, but now CPU1 and CPU2 both stick their
   own data structure into the free_at_online pointer of CPU0. So we leak
   CPU1s data structure.

2) Fix the memory leak described in #1

   Instead of having a single pointer, use a hlist to enqueue the
   superflous data structures which are then freed by the first cpu
   invoking the online callback.

Ideally we should know the sharing _before_ invoking the prepare callback,
but that's way beyond the scope of this bug fix.

[ tglx: Rewrote changelog ]

Fixes: 96b2bd38 ("perf/x86/amd/uncore: Convert to hotplug state machine")
Reported-and-tested-by: default avatarEric Sandeen <sandeen@sandeen.net>
Signed-off-by: default avatarSebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: Borislav Petkov <bp@suse.de>
Link: http://lkml.kernel.org/r/20160909160822.lowgmkdwms2dheyv@linutronix.deSigned-off-by: default avatarThomas Gleixner <tglx@linutronix.de>
parent 79d102cb
...@@ -29,6 +29,8 @@ ...@@ -29,6 +29,8 @@
#define COUNTER_SHIFT 16 #define COUNTER_SHIFT 16
static HLIST_HEAD(uncore_unused_list);
struct amd_uncore { struct amd_uncore {
int id; int id;
int refcnt; int refcnt;
...@@ -39,7 +41,7 @@ struct amd_uncore { ...@@ -39,7 +41,7 @@ struct amd_uncore {
cpumask_t *active_mask; cpumask_t *active_mask;
struct pmu *pmu; struct pmu *pmu;
struct perf_event *events[MAX_COUNTERS]; struct perf_event *events[MAX_COUNTERS];
struct amd_uncore *free_when_cpu_online; struct hlist_node node;
}; };
static struct amd_uncore * __percpu *amd_uncore_nb; static struct amd_uncore * __percpu *amd_uncore_nb;
...@@ -306,6 +308,7 @@ static int amd_uncore_cpu_up_prepare(unsigned int cpu) ...@@ -306,6 +308,7 @@ static int amd_uncore_cpu_up_prepare(unsigned int cpu)
uncore_nb->msr_base = MSR_F15H_NB_PERF_CTL; uncore_nb->msr_base = MSR_F15H_NB_PERF_CTL;
uncore_nb->active_mask = &amd_nb_active_mask; uncore_nb->active_mask = &amd_nb_active_mask;
uncore_nb->pmu = &amd_nb_pmu; uncore_nb->pmu = &amd_nb_pmu;
uncore_nb->id = -1;
*per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb; *per_cpu_ptr(amd_uncore_nb, cpu) = uncore_nb;
} }
...@@ -319,6 +322,7 @@ static int amd_uncore_cpu_up_prepare(unsigned int cpu) ...@@ -319,6 +322,7 @@ static int amd_uncore_cpu_up_prepare(unsigned int cpu)
uncore_l2->msr_base = MSR_F16H_L2I_PERF_CTL; uncore_l2->msr_base = MSR_F16H_L2I_PERF_CTL;
uncore_l2->active_mask = &amd_l2_active_mask; uncore_l2->active_mask = &amd_l2_active_mask;
uncore_l2->pmu = &amd_l2_pmu; uncore_l2->pmu = &amd_l2_pmu;
uncore_l2->id = -1;
*per_cpu_ptr(amd_uncore_l2, cpu) = uncore_l2; *per_cpu_ptr(amd_uncore_l2, cpu) = uncore_l2;
} }
...@@ -348,7 +352,7 @@ amd_uncore_find_online_sibling(struct amd_uncore *this, ...@@ -348,7 +352,7 @@ amd_uncore_find_online_sibling(struct amd_uncore *this,
continue; continue;
if (this->id == that->id) { if (this->id == that->id) {
that->free_when_cpu_online = this; hlist_add_head(&this->node, &uncore_unused_list);
this = that; this = that;
break; break;
} }
...@@ -388,13 +392,23 @@ static int amd_uncore_cpu_starting(unsigned int cpu) ...@@ -388,13 +392,23 @@ static int amd_uncore_cpu_starting(unsigned int cpu)
return 0; return 0;
} }
static void uncore_clean_online(void)
{
struct amd_uncore *uncore;
struct hlist_node *n;
hlist_for_each_entry_safe(uncore, n, &uncore_unused_list, node) {
hlist_del(&uncore->node);
kfree(uncore);
}
}
static void uncore_online(unsigned int cpu, static void uncore_online(unsigned int cpu,
struct amd_uncore * __percpu *uncores) struct amd_uncore * __percpu *uncores)
{ {
struct amd_uncore *uncore = *per_cpu_ptr(uncores, cpu); struct amd_uncore *uncore = *per_cpu_ptr(uncores, cpu);
kfree(uncore->free_when_cpu_online); uncore_clean_online();
uncore->free_when_cpu_online = NULL;
if (cpu == uncore->cpu) if (cpu == uncore->cpu)
cpumask_set_cpu(cpu, uncore->active_mask); cpumask_set_cpu(cpu, uncore->active_mask);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment