326 lines
11 KiB
Diff
326 lines
11 KiB
Diff
From: Lee Schermerhorn <lee.schermerhorn@hp.com>
|
|
Subject: slab - handle memoryless nodes V2a
|
|
References: bnc#436025, bnc#570492
|
|
Patch-mainline: not yet
|
|
|
|
The slab cache, since [apparently] 2.6.21, does not handle memoryless
|
|
nodes well. Specifically, the "fast path"-- ____cache_alloc()--will
|
|
never succeed, but will be called twice: once speculatively [expected
|
|
to succeed] and once in the fallback path. This adds significant
|
|
overhead to all kmem cache allocations, incurring a significant
|
|
regression relative to earlier kernels [from before slab.c was
|
|
reorganized].
|
|
|
|
This patch addresses the regression by modifying slab.c to treat the
|
|
first fallback node in a memoryless node's general zonelist as the "slab
|
|
local node" -- i.e., the local node for the purpose of slab allocations.
|
|
This is, in fact, the node from which all "local" allocations for cpus
|
|
attached to a memoryless node will be satisfied.
|
|
|
|
The new function numa_slab_nid(gfp_t) replaces all calls to
|
|
numa_node_id() in slab.c. numa_slab_id() will simply return
|
|
numa_node_id() for nodes with memory, but will return the first
|
|
node in the local node's zonelist selected by the gfp flags.
|
|
|
|
Effects of the patch:
|
|
|
|
We first noticed the effects of the slab reorganization running the
|
|
AIM benchmark on a distro based on 2.6.27. The effect is even more
|
|
pronounced in the hackbench results. The platform in an HP rx8640
|
|
numa platform, configured with "0% Cell Local Memory". In this
|
|
configuration, all memory appears in a "pseudo-node"--an artifact
|
|
of the firmware--and is interleaved across all the physical nodes'
|
|
memory on a cacheline granularity. All cpus are presented as
|
|
attached to memoryless nodes.
|
|
|
|
Here are the results of running hackbench at various load levels
|
|
with and without the patch on the same platform configured for
|
|
0% CLM and "100% CLM".
|
|
|
|
Command: hackbench N process 100, for N = 10..100 by 10
|
|
|
|
|
|
100% CLM 0% CLM
|
|
Tasks no with no with
|
|
patch patch %diff patch patch %diff
|
|
400 0.246 0.281 14.23% 2.962 0.410 -86.16%
|
|
800 0.418 0.421 0.72% 6.224 0.793 -87.26%
|
|
1200 0.548 0.532 -2.92% 9.058 1.090 -87.97%
|
|
1600 0.654 0.716 9.48% 12.473 1.562 -87.48%
|
|
2000 0.871 0.859 -1.38% 15.484 1.889 -87.80%
|
|
2400 0.994 1.043 4.93% 18.689 2.309 -87.65%
|
|
2800 1.196 1.195 -0.08% 22.069 2.606 -88.19%
|
|
3200 1.322 1.344 1.66% 25.642 2.988 -88.35%
|
|
3600 1.474 1.519 3.05% 28.003 3.418 -87.79%
|
|
4000 1.682 1.750 4.04% 30.887 3.858 -87.51%
|
|
|
|
In the 100% CLM case, the regression does not appear, because
|
|
all nodes have local memory. Note that the patch has >10%
|
|
overhead on the first run, but then varies widely from run
|
|
to run [more below]. For the 0%CLM configuration, the patch
|
|
reduced the run time by 86-88%.
|
|
|
|
|
|
The following runs extend the number of hackbench tasks using:
|
|
|
|
hackbench N process 100, for N = 100 to 400 by 20
|
|
|
|
We didn't run the 0%CLM/no-patch runs as they were taking too
|
|
long for our liking. We wanted to see how the patched kernel
|
|
performed as we extended the range.
|
|
|
|
100% CLM 0% CLM
|
|
Tasks no with no with
|
|
patch patch %diff patch patch %diff
|
|
4800 1.879 2.117 12.67% not-run 4.458
|
|
5600 2.100 2.352 12.00% not-run 5.207
|
|
6400 2.532 2.447 -3.36% not-run 5.669
|
|
8000 2.799 2.792 -0.25% not-run 6.651
|
|
8000 3.244 3.030 -6.60% not-run 7.366
|
|
8800 3.282 3.550 8.17% not-run 8.169
|
|
9600 3.595 3.738 3.98% not-run 8.582
|
|
10400 3.811 4.045 6.14% not-run 9.705
|
|
11200 4.090 4.162 1.76% not-run 9.760
|
|
12000 4.408 4.419 0.25% not-run 10.141
|
|
12800 4.665 4.787 2.62% not-run 11.628
|
|
13600 5.095 5.069 -0.51% not-run 11.735
|
|
14400 5.347 5.464 2.19% not-run 12.621
|
|
15200 5.620 5.831 3.75% not-run 13.452
|
|
16000 5.870 6.161 4.96% not-run 14.069
|
|
|
|
The 0% CLM configuration with the patch performs worse than
|
|
the 100% CLM configuration. In the 0% CLM case we had 64
|
|
ia64 cores beating on a single zone in the interleaved
|
|
memory-only pseudo-node. In the 100% CLM case, we have 16
|
|
cores allocating memory locally to each of 4 nodes,
|
|
demonstating the difference between [pseudo-]SMP and NUMA
|
|
behavior.
|
|
|
|
Note, again, that the first run[s] have higher % difference
|
|
between the patched and unpatched kernels for the 100% CLM
|
|
config, and then vary quite a bit run to run. To get a feel
|
|
for the average overhead, we ran 40 runs at the 16000 task
|
|
load point with more interations to increase the runtime
|
|
per run:
|
|
|
|
hackbench 400 process 200
|
|
|
|
These were run on the 100% CLM configuration, as this best represents
|
|
most NUMA platforms:
|
|
|
|
No patch with Patch %diff
|
|
Average of 40: 9.796 9.857 0.623
|
|
|
|
|
|
Signed-off-by: Lee Schermerhorn <lee.schermerhorn@hp.com>
|
|
Acked-by: Nick Piggin <npiggin@suse.de>
|
|
|
|
mm/slab.c | 91 +++++++++++++++++++++++++++++++++++++++++++++++++++++---------
|
|
1 file changed, 78 insertions(+), 13 deletions(-)
|
|
|
|
--- a/mm/slab.c
|
|
+++ b/mm/slab.c
|
|
@@ -305,7 +305,7 @@ struct kmem_list3 {
|
|
struct array_cache **alien; /* on other nodes */
|
|
unsigned long next_reap; /* updated without locking */
|
|
int free_touched; /* updated without locking */
|
|
-};
|
|
+} __attribute__((aligned(sizeof(long))));
|
|
|
|
/*
|
|
* Need this for bootstrapping a per node allocator.
|
|
@@ -968,6 +968,11 @@ static int transfer_objects(struct array
|
|
#define drain_alien_cache(cachep, alien) do { } while (0)
|
|
#define reap_alien(cachep, l3) do { } while (0)
|
|
|
|
+static inline int numa_slab_nid(struct kmem_cache *cachep, gfp_t flags)
|
|
+{
|
|
+ return 0;
|
|
+}
|
|
+
|
|
static inline struct array_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
|
|
{
|
|
return (struct array_cache **)BAD_ALIEN_MAGIC;
|
|
@@ -999,6 +1004,64 @@ static inline void *____cache_alloc_node
|
|
static void *____cache_alloc_node(struct kmem_cache *, gfp_t, int);
|
|
static void *alternate_node_alloc(struct kmem_cache *, gfp_t);
|
|
|
|
+/*
|
|
+ * slow path for numa_slab_nid(), below
|
|
+ */
|
|
+static noinline int __numa_slab_nid(struct kmem_cache *cachep,
|
|
+ int node, gfp_t flags)
|
|
+{
|
|
+ struct zonelist *zonelist;
|
|
+ struct zone *zone;
|
|
+ enum zone_type highest_zoneidx = gfp_zone(flags);
|
|
+
|
|
+ if (likely(node_state(node, N_NORMAL_MEMORY)))
|
|
+ return node;
|
|
+
|
|
+ /*
|
|
+ * memoryless node: consult its zonelist.
|
|
+ * Cache the fallback node, if cache pointer provided.
|
|
+ */
|
|
+ zonelist = &NODE_DATA(node)->node_zonelists[0];
|
|
+ (void)first_zones_zonelist(zonelist, highest_zoneidx,
|
|
+ NULL,
|
|
+ &zone);
|
|
+ if (cachep)
|
|
+ cachep->nodelists[node] =
|
|
+ (struct kmem_list3 *)((unsigned long)zone->node << 1 | 1);
|
|
+ return zone->node;
|
|
+}
|
|
+
|
|
+/*
|
|
+ * "Local" node for slab is first node in zonelist with memory.
|
|
+ * For nodes with memory this will be the actual local node.
|
|
+ *
|
|
+ * Use nodelist[numa_node_id()] to cache the fallback node for
|
|
+ * memoryless nodes. We'll be loading that member soon anyway,
|
|
+ * or already have, when called for cache refill, ... Use low
|
|
+ * bit of "pointer" as flag for "memoryless_node", indicating
|
|
+ * that the fallback nodes is stored here [<<1].
|
|
+ */
|
|
+#define memoryless_node(L3L) ((L3L) & 1)
|
|
+static inline int numa_slab_nid(struct kmem_cache *cachep, gfp_t flags)
|
|
+{
|
|
+ int node = numa_node_id();
|
|
+
|
|
+ if (likely(cachep)){
|
|
+ unsigned long l3l = (unsigned long)cachep->nodelists[node];
|
|
+
|
|
+ if (likely(l3l)) {
|
|
+ if (unlikely(memoryless_node(l3l)))
|
|
+ node = (int)(l3l >> 1);
|
|
+ return node;
|
|
+ }
|
|
+ }
|
|
+
|
|
+ /*
|
|
+ * !cachep || !l3l - the slow path
|
|
+ */
|
|
+ return __numa_slab_nid(cachep, node, flags);
|
|
+}
|
|
+
|
|
static struct array_cache **alloc_alien_cache(int node, int limit, gfp_t gfp)
|
|
{
|
|
struct array_cache **ac_ptr;
|
|
@@ -1098,7 +1161,7 @@ static inline int cache_free_alien(struc
|
|
struct array_cache *alien = NULL;
|
|
int node;
|
|
|
|
- node = numa_node_id();
|
|
+ node = numa_slab_nid(cachep, GFP_KERNEL);
|
|
|
|
/*
|
|
* Make sure we are not freeing a object from another node to the array
|
|
@@ -1443,7 +1506,7 @@ void __init kmem_cache_init(void)
|
|
* 6) Resize the head arrays of the kmalloc caches to their final sizes.
|
|
*/
|
|
|
|
- node = numa_node_id();
|
|
+ node = numa_slab_nid(NULL, GFP_KERNEL);
|
|
|
|
/* 1) create the cache_cache */
|
|
INIT_LIST_HEAD(&cache_chain);
|
|
@@ -2079,7 +2142,7 @@ static int __init_refok setup_cpu_cache(
|
|
}
|
|
}
|
|
}
|
|
- cachep->nodelists[numa_node_id()]->next_reap =
|
|
+ cachep->nodelists[numa_slab_nid(cachep, GFP_KERNEL)]->next_reap =
|
|
jiffies + REAPTIMEOUT_LIST3 +
|
|
((unsigned long)cachep) % REAPTIMEOUT_LIST3;
|
|
|
|
@@ -2411,7 +2474,7 @@ static void check_spinlock_acquired(stru
|
|
{
|
|
#ifdef CONFIG_SMP
|
|
check_irq_off();
|
|
- assert_spin_locked(&cachep->nodelists[numa_node_id()]->list_lock);
|
|
+ assert_spin_locked(&cachep->nodelists[numa_slab_nid(cachep, GFP_KERNEL)]->list_lock);
|
|
#endif
|
|
}
|
|
|
|
@@ -2438,7 +2501,7 @@ static void do_drain(void *arg)
|
|
{
|
|
struct kmem_cache *cachep = arg;
|
|
struct array_cache *ac;
|
|
- int node = numa_node_id();
|
|
+ int node = numa_slab_nid(cachep, GFP_KERNEL);
|
|
|
|
check_irq_off();
|
|
ac = cpu_cache_get(cachep);
|
|
@@ -2975,7 +3038,7 @@ static void *cache_alloc_refill(struct k
|
|
|
|
retry:
|
|
check_irq_off();
|
|
- node = numa_node_id();
|
|
+ node = numa_slab_nid(cachep, flags);
|
|
if (unlikely(must_refill))
|
|
goto force_grow;
|
|
ac = cpu_cache_get(cachep);
|
|
@@ -3185,7 +3248,7 @@ static void *alternate_node_alloc(struct
|
|
|
|
if (in_interrupt() || (flags & __GFP_THISNODE))
|
|
return NULL;
|
|
- nid_alloc = nid_here = numa_node_id();
|
|
+ nid_alloc = nid_here = numa_slab_nid(cachep, flags);
|
|
if (cpuset_do_slab_mem_spread() && (cachep->flags & SLAB_MEM_SPREAD))
|
|
nid_alloc = cpuset_mem_spread_node();
|
|
else if (current->mempolicy)
|
|
@@ -3360,6 +3423,7 @@ __cache_alloc_node(struct kmem_cache *ca
|
|
{
|
|
unsigned long save_flags;
|
|
void *ptr;
|
|
+ int slab_node = numa_slab_nid(cachep, flags);
|
|
|
|
flags &= gfp_allowed_mask;
|
|
|
|
@@ -3372,7 +3436,7 @@ __cache_alloc_node(struct kmem_cache *ca
|
|
local_irq_save(save_flags);
|
|
|
|
if (nodeid == -1)
|
|
- nodeid = numa_node_id();
|
|
+ nodeid = slab_node;
|
|
|
|
if (unlikely(!cachep->nodelists[nodeid])) {
|
|
/* Node not bootstrapped yet */
|
|
@@ -3380,7 +3444,7 @@ __cache_alloc_node(struct kmem_cache *ca
|
|
goto out;
|
|
}
|
|
|
|
- if (nodeid == numa_node_id()) {
|
|
+ if (nodeid == slab_node) {
|
|
/*
|
|
* Use the locally cached objects if possible.
|
|
* However ____cache_alloc does not allow fallback
|
|
@@ -3425,7 +3489,8 @@ __do_cache_alloc(struct kmem_cache *cach
|
|
* ____cache_alloc_node() knows how to locate memory on other nodes
|
|
*/
|
|
if (!objp)
|
|
- objp = ____cache_alloc_node(cache, flags, numa_node_id());
|
|
+ objp = ____cache_alloc_node(cache, flags,
|
|
+ numa_slab_nid(cache, flags));
|
|
|
|
out:
|
|
return objp;
|
|
@@ -3522,7 +3587,7 @@ static void cache_flusharray(struct kmem
|
|
{
|
|
int batchcount;
|
|
struct kmem_list3 *l3;
|
|
- int node = numa_node_id();
|
|
+ int node = numa_slab_nid(cachep, GFP_KERNEL);
|
|
|
|
batchcount = ac->batchcount;
|
|
#if DEBUG
|
|
@@ -4172,7 +4237,7 @@ static void cache_reap(struct work_struc
|
|
{
|
|
struct kmem_cache *searchp;
|
|
struct kmem_list3 *l3;
|
|
- int node = numa_node_id();
|
|
+ int node = numa_slab_nid(NULL, GFP_KERNEL);
|
|
struct delayed_work *work = to_delayed_work(w);
|
|
|
|
if (!mutex_trylock(&cache_chain_mutex))
|