[PATCH V2 2/3] slub: optimize bulk slowpath free by detached freelist
From: Jesper Dangaard Brouer
Date: Sun Aug 23 2015 - 20:59:11 EST
This change focus on improving the speed of object freeing in the
"slowpath" of kmem_cache_free_bulk.
The slowpath call __slab_free() have been extended with support for
bulk free, which amortize the overhead of the locked cmpxchg_double_slab.
To use the new bulking feature of __slab_free(), we build what I call
a detached freelist. The detached freelist takes advantage of three
properties:
1) the free function call owns the object that is about to be freed,
thus writing into this memory is synchronization-free.
2) many freelist's can co-exist side-by-side in the same page each
with a separate head pointer.
3) it is the visibility of the head pointer that needs synchronization.
Given these properties, the brilliant part is that the detached
freelist can be constructed without any need for synchronization.
The freelist is constructed directly in the page objects, without any
synchronization needed. The detached freelist is allocated on the
stack of the function call kmem_cache_free_bulk. Thus, the freelist
head pointer is not visible to other CPUs.
This implementation is fairly simple, as it only builds the detached
freelist if two consecutive objects belongs to the same page. When
detecting object page does not match, it simply flushes the local
freelist, and starts a new local detached freelist. It will not
look-ahead to see if further opputunities exists in the
The next patch have a more advanced look-ahead approach, but is also
more complicated. Splitting them up, because I want to be able to
benchmark the simple against the advanced approach.
Signed-off-by: Jesper Dangaard Brouer <brouer@xxxxxxxxxx>
---
bulk- Fallback - Bulk API
1 - 64 cycles(tsc) 16.109 ns - 47 cycles(tsc) 11.894 - improved 26.6%
2 - 56 cycles(tsc) 14.158 ns - 45 cycles(tsc) 11.274 - improved 19.6%
3 - 54 cycles(tsc) 13.650 ns - 23 cycles(tsc) 6.001 - improved 57.4%
4 - 53 cycles(tsc) 13.268 ns - 21 cycles(tsc) 5.262 - improved 60.4%
8 - 51 cycles(tsc) 12.841 ns - 18 cycles(tsc) 4.718 - improved 64.7%
16 - 50 cycles(tsc) 12.583 ns - 19 cycles(tsc) 4.896 - improved 62.0%
30 - 85 cycles(tsc) 21.357 ns - 26 cycles(tsc) 6.549 - improved 69.4%
32 - 82 cycles(tsc) 20.690 ns - 25 cycles(tsc) 6.412 - improved 69.5%
34 - 81 cycles(tsc) 20.322 ns - 25 cycles(tsc) 6.365 - improved 69.1%
48 - 93 cycles(tsc) 23.332 ns - 28 cycles(tsc) 7.139 - improved 69.9%
64 - 98 cycles(tsc) 24.544 ns - 62 cycles(tsc) 15.543 - improved 36.7%
128 - 96 cycles(tsc) 24.219 ns - 68 cycles(tsc) 17.143 - improved 29.2%
158 - 107 cycles(tsc) 26.817 ns - 69 cycles(tsc) 17.431 - improved 35.5%
250 - 107 cycles(tsc) 26.824 ns - 70 cycles(tsc) 17.730 - improved 34.6%
---
mm/slub.c | 48 +++++++++++++++++++++++++++++++++++++++++-------
1 file changed, 41 insertions(+), 7 deletions(-)
diff --git a/mm/slub.c b/mm/slub.c
index 10b57a3bb895..40e4b5926311 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -2756,12 +2756,26 @@ void kmem_cache_free(struct kmem_cache *s, void *x)
}
EXPORT_SYMBOL(kmem_cache_free);
+struct detached_freelist {
+ struct page *page;
+ void *freelist;
+ void *tail_object;
+ int cnt;
+};
+
/* Note that interrupts must be enabled when calling this function. */
void kmem_cache_free_bulk(struct kmem_cache *s, size_t size, void **p)
{
struct kmem_cache_cpu *c;
struct page *page;
int i;
+ /* Opportunistically delay updating page->freelist, hoping
+ * next free happen to same page. Start building the freelist
+ * in the page, but keep local stack ptr to freelist. If
+ * successful several object can be transferred to page with a
+ * single cmpxchg_double.
+ */
+ struct detached_freelist df = {0};
local_irq_disable();
c = this_cpu_ptr(s->cpu_slab);
@@ -2778,22 +2792,42 @@ void kmem_cache_free_bulk(struct kmem_cache *s, size_t size, void **p)
page = virt_to_head_page(object);
- if (c->page == page) {
+ if (page == df.page) {
+ /* Oppotunity to delay real free */
+ set_freepointer(s, object, df.freelist);
+ df.freelist = object;
+ df.cnt++;
+ } else if (c->page == page) {
/* Fastpath: local CPU free */
set_freepointer(s, object, c->freelist);
c->freelist = object;
} else {
- c->tid = next_tid(c->tid);
- local_irq_enable();
- /* Slowpath: overhead locked cmpxchg_double_slab */
- __slab_free(s, page, object, _RET_IP_, NULL, 1);
- local_irq_disable();
- c = this_cpu_ptr(s->cpu_slab);
+ /* Slowpath: Flush delayed free */
+ if (df.page) {
+ c->tid = next_tid(c->tid);
+ local_irq_enable();
+ __slab_free(s, df.page, df.tail_object,
+ _RET_IP_, df.freelist, df.cnt);
+ local_irq_disable();
+ c = this_cpu_ptr(s->cpu_slab);
+ }
+ /* Start new round of delayed free */
+ df.page = page;
+ df.tail_object = object;
+ set_freepointer(s, object, NULL);
+ df.freelist = object;
+ df.cnt = 1;
}
}
exit:
c->tid = next_tid(c->tid);
local_irq_enable();
+
+ /* Flush detached freelist */
+ if (df.page) {
+ __slab_free(s, df.page, df.tail_object,
+ _RET_IP_, df.freelist, df.cnt);
+ }
}
EXPORT_SYMBOL(kmem_cache_free_bulk);
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/