[dpdk-dev] [PATCH] malloc: fix rte_free run time in O(n) free blocks

Sanford, Robert rsanford at akamai.com
Thu May 8 23:17:22 CEST 2014


I haven't heard anything regarding this patch.
In general, how does your team keep track of outstanding patch requests or
RFCs?

Thanks,
Robert


>Here is our proposed patch:
>
>Lib rte_malloc stores free blocks in singly-linked lists.
>This results in O(n), i.e., poor performance when freeing memory
>to a heap that contains thousands of free blocks.
>
>A simple solution is to store free blocks on doubly-linked lists.
>- Space-wise, we add one new pointer to struct malloc_elem.
> This does not affect its size, since it is cache-aligned and
> currently has room for another pointer.
>- We perform all free list adds and deletes in functions
> malloc_elem_add_to_free_list and remove_from_free_list.
>- We clean up internal malloc functions that push or
> return pointers to previous blocks in the free list.
>
>
>---
> lib/librte_malloc/malloc_elem.c |   82
>+++++++++++++++++++++++---------------
> lib/librte_malloc/malloc_elem.h |    6 ++-
> lib/librte_malloc/malloc_heap.c |   19 +++------
> 3 files changed, 60 insertions(+), 47 deletions(-)
>
>diff --git a/lib/librte_malloc/malloc_elem.c
>b/lib/librte_malloc/malloc_elem.c
>index f0da640..9a14e23 100644
>--- a/lib/librte_malloc/malloc_elem.c
>+++ b/lib/librte_malloc/malloc_elem.c
>@@ -60,7 +60,7 @@ malloc_elem_init(struct malloc_elem *elem,
> {
> 	elem->heap = heap;
> 	elem->mz = mz;
>-	elem->prev = elem->next_free = NULL;
>+	elem->prev = elem->next_free = elem->prev_free = NULL;
> 	elem->state = ELEM_FREE;
> 	elem->size = size;
> 	elem->pad = 0;
>@@ -125,14 +125,58 @@ split_elem(struct malloc_elem *elem, struct
>malloc_elem *split_pt)
> }
> 
> /*
>+ * Add the specified element to its heap's free list.
>+ */
>+void
>+malloc_elem_add_to_free_list(struct malloc_elem *elem)
>+{
>+	elem->state = ELEM_FREE;
>+	elem->prev_free = NULL;
>+	if (!!(elem->next_free = elem->heap->free_head))
>+		elem->next_free->prev_free = elem;
>+	elem->heap->free_head = elem;
>+}
>+
>+/*
>+ * Remove the specified element from its heap's free list.
>+ */
>+static inline void
>+remove_from_free_list(struct malloc_elem *elem)
>+{
>+	struct malloc_elem *next_free_elem = elem->next_free;
>+	struct malloc_elem *prev_free_elem = elem->prev_free;
>+
>+	if (!!next_free_elem) {
>+#ifdef RTE_LIBRTE_MALLOC_DEBUG
>+		RTE_VERIFY(next_free_elem->prev_free == elem);
>+#endif
>+		next_free_elem->prev_free = prev_free_elem;
>+	}
>+
>+	if (!!prev_free_elem) {
>+#ifdef RTE_LIBRTE_MALLOC_DEBUG
>+		RTE_VERIFY(prev_free_elem->next_free == elem);
>+#endif
>+		prev_free_elem->next_free = next_free_elem;
>+	}
>+	else {
>+		struct malloc_heap *heap = elem->heap;
>+
>+#ifdef RTE_LIBRTE_MALLOC_DEBUG
>+		RTE_VERIFY(heap->free_head == elem);
>+#endif
>+		heap->free_head = next_free_elem;
>+	}
>+}
>+
>+/*
>  * reserve a block of data in an existing malloc_elem. If the malloc_elem
>  * is much larger than the data block requested, we split the element in
>two.
>  * This function is only called from malloc_heap_alloc so parameter
>checking
>  * is not done here, as it's done there previously.
>  */
> struct malloc_elem *
>-malloc_elem_alloc(struct malloc_elem *elem, size_t size,
>-		unsigned align, struct malloc_elem *prev_free)
>+malloc_elem_alloc(struct malloc_elem *elem, size_t size, unsigned align)
> {
> 	struct malloc_elem *new_elem = elem_start_pt(elem, size, align);
> 	const unsigned old_elem_size = (uintptr_t)new_elem - (uintptr_t)elem;
>@@ -151,10 +195,7 @@ malloc_elem_alloc(struct malloc_elem *elem, size_t
>size,
> 			set_header(new_elem);
> 		}
> 		/* remove element from free list */
>-		if (prev_free == NULL)
>-			elem->heap->free_head = elem->next_free;
>-		else
>-			prev_free->next_free = elem->next_free;
>+		remove_from_free_list(elem);
> 
> 		return new_elem;
> 	}
>@@ -182,25 +223,6 @@ join_elem(struct malloc_elem *elem1, struct
>malloc_elem *elem2)
> }
> 
> /*
>- * scan the free list, and remove the request element from that
>- * free list. (Free list to scan is got from heap pointer in element)
>- */
>-static inline void
>-remove_from_free_list(struct malloc_elem *elem)
>-{
>-	if (elem == elem->heap->free_head)
>-		elem->heap->free_head = elem->next_free;
>-	else{
>-		struct malloc_elem *prev_free = elem->heap->free_head;
>-		while (prev_free && prev_free->next_free != elem)
>-			prev_free = prev_free->next_free;
>-		if (!prev_free)
>-			rte_panic("Corrupted free list\n");
>-		prev_free->next_free = elem->next_free;
>-	}
>-}
>-
>-/*
>  * free a malloc_elem block by adding it to the free list. If the
>  * blocks either immediately before or immediately after newly freed
>block
>  * are also free, the blocks are merged together.
>@@ -226,9 +248,7 @@ malloc_elem_free(struct malloc_elem *elem)
> 		join_elem(elem->prev, elem);
> 	/* otherwise add ourselves to the free list */
> 	else {
>-		elem->next_free = elem->heap->free_head;
>-		elem->heap->free_head = elem;
>-		elem->state = ELEM_FREE;
>+		malloc_elem_add_to_free_list(elem);
> 		elem->pad = 0;
> 	}
> 	/* decrease heap's count of allocated elements */
>@@ -269,9 +289,7 @@ malloc_elem_resize(struct malloc_elem *elem, size_t
>size)
> 		struct malloc_elem *split_pt = RTE_PTR_ADD(elem, new_size);
> 		split_pt = RTE_PTR_ALIGN_CEIL(split_pt, CACHE_LINE_SIZE);
> 		split_elem(elem, split_pt);
>-		split_pt->state = ELEM_FREE;
>-		split_pt->next_free = elem->heap->free_head;
>-		elem->heap->free_head = split_pt;
>+		malloc_elem_add_to_free_list(split_pt);
> 	}
> 	rte_spinlock_unlock(&elem->heap->lock);
> 	return 0;
>diff --git a/lib/librte_malloc/malloc_elem.h
>b/lib/librte_malloc/malloc_elem.h
>index eadecf9..6d2d7db 100644
>--- a/lib/librte_malloc/malloc_elem.h
>+++ b/lib/librte_malloc/malloc_elem.h
>@@ -47,6 +47,7 @@ struct malloc_elem {
> 	struct malloc_heap *heap;
> 	struct malloc_elem *volatile prev;      /* points to prev elem in
>memzone */
> 	struct malloc_elem *volatile next_free; /* to make list of free elements
>*/
>+	struct malloc_elem *volatile prev_free; /* make free list doubly-linked
>*/
> 	const struct rte_memzone *mz;
> 	volatile enum elem_state state;
> 	uint32_t pad;
>@@ -156,8 +157,7 @@ malloc_elem_can_hold(struct malloc_elem *elem, size_t
>size, unsigned align);
>  * is much larger than the data block requested, we split the element in
>two.
>  */
> struct malloc_elem *
>-malloc_elem_alloc(struct malloc_elem *elem, size_t size,
>-		unsigned align, struct malloc_elem *prev_free);
>+malloc_elem_alloc(struct malloc_elem *elem, size_t size, unsigned align);
> 
> /*
>  * free a malloc_elem block by adding it to the free list. If the
>@@ -174,4 +174,6 @@ malloc_elem_free(struct malloc_elem *elem);
> int
> malloc_elem_resize(struct malloc_elem *elem, size_t size);
> 
>+void malloc_elem_add_to_free_list(struct malloc_elem *elem);
>+
> #endif /* MALLOC_ELEM_H_ */
>diff --git a/lib/librte_malloc/malloc_heap.c
>b/lib/librte_malloc/malloc_heap.c
>index f4a0294..bc146e1 100644
>--- a/lib/librte_malloc/malloc_heap.c
>+++ b/lib/librte_malloc/malloc_heap.c
>@@ -112,9 +112,8 @@ malloc_heap_add_memzone(struct malloc_heap *heap,
>size_t size, unsigned align)
> 	const unsigned elem_size = (uintptr_t)end_elem - (uintptr_t)start_elem;
> 	malloc_elem_init(start_elem, heap, mz, elem_size);
> 	malloc_elem_mkend(end_elem, start_elem);
>+	malloc_elem_add_to_free_list(start_elem);
> 
>-	start_elem->next_free = heap->free_head;
>-	heap->free_head = start_elem;
> 	/* increase heap total size by size of new memzone */
> 	heap->total_size+=mz_size - MALLOC_ELEM_OVERHEAD;
> 	return 0;
>@@ -160,27 +159,22 @@ malloc_heap_init(struct malloc_heap *heap)
>  * (to make removing the element from the free list faster).
>  */
> static struct malloc_elem *
>-find_suitable_element(struct malloc_heap *heap, size_t size,
>-		unsigned align, struct malloc_elem **prev)
>+find_suitable_element(struct malloc_heap *heap, size_t size, unsigned
>align)
> {
>-	struct malloc_elem *elem, *min_elem, *min_prev;
>+	struct malloc_elem *elem, *min_elem;
> 	size_t min_sz;
> 
> 	elem = heap->free_head;
> 	min_elem = NULL;
>-	min_prev = NULL;
> 	min_sz = (size_t) SIZE_MAX;
>-	*prev = NULL;
> 
> 	while(elem){
> 		if (malloc_elem_can_hold(elem, size, align)) {
> 			if (min_sz > elem->size) {
> 				min_elem = elem;
>-				*prev = min_prev;
> 				min_sz = elem->size;
> 			}
> 		}
>-		min_prev = elem;
> 		elem = elem->next_free;
> 	}
> 	return (min_elem);
>@@ -202,15 +196,14 @@ malloc_heap_alloc(struct malloc_heap *heap,
> 	size = CACHE_LINE_ROUNDUP(size);
> 	align = CACHE_LINE_ROUNDUP(align);
> 	rte_spinlock_lock(&heap->lock);
>-	struct malloc_elem *prev, *elem = find_suitable_element(heap,
>-			size, align, &prev);
>+	struct malloc_elem *elem = find_suitable_element(heap, size, align);
> 	if (elem == NULL){
> 		if ((malloc_heap_add_memzone(heap, size, align)) == 0)
>-			elem = find_suitable_element(heap, size, align, &prev);
>+			elem = find_suitable_element(heap, size, align);
> 	}
> 
> 	if (elem != NULL){
>-		elem = malloc_elem_alloc(elem, size, align, prev);
>+		elem = malloc_elem_alloc(elem, size, align);
> 		/* increase heap's count of allocated elements */
> 		heap->alloc_count++;
> 	}
>-- 
>1.7.1
>
>Signed-off-by: Robert Sanford <rsanford at akamai.com>
>



More information about the dev mailing list