|  | /* | 
|  | * Copyright (C) 2011 STRATO AG | 
|  | * written by Arne Jansen <sensille@gmx.net> | 
|  | * Distributed under the GNU GPL license version 2. | 
|  | */ | 
|  |  | 
|  | #include <linux/slab.h> | 
|  | #include <linux/export.h> | 
|  | #include "ulist.h" | 
|  |  | 
|  | /* | 
|  | * ulist is a generic data structure to hold a collection of unique u64 | 
|  | * values. The only operations it supports is adding to the list and | 
|  | * enumerating it. | 
|  | * It is possible to store an auxiliary value along with the key. | 
|  | * | 
|  | * The implementation is preliminary and can probably be sped up | 
|  | * significantly. A first step would be to store the values in an rbtree | 
|  | * as soon as ULIST_SIZE is exceeded. | 
|  | * | 
|  | * A sample usage for ulists is the enumeration of directed graphs without | 
|  | * visiting a node twice. The pseudo-code could look like this: | 
|  | * | 
|  | * ulist = ulist_alloc(); | 
|  | * ulist_add(ulist, root); | 
|  | * ULIST_ITER_INIT(&uiter); | 
|  | * | 
|  | * while ((elem = ulist_next(ulist, &uiter)) { | 
|  | * 	for (all child nodes n in elem) | 
|  | *		ulist_add(ulist, n); | 
|  | *	do something useful with the node; | 
|  | * } | 
|  | * ulist_free(ulist); | 
|  | * | 
|  | * This assumes the graph nodes are adressable by u64. This stems from the | 
|  | * usage for tree enumeration in btrfs, where the logical addresses are | 
|  | * 64 bit. | 
|  | * | 
|  | * It is also useful for tree enumeration which could be done elegantly | 
|  | * recursively, but is not possible due to kernel stack limitations. The | 
|  | * loop would be similar to the above. | 
|  | */ | 
|  |  | 
|  | /** | 
|  | * ulist_init - freshly initialize a ulist | 
|  | * @ulist:	the ulist to initialize | 
|  | * | 
|  | * Note: don't use this function to init an already used ulist, use | 
|  | * ulist_reinit instead. | 
|  | */ | 
|  | void ulist_init(struct ulist *ulist) | 
|  | { | 
|  | ulist->nnodes = 0; | 
|  | ulist->nodes = ulist->int_nodes; | 
|  | ulist->nodes_alloced = ULIST_SIZE; | 
|  | ulist->root = RB_ROOT; | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_init); | 
|  |  | 
|  | /** | 
|  | * ulist_fini - free up additionally allocated memory for the ulist | 
|  | * @ulist:	the ulist from which to free the additional memory | 
|  | * | 
|  | * This is useful in cases where the base 'struct ulist' has been statically | 
|  | * allocated. | 
|  | */ | 
|  | void ulist_fini(struct ulist *ulist) | 
|  | { | 
|  | /* | 
|  | * The first ULIST_SIZE elements are stored inline in struct ulist. | 
|  | * Only if more elements are alocated they need to be freed. | 
|  | */ | 
|  | if (ulist->nodes_alloced > ULIST_SIZE) | 
|  | kfree(ulist->nodes); | 
|  | ulist->nodes_alloced = 0;	/* in case ulist_fini is called twice */ | 
|  | ulist->root = RB_ROOT; | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_fini); | 
|  |  | 
|  | /** | 
|  | * ulist_reinit - prepare a ulist for reuse | 
|  | * @ulist:	ulist to be reused | 
|  | * | 
|  | * Free up all additional memory allocated for the list elements and reinit | 
|  | * the ulist. | 
|  | */ | 
|  | void ulist_reinit(struct ulist *ulist) | 
|  | { | 
|  | ulist_fini(ulist); | 
|  | ulist_init(ulist); | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_reinit); | 
|  |  | 
|  | /** | 
|  | * ulist_alloc - dynamically allocate a ulist | 
|  | * @gfp_mask:	allocation flags to for base allocation | 
|  | * | 
|  | * The allocated ulist will be returned in an initialized state. | 
|  | */ | 
|  | struct ulist *ulist_alloc(gfp_t gfp_mask) | 
|  | { | 
|  | struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); | 
|  |  | 
|  | if (!ulist) | 
|  | return NULL; | 
|  |  | 
|  | ulist_init(ulist); | 
|  |  | 
|  | return ulist; | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_alloc); | 
|  |  | 
|  | /** | 
|  | * ulist_free - free dynamically allocated ulist | 
|  | * @ulist:	ulist to free | 
|  | * | 
|  | * It is not necessary to call ulist_fini before. | 
|  | */ | 
|  | void ulist_free(struct ulist *ulist) | 
|  | { | 
|  | if (!ulist) | 
|  | return; | 
|  | ulist_fini(ulist); | 
|  | kfree(ulist); | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_free); | 
|  |  | 
|  | static struct ulist_node *ulist_rbtree_search(struct ulist *ulist, u64 val) | 
|  | { | 
|  | struct rb_node *n = ulist->root.rb_node; | 
|  | struct ulist_node *u = NULL; | 
|  |  | 
|  | while (n) { | 
|  | u = rb_entry(n, struct ulist_node, rb_node); | 
|  | if (u->val < val) | 
|  | n = n->rb_right; | 
|  | else if (u->val > val) | 
|  | n = n->rb_left; | 
|  | else | 
|  | return u; | 
|  | } | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static int ulist_rbtree_insert(struct ulist *ulist, struct ulist_node *ins) | 
|  | { | 
|  | struct rb_node **p = &ulist->root.rb_node; | 
|  | struct rb_node *parent = NULL; | 
|  | struct ulist_node *cur = NULL; | 
|  |  | 
|  | while (*p) { | 
|  | parent = *p; | 
|  | cur = rb_entry(parent, struct ulist_node, rb_node); | 
|  |  | 
|  | if (cur->val < ins->val) | 
|  | p = &(*p)->rb_right; | 
|  | else if (cur->val > ins->val) | 
|  | p = &(*p)->rb_left; | 
|  | else | 
|  | return -EEXIST; | 
|  | } | 
|  | rb_link_node(&ins->rb_node, parent, p); | 
|  | rb_insert_color(&ins->rb_node, &ulist->root); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * ulist_add - add an element to the ulist | 
|  | * @ulist:	ulist to add the element to | 
|  | * @val:	value to add to ulist | 
|  | * @aux:	auxiliary value to store along with val | 
|  | * @gfp_mask:	flags to use for allocation | 
|  | * | 
|  | * Note: locking must be provided by the caller. In case of rwlocks write | 
|  | *       locking is needed | 
|  | * | 
|  | * Add an element to a ulist. The @val will only be added if it doesn't | 
|  | * already exist. If it is added, the auxiliary value @aux is stored along with | 
|  | * it. In case @val already exists in the ulist, @aux is ignored, even if | 
|  | * it differs from the already stored value. | 
|  | * | 
|  | * ulist_add returns 0 if @val already exists in ulist and 1 if @val has been | 
|  | * inserted. | 
|  | * In case of allocation failure -ENOMEM is returned and the ulist stays | 
|  | * unaltered. | 
|  | */ | 
|  | int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) | 
|  | { | 
|  | return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); | 
|  | } | 
|  |  | 
|  | int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux, | 
|  | u64 *old_aux, gfp_t gfp_mask) | 
|  | { | 
|  | int ret = 0; | 
|  | struct ulist_node *node = NULL; | 
|  | node = ulist_rbtree_search(ulist, val); | 
|  | if (node) { | 
|  | if (old_aux) | 
|  | *old_aux = node->aux; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (ulist->nnodes >= ulist->nodes_alloced) { | 
|  | u64 new_alloced = ulist->nodes_alloced + 128; | 
|  | struct ulist_node *new_nodes; | 
|  | void *old = NULL; | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < ulist->nnodes; i++) | 
|  | rb_erase(&ulist->nodes[i].rb_node, &ulist->root); | 
|  |  | 
|  | /* | 
|  | * if nodes_alloced == ULIST_SIZE no memory has been allocated | 
|  | * yet, so pass NULL to krealloc | 
|  | */ | 
|  | if (ulist->nodes_alloced > ULIST_SIZE) | 
|  | old = ulist->nodes; | 
|  |  | 
|  | new_nodes = krealloc(old, sizeof(*new_nodes) * new_alloced, | 
|  | gfp_mask); | 
|  | if (!new_nodes) | 
|  | return -ENOMEM; | 
|  |  | 
|  | if (!old) | 
|  | memcpy(new_nodes, ulist->int_nodes, | 
|  | sizeof(ulist->int_nodes)); | 
|  |  | 
|  | ulist->nodes = new_nodes; | 
|  | ulist->nodes_alloced = new_alloced; | 
|  |  | 
|  | /* | 
|  | * krealloc actually uses memcpy, which does not copy rb_node | 
|  | * pointers, so we have to do it ourselves.  Otherwise we may | 
|  | * be bitten by crashes. | 
|  | */ | 
|  | for (i = 0; i < ulist->nnodes; i++) { | 
|  | ret = ulist_rbtree_insert(ulist, &ulist->nodes[i]); | 
|  | if (ret < 0) | 
|  | return ret; | 
|  | } | 
|  | } | 
|  | ulist->nodes[ulist->nnodes].val = val; | 
|  | ulist->nodes[ulist->nnodes].aux = aux; | 
|  | ret = ulist_rbtree_insert(ulist, &ulist->nodes[ulist->nnodes]); | 
|  | BUG_ON(ret); | 
|  | ++ulist->nnodes; | 
|  |  | 
|  | return 1; | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_add); | 
|  |  | 
|  | /** | 
|  | * ulist_next - iterate ulist | 
|  | * @ulist:	ulist to iterate | 
|  | * @uiter:	iterator variable, initialized with ULIST_ITER_INIT(&iterator) | 
|  | * | 
|  | * Note: locking must be provided by the caller. In case of rwlocks only read | 
|  | *       locking is needed | 
|  | * | 
|  | * This function is used to iterate an ulist. | 
|  | * It returns the next element from the ulist or %NULL when the | 
|  | * end is reached. No guarantee is made with respect to the order in which | 
|  | * the elements are returned. They might neither be returned in order of | 
|  | * addition nor in ascending order. | 
|  | * It is allowed to call ulist_add during an enumeration. Newly added items | 
|  | * are guaranteed to show up in the running enumeration. | 
|  | */ | 
|  | struct ulist_node *ulist_next(struct ulist *ulist, struct ulist_iterator *uiter) | 
|  | { | 
|  | if (ulist->nnodes == 0) | 
|  | return NULL; | 
|  | if (uiter->i < 0 || uiter->i >= ulist->nnodes) | 
|  | return NULL; | 
|  |  | 
|  | return &ulist->nodes[uiter->i++]; | 
|  | } | 
|  | EXPORT_SYMBOL(ulist_next); |