From: Heng Li Date: Sat, 14 Apr 2018 22:35:48 +0000 (-0400) Subject: added kavl.h X-Git-Url: http://www.kaiwu.me/postgresql/commit/?a=commitdiff_plain;h=fc769eb46b65cdc4b8109d7c71f465e01fffcfab;p=klib.git added kavl.h Intrusive generic AVL tree in ANSI C --- diff --git a/kavl.h b/kavl.h new file mode 100644 index 0000000..4d0f17e --- /dev/null +++ b/kavl.h @@ -0,0 +1,247 @@ +#ifndef KAVL_H +#define KAVL_H + +#ifdef __STRICT_ANSI__ +#define inline __inline__ +#endif + +#define KAVL_MAX_DEPTH 64 + +#define kavl_size(head, p) ((p)? (p)->head.size : 0) +#define kavl_size_child(head, q, i) ((q)->head.p[(i)]? (q)->head.p[(i)]->head.size : 0) + +#define KAVL_HEAD(__type) \ + struct { \ + __type *p[2]; \ + signed char balance; /* balance factor */ \ + unsigned size; /* #elements in subtree */ \ + } + +#define __KAVL_FIND(suf, __scope, __type, __head, __cmp) \ + __scope __type *kavl_find_##suf(const __type *root, const __type *x, unsigned *cnt_) { \ + const __type *p = root; \ + unsigned cnt = 0; \ + while (p != 0) { \ + int cmp; \ + cmp = __cmp(x, p); \ + if (cmp < 0) p = p->__head.p[0]; \ + else if (cmp > 0) { \ + p = p->__head.p[1]; \ + cnt += kavl_size_child(__head, p, 0) + 1; \ + } else { \ + if (cnt_) *cnt_ = cnt + 1; \ + return (__type*)p; \ + } \ + } \ + if (cnt_) *cnt_ = cnt; \ + return 0; \ + } + +#define __KAVL_ROTATE(suf, __type, __head) \ + /* one rotation: (a,(b,c)q)p => ((a,b)p,c)q */ \ + static inline __type *kavl_rotate1_##suf(__type *p, int dir) { /* dir=0 to left; dir=1 to right */ \ + int opp = 1 - dir; /* opposite direction */ \ + __type *q = p->__head.p[opp]; \ + unsigned size_p = p->__head.size; \ + p->__head.size -= q->__head.size - kavl_size_child(__head, q, dir); \ + q->__head.size = size_p; \ + p->__head.p[opp] = q->__head.p[dir]; \ + q->__head.p[dir] = p; \ + return q; \ + } \ + /* two consecutive rotations: (a,((b,c)r,d)q)p => ((a,b)p,(c,d)q)r */ \ + static inline __type *kavl_rotate2_##suf(__type *p, int dir) { \ + int b1, opp = 1 - dir; \ + __type *q = p->__head.p[opp], *r = q->__head.p[dir]; \ + unsigned size_x_dir = kavl_size_child(__head, r, dir); \ + r->__head.size = p->__head.size; \ + p->__head.size -= q->__head.size - size_x_dir; \ + q->__head.size -= size_x_dir + 1; \ + p->__head.p[opp] = r->__head.p[dir]; \ + r->__head.p[dir] = p; \ + q->__head.p[dir] = r->__head.p[opp]; \ + r->__head.p[opp] = q; \ + b1 = dir == 0? +1 : -1; \ + if (r->__head.balance == b1) q->__head.balance = 0, p->__head.balance = -b1; \ + else if (r->__head.balance == 0) q->__head.balance = p->__head.balance = 0; \ + else q->__head.balance = b1, p->__head.balance = 0; \ + r->__head.balance = 0; \ + return r; \ + } + +#define __KAVL_INSERT(suf, __scope, __type, __head, __cmp) \ + __scope __type *kavl_insert_##suf(__type **root_, __type *x, unsigned *cnt_) { \ + unsigned char stack[KAVL_MAX_DEPTH]; \ + __type *path[KAVL_MAX_DEPTH]; \ + __type *bp, *bq; \ + __type *p, *q, *r = 0; /* _r_ is potentially the new root */ \ + int i, which, top, b1, path_len; \ + unsigned cnt = 0; \ + bp = *root_, bq = 0; \ + /* find the insertion location */ \ + for (p = bp, q = bq, top = path_len = 0; p; q = p, p = p->__head.p[which]) { \ + int cmp; \ + cmp = __cmp(x, p); \ + if (cmp == 0) { \ + if (cnt_) *cnt_ = cnt + 1; \ + return p; \ + } \ + if (p->__head.balance != 0) \ + bq = q, bp = p, top = 0; \ + if (cmp > 0) cnt += kavl_size_child(__head, p, 0) + 1; \ + stack[top++] = which = (cmp > 0); \ + path[path_len++] = p; \ + } \ + if (cnt_) *cnt_ = cnt; \ + x->__head.balance = 0, x->__head.size = 1, x->__head.p[0] = x->__head.p[1] = 0; \ + if (q == 0) *root_ = x; \ + else q->__head.p[which] = x; \ + if (bp == 0) return x; \ + for (i = 0; i < path_len; ++i) ++path[i]->__head.size; \ + for (p = bp, top = 0; p != x; p = p->__head.p[stack[top]], ++top) /* update balance factors */ \ + if (stack[top] == 0) --p->__head.balance; \ + else ++p->__head.balance; \ + if (bp->__head.balance > -2 && bp->__head.balance < 2) return x; /* no re-balance needed */ \ + /* re-balance */ \ + which = (bp->__head.balance < 0); \ + b1 = which == 0? +1 : -1; \ + q = bp->__head.p[1 - which]; \ + if (q->__head.balance == b1) { \ + r = kavl_rotate1_##suf(bp, which); \ + q->__head.balance = bp->__head.balance = 0; \ + } else r = kavl_rotate2_##suf(bp, which); \ + if (bq == 0) *root_ = r; \ + else bq->__head.p[bp != bq->__head.p[0]] = r; \ + return x; \ + } + +#define __KAVL_ERASE(suf, __scope, __type, __head, __cmp) \ + __scope __type *kavl_erase_##suf(__type **root_, const __type *x) { \ + __type *p, *path[KAVL_MAX_DEPTH], fake; \ + unsigned char dir[KAVL_MAX_DEPTH]; \ + int i, d = 0, cmp; \ + fake.__head.p[0] = *root_, fake.__head.p[1] = 0; \ + for (cmp = -1, p = &fake; cmp; cmp = __cmp(x, p)) { \ + int which = (cmp > 0); \ + dir[d] = which; \ + path[d++] = p; \ + p = p->__head.p[which]; \ + if (p == 0) return 0; \ + } \ + for (i = 1; i < d; ++i) --path[i]->__head.size; \ + if (p->__head.p[1] == 0) { /* ((1,.)2,3)4 => (1,3)4; p=2 */ \ + path[d-1]->__head.p[dir[d-1]] = p->__head.p[0]; \ + } else { \ + __type *q = p->__head.p[1]; \ + if (q->__head.p[0] == 0) { /* ((1,2)3,4)5 => ((1)2,4)5; p=3 */ \ + q->__head.p[0] = p->__head.p[0]; \ + q->__head.balance = p->__head.balance; \ + path[d-1]->__head.p[dir[d-1]] = q; \ + path[d] = q, dir[d++] = 1; \ + q->__head.size = p->__head.size - 1; \ + } else { /* ((1,((.,2)3,4)5)6,7)8 => ((1,(2,4)5)3,7)8; p=6 */ \ + __type *r; \ + int e = d++; /* backup _d_ */\ + for (;;) { \ + dir[d] = 0; \ + path[d++] = q; \ + r = q->__head.p[0]; \ + if (r->__head.p[0] == 0) break; \ + q = r; \ + } \ + r->__head.p[0] = p->__head.p[0]; \ + q->__head.p[0] = r->__head.p[1]; \ + r->__head.p[1] = p->__head.p[1]; \ + r->__head.balance = p->__head.balance; \ + path[e-1]->__head.p[dir[e-1]] = r; \ + path[e] = r, dir[e] = 1; \ + for (i = e + 1; i < d; ++i) --path[i]->__head.size; \ + r->__head.size = p->__head.size - 1; \ + } \ + } \ + while (--d > 0) { \ + __type *q = path[d]; \ + int which, other, b1 = 1, b2 = 2; \ + which = dir[d], other = 1 - which; \ + if (which) b1 = -b1, b2 = -b2; \ + q->__head.balance += b1; \ + if (q->__head.balance == b1) break; \ + else if (q->__head.balance == b2) { \ + __type *r = q->__head.p[other]; \ + if (r->__head.balance == -b1) { \ + path[d-1]->__head.p[dir[d-1]] = kavl_rotate2_##suf(q, which); \ + } else { \ + path[d-1]->__head.p[dir[d-1]] = kavl_rotate1_##suf(q, which); \ + if (r->__head.balance == 0) { \ + r->__head.balance = -b1; \ + q->__head.balance = b1; \ + break; \ + } else r->__head.balance = q->__head.balance = 0; \ + } \ + } \ + } \ + *root_ = fake.__head.p[0]; \ + return p; \ + } + +#define __KAVL_ITR(suf, __scope, __type, __head, __cmp) \ + struct kavl_itr_##suf { \ + const __type *stack[KAVL_MAX_DEPTH], **top, *right; /* _right_ points to the right child of *top */ \ + }; \ + __scope void kavl_itr_first_##suf(const __type *root, struct kavl_itr_##suf *itr) { \ + const __type *p; \ + for (itr->top = itr->stack - 1, p = root; p; p = p->__head.p[0]) \ + *++itr->top = p; \ + itr->right = (*itr->top)->__head.p[1]; \ + } \ + __scope int kavl_itr_eq_##suf(const __type *root, const __type *x, struct kavl_itr_##suf *itr) { \ + const __type *p = root; \ + itr->top = itr->stack - 1; \ + while (p != 0) { \ + int cmp; \ + cmp = __cmp(x, p); \ + if (cmp < 0) *++itr->top = p, p = p->__head.p[0]; \ + else if (cmp > 0) p = p->__head.p[1]; \ + else break; \ + } \ + if (p) { \ + *++itr->top = p; \ + itr->right = p->__head.p[1]; \ + return 1; \ + } else { \ + itr->top = itr->stack - 1; \ + return 0; \ + } \ + } \ + __scope int kavl_itr_next_##suf(struct kavl_itr_##suf *itr) { \ + for (;;) { \ + const __type *p; \ + for (p = itr->right, --itr->top; p; p = p->__head.p[0]) \ + *++itr->top = p; \ + if (itr->top < itr->stack) return 0; \ + itr->right = (*itr->top)->__head.p[1]; \ + return 1; \ + } \ + } + +#define kavl_find(suf, root, x, cnt) kavl_find_##suf(root, x, cnt) +#define kavl_insert(suf, root, x, cnt) kavl_insert_##suf(root, x, cnt) +#define kavl_erase(suf, root, x) kavl_erase_##suf(root, x) + +#define kavl_itr_t(suf) struct kavl_itr_##suf +#define kavl_itr_first(suf, root, itr) kavl_itr_first_##suf(root, itr) +#define kavl_itr_eq(suf, root, x, itr) kavl_itr_eq_##suf(root, x, itr) +#define kavl_itr_next(suf, itr) kavl_itr_next_##suf(itr) +#define kavl_at(itr) ((itr)->top < (itr)->stack? 0 : *(itr)->top) + +#define KAVL_INIT2(suf, __scope, __type, __head, __cmp) \ + __KAVL_FIND(suf, __scope, __type, __head, __cmp) \ + __KAVL_ROTATE(suf, __type, __head) \ + __KAVL_INSERT(suf, __scope, __type, __head, __cmp) \ + __KAVL_ERASE(suf, __scope, __type, __head, __cmp) \ + __KAVL_ITR(suf, __scope, __type, __head, __cmp) + +#define KAVL_INIT(suf, __type, __head, __cmp) \ + KAVL_INIT2(suf,, __type, __head, __cmp) + +#endif diff --git a/test/kavl_test.c b/test/kavl_test.c new file mode 100644 index 0000000..6fa23f8 --- /dev/null +++ b/test/kavl_test.c @@ -0,0 +1,98 @@ +#include +#include +#include +#include +#include "kavl.h" + +#define CALLOC(type, num) ((type*)calloc(num, sizeof(type))) + +struct my_node { + int key; + KAVL_HEAD(struct my_node) head; +}; + +#define my_cmp(p, q) ((p)->key - (q)->key) +KAVL_INIT(my, struct my_node, head, my_cmp) + +int check_size(struct my_node *p) +{ + int c = 1; + if (p == 0) return 0; + if (p->head.p[0]) c += check_size(p->head.p[0]); + if (p->head.p[1]) c += check_size(p->head.p[1]); + if (c != (int)p->head.size) + fprintf(stderr, "%d != %d at %c\n", p->head.size, c, p->key); + return c; +} +/* +int print_tree(const struct my_node *p) +{ + int c = 1; + if (p == 0) return 0; + if (p->head.p[0] || p->head.p[1]) { + putchar('('); + if (p->head.p[0]) c += print_tree(p->head.p[0]); + else putchar('.'); + putchar(','); + if (p->head.p[1]) c += print_tree(p->head.p[1]); + else putchar('.'); + putchar(')'); + } + putchar(p->key); + return c; +} + +void check_and_print(struct my_node *root) +{ + check_size(root); + print_tree(root); + putchar('\n'); +} +*/ +void shuffle(int n, char a[]) +{ + int i, j; + for (i = n; i > 1; --i) { + char tmp; + j = (int)(drand48() * i); + tmp = a[j]; a[j] = a[i-1]; a[i-1] = tmp; + } +} + +int main(void) +{ + char buf[256]; + int i, n; + struct my_node *root = 0; + struct my_node *p, *q, t; + kavl_itr_t(my) itr; + unsigned cnt; + + for (i = 33, n = 0; i <= 127; ++i) + if (i != '(' && i != ')' && i != '.' && i != ';') + buf[n++] = i; + shuffle(n, buf); + for (i = 0; i < n; ++i) { + p = CALLOC(struct my_node, 1); + p->key = buf[i]; + q = kavl_insert(my, &root, p, &cnt); + if (p != q) free(p); + check_size(root); + } + shuffle(n, buf); + for (i = 0; i < n/2; ++i) { + t.key = buf[i]; + q = kavl_erase(my, &root, &t); + if (q) free(q); + check_size(root); + } + + kavl_itr_first(my, root, &itr); + do { + const struct my_node *r = kavl_at(&itr); + putchar(r->key); + free((void*)r); + } while (kavl_itr_next(my, &itr)); + putchar('\n'); + return 0; +}