d476592013-06-12Arne Goedeke #include "global.h" #include "pike_error.h" #include "pike_memory.h" #include "block_allocator.h" #include "bitvector.h"
aa432a2013-08-01Arne Goedeke #include <stdlib.h> #define BA_BLOCKN(l, p, n) ((struct ba_block_header *)((char*)(p) + (l).doffset + (n)*((l).block_size))) #define BA_LASTBLOCK(l, p) ((struct ba_block_header*)((char*)(p) + (l).doffset + (l).offset)) #define BA_CHECK_PTR(l, p, ptr) ((size_t)((char*)(ptr) - (char*)(p)) <= (l).offset + (l).doffset)
d476592013-06-12Arne Goedeke  #define BA_ONE ((struct ba_block_header *)1)
ceae362013-10-09Arne Goedeke #define BA_FLAG_SORTED 1u
aa432a2013-08-01Arne Goedeke static void print_allocator(const struct block_allocator * a);
d476592013-06-12Arne Goedeke 
ceae362013-10-09Arne Goedeke static INLINE unsigned INT32 ba_block_number(const struct ba_layout * l, const struct ba_page * p, const void * ptr) { return ((char*)ptr - (char*)BA_BLOCKN(*l, p, 0)) / l->block_size; }
d476592013-06-12Arne Goedeke static INLINE void ba_dec_layout(struct ba_layout * l, int i) { l->blocks >>= i; l->offset += l->block_size; l->offset >>= i; l->offset -= l->block_size; } static INLINE void ba_inc_layout(struct ba_layout * l, int i) { l->blocks <<= i; l->offset += l->block_size; l->offset <<= i; l->offset -= l->block_size; } static INLINE void ba_double_layout(struct ba_layout * l) { ba_inc_layout(l, 1); } static INLINE void ba_half_layout(struct ba_layout * l) { ba_dec_layout(l, 1); } static INLINE struct ba_layout ba_get_layout(const struct block_allocator * a, int i) { struct ba_layout l = a->l; ba_inc_layout(&l, i); return l; } struct ba_block_header { struct ba_block_header * next; };
3036092013-10-09Arne Goedeke static INLINE void ba_clear_page(struct block_allocator * a, struct ba_page * p, struct ba_layout * l) { p->h.used = 0; p->h.flags = BA_FLAG_SORTED; p->h.first = BA_BLOCKN(*l, p, 0); PIKE_MEMPOOL_ALLOC(a, p->h.first, l->block_size); p->h.first->next = BA_ONE; PIKE_MEMPOOL_FREE(a, p->h.first, l->block_size); }
d476592013-06-12Arne Goedeke static struct ba_page * ba_alloc_page(struct block_allocator * a, int i) { struct ba_layout l = ba_get_layout(a, i);
aa432a2013-08-01Arne Goedeke  size_t n = l.offset + l.block_size + l.doffset;
0e26b62013-07-01Arne Goedeke  struct ba_page * p;
aa432a2013-08-01Arne Goedeke  if (l.alignment) { p = (struct ba_page*)aligned_alloc(n, l.alignment); } else {
0e26b62013-07-01Arne Goedeke #ifdef DEBUG_MALLOC
aa432a2013-08-01Arne Goedeke  /* In debug malloc mode, calling xalloc from the block alloc may result * in a deadlock, since xalloc will call ba_alloc, which in turn may call xalloc. */ p = (struct ba_page*)system_malloc(n); if (!p) { fprintf(stderr, "Fatal: Out of memory.\n"); exit(17); }
0e26b62013-07-01Arne Goedeke #else
aa432a2013-08-01Arne Goedeke  p = (struct ba_page*)xalloc(n);
0e26b62013-07-01Arne Goedeke #endif
aa432a2013-08-01Arne Goedeke  }
3036092013-10-09Arne Goedeke  ba_clear_page(a, p, &a->l);
411bb52013-08-02Arne Goedeke  PIKE_MEM_NA_RANGE((char*)p + l.doffset, n - l.doffset);
d476592013-06-12Arne Goedeke  return p; }
3036092013-10-09Arne Goedeke static void ba_free_empty_pages(struct block_allocator * a) { int i = a->size - 1; for (i = a->size - 1; i >= 0; i--) { struct ba_page * p = a->pages[i]; if (p->h.used) break; #ifdef DEBUG_MALLOC system_free(p); #else free(p); #endif a->pages[i] = NULL; } a->size = i+1; a->alloc = a->last_free = MAXIMUM(0, i); }
aa432a2013-08-01Arne Goedeke  PMOD_EXPORT void ba_init_aligned(struct block_allocator * a, unsigned INT32 block_size, unsigned INT32 blocks, unsigned INT32 alignment) {
3036092013-10-09Arne Goedeke  PIKE_MEMPOOL_CREATE(a);
d476592013-06-12Arne Goedeke  block_size = MAXIMUM(block_size, sizeof(struct ba_block_header));
aa432a2013-08-01Arne Goedeke  if (alignment) { if (alignment & (alignment - 1)) Pike_fatal("Block allocator alignment is not a power of 2.\n"); if (block_size & (alignment-1)) Pike_fatal("Block allocator block size is not aligned.\n"); a->l.doffset = PIKE_ALIGNTO(sizeof(struct ba_page), alignment); } else { a->l.doffset = sizeof(struct ba_page); }
d476592013-06-12Arne Goedeke  blocks = round_up32(blocks); a->alloc = a->last_free = 0; a->size = 1; a->l.block_size = block_size; a->l.blocks = blocks; a->l.offset = block_size * (blocks-1);
aa432a2013-08-01Arne Goedeke  a->l.alignment = alignment;
d476592013-06-12Arne Goedeke  memset(a->pages, 0, sizeof(a->pages)); a->pages[0] = ba_alloc_page(a, 0); } PMOD_EXPORT void ba_destroy(struct block_allocator * a) { int i;
3036092013-10-09Arne Goedeke  if (!a->l.offset) return;
d476592013-06-12Arne Goedeke  for (i = 0; i < a->size; i++) { if (a->pages[i]) {
0e26b62013-07-01Arne Goedeke #ifdef DEBUG_MALLOC system_free(a->pages[i]); #else
d476592013-06-12Arne Goedeke  free(a->pages[i]);
0e26b62013-07-01Arne Goedeke #endif
d476592013-06-12Arne Goedeke  a->pages[i] = NULL; } } a->size = 0; a->alloc = 0; a->last_free = 0;
3036092013-10-09Arne Goedeke  PIKE_MEMPOOL_DESTROY(a); } PMOD_EXPORT void ba_free_all(struct block_allocator * a) { int i; struct ba_layout l; if (!a->l.offset) return; if (!a->size) return; l = ba_get_layout(a, 0); for (i = 0; i < a->size; i++) { struct ba_page * page = a->pages[i]; ba_clear_page(a, page, &l); ba_double_layout(&l); } a->alloc = 0; a->last_free = 0;
d476592013-06-12Arne Goedeke } PMOD_EXPORT size_t ba_count(const struct block_allocator * a) { size_t c = 0; unsigned int i; for (i = 0; i < a->size; i++) { c += a->pages[i]->h.used; } return c; } PMOD_EXPORT void ba_count_all(const struct block_allocator * a, size_t * num, size_t * size) {
0ec53e2013-09-13Per Hedbor  size_t n = 0, b = sizeof( struct block_allocator ); unsigned int i; for( i=0; i<a->size; i++ ) { struct ba_layout l = ba_get_layout( a, i ); b += l.offset + l.block_size + l.doffset; n += a->pages[i]->h.used;
0e4e932013-06-13Arne Goedeke  }
0ec53e2013-09-13Per Hedbor  *num = n; *size = b;
d476592013-06-12Arne Goedeke } static void ba_low_alloc(struct block_allocator * a) { if (a->l.offset) { unsigned int i; for (i = 1; i <= a->size; i++) { struct ba_page * p = a->pages[a->size - i]; if (p->h.first) { a->alloc = a->size - i; return; } } if (a->size == (sizeof(a->pages)/sizeof(a->pages[0]))) { Pike_error("Out of memory."); } a->pages[a->size] = ba_alloc_page(a, a->size); a->alloc = a->size; a->size++; } else {
aa432a2013-08-01Arne Goedeke  ba_init_aligned(a, a->l.block_size, a->l.blocks, a->l.alignment);
d476592013-06-12Arne Goedeke  } } ATTRIBUTE((malloc)) PMOD_EXPORT void * ba_alloc(struct block_allocator * a) { struct ba_page * p = a->pages[a->alloc]; struct ba_block_header * ptr; if (!p || !p->h.first) { ba_low_alloc(a); p = a->pages[a->alloc]; } ptr = p->h.first;
411bb52013-08-02Arne Goedeke  PIKE_MEMPOOL_ALLOC(a, ptr, a->l.block_size); PIKE_MEM_RW_RANGE(ptr, sizeof(struct ba_block_header));
d476592013-06-12Arne Goedeke  p->h.used++;
aa432a2013-08-01Arne Goedeke #ifdef PIKE_DEBUG { struct ba_layout l = ba_get_layout(a, a->alloc); if (!BA_CHECK_PTR(l, p, ptr)) { print_allocator(a); Pike_fatal("about to return pointer from hell: %p\n", ptr); } } #endif
d476592013-06-12Arne Goedeke  if (ptr->next == BA_ONE) { struct ba_layout l = ba_get_layout(a, a->alloc); p->h.first = (struct ba_block_header*)((char*)ptr + a->l.block_size);
411bb52013-08-02Arne Goedeke  PIKE_MEMPOOL_ALLOC(a, p->h.first, a->l.block_size);
d476592013-06-12Arne Goedeke  p->h.first->next = (struct ba_block_header*)(ptrdiff_t)!(p->h.first == BA_LASTBLOCK(l, p));
411bb52013-08-02Arne Goedeke  PIKE_MEMPOOL_FREE(a, p->h.first, a->l.block_size);
d476592013-06-12Arne Goedeke  } else { p->h.first = ptr->next; }
411bb52013-08-02Arne Goedeke  PIKE_MEM_WO_RANGE(ptr, sizeof(struct ba_block_header));
aa432a2013-08-01Arne Goedeke  #if PIKE_DEBUG if (a->l.alignment && (size_t)ptr & (a->l.alignment - 1)) { print_allocator(a); Pike_fatal("Returning unaligned pointer.\n"); }
88924f2013-06-16Arne Goedeke #endif
d476592013-06-12Arne Goedeke  return ptr; } PMOD_EXPORT void ba_free(struct block_allocator * a, void * ptr) { int i = a->last_free; struct ba_page * p = a->pages[i]; struct ba_layout l = ba_get_layout(a, i);
aa432a2013-08-01Arne Goedeke #if PIKE_DEBUG if (a->l.alignment && (size_t)ptr & (a->l.alignment - 1)) { print_allocator(a); Pike_fatal("Returning unaligned pointer.\n"); } #endif
d476592013-06-12Arne Goedeke  if (BA_CHECK_PTR(l, p, ptr)) goto found;
aa432a2013-08-01Arne Goedeke #ifdef PIKE_DEBUG
d476592013-06-12Arne Goedeke  p = NULL;
aa432a2013-08-01Arne Goedeke #endif
d476592013-06-12Arne Goedeke  for (i = a->size-1, l = ba_get_layout(a, i); i >= 0; i--, ba_half_layout(&l)) { if (BA_CHECK_PTR(l, a->pages[i], ptr)) { a->last_free = i; p = a->pages[i]; break; } } found: #ifdef PIKE_DEBUG if (p) { #endif { struct ba_block_header * b = (struct ba_block_header*)ptr; b->next = p->h.first; p->h.first = b;
ceae362013-10-09Arne Goedeke  p->h.flags = 0;
d476592013-06-12Arne Goedeke #ifdef PIKE_DEBUG if (!p->h.used) {
aa432a2013-08-01Arne Goedeke  print_allocator(a); Pike_fatal("freeing from empty page %p\n", p);
d476592013-06-12Arne Goedeke  } #endif
3036092013-10-09Arne Goedeke  if (!(--p->h.used)) { if (i+1 == a->size) { ba_free_empty_pages(a); } else { ba_clear_page(a, p, &l); }
d476592013-06-12Arne Goedeke  } } #ifdef PIKE_DEBUG } else {
aa432a2013-08-01Arne Goedeke  print_allocator(a);
d476592013-06-12Arne Goedeke  Pike_fatal("ptr %p not in any page.\n", ptr); } #endif
411bb52013-08-02Arne Goedeke  PIKE_MEMPOOL_FREE(a, ptr, a->l.block_size);
d476592013-06-12Arne Goedeke }
aa432a2013-08-01Arne Goedeke  static void print_allocator(const struct block_allocator * a) { int i; struct ba_layout l; for (i = a->size-1, l = ba_get_layout(a, i); i >= 0; ba_half_layout(&l), i--) { struct ba_page * p = a->pages[i]; fprintf(stderr, "page: %p used: %u/%u last: %p p+offset: %p\n", a->pages[i], p->h.used, l.blocks, BA_BLOCKN(l, p, l.blocks-1), BA_LASTBLOCK(l, p)); } }
ceae362013-10-09Arne Goedeke  #if SIZEOF_LONG == 8 || SIZEOF_LONG_LONG == 8 #define BV_LENGTH 64 #define BV_ONE ((unsigned INT64)1) #define BV_NIL ((unsigned INT64)0) #define BV_CLZ clz64 #define BV_CTZ ctz64 typedef unsigned INT64 bv_int_t; #else #define BV_LENGTH 32 #define BV_ONE ((unsigned INT32)1) #define BV_NIL ((unsigned INT32)0) #define BV_CLZ clz32 #define BV_CTZ ctz32 typedef unsigned INT32 bv_int_t; #endif #define BV_WIDTH (BV_LENGTH/8) struct bitvector { size_t length; bv_int_t * v; }; static INLINE void bv_set_vector(struct bitvector * bv, void * p) { bv->v = (bv_int_t*)p; } static INLINE size_t bv_byte_length(struct bitvector * bv) { size_t bytes = (bv->length >> 3) + !!(bv->length & 7); if (bytes & (BV_LENGTH-1)) { bytes += (BV_LENGTH - (bytes & (BV_LENGTH-1))); } return bytes; } static INLINE void bv_set(struct bitvector * bv, size_t n, int value) { const size_t bit = n&(BV_LENGTH-1); const size_t c = n / BV_LENGTH; bv_int_t * _v = bv->v + c; if (value) *_v |= BV_ONE << bit; else *_v &= ~(BV_ONE << bit); } static INLINE int bv_get(struct bitvector * bv, size_t n) { const size_t bit = n&(BV_LENGTH-1); const size_t c = n / BV_LENGTH; return !!(bv->v[c] & (BV_ONE << bit)); } static INLINE size_t bv_ctz(struct bitvector * bv, size_t n) { size_t bit = n&(BV_LENGTH-1); size_t c = n / BV_LENGTH; bv_int_t * _v = bv->v + c; bv_int_t V = *_v & (~BV_NIL << bit); bit = c * BV_LENGTH; while (!(V)) { if (bit >= bv->length) { bit = (size_t)-1; goto RET; } V = *(++_v); bit += (BV_WIDTH*8); } bit += BV_CTZ(V); if (bit >= bv->length) bit = (size_t)-1; RET: return bit; }
4b8b1b2013-10-09Arne Goedeke #ifdef BA_DEBUG static INLINE void bv_print(struct bitvector * bv) { size_t i; for (i = 0; i < bv->length; i++) { fprintf(stderr, "%d", bv_get(bv, i)); } fprintf(stderr, "\n"); } #endif
ceae362013-10-09Arne Goedeke struct ba_block_header * ba_sort_list(const struct ba_page * p, struct ba_block_header * b, const struct ba_layout * l) { struct bitvector v; size_t i, j; struct ba_block_header ** t = &b; v.length = l->blocks; i = bv_byte_length(&v); /* we should probably reuse an area for this. */ bv_set_vector(&v, alloca(i)); memset(v.v, 0, i); /* * store the position of all blocks in a bitmask */ while (b) { unsigned INT32 n = ba_block_number(l, p, b); bv_set(&v, n, 1); if (b->next == BA_ONE) { v.length = n+1; break; } else b = b->next; } /* * Handle consecutive free blocks in the end, those * we dont need anyway. */ if (v.length) { i = v.length-1; while (i && bv_get(&v, i)) { i--; } v.length = i+1; } j = 0; /* * We now rechain all blocks. */ while ((i = bv_ctz(&v, j)) != (size_t)-1) { *t = BA_BLOCKN(*l, p, i); t = &((*t)->next); j = i+1; } /* * The last one */ if (v.length < l->blocks) { *t = BA_BLOCKN(*l, p, v.length); (*t)->next = BA_ONE; } else *t = NULL; return b; } static INLINE void ba_list_defined(struct block_allocator * a, struct ba_block_header * b) { while (b && b != BA_ONE) { PIKE_MEMPOOL_ALLOC(a, b, a->l.block_size); PIKE_MEM_RW_RANGE(b, sizeof(struct ba_block_header)); b = b->next; } } static INLINE void ba_list_undefined(struct block_allocator * a, struct ba_block_header * b) { while (b && b != BA_ONE) { struct ba_block_header * next = b->next; PIKE_MEMPOOL_FREE(a, b, a->l.block_size); b = next; } } /* * This function allows iteration over all allocated blocks. Some things are not allowed: * - throwing from within the callback * - allocating blocks during iteration * - nested iteration * * - freeing is OK, however some nodes will _still_ beiterated over, when they are freed during the * iteration. * * TODO: if needed, allocation can be fixed. For that to work, the free list of the currently * iterated page has to be removed and restored after iteration. that would guarantee allocation * from a different page * * NOTE * the callback will be called multiple times. for a usage example, see las.c */ PMOD_EXPORT void ba_walk(struct block_allocator * a, ba_walk_callback cb, void * data) { struct ba_iterator it; unsigned INT32 i;
4b8b1b2013-10-09Arne Goedeke  it.l = ba_get_layout(a, 0);
ceae362013-10-09Arne Goedeke  if (!a->size) return; for (i = 0; i < a->size; i++) { struct ba_page * p = a->pages[i]; if (p && p->h.used) {
4b8b1b2013-10-09Arne Goedeke  struct ba_block_header * free_list, * free_block;
ceae362013-10-09Arne Goedeke 
4b8b1b2013-10-09Arne Goedeke  ba_list_defined(a, p->h.first);
ceae362013-10-09Arne Goedeke  if (!(p->h.flags & BA_FLAG_SORTED)) { p->h.first = ba_sort_list(p, p->h.first, &it.l); p->h.flags |= BA_FLAG_SORTED; }
4b8b1b2013-10-09Arne Goedeke  /* we fake an allocation to prevent the page from being freed during iteration */ p->h.used ++; free_list = p->h.first; free_block = free_list;
ceae362013-10-09Arne Goedeke  it.cur = BA_BLOCKN(it.l, p, 0); while(1) { if (free_block == NULL) { it.end = ((char*)BA_LASTBLOCK(it.l, p) + it.l.block_size);
4b8b1b2013-10-09Arne Goedeke  if ((char*)it.end != (char*)it.cur) {
ceae362013-10-09Arne Goedeke  cb(&it, data);
4b8b1b2013-10-09Arne Goedeke  }
ceae362013-10-09Arne Goedeke  break; } else if (free_block == BA_ONE) { /* all consecutive blocks are free, so we are dont */ break; } it.end = free_block; #ifdef PIKE_DEBUG if (free_block >= free_block->next) Pike_fatal("Free list not sorted in ba_walk.\n"); #endif free_block = free_block->next; if ((char*)it.end != (char*)it.cur) cb(&it, data); it.cur = (char*)it.end + it.l.block_size; } /* if the callback throws, this will never happen */ ba_list_undefined(a, free_list); p->h.used--; } ba_double_layout(&it.l); } /* during the iteration blocks might have been freed. The pages will still be there, so we might have * to do some cleanup. */ if (!a->pages[a->size-1]->h.used) ba_free_empty_pages(a); }