Changeset 58fe85a for libcfa/src/heap.cfa
- Timestamp:
- Jan 7, 2021, 3:27:00 PM (5 years ago)
- Branches:
- ADT, arm-eh, ast-experimental, enum, forall-pointer-decay, jacob/cs343-translation, master, new-ast-unique-expr, pthread-emulation, qualifiedEnum
- Children:
- 2b4daf2, 64aeca0
- Parents:
- 3c64c668 (diff), eef8dfb (diff)
Note: this is a merge changeset, the changes displayed below correspond to the merge itself.
Use the(diff)
links above to see all the changes relative to each parent. - File:
-
- 1 edited
Legend:
- Unmodified
- Added
- Removed
-
libcfa/src/heap.cfa
r3c64c668 r58fe85a 5 5 // file "LICENCE" distributed with Cforall. 6 6 // 7 // heap.c --7 // heap.cfa -- 8 8 // 9 9 // Author : Peter A. Buhr 10 10 // Created On : Tue Dec 19 21:58:35 2017 11 11 // Last Modified By : Peter A. Buhr 12 // Last Modified On : Tue Feb 4 10:04:51202013 // Update Count : 64812 // Last Modified On : Wed Dec 16 12:28:25 2020 13 // Update Count : 1023 14 14 // 15 15 16 16 #include <unistd.h> // sbrk, sysconf 17 #include <stdlib.h> // EXIT_FAILURE 17 18 #include <stdbool.h> // true, false 18 19 #include <stdio.h> // snprintf, fileno 19 20 #include <errno.h> // errno 20 21 #include <string.h> // memset, memcpy 21 extern "C" { 22 #include <limits.h> // ULONG_MAX 23 #include <malloc.h> // memalign, malloc_usable_size 22 24 #include <sys/mman.h> // mmap, munmap 23 } // extern "C" 24 25 // #comment TD : Many of these should be merged into math I believe 26 #include "bits/align.hfa" // libPow2 25 26 #include "bits/align.hfa" // libAlign 27 27 #include "bits/defs.hfa" // likely, unlikely 28 28 #include "bits/locks.hfa" // __spinlock_t 29 29 #include "startup.hfa" // STARTUP_PRIORITY_MEMORY 30 //#include "stdlib.hfa" // bsearchl 31 #include "malloc.h" 32 33 #define MIN(x, y) (y > x ? x : y) 30 #include "math.hfa" // ceiling 31 #include "bitmanip.hfa" // is_pow2, ceiling2 34 32 35 33 static bool traceHeap = false; … … 74 72 // Define the default extension heap amount in units of bytes. When the uC++ supplied heap reaches the brk address, 75 73 // the brk address is extended by the extension amount. 76 __CFA_DEFAULT_HEAP_EXPANSION__ = (1 * 1024 * 1024),74 __CFA_DEFAULT_HEAP_EXPANSION__ = (10 * 1024 * 1024), 77 75 78 76 // Define the mmap crossover point during allocation. Allocations less than this amount are allocated from buckets; … … 91 89 92 90 #ifdef __CFA_DEBUG__ 93 static unsigned int allocFree;// running total of allocations minus frees91 static size_t allocUnfreed; // running total of allocations minus frees 94 92 95 93 static void prtUnfreed() { 96 if ( alloc Free!= 0 ) {94 if ( allocUnfreed != 0 ) { 97 95 // DO NOT USE STREAMS AS THEY MAY BE UNAVAILABLE AT THIS POINT. 98 96 char helpText[512]; 99 int len = snprintf( helpText, sizeof(helpText), "CFA warning (UNIX pid:%ld) : program terminating with % u(0x%x) bytes of storage allocated but not freed.\n"97 int len = snprintf( helpText, sizeof(helpText), "CFA warning (UNIX pid:%ld) : program terminating with %zu(0x%zx) bytes of storage allocated but not freed.\n" 100 98 "Possible cause is unfreed storage allocated by the program or system/library routines called from the program.\n", 101 (long int)getpid(), alloc Free, allocFree); // always print the UNIX pid99 (long int)getpid(), allocUnfreed, allocUnfreed ); // always print the UNIX pid 102 100 __cfaabi_bits_write( STDERR_FILENO, helpText, len ); // print debug/nodebug 103 101 } // if … … 106 104 extern "C" { 107 105 void heapAppStart() { // called by __cfaabi_appready_startup 108 alloc Free= 0;106 allocUnfreed = 0; 109 107 } // heapAppStart 110 108 … … 118 116 119 117 // statically allocated variables => zero filled. 120 static size_t pageSize; // architecture pagesize 118 size_t __page_size; // architecture pagesize 119 int __map_prot; // common mmap/mprotect protection 121 120 static size_t heapExpand; // sbrk advance 122 121 static size_t mmapStart; // cross over point for mmap … … 127 126 #define LOCKFREE 1 128 127 #define BUCKETLOCK SPINLOCK 129 #if BUCKETLOCK == LOCKFREE 130 #include <uStackLF.h> 128 #if BUCKETLOCK == SPINLOCK 129 #elif BUCKETLOCK == LOCKFREE 130 #include <stackLockFree.hfa> 131 #else 132 #error undefined lock type for bucket lock 131 133 #endif // LOCKFREE 132 134 … … 136 138 137 139 struct HeapManager { 138 // struct FreeHeader; // forward declaration139 140 140 struct Storage { 141 141 struct Header { // header … … 145 145 struct { // 4-byte word => 8-byte header, 8-byte word => 16-byte header 146 146 #if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ && __SIZEOF_POINTER__ == 4 147 uint 32_t padding; // unused, force home/blocksize to overlay alignment in fake header147 uint64_t padding; // unused, force home/blocksize to overlay alignment in fake header 148 148 #endif // __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ && __SIZEOF_POINTER__ == 4 149 149 150 150 union { 151 // FreeHeader * home; // allocated block points back to home locations (must overlay alignment) 151 // FreeHeader * home; // allocated block points back to home locations (must overlay alignment) 152 // 2nd low-order bit => zero filled 152 153 void * home; // allocated block points back to home locations (must overlay alignment) 153 154 size_t blockSize; // size for munmap (must overlay alignment) 154 #if BUCK LOCK == SPINLOCK155 #if BUCKETLOCK == SPINLOCK 155 156 Storage * next; // freed block points next freed block of same size 156 157 #endif // SPINLOCK 157 158 }; 159 size_t size; // allocation size in bytes 158 160 159 161 #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ && __SIZEOF_POINTER__ == 4 160 uint 32_t padding; // unused, force home/blocksize to overlay alignment in fake header162 uint64_t padding; // unused, force home/blocksize to overlay alignment in fake header 161 163 #endif // __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ && __SIZEOF_POINTER__ == 4 162 164 }; 163 // future code 164 #if BUCKLOCK == LOCKFREE 165 Stack<Storage>::Link next; // freed block points next freed block of same size (double-wide) 165 #if BUCKETLOCK == LOCKFREE 166 Link(Storage) next; // freed block points next freed block of same size (double-wide) 166 167 #endif // LOCKFREE 167 168 }; 168 169 } real; // RealHeader 170 169 171 struct FakeHeader { 170 172 #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ 171 uint32_t alignment; // low-order bits of home/blockSize used for tricks173 uint32_t alignment; // 1st low-order bit => fake header & alignment 172 174 #endif // __ORDER_LITTLE_ENDIAN__ 173 175 … … 187 189 188 190 struct FreeHeader { 189 #if BUCK LOCK == SPINLOCK191 #if BUCKETLOCK == SPINLOCK 190 192 __spinlock_t lock; // must be first field for alignment 191 193 Storage * freeList; 192 #elif BUCKLOCK == LOCKFREE193 // future code194 StackLF<Storage> freeList;195 194 #else 196 #error undefined lock type for bucket lock197 #endif // SPINLOCK195 StackLF(Storage) freeList; 196 #endif // BUCKETLOCK 198 197 size_t blockSize; // size of allocations on this list 199 198 }; // FreeHeader … … 208 207 }; // HeapManager 209 208 209 #if BUCKETLOCK == LOCKFREE 210 static inline { 211 Link(HeapManager.Storage) * ?`next( HeapManager.Storage * this ) { return &this->header.kind.real.next; } 212 void ?{}( HeapManager.FreeHeader & ) {} 213 void ^?{}( HeapManager.FreeHeader & ) {} 214 } // distribution 215 #endif // LOCKFREE 216 210 217 static inline size_t getKey( const HeapManager.FreeHeader & freeheader ) { return freeheader.blockSize; } 211 218 … … 214 221 #define __STATISTICS__ 215 222 216 // Bucket size must be multiple of 16. 217 // Powers of 2 are common allocation sizes, so make powers of 2 generate the minimum required size. 223 // Size of array must harmonize with NoBucketSizes and individual bucket sizes must be multiple of 16. 224 // Smaller multiples of 16 and powers of 2 are common allocation sizes, so make them generate the minimum required bucket size. 225 // malloc(0) returns 0p, so no bucket is necessary for 0 bytes returning an address that can be freed. 218 226 static const unsigned int bucketSizes[] @= { // different bucket sizes 219 16 , 32, 48, 64 + sizeof(HeapManager.Storage), // 4220 96 , 112, 128 + sizeof(HeapManager.Storage), // 3227 16 + sizeof(HeapManager.Storage), 32 + sizeof(HeapManager.Storage), 48 + sizeof(HeapManager.Storage), 64 + sizeof(HeapManager.Storage), // 4 228 96 + sizeof(HeapManager.Storage), 112 + sizeof(HeapManager.Storage), 128 + sizeof(HeapManager.Storage), // 3 221 229 160, 192, 224, 256 + sizeof(HeapManager.Storage), // 4 222 230 320, 384, 448, 512 + sizeof(HeapManager.Storage), // 4 … … 236 244 }; 237 245 238 static_assert( NoBucketSizes == sizeof(bucketSizes) / sizeof(bucketSizes[0] ), "size of bucket array wrong" );246 static_assert( NoBucketSizes == sizeof(bucketSizes) / sizeof(bucketSizes[0] ), "size of bucket array wrong" ); 239 247 240 248 #ifdef FASTLOOKUP … … 243 251 #endif // FASTLOOKUP 244 252 245 static int mmapFd = -1;// fake or actual fd for anonymous file253 static const off_t mmapFd = -1; // fake or actual fd for anonymous file 246 254 #ifdef __CFA_DEBUG__ 247 255 static bool heapBoot = 0; // detect recursion during boot 248 256 #endif // __CFA_DEBUG__ 257 258 // The constructor for heapManager is called explicitly in memory_startup. 249 259 static HeapManager heapManager __attribute__(( aligned (128) )) @= {}; // size of cache line to prevent false sharing 250 260 … … 252 262 #ifdef __STATISTICS__ 253 263 // Heap statistics counters. 264 static unsigned int malloc_calls; 265 static unsigned long long int malloc_storage; 266 static unsigned int aalloc_calls; 267 static unsigned long long int aalloc_storage; 268 static unsigned int calloc_calls; 269 static unsigned long long int calloc_storage; 270 static unsigned int memalign_calls; 271 static unsigned long long int memalign_storage; 272 static unsigned int amemalign_calls; 273 static unsigned long long int amemalign_storage; 274 static unsigned int cmemalign_calls; 275 static unsigned long long int cmemalign_storage; 276 static unsigned int resize_calls; 277 static unsigned long long int resize_storage; 278 static unsigned int realloc_calls; 279 static unsigned long long int realloc_storage; 280 static unsigned int free_calls; 281 static unsigned long long int free_storage; 282 static unsigned int mmap_calls; 254 283 static unsigned long long int mmap_storage; 255 static unsigned int m map_calls;284 static unsigned int munmap_calls; 256 285 static unsigned long long int munmap_storage; 257 static unsigned int munmap_calls;286 static unsigned int sbrk_calls; 258 287 static unsigned long long int sbrk_storage; 259 static unsigned int sbrk_calls;260 static unsigned long long int malloc_storage;261 static unsigned int malloc_calls;262 static unsigned long long int free_storage;263 static unsigned int free_calls;264 static unsigned long long int calloc_storage;265 static unsigned int calloc_calls;266 static unsigned long long int memalign_storage;267 static unsigned int memalign_calls;268 static unsigned long long int cmemalign_storage;269 static unsigned int cmemalign_calls;270 static unsigned long long int realloc_storage;271 static unsigned int realloc_calls;272 288 // Statistics file descriptor (changed by malloc_stats_fd). 273 static int stat fd = STDERR_FILENO; // default stderr289 static int stat_fd = STDERR_FILENO; // default stderr 274 290 275 291 // Use "write" because streams may be shutdown when calls are made. 276 292 static void printStats() { 277 char helpText[ 512];293 char helpText[1024]; 278 294 __cfaabi_bits_print_buffer( STDERR_FILENO, helpText, sizeof(helpText), 279 295 "\nHeap statistics:\n" 280 296 " malloc: calls %u / storage %llu\n" 297 " aalloc: calls %u / storage %llu\n" 281 298 " calloc: calls %u / storage %llu\n" 282 299 " memalign: calls %u / storage %llu\n" 300 " amemalign: calls %u / storage %llu\n" 283 301 " cmemalign: calls %u / storage %llu\n" 302 " resize: calls %u / storage %llu\n" 284 303 " realloc: calls %u / storage %llu\n" 285 304 " free: calls %u / storage %llu\n" … … 288 307 " sbrk: calls %u / storage %llu\n", 289 308 malloc_calls, malloc_storage, 309 aalloc_calls, aalloc_storage, 290 310 calloc_calls, calloc_storage, 291 311 memalign_calls, memalign_storage, 312 amemalign_calls, amemalign_storage, 292 313 cmemalign_calls, cmemalign_storage, 314 resize_calls, resize_storage, 293 315 realloc_calls, realloc_storage, 294 316 free_calls, free_storage, … … 300 322 301 323 static int printStatsXML( FILE * stream ) { // see malloc_info 302 char helpText[ 512];324 char helpText[1024]; 303 325 int len = snprintf( helpText, sizeof(helpText), 304 326 "<malloc version=\"1\">\n" … … 307 329 "</sizes>\n" 308 330 "<total type=\"malloc\" count=\"%u\" size=\"%llu\"/>\n" 331 "<total type=\"aalloc\" count=\"%u\" size=\"%llu\"/>\n" 309 332 "<total type=\"calloc\" count=\"%u\" size=\"%llu\"/>\n" 310 333 "<total type=\"memalign\" count=\"%u\" size=\"%llu\"/>\n" 334 "<total type=\"amemalign\" count=\"%u\" size=\"%llu\"/>\n" 311 335 "<total type=\"cmemalign\" count=\"%u\" size=\"%llu\"/>\n" 336 "<total type=\"resize\" count=\"%u\" size=\"%llu\"/>\n" 312 337 "<total type=\"realloc\" count=\"%u\" size=\"%llu\"/>\n" 313 338 "<total type=\"free\" count=\"%u\" size=\"%llu\"/>\n" … … 317 342 "</malloc>", 318 343 malloc_calls, malloc_storage, 344 aalloc_calls, aalloc_storage, 319 345 calloc_calls, calloc_storage, 320 346 memalign_calls, memalign_storage, 347 amemalign_calls, amemalign_storage, 321 348 cmemalign_calls, cmemalign_storage, 349 resize_calls, resize_storage, 322 350 realloc_calls, realloc_storage, 323 351 free_calls, free_storage, … … 332 360 333 361 334 // static inline void noMemory() {335 // abort( "Heap memory exhausted at %zu bytes.\n"336 // "Possible cause is very large memory allocation and/or large amount of unfreed storage allocated by the program or system/library routines.",337 // ((char *)(sbrk( 0 )) - (char *)(heapManager.heapBegin)) );338 // } // noMemory339 340 341 static inline void checkAlign( size_t alignment ) {342 if ( alignment < libAlign() || ! libPow2( alignment ) ) {343 abort( "Alignment %zu for memory allocation is less than %d and/or not a power of 2.", alignment, libAlign() );344 } // if345 } // checkAlign346 347 348 static inline bool setHeapExpand( size_t value ) {349 if ( heapExpand < pageSize ) return true;350 heapExpand = value;351 return false;352 } // setHeapExpand353 354 355 362 // thunk problem 356 363 size_t Bsearchl( unsigned int key, const unsigned int * vals, size_t dim ) { … … 369 376 370 377 static inline bool setMmapStart( size_t value ) { // true => mmapped, false => sbrk 371 if ( value < pageSize || bucketSizes[NoBucketSizes - 1] < value ) return true;378 if ( value < __page_size || bucketSizes[NoBucketSizes - 1] < value ) return false; 372 379 mmapStart = value; // set global 373 380 … … 376 383 assert( maxBucketsUsed < NoBucketSizes ); // subscript failure ? 377 384 assert( mmapStart <= bucketSizes[maxBucketsUsed] ); // search failure ? 378 return false;385 return true; 379 386 } // setMmapStart 387 388 389 // <-------+----------------------------------------------------> bsize (bucket size) 390 // |header |addr 391 //================================================================================== 392 // align/offset | 393 // <-----------------<------------+-----------------------------> bsize (bucket size) 394 // |fake-header | addr 395 #define headerAddr( addr ) ((HeapManager.Storage.Header *)( (char *)addr - sizeof(HeapManager.Storage) )) 396 #define realHeader( header ) ((HeapManager.Storage.Header *)((char *)header - header->kind.fake.offset)) 397 398 // <-------<<--------------------- dsize ---------------------->> bsize (bucket size) 399 // |header |addr 400 //================================================================================== 401 // align/offset | 402 // <------------------------------<<---------- dsize --------->>> bsize (bucket size) 403 // |fake-header |addr 404 #define dataStorage( bsize, addr, header ) (bsize - ( (char *)addr - (char *)header )) 405 406 407 static inline void checkAlign( size_t alignment ) { 408 if ( alignment < libAlign() || ! is_pow2( alignment ) ) { 409 abort( "Alignment %zu for memory allocation is less than %d and/or not a power of 2.", alignment, libAlign() ); 410 } // if 411 } // checkAlign 380 412 381 413 … … 391 423 static inline void fakeHeader( HeapManager.Storage.Header *& header, size_t & alignment ) { 392 424 if ( unlikely( (header->kind.fake.alignment & 1) == 1 ) ) { // fake header ? 393 size_t offset = header->kind.fake.offset;394 425 alignment = header->kind.fake.alignment & -2; // remove flag from value 395 426 #ifdef __CFA_DEBUG__ 396 427 checkAlign( alignment ); // check alignment 397 428 #endif // __CFA_DEBUG__ 398 header = (HeapManager.Storage.Header *)((char *)header - offset); 429 header = realHeader( header ); // backup from fake to real header 430 } else { 431 alignment = libAlign(); // => no fake header 399 432 } // if 400 433 } // fakeHeader 401 434 402 435 403 // <-------+----------------------------------------------------> bsize (bucket size) 404 // |header |addr 405 //================================================================================== 406 // | alignment 407 // <-----------------<------------+-----------------------------> bsize (bucket size) 408 // |fake-header | addr 409 #define headerAddr( addr ) ((HeapManager.Storage.Header *)( (char *)addr - sizeof(HeapManager.Storage) )) 410 411 // <-------<<--------------------- dsize ---------------------->> bsize (bucket size) 412 // |header |addr 413 //================================================================================== 414 // | alignment 415 // <------------------------------<<---------- dsize --------->>> bsize (bucket size) 416 // |fake-header |addr 417 #define dataStorage( bsize, addr, header ) (bsize - ( (char *)addr - (char *)header )) 418 419 420 static inline bool headers( const char name[] __attribute__(( unused )), void * addr, HeapManager.Storage.Header *& header, HeapManager.FreeHeader *& freeElem, size_t & size, size_t & alignment ) with ( heapManager ) { 436 static inline bool headers( const char name[] __attribute__(( unused )), void * addr, HeapManager.Storage.Header *& header, HeapManager.FreeHeader *& freeElem, 437 size_t & size, size_t & alignment ) with( heapManager ) { 421 438 header = headerAddr( addr ); 422 439 423 if ( unlikely( heapEnd < addr ) ) {// mmapped ?440 if ( unlikely( addr < heapBegin || heapEnd < addr ) ) { // mmapped ? 424 441 fakeHeader( header, alignment ); 425 442 size = header->kind.real.blockSize & -3; // mmap size … … 428 445 429 446 #ifdef __CFA_DEBUG__ 430 checkHeader( addr < heapBegin ||header < (HeapManager.Storage.Header *)heapBegin, name, addr ); // bad low address ?447 checkHeader( header < (HeapManager.Storage.Header *)heapBegin, name, addr ); // bad low address ? 431 448 #endif // __CFA_DEBUG__ 432 449 … … 449 466 } // headers 450 467 451 452 static inline void * extend( size_t size ) with ( heapManager ) { 468 #ifdef __CFA_DEBUG__ 469 #if __SIZEOF_POINTER__ == 4 470 #define MASK 0xdeadbeef 471 #else 472 #define MASK 0xdeadbeefdeadbeef 473 #endif 474 #define STRIDE size_t 475 476 static void * Memset( void * addr, STRIDE size ) { // debug only 477 if ( size % sizeof(STRIDE) != 0 ) abort( "Memset() : internal error, size %zd not multiple of %zd.", size, sizeof(STRIDE) ); 478 if ( (STRIDE)addr % sizeof(STRIDE) != 0 ) abort( "Memset() : internal error, addr %p not multiple of %zd.", addr, sizeof(STRIDE) ); 479 480 STRIDE * end = (STRIDE *)addr + size / sizeof(STRIDE); 481 for ( STRIDE * p = (STRIDE *)addr; p < end; p += 1 ) *p = MASK; 482 return addr; 483 } // Memset 484 #endif // __CFA_DEBUG__ 485 486 487 #define NO_MEMORY_MSG "insufficient heap memory available for allocating %zd new bytes." 488 489 static inline void * extend( size_t size ) with( heapManager ) { 453 490 lock( extlock __cfaabi_dbg_ctx2 ); 454 491 ptrdiff_t rem = heapRemaining - size; … … 456 493 // If the size requested is bigger than the current remaining storage, increase the size of the heap. 457 494 458 size_t increase = libCeiling( size > heapExpand ? size : heapExpand, libAlign() ); 459 if ( sbrk( increase ) == (void *)-1 ) { 495 size_t increase = ceiling2( size > heapExpand ? size : heapExpand, __page_size ); 496 // Do not call abort or strerror( errno ) as they may call malloc. 497 if ( sbrk( increase ) == (void *)-1 ) { // failed, no memory ? 460 498 unlock( extlock ); 461 errno = ENOMEM; 462 return 0p; 499 __cfaabi_bits_print_nolock( STDERR_FILENO, NO_MEMORY_MSG, size ); 500 _exit( EXIT_FAILURE ); 501 } // if 502 if ( mprotect( (char *)heapEnd + heapRemaining, increase, __map_prot ) ) { 503 unlock( extlock ); 504 __cfaabi_bits_print_nolock( STDERR_FILENO, "extend() : internal error, mprotect failure, heapEnd:%p size:%zd, errno:%d.\n", heapEnd, increase, errno ); 505 _exit( EXIT_FAILURE ); 463 506 } // if 464 507 #ifdef __STATISTICS__ … … 468 511 #ifdef __CFA_DEBUG__ 469 512 // Set new memory to garbage so subsequent uninitialized usages might fail. 470 memset( (char *)heapEnd + heapRemaining, '\377', increase ); 513 memset( (char *)heapEnd + heapRemaining, '\xde', increase ); 514 //Memset( (char *)heapEnd + heapRemaining, increase ); 471 515 #endif // __CFA_DEBUG__ 472 516 rem = heapRemaining + increase - size; … … 481 525 482 526 483 static inline void * doMalloc( size_t size ) with 527 static inline void * doMalloc( size_t size ) with( heapManager ) { 484 528 HeapManager.Storage * block; // pointer to new block of storage 485 529 … … 487 531 // along with the block and is a multiple of the alignment size. 488 532 489 if ( unlikely( size > ~0ul- sizeof(HeapManager.Storage) ) ) return 0p;533 if ( unlikely( size > ULONG_MAX - sizeof(HeapManager.Storage) ) ) return 0p; 490 534 size_t tsize = size + sizeof(HeapManager.Storage); 491 535 if ( likely( tsize < mmapStart ) ) { // small size => sbrk … … 497 541 posn = Bsearchl( (unsigned int)tsize, bucketSizes, (size_t)maxBucketsUsed ); 498 542 HeapManager.FreeHeader * freeElem = &freeLists[posn]; 499 // #ifdef FASTLOOKUP 500 // if ( tsize < LookupSizes ) 501 // freeElem = &freeLists[lookup[tsize]]; 502 // else 503 // #endif // FASTLOOKUP 504 // freeElem = bsearchl( tsize, freeLists, (size_t)maxBucketsUsed ); // binary search 505 // HeapManager.FreeHeader * freeElem = 506 // #ifdef FASTLOOKUP 507 // tsize < LookupSizes ? &freeLists[lookup[tsize]] : 508 // #endif // FASTLOOKUP 509 // bsearchl( tsize, freeLists, (size_t)maxBucketsUsed ); // binary search 510 assert( freeElem <= &freeLists[maxBucketsUsed] ); // subscripting error ? 511 assert( tsize <= freeElem->blockSize ); // search failure ? 543 verify( freeElem <= &freeLists[maxBucketsUsed] ); // subscripting error ? 544 verify( tsize <= freeElem->blockSize ); // search failure ? 512 545 tsize = freeElem->blockSize; // total space needed for request 513 546 514 547 // Spin until the lock is acquired for this particular size of block. 515 548 516 #if defined( SPINLOCK )549 #if BUCKETLOCK == SPINLOCK 517 550 lock( freeElem->lock __cfaabi_dbg_ctx2 ); 518 551 block = freeElem->freeList; // remove node from stack 519 552 #else 520 block = freeElem->freeList.pop();521 #endif // SPINLOCK553 block = pop( freeElem->freeList ); 554 #endif // BUCKETLOCK 522 555 if ( unlikely( block == 0p ) ) { // no free block ? 523 #if defined( SPINLOCK )556 #if BUCKETLOCK == SPINLOCK 524 557 unlock( freeElem->lock ); 525 #endif // SPINLOCK558 #endif // BUCKETLOCK 526 559 527 560 // Freelist for that size was empty, so carve it out of the heap if there's enough left, or get some more … … 529 562 530 563 block = (HeapManager.Storage *)extend( tsize ); // mutual exclusion on call 531 if ( unlikely( block == 0p ) ) return 0p; 532 #if defined( SPINLOCK ) 564 #if BUCKETLOCK == SPINLOCK 533 565 } else { 534 566 freeElem->freeList = block->header.kind.real.next; 535 567 unlock( freeElem->lock ); 536 #endif // SPINLOCK568 #endif // BUCKETLOCK 537 569 } // if 538 570 539 571 block->header.kind.real.home = freeElem; // pointer back to free list of apropriate size 540 572 } else { // large size => mmap 541 if ( unlikely( size > ~0ul - pageSize ) ) return 0p;542 tsize = libCeiling( tsize, pageSize ); // must be multiple of page size573 if ( unlikely( size > ULONG_MAX - __page_size ) ) return 0p; 574 tsize = ceiling2( tsize, __page_size ); // must be multiple of page size 543 575 #ifdef __STATISTICS__ 544 576 __atomic_add_fetch( &mmap_calls, 1, __ATOMIC_SEQ_CST ); 545 577 __atomic_add_fetch( &mmap_storage, tsize, __ATOMIC_SEQ_CST ); 546 578 #endif // __STATISTICS__ 547 block = (HeapManager.Storage *)mmap( 0, tsize, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS, mmapFd, 0 ); 548 if ( block == (HeapManager.Storage *)MAP_FAILED ) { 579 580 block = (HeapManager.Storage *)mmap( 0, tsize, __map_prot, MAP_PRIVATE | MAP_ANONYMOUS, mmapFd, 0 ); 581 if ( block == (HeapManager.Storage *)MAP_FAILED ) { // failed ? 582 if ( errno == ENOMEM ) abort( NO_MEMORY_MSG, tsize ); // no memory 549 583 // Do not call strerror( errno ) as it may call malloc. 550 abort( "(HeapManager &)0x%p.doMalloc() : internal error, mmap failure, size:%zu err or:%d.", &heapManager, tsize, errno );551 } // 584 abort( "(HeapManager &)0x%p.doMalloc() : internal error, mmap failure, size:%zu errno:%d.", &heapManager, tsize, errno ); 585 } //if 552 586 #ifdef __CFA_DEBUG__ 553 587 // Set new memory to garbage so subsequent uninitialized usages might fail. 554 memset( block, '\377', tsize ); 588 memset( block, '\xde', tsize ); 589 //Memset( block, tsize ); 555 590 #endif // __CFA_DEBUG__ 556 591 block->header.kind.real.blockSize = tsize; // storage size for munmap 557 592 } // if 558 593 594 block->header.kind.real.size = size; // store allocation size 559 595 void * addr = &(block->data); // adjust off header to user bytes 596 verify( ((uintptr_t)addr & (libAlign() - 1)) == 0 ); // minimum alignment ? 560 597 561 598 #ifdef __CFA_DEBUG__ 562 assert( ((uintptr_t)addr & (libAlign() - 1)) == 0 ); // minimum alignment ? 563 __atomic_add_fetch( &allocFree, tsize, __ATOMIC_SEQ_CST ); 599 __atomic_add_fetch( &allocUnfreed, tsize, __ATOMIC_SEQ_CST ); 564 600 if ( traceHeap() ) { 565 601 enum { BufferSize = 64 }; 566 602 char helpText[BufferSize]; 567 603 int len = snprintf( helpText, BufferSize, "%p = Malloc( %zu ) (allocated %zu)\n", addr, size, tsize ); 568 // int len = snprintf( helpText, BufferSize, "Malloc %p %zu\n", addr, size );569 604 __cfaabi_bits_write( STDERR_FILENO, helpText, len ); // print debug/nodebug 570 605 } // if … … 575 610 576 611 577 static inline void doFree( void * addr ) with 612 static inline void doFree( void * addr ) with( heapManager ) { 578 613 #ifdef __CFA_DEBUG__ 579 614 if ( unlikely( heapManager.heapBegin == 0p ) ) { … … 592 627 #endif // __STATISTICS__ 593 628 if ( munmap( header, size ) == -1 ) { 594 #ifdef __CFA_DEBUG__595 629 abort( "Attempt to deallocate storage %p not allocated or with corrupt header.\n" 596 630 "Possible cause is invalid pointer.", 597 631 addr ); 598 #endif // __CFA_DEBUG__599 632 } // if 600 633 } else { 601 634 #ifdef __CFA_DEBUG__ 602 635 // Set free memory to garbage so subsequent usages might fail. 603 memset( ((HeapManager.Storage *)header)->data, '\377', freeElem->blockSize - sizeof( HeapManager.Storage ) ); 636 memset( ((HeapManager.Storage *)header)->data, '\xde', freeElem->blockSize - sizeof( HeapManager.Storage ) ); 637 //Memset( ((HeapManager.Storage *)header)->data, freeElem->blockSize - sizeof( HeapManager.Storage ) ); 604 638 #endif // __CFA_DEBUG__ 605 639 … … 607 641 free_storage += size; 608 642 #endif // __STATISTICS__ 609 #if defined( SPINLOCK )643 #if BUCKETLOCK == SPINLOCK 610 644 lock( freeElem->lock __cfaabi_dbg_ctx2 ); // acquire spin lock 611 645 header->kind.real.next = freeElem->freeList; // push on stack … … 613 647 unlock( freeElem->lock ); // release spin lock 614 648 #else 615 freeElem->freeList.push(*(HeapManager.Storage *)header );616 #endif // SPINLOCK649 push( freeElem->freeList, *(HeapManager.Storage *)header ); 650 #endif // BUCKETLOCK 617 651 } // if 618 652 619 653 #ifdef __CFA_DEBUG__ 620 __atomic_add_fetch( &alloc Free, -size, __ATOMIC_SEQ_CST );654 __atomic_add_fetch( &allocUnfreed, -size, __ATOMIC_SEQ_CST ); 621 655 if ( traceHeap() ) { 622 enum { BufferSize = 64 }; 623 char helpText[BufferSize]; 656 char helpText[64]; 624 657 int len = snprintf( helpText, sizeof(helpText), "Free( %p ) size:%zu\n", addr, size ); 625 658 __cfaabi_bits_write( STDERR_FILENO, helpText, len ); // print debug/nodebug … … 629 662 630 663 631 size_t prtFree( HeapManager & manager ) with 664 size_t prtFree( HeapManager & manager ) with( manager ) { 632 665 size_t total = 0; 633 666 #ifdef __STATISTICS__ … … 641 674 #endif // __STATISTICS__ 642 675 643 #if defined( SPINLOCK )676 #if BUCKETLOCK == SPINLOCK 644 677 for ( HeapManager.Storage * p = freeLists[i].freeList; p != 0p; p = p->header.kind.real.next ) { 645 678 #else 646 for ( HeapManager.Storage * p = freeLists[i].freeList.top(); p != 0p; p = p->header.kind.real.next.top ) { 647 #endif // SPINLOCK 679 for(;;) { 680 // for ( HeapManager.Storage * p = top( freeLists[i].freeList ); p != 0p; p = (p)`next->top ) { 681 // for ( HeapManager.Storage * p = top( freeLists[i].freeList ); p != 0p; /* p = getNext( p )->top */) { 682 // HeapManager.Storage * temp = p->header.kind.real.next.top; // FIX ME: direct assignent fails, initialization works` 683 // typeof(p) temp = (( p )`next)->top; // FIX ME: direct assignent fails, initialization works` 684 // p = temp; 685 #endif // BUCKETLOCK 648 686 total += size; 649 687 #ifdef __STATISTICS__ … … 665 703 666 704 667 static void ?{}( HeapManager & manager ) with ( manager ) { 668 pageSize = sysconf( _SC_PAGESIZE ); 705 static void ?{}( HeapManager & manager ) with( manager ) { 706 __page_size = sysconf( _SC_PAGESIZE ); 707 __map_prot = PROT_READ | PROT_WRITE | PROT_EXEC; 669 708 670 709 for ( unsigned int i = 0; i < NoBucketSizes; i += 1 ) { // initialize the free lists … … 680 719 #endif // FASTLOOKUP 681 720 682 if ( setMmapStart( default_mmap_start() ) ) {721 if ( ! setMmapStart( default_mmap_start() ) ) { 683 722 abort( "HeapManager : internal error, mmap start initialization failure." ); 684 723 } // if … … 686 725 687 726 char * end = (char *)sbrk( 0 ); 688 sbrk( (char *)libCeiling( (long unsigned int)end, libAlign() ) - end ); // move start of heap to multiple of alignment 689 heapBegin = heapEnd = sbrk( 0 ); // get new start point 727 heapBegin = heapEnd = sbrk( (char *)ceiling2( (long unsigned int)end, __page_size ) - end ); // move start of heap to multiple of alignment 690 728 } // HeapManager 691 729 … … 695 733 if ( traceHeapTerm() ) { 696 734 printStats(); 697 // if ( prtfree() ) prtFree( heapManager, true );735 // prtUnfreed() called in heapAppStop() 698 736 } // if 699 737 #endif // __STATISTICS__ … … 704 742 void memory_startup( void ) { 705 743 #ifdef __CFA_DEBUG__ 706 if ( unlikely( heapBoot ) ) { // check for recursion during system boot 707 // DO NOT USE STREAMS AS THEY MAY BE UNAVAILABLE AT THIS POINT. 744 if ( heapBoot ) { // check for recursion during system boot 708 745 abort( "boot() : internal error, recursively invoked during system boot." ); 709 746 } // if … … 711 748 #endif // __CFA_DEBUG__ 712 749 713 // assert( heapManager.heapBegin != 0 );750 //verify( heapManager.heapBegin != 0 ); 714 751 //heapManager{}; 715 if ( heapManager.heapBegin == 0p ) heapManager{}; 752 if ( heapManager.heapBegin == 0p ) heapManager{}; // sanity check 716 753 } // memory_startup 717 754 … … 723 760 724 761 static inline void * mallocNoStats( size_t size ) { // necessary for malloc statistics 725 //assert( heapManager.heapBegin != 0 ); 726 if ( unlikely( heapManager.heapBegin == 0p ) ) heapManager{}; // called before memory_startup ? 727 void * addr = doMalloc( size ); 728 if ( unlikely( addr == 0p ) ) errno = ENOMEM; // POSIX 729 return addr; 762 verify( heapManager.heapBegin != 0p ); // called before memory_startup ? 763 if ( unlikely( size ) == 0 ) return 0p; // 0 BYTE ALLOCATION RETURNS NULL POINTER 764 765 #if __SIZEOF_POINTER__ == 8 766 verify( size < ((typeof(size_t))1 << 48) ); 767 #endif // __SIZEOF_POINTER__ == 8 768 return doMalloc( size ); 730 769 } // mallocNoStats 731 770 732 771 733 static inline void * callocNoStats( size_t noOfElems, size_t elemSize ) { 734 size_t size = noOfElems * elemSize; 772 static inline void * callocNoStats( size_t dim, size_t elemSize ) { 773 size_t size = dim * elemSize; 774 if ( unlikely( size ) == 0 ) return 0p; // 0 BYTE ALLOCATION RETURNS NULL POINTER 735 775 char * addr = (char *)mallocNoStats( size ); 736 if ( unlikely( addr == 0p ) ) return 0p;737 776 738 777 HeapManager.Storage.Header * header; 739 778 HeapManager.FreeHeader * freeElem; 740 779 size_t bsize, alignment; 741 bool mapped __attribute__(( unused )) = headers( "calloc", addr, header, freeElem, bsize, alignment );742 780 #ifndef __CFA_DEBUG__ 781 bool mapped = 782 #endif // __CFA_DEBUG__ 783 headers( "calloc", addr, header, freeElem, bsize, alignment ); 784 #ifndef __CFA_DEBUG__ 785 743 786 // Mapped storage is zero filled, but in debug mode mapped memory is scrubbed in doMalloc, so it has to be reset to zero. 744 787 if ( ! mapped ) 745 788 #endif // __CFA_DEBUG__ 746 // Zero entire data space even when > than size => realloc without a new allocation and zero fill works. 747 // <-------00000000000000000000000000000000000000000000000000000> bsize (bucket size) 789 // <-------0000000000000000000000000000UUUUUUUUUUUUUUUUUUUUUUUUU> bsize (bucket size) U => undefined 748 790 // `-header`-addr `-size 749 memset( addr, '\0', bsize - sizeof(HeapManager.Storage) );// set to zeros791 memset( addr, '\0', size ); // set to zeros 750 792 751 793 header->kind.real.blockSize |= 2; // mark as zero filled … … 754 796 755 797 756 static inline void * memalignNoStats( size_t alignment, size_t size ) { // necessary for malloc statistics 798 static inline void * memalignNoStats( size_t alignment, size_t size ) { 799 if ( unlikely( size ) == 0 ) return 0p; // 0 BYTE ALLOCATION RETURNS NULL POINTER 800 757 801 #ifdef __CFA_DEBUG__ 758 802 checkAlign( alignment ); // check alignment … … 772 816 // add sizeof(Storage) for fake header 773 817 char * addr = (char *)mallocNoStats( size + alignment - libAlign() + sizeof(HeapManager.Storage) ); 774 if ( unlikely( addr == 0p ) ) return addr;775 818 776 819 // address in the block of the "next" alignment address 777 char * user = (char *) libCeiling( (uintptr_t)(addr + sizeof(HeapManager.Storage)), alignment );820 char * user = (char *)ceiling2( (uintptr_t)(addr + sizeof(HeapManager.Storage)), alignment ); 778 821 779 822 // address of header from malloc 780 823 HeapManager.Storage.Header * realHeader = headerAddr( addr ); 824 realHeader->kind.real.size = size; // correct size to eliminate above alignment offset 781 825 // address of fake header * before* the alignment location 782 826 HeapManager.Storage.Header * fakeHeader = headerAddr( user ); … … 790 834 791 835 792 static inline void * cmemalignNoStats( size_t alignment, size_t noOfElems, size_t elemSize ) { 793 size_t size = noOfElems * elemSize; 836 static inline void * cmemalignNoStats( size_t alignment, size_t dim, size_t elemSize ) { 837 size_t size = dim * elemSize; 838 if ( unlikely( size ) == 0 ) return 0p; // 0 BYTE ALLOCATION RETURNS NULL POINTER 794 839 char * addr = (char *)memalignNoStats( alignment, size ); 795 if ( unlikely( addr == 0p ) ) return 0p; 840 796 841 HeapManager.Storage.Header * header; 797 842 HeapManager.FreeHeader * freeElem; 798 843 size_t bsize; 799 bool mapped __attribute__(( unused )) = headers( "cmemalign", addr, header, freeElem, bsize, alignment );800 844 #ifndef __CFA_DEBUG__ 845 bool mapped = 846 #endif // __CFA_DEBUG__ 847 headers( "cmemalign", addr, header, freeElem, bsize, alignment ); 848 801 849 // Mapped storage is zero filled, but in debug mode mapped memory is scrubbed in doMalloc, so it has to be reset to zero. 850 #ifndef __CFA_DEBUG__ 802 851 if ( ! mapped ) 803 852 #endif // __CFA_DEBUG__ 804 memset( addr, '\0', dataStorage( bsize, addr, header ) ); // set to zeros 805 header->kind.real.blockSize |= 2; // mark as zero filled 806 853 // <-------0000000000000000000000000000UUUUUUUUUUUUUUUUUUUUUUUUU> bsize (bucket size) U => undefined 854 // `-header`-addr `-size 855 memset( addr, '\0', size ); // set to zeros 856 857 header->kind.real.blockSize |= 2; // mark as zero filled 807 858 return addr; 808 859 } // cmemalignNoStats 809 860 810 861 811 // supported mallopt options812 #ifndef M_MMAP_THRESHOLD813 #define M_MMAP_THRESHOLD (-1)814 #endif // M_TOP_PAD815 #ifndef M_TOP_PAD816 #define M_TOP_PAD (-2)817 #endif // M_TOP_PAD818 819 820 862 extern "C" { 821 // The malloc() function allocates size bytes and returns a pointer to the allocated memory. The memory is not 822 // initialized. If size is 0, then malloc() returns either 0p, or a unique pointer value that can later be 823 // successfully passed to free(). 863 // Allocates size bytes and returns a pointer to the allocated memory. The contents are undefined. If size is 0, 864 // then malloc() returns a unique pointer value that can later be successfully passed to free(). 824 865 void * malloc( size_t size ) { 825 866 #ifdef __STATISTICS__ … … 831 872 } // malloc 832 873 833 // The calloc() function allocates memory for an array of nmemb elements of size bytes each and returns a pointer to 834 // the allocated memory. The memory is set to zero. If nmemb or size is 0, then calloc() returns either 0p, or a 835 // unique pointer value that can later be successfully passed to free(). 836 void * calloc( size_t noOfElems, size_t elemSize ) { 874 875 // Same as malloc() except size bytes is an array of dim elements each of elemSize bytes. 876 void * aalloc( size_t dim, size_t elemSize ) { 877 size_t size = dim * elemSize; 878 #ifdef __STATISTICS__ 879 __atomic_add_fetch( &aalloc_calls, 1, __ATOMIC_SEQ_CST ); 880 __atomic_add_fetch( &aalloc_storage, size, __ATOMIC_SEQ_CST ); 881 #endif // __STATISTICS__ 882 883 return mallocNoStats( size ); 884 } // aalloc 885 886 887 // Same as aalloc() with memory set to zero. 888 void * calloc( size_t dim, size_t elemSize ) { 837 889 #ifdef __STATISTICS__ 838 890 __atomic_add_fetch( &calloc_calls, 1, __ATOMIC_SEQ_CST ); 839 __atomic_add_fetch( &calloc_storage, noOfElems* elemSize, __ATOMIC_SEQ_CST );840 #endif // __STATISTICS__ 841 842 return callocNoStats( noOfElems, elemSize );891 __atomic_add_fetch( &calloc_storage, dim * elemSize, __ATOMIC_SEQ_CST ); 892 #endif // __STATISTICS__ 893 894 return callocNoStats( dim, elemSize ); 843 895 } // calloc 844 896 845 // The realloc() function changes the size of the memory block pointed to by ptr to size bytes. The contents will be 846 // unchanged in the range from the start of the region up to the minimum of the old and new sizes. If the new size 847 // is larger than the old size, the added memory will not be initialized. If ptr is 0p, then the call is 848 // equivalent to malloc(size), for all values of size; if size is equal to zero, and ptr is not 0p, then the call 849 // is equivalent to free(ptr). Unless ptr is 0p, it must have been returned by an earlier call to malloc(), 850 // calloc() or realloc(). If the area pointed to was moved, a free(ptr) is done. 851 void * realloc( void * oaddr, size_t size ) { 852 #ifdef __STATISTICS__ 853 __atomic_add_fetch( &realloc_calls, 1, __ATOMIC_SEQ_CST ); 897 898 // Change the size of the memory block pointed to by oaddr to size bytes. The contents are undefined. If oaddr is 899 // 0p, then the call is equivalent to malloc(size), for all values of size; if size is equal to zero, and oaddr is 900 // not 0p, then the call is equivalent to free(oaddr). Unless oaddr is 0p, it must have been returned by an earlier 901 // call to malloc(), alloc(), calloc() or realloc(). If the area pointed to was moved, a free(oaddr) is done. 902 void * resize( void * oaddr, size_t size ) { 903 #ifdef __STATISTICS__ 904 __atomic_add_fetch( &resize_calls, 1, __ATOMIC_SEQ_CST ); 854 905 #endif // __STATISTICS__ 855 906 856 907 // If size is equal to 0, either NULL or a pointer suitable to be passed to free() is returned. 857 if ( unlikely( size == 0 ) ) { free( oaddr ); return mallocNoStats( size ); } // special cases 858 if ( unlikely( oaddr == 0p ) ) return mallocNoStats( size ); 908 if ( unlikely( size == 0 ) ) { free( oaddr ); return 0p; } // special cases 909 if ( unlikely( oaddr == 0p ) ) { 910 #ifdef __STATISTICS__ 911 __atomic_add_fetch( &resize_storage, size, __ATOMIC_SEQ_CST ); 912 #endif // __STATISTICS__ 913 return mallocNoStats( size ); 914 } // if 859 915 860 916 HeapManager.Storage.Header * header; 861 917 HeapManager.FreeHeader * freeElem; 862 size_t bsize, oalign = 0; 918 size_t bsize, oalign; 919 headers( "resize", oaddr, header, freeElem, bsize, oalign ); 920 size_t odsize = dataStorage( bsize, oaddr, header ); // data storage available in bucket 921 922 // same size, DO NOT preserve STICKY PROPERTIES. 923 if ( oalign == libAlign() && size <= odsize && odsize <= size * 2 ) { // allow 50% wasted storage for smaller size 924 header->kind.real.blockSize &= -2; // no alignment and turn off 0 fill 925 header->kind.real.size = size; // reset allocation size 926 return oaddr; 927 } // if 928 929 #ifdef __STATISTICS__ 930 __atomic_add_fetch( &resize_storage, size, __ATOMIC_SEQ_CST ); 931 #endif // __STATISTICS__ 932 933 // change size, DO NOT preserve STICKY PROPERTIES. 934 free( oaddr ); 935 return mallocNoStats( size ); // create new area 936 } // resize 937 938 939 // Same as resize() but the contents are unchanged in the range from the start of the region up to the minimum of 940 // the old and new sizes. 941 void * realloc( void * oaddr, size_t size ) { 942 #ifdef __STATISTICS__ 943 __atomic_add_fetch( &realloc_calls, 1, __ATOMIC_SEQ_CST ); 944 #endif // __STATISTICS__ 945 946 // If size is equal to 0, either NULL or a pointer suitable to be passed to free() is returned. 947 if ( unlikely( size == 0 ) ) { free( oaddr ); return 0p; } // special cases 948 if ( unlikely( oaddr == 0p ) ) { 949 #ifdef __STATISTICS__ 950 __atomic_add_fetch( &realloc_storage, size, __ATOMIC_SEQ_CST ); 951 #endif // __STATISTICS__ 952 return mallocNoStats( size ); 953 } // if 954 955 HeapManager.Storage.Header * header; 956 HeapManager.FreeHeader * freeElem; 957 size_t bsize, oalign; 863 958 headers( "realloc", oaddr, header, freeElem, bsize, oalign ); 864 959 865 960 size_t odsize = dataStorage( bsize, oaddr, header ); // data storage available in bucket 866 if ( size <= odsize && odsize <= size * 2 ) { // allow up to 50% wasted storage in smaller size 867 // Do not know size of original allocation => cannot do 0 fill for any additional space because do not know 868 // where to start filling, i.e., do not overwrite existing values in space. 869 // 870 // This case does not result in a new profiler entry because the previous one still exists and it must match with 871 // the free for this memory. Hence, this realloc does not appear in the profiler output. 961 size_t osize = header->kind.real.size; // old allocation size 962 bool ozfill = (header->kind.real.blockSize & 2); // old allocation zero filled 963 if ( unlikely( size <= odsize ) && odsize <= size * 2 ) { // allow up to 50% wasted storage 964 header->kind.real.size = size; // reset allocation size 965 if ( unlikely( ozfill ) && size > osize ) { // previous request zero fill and larger ? 966 memset( (char *)oaddr + osize, '\0', size - osize ); // initialize added storage 967 } // if 872 968 return oaddr; 873 969 } // if 874 970 875 971 #ifdef __STATISTICS__ 876 __atomic_add_fetch( &realloc_storage, size, __ATOMIC_SEQ_CST );972 __atomic_add_fetch( &realloc_storage, size, __ATOMIC_SEQ_CST ); 877 973 #endif // __STATISTICS__ 878 974 … … 880 976 881 977 void * naddr; 882 if ( unlikely( oalign != 0 ) ) { // previous request memalign? 883 if ( unlikely( header->kind.real.blockSize & 2 ) ) { // previous request zero fill 884 naddr = cmemalignNoStats( oalign, 1, size ); // create new aligned area 885 } else { 886 naddr = memalignNoStats( oalign, size ); // create new aligned area 978 if ( likely( oalign == libAlign() ) ) { // previous request not aligned ? 979 naddr = mallocNoStats( size ); // create new area 980 } else { 981 naddr = memalignNoStats( oalign, size ); // create new aligned area 982 } // if 983 984 headers( "realloc", naddr, header, freeElem, bsize, oalign ); 985 memcpy( naddr, oaddr, min( osize, size ) ); // copy bytes 986 free( oaddr ); 987 988 if ( unlikely( ozfill ) ) { // previous request zero fill ? 989 header->kind.real.blockSize |= 2; // mark new request as zero filled 990 if ( size > osize ) { // previous request larger ? 991 memset( (char *)naddr + osize, '\0', size - osize ); // initialize added storage 887 992 } // if 888 } else { 889 if ( unlikely( header->kind.real.blockSize & 2 ) ) { // previous request zero fill 890 naddr = callocNoStats( 1, size ); // create new area 891 } else { 892 naddr = mallocNoStats( size ); // create new area 893 } // if 894 } // if 895 if ( unlikely( naddr == 0p ) ) return 0p; 896 897 headers( "realloc", naddr, header, freeElem, bsize, oalign ); 898 size_t ndsize = dataStorage( bsize, naddr, header ); // data storage avilable in bucket 899 // To preserve prior fill, the entire bucket must be copied versus the size. 900 memcpy( naddr, oaddr, MIN( odsize, ndsize ) ); // copy bytes 901 free( oaddr ); 993 } // if 902 994 return naddr; 903 995 } // realloc 904 996 905 // The obsolete function memalign() allocates size bytes and returns a pointer to the allocated memory. The memory 906 // address will be a multiple of alignment, which must be a power of two.997 998 // Same as malloc() except the memory address is a multiple of alignment, which must be a power of two. (obsolete) 907 999 void * memalign( size_t alignment, size_t size ) { 908 1000 #ifdef __STATISTICS__ … … 915 1007 916 1008 917 // The cmemalign() function is the same as calloc() with memory alignment. 918 void * cmemalign( size_t alignment, size_t noOfElems, size_t elemSize ) { 1009 // Same as aalloc() with memory alignment. 1010 void * amemalign( size_t alignment, size_t dim, size_t elemSize ) { 1011 size_t size = dim * elemSize; 919 1012 #ifdef __STATISTICS__ 920 1013 __atomic_add_fetch( &cmemalign_calls, 1, __ATOMIC_SEQ_CST ); 921 __atomic_add_fetch( &cmemalign_storage, noOfElems * elemSize, __ATOMIC_SEQ_CST ); 922 #endif // __STATISTICS__ 923 924 return cmemalignNoStats( alignment, noOfElems, elemSize ); 1014 __atomic_add_fetch( &cmemalign_storage, size, __ATOMIC_SEQ_CST ); 1015 #endif // __STATISTICS__ 1016 1017 return memalignNoStats( alignment, size ); 1018 } // amemalign 1019 1020 1021 // Same as calloc() with memory alignment. 1022 void * cmemalign( size_t alignment, size_t dim, size_t elemSize ) { 1023 #ifdef __STATISTICS__ 1024 __atomic_add_fetch( &cmemalign_calls, 1, __ATOMIC_SEQ_CST ); 1025 __atomic_add_fetch( &cmemalign_storage, dim * elemSize, __ATOMIC_SEQ_CST ); 1026 #endif // __STATISTICS__ 1027 1028 return cmemalignNoStats( alignment, dim, elemSize ); 925 1029 } // cmemalign 926 1030 927 // The function aligned_alloc() is the same as memalign(), except for the added restriction that size should be a 928 // multiple of alignment. 1031 1032 // Same as memalign(), but ISO/IEC 2011 C11 Section 7.22.2 states: the value of size shall be an integral multiple 1033 // of alignment. This requirement is universally ignored. 929 1034 void * aligned_alloc( size_t alignment, size_t size ) { 930 1035 return memalign( alignment, size ); … … 932 1037 933 1038 934 // The function posix_memalign() allocates size bytes and places the address of the allocated memory in *memptr. The935 // address of the allocated memory will be a multiple of alignment, which must be a power of two and a multiple of936 // sizeof(void *). If size is 0, then posix_memalign() returns either 0p, or a unique pointer value that can later937 // be successfully passed tofree(3).1039 // Allocates size bytes and places the address of the allocated memory in *memptr. The address of the allocated 1040 // memory shall be a multiple of alignment, which must be a power of two and a multiple of sizeof(void *). If size 1041 // is 0, then posix_memalign() returns either 0p, or a unique pointer value that can later be successfully passed to 1042 // free(3). 938 1043 int posix_memalign( void ** memptr, size_t alignment, size_t size ) { 939 if ( alignment < sizeof(void *) || ! libPow2( alignment ) ) return EINVAL; // check alignment1044 if ( alignment < libAlign() || ! is_pow2( alignment ) ) return EINVAL; // check alignment 940 1045 * memptr = memalign( alignment, size ); 941 if ( unlikely( * memptr == 0p ) ) return ENOMEM;942 1046 return 0; 943 1047 } // posix_memalign 944 1048 945 // The obsolete function valloc() allocates size bytes and returns a pointer to the allocated memory. The memory 946 // address will be a multiple of the page size. It is equivalent to memalign(sysconf(_SC_PAGESIZE),size). 1049 1050 // Allocates size bytes and returns a pointer to the allocated memory. The memory address shall be a multiple of the 1051 // page size. It is equivalent to memalign(sysconf(_SC_PAGESIZE),size). 947 1052 void * valloc( size_t size ) { 948 return memalign( pageSize, size );1053 return memalign( __page_size, size ); 949 1054 } // valloc 950 1055 951 1056 952 // The free() function frees the memory space pointed to by ptr, which must have been returned by a previous call to 953 // malloc(), calloc() or realloc(). Otherwise, or if free(ptr) has already been called before, undefined behavior 954 // occurs. If ptr is 0p, no operation is performed. 1057 // Same as valloc but rounds size to multiple of page size. 1058 void * pvalloc( size_t size ) { 1059 return memalign( __page_size, ceiling2( size, __page_size ) ); 1060 } // pvalloc 1061 1062 1063 // Frees the memory space pointed to by ptr, which must have been returned by a previous call to malloc(), calloc() 1064 // or realloc(). Otherwise, or if free(ptr) has already been called before, undefined behaviour occurs. If ptr is 1065 // 0p, no operation is performed. 955 1066 void free( void * addr ) { 956 1067 #ifdef __STATISTICS__ … … 973 1084 974 1085 975 // The malloc_alignment() function returns the alignment of theallocation.1086 // Returns the alignment of an allocation. 976 1087 size_t malloc_alignment( void * addr ) { 977 1088 if ( unlikely( addr == 0p ) ) return libAlign(); // minimum alignment … … 980 1091 return header->kind.fake.alignment & -2; // remove flag from value 981 1092 } else { 982 return libAlign 1093 return libAlign(); // minimum alignment 983 1094 } // if 984 1095 } // malloc_alignment 985 1096 986 1097 987 // The malloc_zero_fill() function returns true if the allocation is zero filled, i.e., initially allocated by calloc(). 1098 // Set the alignment for an the allocation and return previous alignment or 0 if no alignment. 1099 size_t $malloc_alignment_set( void * addr, size_t alignment ) { 1100 if ( unlikely( addr == 0p ) ) return libAlign(); // minimum alignment 1101 size_t ret; 1102 HeapManager.Storage.Header * header = headerAddr( addr ); 1103 if ( (header->kind.fake.alignment & 1) == 1 ) { // fake header ? 1104 ret = header->kind.fake.alignment & -2; // remove flag from old value 1105 header->kind.fake.alignment = alignment | 1; // add flag to new value 1106 } else { 1107 ret = 0; // => no alignment to change 1108 } // if 1109 return ret; 1110 } // $malloc_alignment_set 1111 1112 1113 // Returns true if the allocation is zero filled, e.g., allocated by calloc(). 988 1114 bool malloc_zero_fill( void * addr ) { 989 1115 if ( unlikely( addr == 0p ) ) return false; // null allocation is not zero fill 990 1116 HeapManager.Storage.Header * header = headerAddr( addr ); 991 1117 if ( (header->kind.fake.alignment & 1) == 1 ) { // fake header ? 992 header = (HeapManager.Storage.Header *)((char *)header - header->kind.fake.offset);993 } // if 994 return (header->kind.real.blockSize & 2) != 0; // zero filled (calloc/cmemalign)?1118 header = realHeader( header ); // backup from fake to real header 1119 } // if 1120 return (header->kind.real.blockSize & 2) != 0; // zero filled ? 995 1121 } // malloc_zero_fill 996 1122 997 998 // The malloc_usable_size() function returns the number of usable bytes in the block pointed to by ptr, a pointer to 999 // a block of memory allocated by malloc(3) or a related function. 1123 // Set allocation is zero filled and return previous zero filled. 1124 bool $malloc_zero_fill_set( void * addr ) { 1125 if ( unlikely( addr == 0p ) ) return false; // null allocation is not zero fill 1126 HeapManager.Storage.Header * header = headerAddr( addr ); 1127 if ( (header->kind.fake.alignment & 1) == 1 ) { // fake header ? 1128 header = realHeader( header ); // backup from fake to real header 1129 } // if 1130 bool ret = (header->kind.real.blockSize & 2) != 0; // zero filled ? 1131 header->kind.real.blockSize |= 2; // mark as zero filled 1132 return ret; 1133 } // $malloc_zero_fill_set 1134 1135 1136 // Returns original total allocation size (not bucket size) => array size is dimension * sizeif(T). 1137 size_t malloc_size( void * addr ) { 1138 if ( unlikely( addr == 0p ) ) return 0; // null allocation has zero size 1139 HeapManager.Storage.Header * header = headerAddr( addr ); 1140 if ( (header->kind.fake.alignment & 1) == 1 ) { // fake header ? 1141 header = realHeader( header ); // backup from fake to real header 1142 } // if 1143 return header->kind.real.size; 1144 } // malloc_size 1145 1146 // Set allocation size and return previous size. 1147 size_t $malloc_size_set( void * addr, size_t size ) { 1148 if ( unlikely( addr == 0p ) ) return 0; // null allocation has 0 size 1149 HeapManager.Storage.Header * header = headerAddr( addr ); 1150 if ( (header->kind.fake.alignment & 1) == 1 ) { // fake header ? 1151 header = realHeader( header ); // backup from fake to real header 1152 } // if 1153 size_t ret = header->kind.real.size; 1154 header->kind.real.size = size; 1155 return ret; 1156 } // $malloc_size_set 1157 1158 1159 // Returns the number of usable bytes in the block pointed to by ptr, a pointer to a block of memory allocated by 1160 // malloc or a related function. 1000 1161 size_t malloc_usable_size( void * addr ) { 1001 1162 if ( unlikely( addr == 0p ) ) return 0; // null allocation has 0 size … … 1005 1166 1006 1167 headers( "malloc_usable_size", addr, header, freeElem, bsize, alignment ); 1007 return dataStorage( bsize, addr, header ); // data storage in bucket1168 return dataStorage( bsize, addr, header ); // data storage in bucket 1008 1169 } // malloc_usable_size 1009 1170 1010 1171 1011 // The malloc_stats() function prints (on default standard error) statistics about memory allocated by malloc(3) and 1012 // related functions. 1172 // Prints (on default standard error) statistics about memory allocated by malloc and related functions. 1013 1173 void malloc_stats( void ) { 1014 1174 #ifdef __STATISTICS__ … … 1018 1178 } // malloc_stats 1019 1179 1020 // The malloc_stats_fd() function changes the file descripter where malloc_stats() writes the statistics. 1180 1181 // Changes the file descripter where malloc_stats() writes statistics. 1021 1182 int malloc_stats_fd( int fd __attribute__(( unused )) ) { 1022 1183 #ifdef __STATISTICS__ 1023 int temp = stat fd;1024 stat fd = fd;1184 int temp = stat_fd; 1185 stat_fd = fd; 1025 1186 return temp; 1026 1187 #else … … 1030 1191 1031 1192 1032 // The mallopt() function adjusts parameters that control the behavior of the memory-allocation functions (see 1033 // malloc(3)). The param argument specifies the parameter to be modified, and value specifies the new value for that 1034 // parameter. 1193 // Adjusts parameters that control the behaviour of the memory-allocation functions (see malloc). The param argument 1194 // specifies the parameter to be modified, and value specifies the new value for that parameter. 1035 1195 int mallopt( int option, int value ) { 1036 1196 choose( option ) { 1037 1197 case M_TOP_PAD: 1038 if ( setHeapExpand( value ) )return 1;1198 heapExpand = ceiling2( value, __page_size ); return 1; 1039 1199 case M_MMAP_THRESHOLD: 1040 1200 if ( setMmapStart( value ) ) return 1; 1201 break; 1041 1202 } // switch 1042 1203 return 0; // error, unsupported 1043 1204 } // mallopt 1044 1205 1045 // The malloc_trim() function attempts to release free memory at the top of the heap (by calling sbrk(2) with a 1046 // suitable argument).1206 1207 // Attempt to release free memory at the top of the heap (by calling sbrk with a suitable argument). 1047 1208 int malloc_trim( size_t ) { 1048 1209 return 0; // => impossible to release memory … … 1050 1211 1051 1212 1052 // The malloc_info() function exports an XML string that describes the current state of the memory-allocation1053 // implementation in the caller. The string is printed on the file stream stream. The exported string includes1054 // information about all arenas (see malloc(3)).1213 // Exports an XML string that describes the current state of the memory-allocation implementation in the caller. 1214 // The string is printed on the file stream stream. The exported string includes information about all arenas (see 1215 // malloc). 1055 1216 int malloc_info( int options, FILE * stream ) { 1056 if ( options != 0 ) { errno = EINVAL; return -1; } 1217 if ( options != 0 ) { errno = EINVAL; return -1; } 1218 #ifdef __STATISTICS__ 1057 1219 return printStatsXML( stream ); 1220 #else 1221 return 0; // unsupported 1222 #endif // __STATISTICS__ 1058 1223 } // malloc_info 1059 1224 1060 1225 1061 // The malloc_get_state() function records the current state of all malloc(3) internal bookkeeping variables (but1062 // not the actual contents of the heap or the state of malloc_hook(3) functions pointers). The state is recorded in1063 // a system-dependent opaque data structure dynamically allocated via malloc(3), and a pointer to that data1064 // structure is returned as the function result. (It is the caller's responsibility to free(3)this memory.)1226 // Records the current state of all malloc internal bookkeeping variables (but not the actual contents of the heap 1227 // or the state of malloc_hook functions pointers). The state is recorded in a system-dependent opaque data 1228 // structure dynamically allocated via malloc, and a pointer to that data structure is returned as the function 1229 // result. (The caller must free this memory.) 1065 1230 void * malloc_get_state( void ) { 1066 1231 return 0p; // unsupported … … 1068 1233 1069 1234 1070 // The malloc_set_state() function restores the state of all malloc(3) internal bookkeeping variables to the values1071 // recorded in the opaque datastructure pointed to by state.1072 int malloc_set_state( void * ptr) {1235 // Restores the state of all malloc internal bookkeeping variables to the values recorded in the opaque data 1236 // structure pointed to by state. 1237 int malloc_set_state( void * ) { 1073 1238 return 0; // unsupported 1074 1239 } // malloc_set_state … … 1077 1242 1078 1243 // Must have CFA linkage to overload with C linkage realloc. 1079 void * re alloc( void * oaddr, size_t nalign, size_t size ) {1244 void * resize( void * oaddr, size_t nalign, size_t size ) { 1080 1245 #ifdef __STATISTICS__ 1081 __atomic_add_fetch( &re alloc_calls, 1, __ATOMIC_SEQ_CST );1246 __atomic_add_fetch( &resize_calls, 1, __ATOMIC_SEQ_CST ); 1082 1247 #endif // __STATISTICS__ 1083 1248 1084 // If size is equal to 0, either NULL or a pointer suitable to be passed to free() is returned. 1085 if ( unlikely( size == 0 ) ) { free( oaddr ); return mallocNoStats( size ); } // special cases 1086 if ( unlikely( oaddr == 0p ) ) return mallocNoStats( size ); 1087 1088 if ( unlikely( nalign == 0 ) ) nalign = libAlign(); // reset alignment to minimum 1249 if ( unlikely( nalign < libAlign() ) ) nalign = libAlign(); // reset alignment to minimum 1089 1250 #ifdef __CFA_DEBUG__ 1090 1251 else … … 1092 1253 #endif // __CFA_DEBUG__ 1093 1254 1094 HeapManager.Storage.Header * header; 1095 HeapManager.FreeHeader * freeElem; 1096 size_t bsize, oalign = 0; 1097 headers( "realloc", oaddr, header, freeElem, bsize, oalign ); 1098 size_t odsize = dataStorage( bsize, oaddr, header ); // data storage available in bucket 1099 1100 if ( oalign != 0 && (uintptr_t)oaddr % nalign == 0 ) { // has alignment and just happens to work out 1101 headerAddr( oaddr )->kind.fake.alignment = nalign | 1; // update alignment (could be the same) 1102 return realloc( oaddr, size ); 1255 // If size is equal to 0, either NULL or a pointer suitable to be passed to free() is returned. 1256 if ( unlikely( size == 0 ) ) { free( oaddr ); return 0p; } // special cases 1257 if ( unlikely( oaddr == 0p ) ) { 1258 #ifdef __STATISTICS__ 1259 __atomic_add_fetch( &resize_storage, size, __ATOMIC_SEQ_CST ); 1260 #endif // __STATISTICS__ 1261 return memalignNoStats( nalign, size ); 1262 } // if 1263 1264 // Attempt to reuse existing alignment. 1265 HeapManager.Storage.Header * header = headerAddr( oaddr ); 1266 bool isFakeHeader = header->kind.fake.alignment & 1; // old fake header ? 1267 size_t oalign; 1268 if ( isFakeHeader ) { 1269 oalign = header->kind.fake.alignment & -2; // old alignment 1270 if ( (uintptr_t)oaddr % nalign == 0 // lucky match ? 1271 && ( oalign <= nalign // going down 1272 || (oalign >= nalign && oalign <= 256) ) // little alignment storage wasted ? 1273 ) { 1274 headerAddr( oaddr )->kind.fake.alignment = nalign | 1; // update alignment (could be the same) 1275 HeapManager.FreeHeader * freeElem; 1276 size_t bsize, oalign; 1277 headers( "resize", oaddr, header, freeElem, bsize, oalign ); 1278 size_t odsize = dataStorage( bsize, oaddr, header ); // data storage available in bucket 1279 1280 if ( size <= odsize && odsize <= size * 2 ) { // allow 50% wasted data storage 1281 headerAddr( oaddr )->kind.fake.alignment = nalign | 1; // update alignment (could be the same) 1282 1283 header->kind.real.blockSize &= -2; // turn off 0 fill 1284 header->kind.real.size = size; // reset allocation size 1285 return oaddr; 1286 } // if 1287 } // if 1288 } else if ( ! isFakeHeader // old real header (aligned on libAlign) ? 1289 && nalign == libAlign() ) { // new alignment also on libAlign => no fake header needed 1290 return resize( oaddr, size ); // duplicate special case checks 1103 1291 } // if 1104 1292 1105 1293 #ifdef __STATISTICS__ 1294 __atomic_add_fetch( &resize_storage, size, __ATOMIC_SEQ_CST ); 1295 #endif // __STATISTICS__ 1296 1297 // change size, DO NOT preserve STICKY PROPERTIES. 1298 free( oaddr ); 1299 return memalignNoStats( nalign, size ); // create new aligned area 1300 } // resize 1301 1302 1303 void * realloc( void * oaddr, size_t nalign, size_t size ) { 1304 if ( unlikely( nalign < libAlign() ) ) nalign = libAlign(); // reset alignment to minimum 1305 #ifdef __CFA_DEBUG__ 1306 else 1307 checkAlign( nalign ); // check alignment 1308 #endif // __CFA_DEBUG__ 1309 1310 // If size is equal to 0, either NULL or a pointer suitable to be passed to free() is returned. 1311 if ( unlikely( size == 0 ) ) { free( oaddr ); return 0p; } // special cases 1312 if ( unlikely( oaddr == 0p ) ) { 1313 #ifdef __STATISTICS__ 1314 __atomic_add_fetch( &realloc_calls, 1, __ATOMIC_SEQ_CST ); 1315 __atomic_add_fetch( &realloc_storage, size, __ATOMIC_SEQ_CST ); 1316 #endif // __STATISTICS__ 1317 return memalignNoStats( nalign, size ); 1318 } // if 1319 1320 // Attempt to reuse existing alignment. 1321 HeapManager.Storage.Header * header = headerAddr( oaddr ); 1322 bool isFakeHeader = header->kind.fake.alignment & 1; // old fake header ? 1323 size_t oalign; 1324 if ( isFakeHeader ) { 1325 oalign = header->kind.fake.alignment & -2; // old alignment 1326 if ( (uintptr_t)oaddr % nalign == 0 // lucky match ? 1327 && ( oalign <= nalign // going down 1328 || (oalign >= nalign && oalign <= 256) ) // little alignment storage wasted ? 1329 ) { 1330 headerAddr( oaddr )->kind.fake.alignment = nalign | 1; // update alignment (could be the same) 1331 return realloc( oaddr, size ); // duplicate alignment and special case checks 1332 } // if 1333 } else if ( ! isFakeHeader // old real header (aligned on libAlign) ? 1334 && nalign == libAlign() ) // new alignment also on libAlign => no fake header needed 1335 return realloc( oaddr, size ); // duplicate alignment and special case checks 1336 1337 #ifdef __STATISTICS__ 1338 __atomic_add_fetch( &realloc_calls, 1, __ATOMIC_SEQ_CST ); 1106 1339 __atomic_add_fetch( &realloc_storage, size, __ATOMIC_SEQ_CST ); 1107 1340 #endif // __STATISTICS__ 1108 1341 1342 HeapManager.FreeHeader * freeElem; 1343 size_t bsize; 1344 headers( "realloc", oaddr, header, freeElem, bsize, oalign ); 1345 1109 1346 // change size and copy old content to new storage 1110 1347 1111 void * naddr; 1112 if ( unlikely( header->kind.real.blockSize & 2 ) ) { // previous request zero fill 1113 naddr = cmemalignNoStats( nalign, 1, size ); // create new aligned area 1114 } else { 1115 naddr = memalignNoStats( nalign, size ); // create new aligned area 1116 } // if 1348 size_t osize = header->kind.real.size; // old allocation size 1349 bool ozfill = (header->kind.real.blockSize & 2); // old allocation zero filled 1350 1351 void * naddr = memalignNoStats( nalign, size ); // create new aligned area 1117 1352 1118 1353 headers( "realloc", naddr, header, freeElem, bsize, oalign ); 1119 size_t ndsize = dataStorage( bsize, naddr, header ); // data storage avilable in bucket 1120 // To preserve prior fill, the entire bucket must be copied versus the size. 1121 memcpy( naddr, oaddr, MIN( odsize, ndsize ) ); // copy bytes 1354 memcpy( naddr, oaddr, min( osize, size ) ); // copy bytes 1122 1355 free( oaddr ); 1356 1357 if ( unlikely( ozfill ) ) { // previous request zero fill ? 1358 header->kind.real.blockSize |= 2; // mark new request as zero filled 1359 if ( size > osize ) { // previous request larger ? 1360 memset( (char *)naddr + osize, '\0', size - osize ); // initialize added storage 1361 } // if 1362 } // if 1123 1363 return naddr; 1124 1364 } // realloc
Note:
See TracChangeset
for help on using the changeset viewer.