1 /* 2 Default header file for malloc-2.8.x, written by Doug Lea 3 and released to the public domain, as explained at 4 http://creativecommons.org/publicdomain/zero/1.0/ 5 6 This header is for ANSI C/C++ only. You can set any of 7 the following #defines before including: 8 9 * If USE_DL_PREFIX is defined, it is assumed that malloc.c 10 was also compiled with this option, so all routines 11 have names starting with "dl". 12 13 * If HAVE_USR_INCLUDE_MALLOC_H is defined, it is assumed that this 14 file will be #included AFTER <malloc.h>. This is needed only if 15 your system defines a struct mallinfo that is incompatible with the 16 standard one declared here. Otherwise, you can include this file 17 INSTEAD of your system system <malloc.h>. At least on ANSI, all 18 declarations should be compatible with system versions 19 20 * If MSPACES is defined, declarations for mspace versions are included. 21 */ 22 23 #ifndef MALLOC_280_H 24 #define MALLOC_280_H 25 26 #ifdef __cplusplus 27 extern "C" { 28 #endif 29 30 #include <stddef.h> /* for size_t */ 31 32 #ifndef ONLY_MSPACES 33 #define ONLY_MSPACES 0 /* define to a value */ 34 #elif ONLY_MSPACES != 0 35 #define ONLY_MSPACES 1 36 #endif /* ONLY_MSPACES */ 37 #ifndef NO_MALLINFO 38 #define NO_MALLINFO 0 39 #endif /* NO_MALLINFO */ 40 41 #ifndef MSPACES 42 #if ONLY_MSPACES 43 #define MSPACES 1 44 #else /* ONLY_MSPACES */ 45 #define MSPACES 0 46 #endif /* ONLY_MSPACES */ 47 #endif /* MSPACES */ 48 49 #if !ONLY_MSPACES 50 51 #ifndef USE_DL_PREFIX 52 #define dlcalloc calloc 53 #define dlfree free 54 #define dlmalloc malloc 55 #define dlmemalign memalign 56 #define dlposix_memalign posix_memalign 57 #define dlrealloc realloc 58 #define dlvalloc valloc 59 #define dlpvalloc pvalloc 60 #define dlmallinfo mallinfo 61 #define dlmallopt mallopt 62 #define dlmalloc_trim malloc_trim 63 #define dlmalloc_stats malloc_stats 64 #define dlmalloc_usable_size malloc_usable_size 65 #define dlmalloc_footprint malloc_footprint 66 #define dlmalloc_max_footprint malloc_max_footprint 67 #define dlmalloc_footprint_limit malloc_footprint_limit 68 #define dlmalloc_set_footprint_limit malloc_set_footprint_limit 69 #define dlmalloc_inspect_all malloc_inspect_all 70 #define dlindependent_calloc independent_calloc 71 #define dlindependent_comalloc independent_comalloc 72 #define dlbulk_free bulk_free 73 #endif /* USE_DL_PREFIX */ 74 75 #if !NO_MALLINFO 76 #ifndef HAVE_USR_INCLUDE_MALLOC_H 77 #ifndef _MALLOC_H 78 #ifndef MALLINFO_FIELD_TYPE 79 #define MALLINFO_FIELD_TYPE size_t 80 #endif /* MALLINFO_FIELD_TYPE */ 81 #ifndef STRUCT_MALLINFO_DECLARED 82 #define STRUCT_MALLINFO_DECLARED 1 83 struct mallinfo { 84 MALLINFO_FIELD_TYPE arena; /* non-mmapped space allocated from system */ 85 MALLINFO_FIELD_TYPE ordblks; /* number of free chunks */ 86 MALLINFO_FIELD_TYPE smblks; /* always 0 */ 87 MALLINFO_FIELD_TYPE hblks; /* always 0 */ 88 MALLINFO_FIELD_TYPE hblkhd; /* space in mmapped regions */ 89 MALLINFO_FIELD_TYPE usmblks; /* maximum total allocated space */ 90 MALLINFO_FIELD_TYPE fsmblks; /* always 0 */ 91 MALLINFO_FIELD_TYPE uordblks; /* total allocated space */ 92 MALLINFO_FIELD_TYPE fordblks; /* total free space */ 93 MALLINFO_FIELD_TYPE keepcost; /* releasable (via malloc_trim) space */ 94 }; 95 #endif /* STRUCT_MALLINFO_DECLARED */ 96 #endif /* _MALLOC_H */ 97 #endif /* HAVE_USR_INCLUDE_MALLOC_H */ 98 #endif /* !NO_MALLINFO */ 99 100 /* 101 malloc(size_t n) 102 Returns a pointer to a newly allocated chunk of at least n bytes, or 103 null if no space is available, in which case errno is set to ENOMEM 104 on ANSI C systems. 105 106 If n is zero, malloc returns a minimum-sized chunk. (The minimum 107 size is 16 bytes on most 32bit systems, and 32 bytes on 64bit 108 systems.) Note that size_t is an unsigned type, so calls with 109 arguments that would be negative if signed are interpreted as 110 requests for huge amounts of space, which will often fail. The 111 maximum supported value of n differs across systems, but is in all 112 cases less than the maximum representable value of a size_t. 113 */ 114 void* dlmalloc(size_t); 115 116 /* 117 free(void* p) 118 Releases the chunk of memory pointed to by p, that had been previously 119 allocated using malloc or a related routine such as realloc. 120 It has no effect if p is null. If p was not malloced or already 121 freed, free(p) will by default cuase the current program to abort. 122 */ 123 void dlfree(void*); 124 125 /* 126 calloc(size_t n_elements, size_t element_size); 127 Returns a pointer to n_elements * element_size bytes, with all locations 128 set to zero. 129 */ 130 void* dlcalloc(size_t, size_t); 131 132 /* 133 realloc(void* p, size_t n) 134 Returns a pointer to a chunk of size n that contains the same data 135 as does chunk p up to the minimum of (n, p's size) bytes, or null 136 if no space is available. 137 138 The returned pointer may or may not be the same as p. The algorithm 139 prefers extending p in most cases when possible, otherwise it 140 employs the equivalent of a malloc-copy-free sequence. 141 142 If p is null, realloc is equivalent to malloc. 143 144 If space is not available, realloc returns null, errno is set (if on 145 ANSI) and p is NOT freed. 146 147 if n is for fewer bytes than already held by p, the newly unused 148 space is lopped off and freed if possible. realloc with a size 149 argument of zero (re)allocates a minimum-sized chunk. 150 151 The old unix realloc convention of allowing the last-free'd chunk 152 to be used as an argument to realloc is not supported. 153 */ 154 void* dlrealloc(void*, size_t); 155 156 /* 157 realloc_in_place(void* p, size_t n) 158 Resizes the space allocated for p to size n, only if this can be 159 done without moving p (i.e., only if there is adjacent space 160 available if n is greater than p's current allocated size, or n is 161 less than or equal to p's size). This may be used instead of plain 162 realloc if an alternative allocation strategy is needed upon failure 163 to expand space; for example, reallocation of a buffer that must be 164 memory-aligned or cleared. You can use realloc_in_place to trigger 165 these alternatives only when needed. 166 167 Returns p if successful; otherwise null. 168 */ 169 void* dlrealloc_in_place(void*, size_t); 170 171 /* 172 memalign(size_t alignment, size_t n); 173 Returns a pointer to a newly allocated chunk of n bytes, aligned 174 in accord with the alignment argument. 175 176 The alignment argument should be a power of two. If the argument is 177 not a power of two, the nearest greater power is used. 178 8-byte alignment is guaranteed by normal malloc calls, so don't 179 bother calling memalign with an argument of 8 or less. 180 181 Overreliance on memalign is a sure way to fragment space. 182 */ 183 void* dlmemalign(size_t, size_t); 184 185 /* 186 int posix_memalign(void** pp, size_t alignment, size_t n); 187 Allocates a chunk of n bytes, aligned in accord with the alignment 188 argument. Differs from memalign only in that it (1) assigns the 189 allocated memory to *pp rather than returning it, (2) fails and 190 returns EINVAL if the alignment is not a power of two (3) fails and 191 returns ENOMEM if memory cannot be allocated. 192 */ 193 int dlposix_memalign(void**, size_t, size_t); 194 195 /* 196 valloc(size_t n); 197 Equivalent to memalign(pagesize, n), where pagesize is the page 198 size of the system. If the pagesize is unknown, 4096 is used. 199 */ 200 void* dlvalloc(size_t); 201 202 /* 203 mallopt(int parameter_number, int parameter_value) 204 Sets tunable parameters The format is to provide a 205 (parameter-number, parameter-value) pair. mallopt then sets the 206 corresponding parameter to the argument value if it can (i.e., so 207 long as the value is meaningful), and returns 1 if successful else 208 0. SVID/XPG/ANSI defines four standard param numbers for mallopt, 209 normally defined in malloc.h. None of these are use in this malloc, 210 so setting them has no effect. But this malloc also supports other 211 options in mallopt: 212 213 Symbol param # default allowed param values 214 M_TRIM_THRESHOLD -1 2*1024*1024 any (-1U disables trimming) 215 M_GRANULARITY -2 page size any power of 2 >= page size 216 M_MMAP_THRESHOLD -3 256*1024 any (or 0 if no MMAP support) 217 */ 218 int dlmallopt(int, int); 219 220 #define M_TRIM_THRESHOLD (-1) 221 #define M_GRANULARITY (-2) 222 #define M_MMAP_THRESHOLD (-3) 223 224 225 /* 226 malloc_footprint(); 227 Returns the number of bytes obtained from the system. The total 228 number of bytes allocated by malloc, realloc etc., is less than this 229 value. Unlike mallinfo, this function returns only a precomputed 230 result, so can be called frequently to monitor memory consumption. 231 Even if locks are otherwise defined, this function does not use them, 232 so results might not be up to date. 233 */ 234 size_t dlmalloc_footprint(void); 235 236 /* 237 malloc_max_footprint(); 238 Returns the maximum number of bytes obtained from the system. This 239 value will be greater than current footprint if deallocated space 240 has been reclaimed by the system. The peak number of bytes allocated 241 by malloc, realloc etc., is less than this value. Unlike mallinfo, 242 this function returns only a precomputed result, so can be called 243 frequently to monitor memory consumption. Even if locks are 244 otherwise defined, this function does not use them, so results might 245 not be up to date. 246 */ 247 size_t dlmalloc_max_footprint(void); 248 249 /* 250 malloc_footprint_limit(); 251 Returns the number of bytes that the heap is allowed to obtain from 252 the system, returning the last value returned by 253 malloc_set_footprint_limit, or the maximum size_t value if 254 never set. The returned value reflects a permission. There is no 255 guarantee that this number of bytes can actually be obtained from 256 the system. 257 */ 258 size_t dlmalloc_footprint_limit(void); 259 260 /* 261 malloc_set_footprint_limit(); 262 Sets the maximum number of bytes to obtain from the system, causing 263 failure returns from malloc and related functions upon attempts to 264 exceed this value. The argument value may be subject to page 265 rounding to an enforceable limit; this actual value is returned. 266 Using an argument of the maximum possible size_t effectively 267 disables checks. If the argument is less than or equal to the 268 current malloc_footprint, then all future allocations that require 269 additional system memory will fail. However, invocation cannot 270 retroactively deallocate existing used memory. 271 */ 272 size_t dlmalloc_set_footprint_limit(size_t bytes); 273 274 /* 275 malloc_inspect_all(void(*handler)(void *start, 276 void *end, 277 size_t used_bytes, 278 void* callback_arg), 279 void* arg); 280 Traverses the heap and calls the given handler for each managed 281 region, skipping all bytes that are (or may be) used for bookkeeping 282 purposes. Traversal does not include include chunks that have been 283 directly memory mapped. Each reported region begins at the start 284 address, and continues up to but not including the end address. The 285 first used_bytes of the region contain allocated data. If 286 used_bytes is zero, the region is unallocated. The handler is 287 invoked with the given callback argument. If locks are defined, they 288 are held during the entire traversal. It is a bad idea to invoke 289 other malloc functions from within the handler. 290 291 For example, to count the number of in-use chunks with size greater 292 than 1000, you could write: 293 static int count = 0; 294 void count_chunks(void* start, void* end, size_t used, void* arg) { 295 if (used >= 1000) ++count; 296 } 297 then: 298 malloc_inspect_all(count_chunks, NULL); 299 300 malloc_inspect_all is compiled only if MALLOC_INSPECT_ALL is defined. 301 */ 302 void dlmalloc_inspect_all(void(*handler)(void*, void *, size_t, void*), 303 void* arg); 304 305 #if !NO_MALLINFO 306 /* 307 mallinfo() 308 Returns (by copy) a struct containing various summary statistics: 309 310 arena: current total non-mmapped bytes allocated from system 311 ordblks: the number of free chunks 312 smblks: always zero. 313 hblks: current number of mmapped regions 314 hblkhd: total bytes held in mmapped regions 315 usmblks: the maximum total allocated space. This will be greater 316 than current total if trimming has occurred. 317 fsmblks: always zero 318 uordblks: current total allocated space (normal or mmapped) 319 fordblks: total free space 320 keepcost: the maximum number of bytes that could ideally be released 321 back to system via malloc_trim. ("ideally" means that 322 it ignores page restrictions etc.) 323 324 Because these fields are ints, but internal bookkeeping may 325 be kept as longs, the reported values may wrap around zero and 326 thus be inaccurate. 327 */ 328 329 struct mallinfo dlmallinfo(void); 330 #endif /* NO_MALLINFO */ 331 332 /* 333 independent_calloc(size_t n_elements, size_t element_size, void* chunks[]); 334 335 independent_calloc is similar to calloc, but instead of returning a 336 single cleared space, it returns an array of pointers to n_elements 337 independent elements that can hold contents of size elem_size, each 338 of which starts out cleared, and can be independently freed, 339 realloc'ed etc. The elements are guaranteed to be adjacently 340 allocated (this is not guaranteed to occur with multiple callocs or 341 mallocs), which may also improve cache locality in some 342 applications. 343 344 The "chunks" argument is optional (i.e., may be null, which is 345 probably the most typical usage). If it is null, the returned array 346 is itself dynamically allocated and should also be freed when it is 347 no longer needed. Otherwise, the chunks array must be of at least 348 n_elements in length. It is filled in with the pointers to the 349 chunks. 350 351 In either case, independent_calloc returns this pointer array, or 352 null if the allocation failed. If n_elements is zero and "chunks" 353 is null, it returns a chunk representing an array with zero elements 354 (which should be freed if not wanted). 355 356 Each element must be freed when it is no longer needed. This can be 357 done all at once using bulk_free. 358 359 independent_calloc simplifies and speeds up implementations of many 360 kinds of pools. It may also be useful when constructing large data 361 structures that initially have a fixed number of fixed-sized nodes, 362 but the number is not known at compile time, and some of the nodes 363 may later need to be freed. For example: 364 365 struct Node { int item; struct Node* next; }; 366 367 struct Node* build_list() { 368 struct Node** pool; 369 int n = read_number_of_nodes_needed(); 370 if (n <= 0) return 0; 371 pool = (struct Node**)(independent_calloc(n, sizeof(struct Node), 0); 372 if (pool == 0) die(); 373 // organize into a linked list... 374 struct Node* first = pool[0]; 375 for (i = 0; i < n-1; ++i) 376 pool[i]->next = pool[i+1]; 377 free(pool); // Can now free the array (or not, if it is needed later) 378 return first; 379 } 380 */ 381 void** dlindependent_calloc(size_t, size_t, void**); 382 383 /* 384 independent_comalloc(size_t n_elements, size_t sizes[], void* chunks[]); 385 386 independent_comalloc allocates, all at once, a set of n_elements 387 chunks with sizes indicated in the "sizes" array. It returns 388 an array of pointers to these elements, each of which can be 389 independently freed, realloc'ed etc. The elements are guaranteed to 390 be adjacently allocated (this is not guaranteed to occur with 391 multiple callocs or mallocs), which may also improve cache locality 392 in some applications. 393 394 The "chunks" argument is optional (i.e., may be null). If it is null 395 the returned array is itself dynamically allocated and should also 396 be freed when it is no longer needed. Otherwise, the chunks array 397 must be of at least n_elements in length. It is filled in with the 398 pointers to the chunks. 399 400 In either case, independent_comalloc returns this pointer array, or 401 null if the allocation failed. If n_elements is zero and chunks is 402 null, it returns a chunk representing an array with zero elements 403 (which should be freed if not wanted). 404 405 Each element must be freed when it is no longer needed. This can be 406 done all at once using bulk_free. 407 408 independent_comallac differs from independent_calloc in that each 409 element may have a different size, and also that it does not 410 automatically clear elements. 411 412 independent_comalloc can be used to speed up allocation in cases 413 where several structs or objects must always be allocated at the 414 same time. For example: 415 416 struct Head { ... } 417 struct Foot { ... } 418 419 void send_message(char* msg) { 420 int msglen = strlen(msg); 421 size_t sizes[3] = { sizeof(struct Head), msglen, sizeof(struct Foot) }; 422 void* chunks[3]; 423 if (independent_comalloc(3, sizes, chunks) == 0) 424 die(); 425 struct Head* head = (struct Head*)(chunks[0]); 426 char* body = (char*)(chunks[1]); 427 struct Foot* foot = (struct Foot*)(chunks[2]); 428 // ... 429 } 430 431 In general though, independent_comalloc is worth using only for 432 larger values of n_elements. For small values, you probably won't 433 detect enough difference from series of malloc calls to bother. 434 435 Overuse of independent_comalloc can increase overall memory usage, 436 since it cannot reuse existing noncontiguous small chunks that 437 might be available for some of the elements. 438 */ 439 void** dlindependent_comalloc(size_t, size_t*, void**); 440 441 /* 442 bulk_free(void* array[], size_t n_elements) 443 Frees and clears (sets to null) each non-null pointer in the given 444 array. This is likely to be faster than freeing them one-by-one. 445 If footers are used, pointers that have been allocated in different 446 mspaces are not freed or cleared, and the count of all such pointers 447 is returned. For large arrays of pointers with poor locality, it 448 may be worthwhile to sort this array before calling bulk_free. 449 */ 450 size_t dlbulk_free(void**, size_t n_elements); 451 452 /* 453 pvalloc(size_t n); 454 Equivalent to valloc(minimum-page-that-holds(n)), that is, 455 round up n to nearest pagesize. 456 */ 457 void* dlpvalloc(size_t); 458 459 /* 460 malloc_trim(size_t pad); 461 462 If possible, gives memory back to the system (via negative arguments 463 to sbrk) if there is unused memory at the `high' end of the malloc 464 pool or in unused MMAP segments. You can call this after freeing 465 large blocks of memory to potentially reduce the system-level memory 466 requirements of a program. However, it cannot guarantee to reduce 467 memory. Under some allocation patterns, some large free blocks of 468 memory will be locked between two used chunks, so they cannot be 469 given back to the system. 470 471 The `pad' argument to malloc_trim represents the amount of free 472 trailing space to leave untrimmed. If this argument is zero, only 473 the minimum amount of memory to maintain internal data structures 474 will be left. Non-zero arguments can be supplied to maintain enough 475 trailing space to service future expected allocations without having 476 to re-obtain memory from the system. 477 478 Malloc_trim returns 1 if it actually released any memory, else 0. 479 */ 480 int dlmalloc_trim(size_t); 481 482 /* 483 malloc_stats(); 484 Prints on stderr the amount of space obtained from the system (both 485 via sbrk and mmap), the maximum amount (which may be more than 486 current if malloc_trim and/or munmap got called), and the current 487 number of bytes allocated via malloc (or realloc, etc) but not yet 488 freed. Note that this is the number of bytes allocated, not the 489 number requested. It will be larger than the number requested 490 because of alignment and bookkeeping overhead. Because it includes 491 alignment wastage as being in use, this figure may be greater than 492 zero even when no user-level chunks are allocated. 493 494 The reported current and maximum system memory can be inaccurate if 495 a program makes other calls to system memory allocation functions 496 (normally sbrk) outside of malloc. 497 498 malloc_stats prints only the most commonly interesting statistics. 499 More information can be obtained by calling mallinfo. 500 501 malloc_stats is not compiled if NO_MALLOC_STATS is defined. 502 */ 503 void dlmalloc_stats(void); 504 505 #endif /* !ONLY_MSPACES */ 506 507 /* 508 malloc_usable_size(void* p); 509 510 Returns the number of bytes you can actually use in 511 an allocated chunk, which may be more than you requested (although 512 often not) due to alignment and minimum size constraints. 513 You can use this many bytes without worrying about 514 overwriting other allocated objects. This is not a particularly great 515 programming practice. malloc_usable_size can be more useful in 516 debugging and assertions, for example: 517 518 p = malloc(n); 519 assert(malloc_usable_size(p) >= 256); 520 */ 521 size_t dlmalloc_usable_size(const void*); 522 523 #if MSPACES 524 525 /* 526 mspace is an opaque type representing an independent 527 region of space that supports mspace_malloc, etc. 528 */ 529 typedef void* mspace; 530 531 /* 532 create_mspace creates and returns a new independent space with the 533 given initial capacity, or, if 0, the default granularity size. It 534 returns null if there is no system memory available to create the 535 space. If argument locked is non-zero, the space uses a separate 536 lock to control access. The capacity of the space will grow 537 dynamically as needed to service mspace_malloc requests. You can 538 control the sizes of incremental increases of this space by 539 compiling with a different DEFAULT_GRANULARITY or dynamically 540 setting with mallopt(M_GRANULARITY, value). 541 */ 542 mspace create_mspace(size_t capacity, int locked); 543 544 /* 545 destroy_mspace destroys the given space, and attempts to return all 546 of its memory back to the system, returning the total number of 547 bytes freed. After destruction, the results of access to all memory 548 used by the space become undefined. 549 */ 550 size_t destroy_mspace(mspace msp); 551 552 /* 553 create_mspace_with_base uses the memory supplied as the initial base 554 of a new mspace. Part (less than 128*sizeof(size_t) bytes) of this 555 space is used for bookkeeping, so the capacity must be at least this 556 large. (Otherwise 0 is returned.) When this initial space is 557 exhausted, additional memory will be obtained from the system. 558 Destroying this space will deallocate all additionally allocated 559 space (if possible) but not the initial base. 560 */ 561 mspace create_mspace_with_base(void* base, size_t capacity, int locked); 562 563 /* 564 mspace_track_large_chunks controls whether requests for large chunks 565 are allocated in their own untracked mmapped regions, separate from 566 others in this mspace. By default large chunks are not tracked, 567 which reduces fragmentation. However, such chunks are not 568 necessarily released to the system upon destroy_mspace. Enabling 569 tracking by setting to true may increase fragmentation, but avoids 570 leakage when relying on destroy_mspace to release all memory 571 allocated using this space. The function returns the previous 572 setting. 573 */ 574 int mspace_track_large_chunks(mspace msp, int enable); 575 576 #if !NO_MALLINFO 577 /* 578 mspace_mallinfo behaves as mallinfo, but reports properties of 579 the given space. 580 */ 581 struct mallinfo mspace_mallinfo(mspace msp); 582 #endif /* NO_MALLINFO */ 583 584 /* 585 An alias for mallopt. 586 */ 587 int mspace_mallopt(int, int); 588 589 /* 590 The following operate identically to their malloc counterparts 591 but operate only for the given mspace argument 592 */ 593 void* mspace_malloc(mspace msp, size_t bytes); 594 void mspace_free(mspace msp, void* mem); 595 void* mspace_calloc(mspace msp, size_t n_elements, size_t elem_size); 596 void* mspace_realloc(mspace msp, void* mem, size_t newsize); 597 void* mspace_realloc_in_place(mspace msp, void* mem, size_t newsize); 598 void* mspace_memalign(mspace msp, size_t alignment, size_t bytes); 599 void** mspace_independent_calloc(mspace msp, size_t n_elements, 600 size_t elem_size, void* chunks[]); 601 void** mspace_independent_comalloc(mspace msp, size_t n_elements, 602 size_t sizes[], void* chunks[]); 603 size_t mspace_bulk_free(mspace msp, void**, size_t n_elements); 604 size_t mspace_usable_size(const void* mem); 605 void mspace_malloc_stats(mspace msp); 606 int mspace_trim(mspace msp, size_t pad); 607 size_t mspace_footprint(mspace msp); 608 size_t mspace_max_footprint(mspace msp); 609 size_t mspace_footprint_limit(mspace msp); 610 size_t mspace_set_footprint_limit(mspace msp, size_t bytes); 611 void mspace_inspect_all(mspace msp, 612 void(*handler)(void *, void *, size_t, void*), 613 void* arg); 614 #endif /* MSPACES */ 615 616 #ifdef __cplusplus 617 }; /* end of extern "C" */ 618 #endif 619 620 #endif /* MALLOC_280_H */ 621