• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1  /*
2    Default header file for malloc-2.8.x, written by Doug Lea
3    and released to the public domain, as explained at
4    http://creativecommons.org/publicdomain/zero/1.0/
5  
6    This header is for ANSI C/C++ only.  You can set any of
7    the following #defines before including:
8  
9    * If USE_DL_PREFIX is defined, it is assumed that malloc.c
10      was also compiled with this option, so all routines
11      have names starting with "dl".
12  
13    * If HAVE_USR_INCLUDE_MALLOC_H is defined, it is assumed that this
14      file will be #included AFTER <malloc.h>. This is needed only if
15      your system defines a struct mallinfo that is incompatible with the
16      standard one declared here.  Otherwise, you can include this file
17      INSTEAD of your system system <malloc.h>.  At least on ANSI, all
18      declarations should be compatible with system versions
19  
20    * If MSPACES is defined, declarations for mspace versions are included.
21  */
22  
23  #ifndef MALLOC_280_H
24  #define MALLOC_280_H
25  
26  #ifdef __cplusplus
27  extern "C" {
28  #endif
29  
30  #include <stddef.h>   /* for size_t */
31  
32  #ifndef ONLY_MSPACES
33  #define ONLY_MSPACES 0     /* define to a value */
34  #elif ONLY_MSPACES != 0
35  #define ONLY_MSPACES 1
36  #endif  /* ONLY_MSPACES */
37  #ifndef NO_MALLINFO
38  #define NO_MALLINFO 0
39  #endif  /* NO_MALLINFO */
40  
41  #ifndef MSPACES
42  #if ONLY_MSPACES
43  #define MSPACES 1
44  #else   /* ONLY_MSPACES */
45  #define MSPACES 0
46  #endif  /* ONLY_MSPACES */
47  #endif  /* MSPACES */
48  
49  #if !ONLY_MSPACES
50  
51  #ifndef USE_DL_PREFIX
52  #define dlcalloc               calloc
53  #define dlfree                 free
54  #define dlmalloc               malloc
55  #define dlmemalign             memalign
56  #define dlposix_memalign       posix_memalign
57  #define dlrealloc              realloc
58  #define dlvalloc               valloc
59  #define dlpvalloc              pvalloc
60  #define dlmallinfo             mallinfo
61  #define dlmallopt              mallopt
62  #define dlmalloc_trim          malloc_trim
63  #define dlmalloc_stats         malloc_stats
64  #define dlmalloc_usable_size   malloc_usable_size
65  #define dlmalloc_footprint     malloc_footprint
66  #define dlmalloc_max_footprint malloc_max_footprint
67  #define dlmalloc_footprint_limit malloc_footprint_limit
68  #define dlmalloc_set_footprint_limit malloc_set_footprint_limit
69  #define dlmalloc_inspect_all   malloc_inspect_all
70  #define dlindependent_calloc   independent_calloc
71  #define dlindependent_comalloc independent_comalloc
72  #define dlbulk_free            bulk_free
73  #endif /* USE_DL_PREFIX */
74  
75  #if !NO_MALLINFO
76  #ifndef HAVE_USR_INCLUDE_MALLOC_H
77  #ifndef _MALLOC_H
78  #ifndef MALLINFO_FIELD_TYPE
79  #define MALLINFO_FIELD_TYPE size_t
80  #endif /* MALLINFO_FIELD_TYPE */
81  #ifndef STRUCT_MALLINFO_DECLARED
82  #define STRUCT_MALLINFO_DECLARED 1
83  struct mallinfo {
84    MALLINFO_FIELD_TYPE arena;    /* non-mmapped space allocated from system */
85    MALLINFO_FIELD_TYPE ordblks;  /* number of free chunks */
86    MALLINFO_FIELD_TYPE smblks;   /* always 0 */
87    MALLINFO_FIELD_TYPE hblks;    /* always 0 */
88    MALLINFO_FIELD_TYPE hblkhd;   /* space in mmapped regions */
89    MALLINFO_FIELD_TYPE usmblks;  /* maximum total allocated space */
90    MALLINFO_FIELD_TYPE fsmblks;  /* always 0 */
91    MALLINFO_FIELD_TYPE uordblks; /* total allocated space */
92    MALLINFO_FIELD_TYPE fordblks; /* total free space */
93    MALLINFO_FIELD_TYPE keepcost; /* releasable (via malloc_trim) space */
94  };
95  #endif /* STRUCT_MALLINFO_DECLARED */
96  #endif  /* _MALLOC_H */
97  #endif  /* HAVE_USR_INCLUDE_MALLOC_H */
98  #endif  /* !NO_MALLINFO */
99  
100  /*
101    malloc(size_t n)
102    Returns a pointer to a newly allocated chunk of at least n bytes, or
103    null if no space is available, in which case errno is set to ENOMEM
104    on ANSI C systems.
105  
106    If n is zero, malloc returns a minimum-sized chunk. (The minimum
107    size is 16 bytes on most 32bit systems, and 32 bytes on 64bit
108    systems.)  Note that size_t is an unsigned type, so calls with
109    arguments that would be negative if signed are interpreted as
110    requests for huge amounts of space, which will often fail. The
111    maximum supported value of n differs across systems, but is in all
112    cases less than the maximum representable value of a size_t.
113  */
114  void* dlmalloc(size_t);
115  
116  /*
117    free(void* p)
118    Releases the chunk of memory pointed to by p, that had been previously
119    allocated using malloc or a related routine such as realloc.
120    It has no effect if p is null. If p was not malloced or already
121    freed, free(p) will by default cuase the current program to abort.
122  */
123  void  dlfree(void*);
124  
125  /*
126    calloc(size_t n_elements, size_t element_size);
127    Returns a pointer to n_elements * element_size bytes, with all locations
128    set to zero.
129  */
130  void* dlcalloc(size_t, size_t);
131  
132  /*
133    realloc(void* p, size_t n)
134    Returns a pointer to a chunk of size n that contains the same data
135    as does chunk p up to the minimum of (n, p's size) bytes, or null
136    if no space is available.
137  
138    The returned pointer may or may not be the same as p. The algorithm
139    prefers extending p in most cases when possible, otherwise it
140    employs the equivalent of a malloc-copy-free sequence.
141  
142    If p is null, realloc is equivalent to malloc.
143  
144    If space is not available, realloc returns null, errno is set (if on
145    ANSI) and p is NOT freed.
146  
147    if n is for fewer bytes than already held by p, the newly unused
148    space is lopped off and freed if possible.  realloc with a size
149    argument of zero (re)allocates a minimum-sized chunk.
150  
151    The old unix realloc convention of allowing the last-free'd chunk
152    to be used as an argument to realloc is not supported.
153  */
154  void* dlrealloc(void*, size_t);
155  
156  /*
157    realloc_in_place(void* p, size_t n)
158    Resizes the space allocated for p to size n, only if this can be
159    done without moving p (i.e., only if there is adjacent space
160    available if n is greater than p's current allocated size, or n is
161    less than or equal to p's size). This may be used instead of plain
162    realloc if an alternative allocation strategy is needed upon failure
163    to expand space; for example, reallocation of a buffer that must be
164    memory-aligned or cleared. You can use realloc_in_place to trigger
165    these alternatives only when needed.
166  
167    Returns p if successful; otherwise null.
168  */
169  void* dlrealloc_in_place(void*, size_t);
170  
171  /*
172    memalign(size_t alignment, size_t n);
173    Returns a pointer to a newly allocated chunk of n bytes, aligned
174    in accord with the alignment argument.
175  
176    The alignment argument should be a power of two. If the argument is
177    not a power of two, the nearest greater power is used.
178    8-byte alignment is guaranteed by normal malloc calls, so don't
179    bother calling memalign with an argument of 8 or less.
180  
181    Overreliance on memalign is a sure way to fragment space.
182  */
183  void* dlmemalign(size_t, size_t);
184  
185  /*
186    int posix_memalign(void** pp, size_t alignment, size_t n);
187    Allocates a chunk of n bytes, aligned in accord with the alignment
188    argument. Differs from memalign only in that it (1) assigns the
189    allocated memory to *pp rather than returning it, (2) fails and
190    returns EINVAL if the alignment is not a power of two (3) fails and
191    returns ENOMEM if memory cannot be allocated.
192  */
193  int dlposix_memalign(void**, size_t, size_t);
194  
195  /*
196    valloc(size_t n);
197    Equivalent to memalign(pagesize, n), where pagesize is the page
198    size of the system. If the pagesize is unknown, 4096 is used.
199  */
200  void* dlvalloc(size_t);
201  
202  /*
203    mallopt(int parameter_number, int parameter_value)
204    Sets tunable parameters The format is to provide a
205    (parameter-number, parameter-value) pair.  mallopt then sets the
206    corresponding parameter to the argument value if it can (i.e., so
207    long as the value is meaningful), and returns 1 if successful else
208    0.  SVID/XPG/ANSI defines four standard param numbers for mallopt,
209    normally defined in malloc.h.  None of these are use in this malloc,
210    so setting them has no effect. But this malloc also supports other
211    options in mallopt:
212  
213    Symbol            param #  default    allowed param values
214    M_TRIM_THRESHOLD     -1   2*1024*1024   any   (-1U disables trimming)
215    M_GRANULARITY        -2     page size   any power of 2 >= page size
216    M_MMAP_THRESHOLD     -3      256*1024   any   (or 0 if no MMAP support)
217  */
218  int dlmallopt(int, int);
219  
220  #define M_TRIM_THRESHOLD     (-1)
221  #define M_GRANULARITY        (-2)
222  #define M_MMAP_THRESHOLD     (-3)
223  
224  
225  /*
226    malloc_footprint();
227    Returns the number of bytes obtained from the system.  The total
228    number of bytes allocated by malloc, realloc etc., is less than this
229    value. Unlike mallinfo, this function returns only a precomputed
230    result, so can be called frequently to monitor memory consumption.
231    Even if locks are otherwise defined, this function does not use them,
232    so results might not be up to date.
233  */
234  size_t dlmalloc_footprint(void);
235  
236  /*
237    malloc_max_footprint();
238    Returns the maximum number of bytes obtained from the system. This
239    value will be greater than current footprint if deallocated space
240    has been reclaimed by the system. The peak number of bytes allocated
241    by malloc, realloc etc., is less than this value. Unlike mallinfo,
242    this function returns only a precomputed result, so can be called
243    frequently to monitor memory consumption.  Even if locks are
244    otherwise defined, this function does not use them, so results might
245    not be up to date.
246  */
247  size_t dlmalloc_max_footprint(void);
248  
249  /*
250    malloc_footprint_limit();
251    Returns the number of bytes that the heap is allowed to obtain from
252    the system, returning the last value returned by
253    malloc_set_footprint_limit, or the maximum size_t value if
254    never set. The returned value reflects a permission. There is no
255    guarantee that this number of bytes can actually be obtained from
256    the system.
257  */
258  size_t dlmalloc_footprint_limit(void);
259  
260  /*
261    malloc_set_footprint_limit();
262    Sets the maximum number of bytes to obtain from the system, causing
263    failure returns from malloc and related functions upon attempts to
264    exceed this value. The argument value may be subject to page
265    rounding to an enforceable limit; this actual value is returned.
266    Using an argument of the maximum possible size_t effectively
267    disables checks. If the argument is less than or equal to the
268    current malloc_footprint, then all future allocations that require
269    additional system memory will fail. However, invocation cannot
270    retroactively deallocate existing used memory.
271  */
272  size_t dlmalloc_set_footprint_limit(size_t bytes);
273  
274  /*
275    malloc_inspect_all(void(*handler)(void *start,
276                                      void *end,
277                                      size_t used_bytes,
278                                      void* callback_arg),
279                        void* arg);
280    Traverses the heap and calls the given handler for each managed
281    region, skipping all bytes that are (or may be) used for bookkeeping
282    purposes.  Traversal does not include include chunks that have been
283    directly memory mapped. Each reported region begins at the start
284    address, and continues up to but not including the end address.  The
285    first used_bytes of the region contain allocated data. If
286    used_bytes is zero, the region is unallocated. The handler is
287    invoked with the given callback argument. If locks are defined, they
288    are held during the entire traversal. It is a bad idea to invoke
289    other malloc functions from within the handler.
290  
291    For example, to count the number of in-use chunks with size greater
292    than 1000, you could write:
293    static int count = 0;
294    void count_chunks(void* start, void* end, size_t used, void* arg) {
295      if (used >= 1000) ++count;
296    }
297    then:
298      malloc_inspect_all(count_chunks, NULL);
299  
300    malloc_inspect_all is compiled only if MALLOC_INSPECT_ALL is defined.
301  */
302  void dlmalloc_inspect_all(void(*handler)(void*, void *, size_t, void*),
303                             void* arg);
304  
305  #if !NO_MALLINFO
306  /*
307    mallinfo()
308    Returns (by copy) a struct containing various summary statistics:
309  
310    arena:     current total non-mmapped bytes allocated from system
311    ordblks:   the number of free chunks
312    smblks:    always zero.
313    hblks:     current number of mmapped regions
314    hblkhd:    total bytes held in mmapped regions
315    usmblks:   the maximum total allocated space. This will be greater
316                  than current total if trimming has occurred.
317    fsmblks:   always zero
318    uordblks:  current total allocated space (normal or mmapped)
319    fordblks:  total free space
320    keepcost:  the maximum number of bytes that could ideally be released
321                 back to system via malloc_trim. ("ideally" means that
322                 it ignores page restrictions etc.)
323  
324    Because these fields are ints, but internal bookkeeping may
325    be kept as longs, the reported values may wrap around zero and
326    thus be inaccurate.
327  */
328  
329  struct mallinfo dlmallinfo(void);
330  #endif  /* NO_MALLINFO */
331  
332  /*
333    independent_calloc(size_t n_elements, size_t element_size, void* chunks[]);
334  
335    independent_calloc is similar to calloc, but instead of returning a
336    single cleared space, it returns an array of pointers to n_elements
337    independent elements that can hold contents of size elem_size, each
338    of which starts out cleared, and can be independently freed,
339    realloc'ed etc. The elements are guaranteed to be adjacently
340    allocated (this is not guaranteed to occur with multiple callocs or
341    mallocs), which may also improve cache locality in some
342    applications.
343  
344    The "chunks" argument is optional (i.e., may be null, which is
345    probably the most typical usage). If it is null, the returned array
346    is itself dynamically allocated and should also be freed when it is
347    no longer needed. Otherwise, the chunks array must be of at least
348    n_elements in length. It is filled in with the pointers to the
349    chunks.
350  
351    In either case, independent_calloc returns this pointer array, or
352    null if the allocation failed.  If n_elements is zero and "chunks"
353    is null, it returns a chunk representing an array with zero elements
354    (which should be freed if not wanted).
355  
356    Each element must be freed when it is no longer needed. This can be
357    done all at once using bulk_free.
358  
359    independent_calloc simplifies and speeds up implementations of many
360    kinds of pools.  It may also be useful when constructing large data
361    structures that initially have a fixed number of fixed-sized nodes,
362    but the number is not known at compile time, and some of the nodes
363    may later need to be freed. For example:
364  
365    struct Node { int item; struct Node* next; };
366  
367    struct Node* build_list() {
368      struct Node** pool;
369      int n = read_number_of_nodes_needed();
370      if (n <= 0) return 0;
371      pool = (struct Node**)(independent_calloc(n, sizeof(struct Node), 0);
372      if (pool == 0) die();
373      // organize into a linked list...
374      struct Node* first = pool[0];
375      for (i = 0; i < n-1; ++i)
376        pool[i]->next = pool[i+1];
377      free(pool);     // Can now free the array (or not, if it is needed later)
378      return first;
379    }
380  */
381  void** dlindependent_calloc(size_t, size_t, void**);
382  
383  /*
384    independent_comalloc(size_t n_elements, size_t sizes[], void* chunks[]);
385  
386    independent_comalloc allocates, all at once, a set of n_elements
387    chunks with sizes indicated in the "sizes" array.    It returns
388    an array of pointers to these elements, each of which can be
389    independently freed, realloc'ed etc. The elements are guaranteed to
390    be adjacently allocated (this is not guaranteed to occur with
391    multiple callocs or mallocs), which may also improve cache locality
392    in some applications.
393  
394    The "chunks" argument is optional (i.e., may be null). If it is null
395    the returned array is itself dynamically allocated and should also
396    be freed when it is no longer needed. Otherwise, the chunks array
397    must be of at least n_elements in length. It is filled in with the
398    pointers to the chunks.
399  
400    In either case, independent_comalloc returns this pointer array, or
401    null if the allocation failed.  If n_elements is zero and chunks is
402    null, it returns a chunk representing an array with zero elements
403    (which should be freed if not wanted).
404  
405    Each element must be freed when it is no longer needed. This can be
406    done all at once using bulk_free.
407  
408    independent_comallac differs from independent_calloc in that each
409    element may have a different size, and also that it does not
410    automatically clear elements.
411  
412    independent_comalloc can be used to speed up allocation in cases
413    where several structs or objects must always be allocated at the
414    same time.  For example:
415  
416    struct Head { ... }
417    struct Foot { ... }
418  
419    void send_message(char* msg) {
420      int msglen = strlen(msg);
421      size_t sizes[3] = { sizeof(struct Head), msglen, sizeof(struct Foot) };
422      void* chunks[3];
423      if (independent_comalloc(3, sizes, chunks) == 0)
424        die();
425      struct Head* head = (struct Head*)(chunks[0]);
426      char*        body = (char*)(chunks[1]);
427      struct Foot* foot = (struct Foot*)(chunks[2]);
428      // ...
429    }
430  
431    In general though, independent_comalloc is worth using only for
432    larger values of n_elements. For small values, you probably won't
433    detect enough difference from series of malloc calls to bother.
434  
435    Overuse of independent_comalloc can increase overall memory usage,
436    since it cannot reuse existing noncontiguous small chunks that
437    might be available for some of the elements.
438  */
439  void** dlindependent_comalloc(size_t, size_t*, void**);
440  
441  /*
442    bulk_free(void* array[], size_t n_elements)
443    Frees and clears (sets to null) each non-null pointer in the given
444    array.  This is likely to be faster than freeing them one-by-one.
445    If footers are used, pointers that have been allocated in different
446    mspaces are not freed or cleared, and the count of all such pointers
447    is returned.  For large arrays of pointers with poor locality, it
448    may be worthwhile to sort this array before calling bulk_free.
449  */
450  size_t  dlbulk_free(void**, size_t n_elements);
451  
452  /*
453    pvalloc(size_t n);
454    Equivalent to valloc(minimum-page-that-holds(n)), that is,
455    round up n to nearest pagesize.
456   */
457  void*  dlpvalloc(size_t);
458  
459  /*
460    malloc_trim(size_t pad);
461  
462    If possible, gives memory back to the system (via negative arguments
463    to sbrk) if there is unused memory at the `high' end of the malloc
464    pool or in unused MMAP segments. You can call this after freeing
465    large blocks of memory to potentially reduce the system-level memory
466    requirements of a program. However, it cannot guarantee to reduce
467    memory. Under some allocation patterns, some large free blocks of
468    memory will be locked between two used chunks, so they cannot be
469    given back to the system.
470  
471    The `pad' argument to malloc_trim represents the amount of free
472    trailing space to leave untrimmed. If this argument is zero, only
473    the minimum amount of memory to maintain internal data structures
474    will be left. Non-zero arguments can be supplied to maintain enough
475    trailing space to service future expected allocations without having
476    to re-obtain memory from the system.
477  
478    Malloc_trim returns 1 if it actually released any memory, else 0.
479  */
480  int  dlmalloc_trim(size_t);
481  
482  /*
483    malloc_stats();
484    Prints on stderr the amount of space obtained from the system (both
485    via sbrk and mmap), the maximum amount (which may be more than
486    current if malloc_trim and/or munmap got called), and the current
487    number of bytes allocated via malloc (or realloc, etc) but not yet
488    freed. Note that this is the number of bytes allocated, not the
489    number requested. It will be larger than the number requested
490    because of alignment and bookkeeping overhead. Because it includes
491    alignment wastage as being in use, this figure may be greater than
492    zero even when no user-level chunks are allocated.
493  
494    The reported current and maximum system memory can be inaccurate if
495    a program makes other calls to system memory allocation functions
496    (normally sbrk) outside of malloc.
497  
498    malloc_stats prints only the most commonly interesting statistics.
499    More information can be obtained by calling mallinfo.
500  
501    malloc_stats is not compiled if NO_MALLOC_STATS is defined.
502  */
503  void  dlmalloc_stats(void);
504  
505  #endif /* !ONLY_MSPACES */
506  
507  /*
508    malloc_usable_size(void* p);
509  
510    Returns the number of bytes you can actually use in
511    an allocated chunk, which may be more than you requested (although
512    often not) due to alignment and minimum size constraints.
513    You can use this many bytes without worrying about
514    overwriting other allocated objects. This is not a particularly great
515    programming practice. malloc_usable_size can be more useful in
516    debugging and assertions, for example:
517  
518    p = malloc(n);
519    assert(malloc_usable_size(p) >= 256);
520  */
521  size_t dlmalloc_usable_size(const void*);
522  
523  #if MSPACES
524  
525  /*
526    mspace is an opaque type representing an independent
527    region of space that supports mspace_malloc, etc.
528  */
529  typedef void* mspace;
530  
531  /*
532    create_mspace creates and returns a new independent space with the
533    given initial capacity, or, if 0, the default granularity size.  It
534    returns null if there is no system memory available to create the
535    space.  If argument locked is non-zero, the space uses a separate
536    lock to control access. The capacity of the space will grow
537    dynamically as needed to service mspace_malloc requests.  You can
538    control the sizes of incremental increases of this space by
539    compiling with a different DEFAULT_GRANULARITY or dynamically
540    setting with mallopt(M_GRANULARITY, value).
541  */
542  mspace create_mspace(size_t capacity, int locked);
543  
544  /*
545    destroy_mspace destroys the given space, and attempts to return all
546    of its memory back to the system, returning the total number of
547    bytes freed. After destruction, the results of access to all memory
548    used by the space become undefined.
549  */
550  size_t destroy_mspace(mspace msp);
551  
552  /*
553    create_mspace_with_base uses the memory supplied as the initial base
554    of a new mspace. Part (less than 128*sizeof(size_t) bytes) of this
555    space is used for bookkeeping, so the capacity must be at least this
556    large. (Otherwise 0 is returned.) When this initial space is
557    exhausted, additional memory will be obtained from the system.
558    Destroying this space will deallocate all additionally allocated
559    space (if possible) but not the initial base.
560  */
561  mspace create_mspace_with_base(void* base, size_t capacity, int locked);
562  
563  /*
564    mspace_track_large_chunks controls whether requests for large chunks
565    are allocated in their own untracked mmapped regions, separate from
566    others in this mspace. By default large chunks are not tracked,
567    which reduces fragmentation. However, such chunks are not
568    necessarily released to the system upon destroy_mspace.  Enabling
569    tracking by setting to true may increase fragmentation, but avoids
570    leakage when relying on destroy_mspace to release all memory
571    allocated using this space.  The function returns the previous
572    setting.
573  */
574  int mspace_track_large_chunks(mspace msp, int enable);
575  
576  #if !NO_MALLINFO
577  /*
578    mspace_mallinfo behaves as mallinfo, but reports properties of
579    the given space.
580  */
581  struct mallinfo mspace_mallinfo(mspace msp);
582  #endif /* NO_MALLINFO */
583  
584  /*
585    An alias for mallopt.
586  */
587  int mspace_mallopt(int, int);
588  
589  /*
590    The following operate identically to their malloc counterparts
591    but operate only for the given mspace argument
592  */
593  void* mspace_malloc(mspace msp, size_t bytes);
594  void mspace_free(mspace msp, void* mem);
595  void* mspace_calloc(mspace msp, size_t n_elements, size_t elem_size);
596  void* mspace_realloc(mspace msp, void* mem, size_t newsize);
597  void* mspace_realloc_in_place(mspace msp, void* mem, size_t newsize);
598  void* mspace_memalign(mspace msp, size_t alignment, size_t bytes);
599  void** mspace_independent_calloc(mspace msp, size_t n_elements,
600                                   size_t elem_size, void* chunks[]);
601  void** mspace_independent_comalloc(mspace msp, size_t n_elements,
602                                     size_t sizes[], void* chunks[]);
603  size_t mspace_bulk_free(mspace msp, void**, size_t n_elements);
604  size_t mspace_usable_size(const void* mem);
605  void mspace_malloc_stats(mspace msp);
606  int mspace_trim(mspace msp, size_t pad);
607  size_t mspace_footprint(mspace msp);
608  size_t mspace_max_footprint(mspace msp);
609  size_t mspace_footprint_limit(mspace msp);
610  size_t mspace_set_footprint_limit(mspace msp, size_t bytes);
611  void mspace_inspect_all(mspace msp,
612                          void(*handler)(void *, void *, size_t, void*),
613                          void* arg);
614  #endif  /* MSPACES */
615  
616  #ifdef __cplusplus
617  };  /* end of extern "C" */
618  #endif
619  
620  #endif /* MALLOC_280_H */
621