| Index: third_party/android_debug_dlmalloc/chromium/malloc.h
 | 
| diff --git a/third_party/android_debug_dlmalloc/chromium/malloc.h b/third_party/android_debug_dlmalloc/chromium/malloc.h
 | 
| new file mode 100644
 | 
| index 0000000000000000000000000000000000000000..2de8791ce0f07f5d4b505aadd126fa960114febe
 | 
| --- /dev/null
 | 
| +++ b/third_party/android_debug_dlmalloc/chromium/malloc.h
 | 
| @@ -0,0 +1,511 @@
 | 
| +/*
 | 
| +  Default header file for malloc-2.8.x, written by Doug Lea
 | 
| +  and released to the public domain, as explained at
 | 
| +  http://creativecommons.org/publicdomain/zero/1.0/ 
 | 
| + 
 | 
| +  This header is for ANSI C/C++ only.  You can set any of
 | 
| +  the following #defines before including:
 | 
| +
 | 
| +  * If USE_DL_PREFIX is defined, it is assumed that malloc.c 
 | 
| +    was also compiled with this option, so all routines
 | 
| +    have names starting with "dl".
 | 
| +
 | 
| +  * If HAVE_USR_INCLUDE_MALLOC_H is defined, it is assumed that this
 | 
| +    file will be #included AFTER <malloc.h>. This is needed only if
 | 
| +    your system defines a struct mallinfo that is incompatible with the
 | 
| +    standard one declared here.  Otherwise, you can include this file
 | 
| +    INSTEAD of your system system <malloc.h>.  At least on ANSI, all
 | 
| +    declarations should be compatible with system versions
 | 
| +
 | 
| +  * If MSPACES is defined, declarations for mspace versions are included.
 | 
| +*/
 | 
| +
 | 
| +#ifndef MALLOC_280_H
 | 
| +#define MALLOC_280_H
 | 
| +
 | 
| +#ifdef __cplusplus
 | 
| +extern "C" {
 | 
| +#endif
 | 
| +
 | 
| +#include <stddef.h>   /* for size_t */
 | 
| +
 | 
| +#ifndef ONLY_MSPACES
 | 
| +#define ONLY_MSPACES 0     /* define to a value */
 | 
| +#elif ONLY_MSPACES != 0
 | 
| +#define ONLY_MSPACES 1
 | 
| +#endif  /* ONLY_MSPACES */
 | 
| +#ifndef NO_MALLINFO
 | 
| +#define NO_MALLINFO 0
 | 
| +#endif  /* NO_MALLINFO */
 | 
| +
 | 
| +#ifndef MSPACES
 | 
| +#if ONLY_MSPACES
 | 
| +#define MSPACES 1
 | 
| +#else   /* ONLY_MSPACES */
 | 
| +#define MSPACES 0
 | 
| +#endif  /* ONLY_MSPACES */
 | 
| +#endif  /* MSPACES */
 | 
| +
 | 
| +#if !ONLY_MSPACES
 | 
| +
 | 
| +#ifndef USE_DL_PREFIX
 | 
| +#define dlcalloc               calloc
 | 
| +#define dlfree                 free
 | 
| +#define dlmalloc               malloc
 | 
| +#define dlmemalign             memalign
 | 
| +#define dlposix_memalign       posix_memalign
 | 
| +#define dlrealloc              realloc
 | 
| +#define dlvalloc               valloc
 | 
| +#define dlpvalloc              pvalloc
 | 
| +#define dlmallinfo             mallinfo
 | 
| +#define dlmallopt              mallopt
 | 
| +#define dlmalloc_trim          malloc_trim
 | 
| +#define dlmalloc_stats         malloc_stats
 | 
| +#define dlmalloc_usable_size   malloc_usable_size
 | 
| +#define dlmalloc_footprint     malloc_footprint
 | 
| +#define dlmalloc_max_footprint malloc_max_footprint
 | 
| +#define dlmalloc_footprint_limit malloc_footprint_limit
 | 
| +#define dlmalloc_set_footprint_limit malloc_set_footprint_limit
 | 
| +#define dlmalloc_inspect_all   malloc_inspect_all
 | 
| +#define dlindependent_calloc   independent_calloc
 | 
| +#define dlindependent_comalloc independent_comalloc
 | 
| +#define dlbulk_free            bulk_free
 | 
| +#endif /* USE_DL_PREFIX */
 | 
| +
 | 
| +#if !NO_MALLINFO 
 | 
| +#ifndef HAVE_USR_INCLUDE_MALLOC_H
 | 
| +#ifndef _MALLOC_H
 | 
| +#ifndef MALLINFO_FIELD_TYPE
 | 
| +#define MALLINFO_FIELD_TYPE size_t
 | 
| +#endif /* MALLINFO_FIELD_TYPE */
 | 
| +#ifndef STRUCT_MALLINFO_DECLARED
 | 
| +#define STRUCT_MALLINFO_DECLARED 1
 | 
| +struct mallinfo {
 | 
| +  MALLINFO_FIELD_TYPE arena;    /* non-mmapped space allocated from system */
 | 
| +  MALLINFO_FIELD_TYPE ordblks;  /* number of free chunks */
 | 
| +  MALLINFO_FIELD_TYPE smblks;   /* always 0 */
 | 
| +  MALLINFO_FIELD_TYPE hblks;    /* always 0 */
 | 
| +  MALLINFO_FIELD_TYPE hblkhd;   /* space in mmapped regions */
 | 
| +  MALLINFO_FIELD_TYPE usmblks;  /* maximum total allocated space */
 | 
| +  MALLINFO_FIELD_TYPE fsmblks;  /* always 0 */
 | 
| +  MALLINFO_FIELD_TYPE uordblks; /* total allocated space */
 | 
| +  MALLINFO_FIELD_TYPE fordblks; /* total free space */
 | 
| +  MALLINFO_FIELD_TYPE keepcost; /* releasable (via malloc_trim) space */
 | 
| +};
 | 
| +#endif /* STRUCT_MALLINFO_DECLARED */
 | 
| +#endif  /* _MALLOC_H */
 | 
| +#endif  /* HAVE_USR_INCLUDE_MALLOC_H */
 | 
| +#endif  /* !NO_MALLINFO */
 | 
| +
 | 
| +/*
 | 
| +  malloc(size_t n)
 | 
| +  Returns a pointer to a newly allocated chunk of at least n bytes, or
 | 
| +  null if no space is available, in which case errno is set to ENOMEM
 | 
| +  on ANSI C systems.
 | 
| +
 | 
| +  If n is zero, malloc returns a minimum-sized chunk. (The minimum
 | 
| +  size is 16 bytes on most 32bit systems, and 32 bytes on 64bit
 | 
| +  systems.)  Note that size_t is an unsigned type, so calls with
 | 
| +  arguments that would be negative if signed are interpreted as
 | 
| +  requests for huge amounts of space, which will often fail. The
 | 
| +  maximum supported value of n differs across systems, but is in all
 | 
| +  cases less than the maximum representable value of a size_t.
 | 
| +*/
 | 
| +void* dlmalloc(size_t);
 | 
| +
 | 
| +/*
 | 
| +  free(void* p)
 | 
| +  Releases the chunk of memory pointed to by p, that had been previously
 | 
| +  allocated using malloc or a related routine such as realloc.
 | 
| +  It has no effect if p is null. If p was not malloced or already
 | 
| +  freed, free(p) will by default cuase the current program to abort.
 | 
| +*/
 | 
| +void  dlfree(void*);
 | 
| +
 | 
| +/*
 | 
| +  calloc(size_t n_elements, size_t element_size);
 | 
| +  Returns a pointer to n_elements * element_size bytes, with all locations
 | 
| +  set to zero.
 | 
| +*/
 | 
| +void* dlcalloc(size_t, size_t);
 | 
| +
 | 
| +/*
 | 
| +  realloc(void* p, size_t n)
 | 
| +  Returns a pointer to a chunk of size n that contains the same data
 | 
| +  as does chunk p up to the minimum of (n, p's size) bytes, or null
 | 
| +  if no space is available.
 | 
| +
 | 
| +  The returned pointer may or may not be the same as p. The algorithm
 | 
| +  prefers extending p in most cases when possible, otherwise it
 | 
| +  employs the equivalent of a malloc-copy-free sequence.
 | 
| +
 | 
| +  If p is null, realloc is equivalent to malloc.
 | 
| +
 | 
| +  If space is not available, realloc returns null, errno is set (if on
 | 
| +  ANSI) and p is NOT freed.
 | 
| +
 | 
| +  if n is for fewer bytes than already held by p, the newly unused
 | 
| +  space is lopped off and freed if possible.  realloc with a size
 | 
| +  argument of zero (re)allocates a minimum-sized chunk.
 | 
| +
 | 
| +  The old unix realloc convention of allowing the last-free'd chunk
 | 
| +  to be used as an argument to realloc is not supported.
 | 
| +*/
 | 
| +void* dlrealloc(void*, size_t);
 | 
| +
 | 
| +/*
 | 
| +  realloc_in_place(void* p, size_t n)
 | 
| +  Resizes the space allocated for p to size n, only if this can be
 | 
| +  done without moving p (i.e., only if there is adjacent space
 | 
| +  available if n is greater than p's current allocated size, or n is
 | 
| +  less than or equal to p's size). This may be used instead of plain
 | 
| +  realloc if an alternative allocation strategy is needed upon failure
 | 
| +  to expand space; for example, reallocation of a buffer that must be
 | 
| +  memory-aligned or cleared. You can use realloc_in_place to trigger
 | 
| +  these alternatives only when needed.
 | 
| +
 | 
| +  Returns p if successful; otherwise null.
 | 
| +*/
 | 
| +void* dlrealloc_in_place(void*, size_t);
 | 
| +
 | 
| +/*
 | 
| +  memalign(size_t alignment, size_t n);
 | 
| +  Returns a pointer to a newly allocated chunk of n bytes, aligned
 | 
| +  in accord with the alignment argument.
 | 
| +
 | 
| +  The alignment argument should be a power of two. If the argument is
 | 
| +  not a power of two, the nearest greater power is used.
 | 
| +  8-byte alignment is guaranteed by normal malloc calls, so don't
 | 
| +  bother calling memalign with an argument of 8 or less.
 | 
| +
 | 
| +  Overreliance on memalign is a sure way to fragment space.
 | 
| +*/
 | 
| +void* dlmemalign(size_t, size_t);
 | 
| +
 | 
| +/*
 | 
| +  int posix_memalign(void** pp, size_t alignment, size_t n);
 | 
| +  Allocates a chunk of n bytes, aligned in accord with the alignment
 | 
| +  argument. Differs from memalign only in that it (1) assigns the
 | 
| +  allocated memory to *pp rather than returning it, (2) fails and
 | 
| +  returns EINVAL if the alignment is not a power of two (3) fails and
 | 
| +  returns ENOMEM if memory cannot be allocated.
 | 
| +*/
 | 
| +int dlposix_memalign(void**, size_t, size_t);
 | 
| +
 | 
| +/*
 | 
| +  valloc(size_t n);
 | 
| +  Equivalent to memalign(pagesize, n), where pagesize is the page
 | 
| +  size of the system. If the pagesize is unknown, 4096 is used.
 | 
| +*/
 | 
| +void* dlvalloc(size_t);
 | 
| +
 | 
| +/*
 | 
| +  mallopt(int parameter_number, int parameter_value)
 | 
| +  Sets tunable parameters The format is to provide a
 | 
| +  (parameter-number, parameter-value) pair.  mallopt then sets the
 | 
| +  corresponding parameter to the argument value if it can (i.e., so
 | 
| +  long as the value is meaningful), and returns 1 if successful else
 | 
| +  0.  SVID/XPG/ANSI defines four standard param numbers for mallopt,
 | 
| +  normally defined in malloc.h.  None of these are use in this malloc,
 | 
| +  so setting them has no effect. But this malloc also supports other
 | 
| +  options in mallopt:
 | 
| +
 | 
| +  Symbol            param #  default    allowed param values
 | 
| +  M_TRIM_THRESHOLD     -1   2*1024*1024   any   (-1U disables trimming)
 | 
| +  M_GRANULARITY        -2     page size   any power of 2 >= page size
 | 
| +  M_MMAP_THRESHOLD     -3      256*1024   any   (or 0 if no MMAP support)
 | 
| +*/
 | 
| +int dlmallopt(int, int);
 | 
| +
 | 
| +#define M_TRIM_THRESHOLD     (-1)
 | 
| +#define M_GRANULARITY        (-2)
 | 
| +#define M_MMAP_THRESHOLD     (-3)
 | 
| +
 | 
| +
 | 
| +/*
 | 
| +  malloc_footprint();
 | 
| +  Returns the number of bytes obtained from the system.  The total
 | 
| +  number of bytes allocated by malloc, realloc etc., is less than this
 | 
| +  value. Unlike mallinfo, this function returns only a precomputed
 | 
| +  result, so can be called frequently to monitor memory consumption.
 | 
| +  Even if locks are otherwise defined, this function does not use them,
 | 
| +  so results might not be up to date.
 | 
| +*/
 | 
| +size_t dlmalloc_footprint(void);
 | 
| +
 | 
| +/*
 | 
| +  malloc_max_footprint();
 | 
| +  Returns the maximum number of bytes obtained from the system. This
 | 
| +  value will be greater than current footprint if deallocated space
 | 
| +  has been reclaimed by the system. The peak number of bytes allocated
 | 
| +  by malloc, realloc etc., is less than this value. Unlike mallinfo,
 | 
| +  this function returns only a precomputed result, so can be called
 | 
| +  frequently to monitor memory consumption.  Even if locks are
 | 
| +  otherwise defined, this function does not use them, so results might
 | 
| +  not be up to date.
 | 
| +*/
 | 
| +size_t dlmalloc_max_footprint(void);
 | 
| +
 | 
| +/*
 | 
| +  malloc_footprint_limit();
 | 
| +  Returns the number of bytes that the heap is allowed to obtain from
 | 
| +  the system, returning the last value returned by
 | 
| +  malloc_set_footprint_limit, or the maximum size_t value if
 | 
| +  never set. The returned value reflects a permission. There is no
 | 
| +  guarantee that this number of bytes can actually be obtained from
 | 
| +  the system.  
 | 
| +*/
 | 
| +size_t dlmalloc_footprint_limit(void);
 | 
| +
 | 
| +/*
 | 
| +  malloc_set_footprint_limit();
 | 
| +  Sets the maximum number of bytes to obtain from the system, causing
 | 
| +  failure returns from malloc and related functions upon attempts to
 | 
| +  exceed this value. The argument value may be subject to page
 | 
| +  rounding to an enforceable limit; this actual value is returned.
 | 
| +  Using an argument of the maximum possible size_t effectively
 | 
| +  disables checks. If the argument is less than or equal to the
 | 
| +  current malloc_footprint, then all future allocations that require
 | 
| +  additional system memory will fail. However, invocation cannot
 | 
| +  retroactively deallocate existing used memory.
 | 
| +*/
 | 
| +size_t dlmalloc_set_footprint_limit(size_t bytes);
 | 
| +
 | 
| +/*
 | 
| +  malloc_inspect_all(void(*handler)(void *start,
 | 
| +                                    void *end,
 | 
| +                                    size_t used_bytes,
 | 
| +                                    void* callback_arg),
 | 
| +                      void* arg);
 | 
| +  Traverses the heap and calls the given handler for each managed
 | 
| +  region, skipping all bytes that are (or may be) used for bookkeeping
 | 
| +  purposes.  Traversal does not include include chunks that have been
 | 
| +  directly memory mapped. Each reported region begins at the start
 | 
| +  address, and continues up to but not including the end address.  The
 | 
| +  first used_bytes of the region contain allocated data. If
 | 
| +  used_bytes is zero, the region is unallocated. The handler is
 | 
| +  invoked with the given callback argument. If locks are defined, they
 | 
| +  are held during the entire traversal. It is a bad idea to invoke
 | 
| +  other malloc functions from within the handler.
 | 
| +
 | 
| +  For example, to count the number of in-use chunks with size greater
 | 
| +  than 1000, you could write:
 | 
| +  static int count = 0;
 | 
| +  void count_chunks(void* start, void* end, size_t used, void* arg) {
 | 
| +    if (used >= 1000) ++count;
 | 
| +  }
 | 
| +  then:
 | 
| +    malloc_inspect_all(count_chunks, NULL);
 | 
| +
 | 
| +  malloc_inspect_all is compiled only if MALLOC_INSPECT_ALL is defined.
 | 
| +*/
 | 
| +void dlmalloc_inspect_all(void(*handler)(void*, void *, size_t, void*),
 | 
| +                           void* arg);
 | 
| +
 | 
| +#if !NO_MALLINFO
 | 
| +/*
 | 
| +  mallinfo()
 | 
| +  Returns (by copy) a struct containing various summary statistics:
 | 
| +
 | 
| +  arena:     current total non-mmapped bytes allocated from system
 | 
| +  ordblks:   the number of free chunks
 | 
| +  smblks:    always zero.
 | 
| +  hblks:     current number of mmapped regions
 | 
| +  hblkhd:    total bytes held in mmapped regions
 | 
| +  usmblks:   the maximum total allocated space. This will be greater
 | 
| +                than current total if trimming has occurred.
 | 
| +  fsmblks:   always zero
 | 
| +  uordblks:  current total allocated space (normal or mmapped)
 | 
| +  fordblks:  total free space
 | 
| +  keepcost:  the maximum number of bytes that could ideally be released
 | 
| +               back to system via malloc_trim. ("ideally" means that
 | 
| +               it ignores page restrictions etc.)
 | 
| +
 | 
| +  Because these fields are ints, but internal bookkeeping may
 | 
| +  be kept as longs, the reported values may wrap around zero and
 | 
| +  thus be inaccurate.
 | 
| +*/
 | 
| +
 | 
| +struct mallinfo dlmallinfo(void);
 | 
| +#endif  /* NO_MALLINFO */
 | 
| +
 | 
| +/*
 | 
| +  independent_calloc(size_t n_elements, size_t element_size, void* chunks[]);
 | 
| +
 | 
| +  independent_calloc is similar to calloc, but instead of returning a
 | 
| +  single cleared space, it returns an array of pointers to n_elements
 | 
| +  independent elements that can hold contents of size elem_size, each
 | 
| +  of which starts out cleared, and can be independently freed,
 | 
| +  realloc'ed etc. The elements are guaranteed to be adjacently
 | 
| +  allocated (this is not guaranteed to occur with multiple callocs or
 | 
| +  mallocs), which may also improve cache locality in some
 | 
| +  applications.
 | 
| +
 | 
| +  The "chunks" argument is optional (i.e., may be null, which is
 | 
| +  probably the most typical usage). If it is null, the returned array
 | 
| +  is itself dynamically allocated and should also be freed when it is
 | 
| +  no longer needed. Otherwise, the chunks array must be of at least
 | 
| +  n_elements in length. It is filled in with the pointers to the
 | 
| +  chunks.
 | 
| +
 | 
| +  In either case, independent_calloc returns this pointer array, or
 | 
| +  null if the allocation failed.  If n_elements is zero and "chunks"
 | 
| +  is null, it returns a chunk representing an array with zero elements
 | 
| +  (which should be freed if not wanted).
 | 
| +
 | 
| +  Each element must be freed when it is no longer needed. This can be
 | 
| +  done all at once using bulk_free.
 | 
| +
 | 
| +  independent_calloc simplifies and speeds up implementations of many
 | 
| +  kinds of pools.  It may also be useful when constructing large data
 | 
| +  structures that initially have a fixed number of fixed-sized nodes,
 | 
| +  but the number is not known at compile time, and some of the nodes
 | 
| +  may later need to be freed. For example:
 | 
| +
 | 
| +  struct Node { int item; struct Node* next; };
 | 
| +
 | 
| +  struct Node* build_list() {
 | 
| +    struct Node** pool;
 | 
| +    int n = read_number_of_nodes_needed();
 | 
| +    if (n <= 0) return 0;
 | 
| +    pool = (struct Node**)(independent_calloc(n, sizeof(struct Node), 0);
 | 
| +    if (pool == 0) die();
 | 
| +    // organize into a linked list...
 | 
| +    struct Node* first = pool[0];
 | 
| +    for (i = 0; i < n-1; ++i)
 | 
| +      pool[i]->next = pool[i+1];
 | 
| +    free(pool);     // Can now free the array (or not, if it is needed later)
 | 
| +    return first;
 | 
| +  }
 | 
| +*/
 | 
| +void** dlindependent_calloc(size_t, size_t, void**);
 | 
| +
 | 
| +/*
 | 
| +  independent_comalloc(size_t n_elements, size_t sizes[], void* chunks[]);
 | 
| +
 | 
| +  independent_comalloc allocates, all at once, a set of n_elements
 | 
| +  chunks with sizes indicated in the "sizes" array.    It returns
 | 
| +  an array of pointers to these elements, each of which can be
 | 
| +  independently freed, realloc'ed etc. The elements are guaranteed to
 | 
| +  be adjacently allocated (this is not guaranteed to occur with
 | 
| +  multiple callocs or mallocs), which may also improve cache locality
 | 
| +  in some applications.
 | 
| +
 | 
| +  The "chunks" argument is optional (i.e., may be null). If it is null
 | 
| +  the returned array is itself dynamically allocated and should also
 | 
| +  be freed when it is no longer needed. Otherwise, the chunks array
 | 
| +  must be of at least n_elements in length. It is filled in with the
 | 
| +  pointers to the chunks.
 | 
| +
 | 
| +  In either case, independent_comalloc returns this pointer array, or
 | 
| +  null if the allocation failed.  If n_elements is zero and chunks is
 | 
| +  null, it returns a chunk representing an array with zero elements
 | 
| +  (which should be freed if not wanted).
 | 
| +
 | 
| +  Each element must be freed when it is no longer needed. This can be
 | 
| +  done all at once using bulk_free.
 | 
| +
 | 
| +  independent_comallac differs from independent_calloc in that each
 | 
| +  element may have a different size, and also that it does not
 | 
| +  automatically clear elements.
 | 
| +
 | 
| +  independent_comalloc can be used to speed up allocation in cases
 | 
| +  where several structs or objects must always be allocated at the
 | 
| +  same time.  For example:
 | 
| +
 | 
| +  struct Head { ... }
 | 
| +  struct Foot { ... }
 | 
| +
 | 
| +  void send_message(char* msg) {
 | 
| +    int msglen = strlen(msg);
 | 
| +    size_t sizes[3] = { sizeof(struct Head), msglen, sizeof(struct Foot) };
 | 
| +    void* chunks[3];
 | 
| +    if (independent_comalloc(3, sizes, chunks) == 0)
 | 
| +      die();
 | 
| +    struct Head* head = (struct Head*)(chunks[0]);
 | 
| +    char*        body = (char*)(chunks[1]);
 | 
| +    struct Foot* foot = (struct Foot*)(chunks[2]);
 | 
| +    // ...
 | 
| +  }
 | 
| +
 | 
| +  In general though, independent_comalloc is worth using only for
 | 
| +  larger values of n_elements. For small values, you probably won't
 | 
| +  detect enough difference from series of malloc calls to bother.
 | 
| +
 | 
| +  Overuse of independent_comalloc can increase overall memory usage,
 | 
| +  since it cannot reuse existing noncontiguous small chunks that
 | 
| +  might be available for some of the elements.
 | 
| +*/
 | 
| +void** dlindependent_comalloc(size_t, size_t*, void**);
 | 
| +
 | 
| +/*
 | 
| +  bulk_free(void* array[], size_t n_elements)
 | 
| +  Frees and clears (sets to null) each non-null pointer in the given
 | 
| +  array.  This is likely to be faster than freeing them one-by-one.
 | 
| +  If footers are used, pointers that have been allocated in different
 | 
| +  mspaces are not freed or cleared, and the count of all such pointers
 | 
| +  is returned.  For large arrays of pointers with poor locality, it
 | 
| +  may be worthwhile to sort this array before calling bulk_free.
 | 
| +*/
 | 
| +size_t  dlbulk_free(void**, size_t n_elements);
 | 
| +
 | 
| +/*
 | 
| +  pvalloc(size_t n);
 | 
| +  Equivalent to valloc(minimum-page-that-holds(n)), that is,
 | 
| +  round up n to nearest pagesize.
 | 
| + */
 | 
| +void*  dlpvalloc(size_t);
 | 
| +
 | 
| +/*
 | 
| +  malloc_trim(size_t pad);
 | 
| +
 | 
| +  If possible, gives memory back to the system (via negative arguments
 | 
| +  to sbrk) if there is unused memory at the `high' end of the malloc
 | 
| +  pool or in unused MMAP segments. You can call this after freeing
 | 
| +  large blocks of memory to potentially reduce the system-level memory
 | 
| +  requirements of a program. However, it cannot guarantee to reduce
 | 
| +  memory. Under some allocation patterns, some large free blocks of
 | 
| +  memory will be locked between two used chunks, so they cannot be
 | 
| +  given back to the system.
 | 
| +
 | 
| +  The `pad' argument to malloc_trim represents the amount of free
 | 
| +  trailing space to leave untrimmed. If this argument is zero, only
 | 
| +  the minimum amount of memory to maintain internal data structures
 | 
| +  will be left. Non-zero arguments can be supplied to maintain enough
 | 
| +  trailing space to service future expected allocations without having
 | 
| +  to re-obtain memory from the system.
 | 
| +
 | 
| +  Malloc_trim returns 1 if it actually released any memory, else 0.
 | 
| +*/
 | 
| +int  dlmalloc_trim(size_t);
 | 
| +
 | 
| +/*
 | 
| +  malloc_stats();
 | 
| +  Prints on stderr the amount of space obtained from the system (both
 | 
| +  via sbrk and mmap), the maximum amount (which may be more than
 | 
| +  current if malloc_trim and/or munmap got called), and the current
 | 
| +  number of bytes allocated via malloc (or realloc, etc) but not yet
 | 
| +  freed. Note that this is the number of bytes allocated, not the
 | 
| +  number requested. It will be larger than the number requested
 | 
| +  because of alignment and bookkeeping overhead. Because it includes
 | 
| +  alignment wastage as being in use, this figure may be greater than
 | 
| +  zero even when no user-level chunks are allocated.
 | 
| +
 | 
| +  The reported current and maximum system memory can be inaccurate if
 | 
| +  a program makes other calls to system memory allocation functions
 | 
| +  (normally sbrk) outside of malloc.
 | 
| +
 | 
| +  malloc_stats prints only the most commonly interesting statistics.
 | 
| +  More information can be obtained by calling mallinfo.
 | 
| +  
 | 
| +  malloc_stats is not compiled if NO_MALLOC_STATS is defined.
 | 
| +*/
 | 
| +void  dlmalloc_stats(void);
 | 
| +
 | 
| +#endif /* !ONLY_MSPACES */
 | 
| +
 | 
| +#ifdef __cplusplus
 | 
| +};  /* end of extern "C" */
 | 
| +#endif
 | 
| +
 | 
| +#endif /* MALLOC_280_H */
 | 
| 
 |