| Index: third_party/android_debug_dlmalloc/chromium/malloc.h
|
| diff --git a/third_party/android_debug_dlmalloc/chromium/malloc.h b/third_party/android_debug_dlmalloc/chromium/malloc.h
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..2de8791ce0f07f5d4b505aadd126fa960114febe
|
| --- /dev/null
|
| +++ b/third_party/android_debug_dlmalloc/chromium/malloc.h
|
| @@ -0,0 +1,511 @@
|
| +/*
|
| + Default header file for malloc-2.8.x, written by Doug Lea
|
| + and released to the public domain, as explained at
|
| + http://creativecommons.org/publicdomain/zero/1.0/
|
| +
|
| + This header is for ANSI C/C++ only. You can set any of
|
| + the following #defines before including:
|
| +
|
| + * If USE_DL_PREFIX is defined, it is assumed that malloc.c
|
| + was also compiled with this option, so all routines
|
| + have names starting with "dl".
|
| +
|
| + * If HAVE_USR_INCLUDE_MALLOC_H is defined, it is assumed that this
|
| + file will be #included AFTER <malloc.h>. This is needed only if
|
| + your system defines a struct mallinfo that is incompatible with the
|
| + standard one declared here. Otherwise, you can include this file
|
| + INSTEAD of your system system <malloc.h>. At least on ANSI, all
|
| + declarations should be compatible with system versions
|
| +
|
| + * If MSPACES is defined, declarations for mspace versions are included.
|
| +*/
|
| +
|
| +#ifndef MALLOC_280_H
|
| +#define MALLOC_280_H
|
| +
|
| +#ifdef __cplusplus
|
| +extern "C" {
|
| +#endif
|
| +
|
| +#include <stddef.h> /* for size_t */
|
| +
|
| +#ifndef ONLY_MSPACES
|
| +#define ONLY_MSPACES 0 /* define to a value */
|
| +#elif ONLY_MSPACES != 0
|
| +#define ONLY_MSPACES 1
|
| +#endif /* ONLY_MSPACES */
|
| +#ifndef NO_MALLINFO
|
| +#define NO_MALLINFO 0
|
| +#endif /* NO_MALLINFO */
|
| +
|
| +#ifndef MSPACES
|
| +#if ONLY_MSPACES
|
| +#define MSPACES 1
|
| +#else /* ONLY_MSPACES */
|
| +#define MSPACES 0
|
| +#endif /* ONLY_MSPACES */
|
| +#endif /* MSPACES */
|
| +
|
| +#if !ONLY_MSPACES
|
| +
|
| +#ifndef USE_DL_PREFIX
|
| +#define dlcalloc calloc
|
| +#define dlfree free
|
| +#define dlmalloc malloc
|
| +#define dlmemalign memalign
|
| +#define dlposix_memalign posix_memalign
|
| +#define dlrealloc realloc
|
| +#define dlvalloc valloc
|
| +#define dlpvalloc pvalloc
|
| +#define dlmallinfo mallinfo
|
| +#define dlmallopt mallopt
|
| +#define dlmalloc_trim malloc_trim
|
| +#define dlmalloc_stats malloc_stats
|
| +#define dlmalloc_usable_size malloc_usable_size
|
| +#define dlmalloc_footprint malloc_footprint
|
| +#define dlmalloc_max_footprint malloc_max_footprint
|
| +#define dlmalloc_footprint_limit malloc_footprint_limit
|
| +#define dlmalloc_set_footprint_limit malloc_set_footprint_limit
|
| +#define dlmalloc_inspect_all malloc_inspect_all
|
| +#define dlindependent_calloc independent_calloc
|
| +#define dlindependent_comalloc independent_comalloc
|
| +#define dlbulk_free bulk_free
|
| +#endif /* USE_DL_PREFIX */
|
| +
|
| +#if !NO_MALLINFO
|
| +#ifndef HAVE_USR_INCLUDE_MALLOC_H
|
| +#ifndef _MALLOC_H
|
| +#ifndef MALLINFO_FIELD_TYPE
|
| +#define MALLINFO_FIELD_TYPE size_t
|
| +#endif /* MALLINFO_FIELD_TYPE */
|
| +#ifndef STRUCT_MALLINFO_DECLARED
|
| +#define STRUCT_MALLINFO_DECLARED 1
|
| +struct mallinfo {
|
| + MALLINFO_FIELD_TYPE arena; /* non-mmapped space allocated from system */
|
| + MALLINFO_FIELD_TYPE ordblks; /* number of free chunks */
|
| + MALLINFO_FIELD_TYPE smblks; /* always 0 */
|
| + MALLINFO_FIELD_TYPE hblks; /* always 0 */
|
| + MALLINFO_FIELD_TYPE hblkhd; /* space in mmapped regions */
|
| + MALLINFO_FIELD_TYPE usmblks; /* maximum total allocated space */
|
| + MALLINFO_FIELD_TYPE fsmblks; /* always 0 */
|
| + MALLINFO_FIELD_TYPE uordblks; /* total allocated space */
|
| + MALLINFO_FIELD_TYPE fordblks; /* total free space */
|
| + MALLINFO_FIELD_TYPE keepcost; /* releasable (via malloc_trim) space */
|
| +};
|
| +#endif /* STRUCT_MALLINFO_DECLARED */
|
| +#endif /* _MALLOC_H */
|
| +#endif /* HAVE_USR_INCLUDE_MALLOC_H */
|
| +#endif /* !NO_MALLINFO */
|
| +
|
| +/*
|
| + malloc(size_t n)
|
| + Returns a pointer to a newly allocated chunk of at least n bytes, or
|
| + null if no space is available, in which case errno is set to ENOMEM
|
| + on ANSI C systems.
|
| +
|
| + If n is zero, malloc returns a minimum-sized chunk. (The minimum
|
| + size is 16 bytes on most 32bit systems, and 32 bytes on 64bit
|
| + systems.) Note that size_t is an unsigned type, so calls with
|
| + arguments that would be negative if signed are interpreted as
|
| + requests for huge amounts of space, which will often fail. The
|
| + maximum supported value of n differs across systems, but is in all
|
| + cases less than the maximum representable value of a size_t.
|
| +*/
|
| +void* dlmalloc(size_t);
|
| +
|
| +/*
|
| + free(void* p)
|
| + Releases the chunk of memory pointed to by p, that had been previously
|
| + allocated using malloc or a related routine such as realloc.
|
| + It has no effect if p is null. If p was not malloced or already
|
| + freed, free(p) will by default cuase the current program to abort.
|
| +*/
|
| +void dlfree(void*);
|
| +
|
| +/*
|
| + calloc(size_t n_elements, size_t element_size);
|
| + Returns a pointer to n_elements * element_size bytes, with all locations
|
| + set to zero.
|
| +*/
|
| +void* dlcalloc(size_t, size_t);
|
| +
|
| +/*
|
| + realloc(void* p, size_t n)
|
| + Returns a pointer to a chunk of size n that contains the same data
|
| + as does chunk p up to the minimum of (n, p's size) bytes, or null
|
| + if no space is available.
|
| +
|
| + The returned pointer may or may not be the same as p. The algorithm
|
| + prefers extending p in most cases when possible, otherwise it
|
| + employs the equivalent of a malloc-copy-free sequence.
|
| +
|
| + If p is null, realloc is equivalent to malloc.
|
| +
|
| + If space is not available, realloc returns null, errno is set (if on
|
| + ANSI) and p is NOT freed.
|
| +
|
| + if n is for fewer bytes than already held by p, the newly unused
|
| + space is lopped off and freed if possible. realloc with a size
|
| + argument of zero (re)allocates a minimum-sized chunk.
|
| +
|
| + The old unix realloc convention of allowing the last-free'd chunk
|
| + to be used as an argument to realloc is not supported.
|
| +*/
|
| +void* dlrealloc(void*, size_t);
|
| +
|
| +/*
|
| + realloc_in_place(void* p, size_t n)
|
| + Resizes the space allocated for p to size n, only if this can be
|
| + done without moving p (i.e., only if there is adjacent space
|
| + available if n is greater than p's current allocated size, or n is
|
| + less than or equal to p's size). This may be used instead of plain
|
| + realloc if an alternative allocation strategy is needed upon failure
|
| + to expand space; for example, reallocation of a buffer that must be
|
| + memory-aligned or cleared. You can use realloc_in_place to trigger
|
| + these alternatives only when needed.
|
| +
|
| + Returns p if successful; otherwise null.
|
| +*/
|
| +void* dlrealloc_in_place(void*, size_t);
|
| +
|
| +/*
|
| + memalign(size_t alignment, size_t n);
|
| + Returns a pointer to a newly allocated chunk of n bytes, aligned
|
| + in accord with the alignment argument.
|
| +
|
| + The alignment argument should be a power of two. If the argument is
|
| + not a power of two, the nearest greater power is used.
|
| + 8-byte alignment is guaranteed by normal malloc calls, so don't
|
| + bother calling memalign with an argument of 8 or less.
|
| +
|
| + Overreliance on memalign is a sure way to fragment space.
|
| +*/
|
| +void* dlmemalign(size_t, size_t);
|
| +
|
| +/*
|
| + int posix_memalign(void** pp, size_t alignment, size_t n);
|
| + Allocates a chunk of n bytes, aligned in accord with the alignment
|
| + argument. Differs from memalign only in that it (1) assigns the
|
| + allocated memory to *pp rather than returning it, (2) fails and
|
| + returns EINVAL if the alignment is not a power of two (3) fails and
|
| + returns ENOMEM if memory cannot be allocated.
|
| +*/
|
| +int dlposix_memalign(void**, size_t, size_t);
|
| +
|
| +/*
|
| + valloc(size_t n);
|
| + Equivalent to memalign(pagesize, n), where pagesize is the page
|
| + size of the system. If the pagesize is unknown, 4096 is used.
|
| +*/
|
| +void* dlvalloc(size_t);
|
| +
|
| +/*
|
| + mallopt(int parameter_number, int parameter_value)
|
| + Sets tunable parameters The format is to provide a
|
| + (parameter-number, parameter-value) pair. mallopt then sets the
|
| + corresponding parameter to the argument value if it can (i.e., so
|
| + long as the value is meaningful), and returns 1 if successful else
|
| + 0. SVID/XPG/ANSI defines four standard param numbers for mallopt,
|
| + normally defined in malloc.h. None of these are use in this malloc,
|
| + so setting them has no effect. But this malloc also supports other
|
| + options in mallopt:
|
| +
|
| + Symbol param # default allowed param values
|
| + M_TRIM_THRESHOLD -1 2*1024*1024 any (-1U disables trimming)
|
| + M_GRANULARITY -2 page size any power of 2 >= page size
|
| + M_MMAP_THRESHOLD -3 256*1024 any (or 0 if no MMAP support)
|
| +*/
|
| +int dlmallopt(int, int);
|
| +
|
| +#define M_TRIM_THRESHOLD (-1)
|
| +#define M_GRANULARITY (-2)
|
| +#define M_MMAP_THRESHOLD (-3)
|
| +
|
| +
|
| +/*
|
| + malloc_footprint();
|
| + Returns the number of bytes obtained from the system. The total
|
| + number of bytes allocated by malloc, realloc etc., is less than this
|
| + value. Unlike mallinfo, this function returns only a precomputed
|
| + result, so can be called frequently to monitor memory consumption.
|
| + Even if locks are otherwise defined, this function does not use them,
|
| + so results might not be up to date.
|
| +*/
|
| +size_t dlmalloc_footprint(void);
|
| +
|
| +/*
|
| + malloc_max_footprint();
|
| + Returns the maximum number of bytes obtained from the system. This
|
| + value will be greater than current footprint if deallocated space
|
| + has been reclaimed by the system. The peak number of bytes allocated
|
| + by malloc, realloc etc., is less than this value. Unlike mallinfo,
|
| + this function returns only a precomputed result, so can be called
|
| + frequently to monitor memory consumption. Even if locks are
|
| + otherwise defined, this function does not use them, so results might
|
| + not be up to date.
|
| +*/
|
| +size_t dlmalloc_max_footprint(void);
|
| +
|
| +/*
|
| + malloc_footprint_limit();
|
| + Returns the number of bytes that the heap is allowed to obtain from
|
| + the system, returning the last value returned by
|
| + malloc_set_footprint_limit, or the maximum size_t value if
|
| + never set. The returned value reflects a permission. There is no
|
| + guarantee that this number of bytes can actually be obtained from
|
| + the system.
|
| +*/
|
| +size_t dlmalloc_footprint_limit(void);
|
| +
|
| +/*
|
| + malloc_set_footprint_limit();
|
| + Sets the maximum number of bytes to obtain from the system, causing
|
| + failure returns from malloc and related functions upon attempts to
|
| + exceed this value. The argument value may be subject to page
|
| + rounding to an enforceable limit; this actual value is returned.
|
| + Using an argument of the maximum possible size_t effectively
|
| + disables checks. If the argument is less than or equal to the
|
| + current malloc_footprint, then all future allocations that require
|
| + additional system memory will fail. However, invocation cannot
|
| + retroactively deallocate existing used memory.
|
| +*/
|
| +size_t dlmalloc_set_footprint_limit(size_t bytes);
|
| +
|
| +/*
|
| + malloc_inspect_all(void(*handler)(void *start,
|
| + void *end,
|
| + size_t used_bytes,
|
| + void* callback_arg),
|
| + void* arg);
|
| + Traverses the heap and calls the given handler for each managed
|
| + region, skipping all bytes that are (or may be) used for bookkeeping
|
| + purposes. Traversal does not include include chunks that have been
|
| + directly memory mapped. Each reported region begins at the start
|
| + address, and continues up to but not including the end address. The
|
| + first used_bytes of the region contain allocated data. If
|
| + used_bytes is zero, the region is unallocated. The handler is
|
| + invoked with the given callback argument. If locks are defined, they
|
| + are held during the entire traversal. It is a bad idea to invoke
|
| + other malloc functions from within the handler.
|
| +
|
| + For example, to count the number of in-use chunks with size greater
|
| + than 1000, you could write:
|
| + static int count = 0;
|
| + void count_chunks(void* start, void* end, size_t used, void* arg) {
|
| + if (used >= 1000) ++count;
|
| + }
|
| + then:
|
| + malloc_inspect_all(count_chunks, NULL);
|
| +
|
| + malloc_inspect_all is compiled only if MALLOC_INSPECT_ALL is defined.
|
| +*/
|
| +void dlmalloc_inspect_all(void(*handler)(void*, void *, size_t, void*),
|
| + void* arg);
|
| +
|
| +#if !NO_MALLINFO
|
| +/*
|
| + mallinfo()
|
| + Returns (by copy) a struct containing various summary statistics:
|
| +
|
| + arena: current total non-mmapped bytes allocated from system
|
| + ordblks: the number of free chunks
|
| + smblks: always zero.
|
| + hblks: current number of mmapped regions
|
| + hblkhd: total bytes held in mmapped regions
|
| + usmblks: the maximum total allocated space. This will be greater
|
| + than current total if trimming has occurred.
|
| + fsmblks: always zero
|
| + uordblks: current total allocated space (normal or mmapped)
|
| + fordblks: total free space
|
| + keepcost: the maximum number of bytes that could ideally be released
|
| + back to system via malloc_trim. ("ideally" means that
|
| + it ignores page restrictions etc.)
|
| +
|
| + Because these fields are ints, but internal bookkeeping may
|
| + be kept as longs, the reported values may wrap around zero and
|
| + thus be inaccurate.
|
| +*/
|
| +
|
| +struct mallinfo dlmallinfo(void);
|
| +#endif /* NO_MALLINFO */
|
| +
|
| +/*
|
| + independent_calloc(size_t n_elements, size_t element_size, void* chunks[]);
|
| +
|
| + independent_calloc is similar to calloc, but instead of returning a
|
| + single cleared space, it returns an array of pointers to n_elements
|
| + independent elements that can hold contents of size elem_size, each
|
| + of which starts out cleared, and can be independently freed,
|
| + realloc'ed etc. The elements are guaranteed to be adjacently
|
| + allocated (this is not guaranteed to occur with multiple callocs or
|
| + mallocs), which may also improve cache locality in some
|
| + applications.
|
| +
|
| + The "chunks" argument is optional (i.e., may be null, which is
|
| + probably the most typical usage). If it is null, the returned array
|
| + is itself dynamically allocated and should also be freed when it is
|
| + no longer needed. Otherwise, the chunks array must be of at least
|
| + n_elements in length. It is filled in with the pointers to the
|
| + chunks.
|
| +
|
| + In either case, independent_calloc returns this pointer array, or
|
| + null if the allocation failed. If n_elements is zero and "chunks"
|
| + is null, it returns a chunk representing an array with zero elements
|
| + (which should be freed if not wanted).
|
| +
|
| + Each element must be freed when it is no longer needed. This can be
|
| + done all at once using bulk_free.
|
| +
|
| + independent_calloc simplifies and speeds up implementations of many
|
| + kinds of pools. It may also be useful when constructing large data
|
| + structures that initially have a fixed number of fixed-sized nodes,
|
| + but the number is not known at compile time, and some of the nodes
|
| + may later need to be freed. For example:
|
| +
|
| + struct Node { int item; struct Node* next; };
|
| +
|
| + struct Node* build_list() {
|
| + struct Node** pool;
|
| + int n = read_number_of_nodes_needed();
|
| + if (n <= 0) return 0;
|
| + pool = (struct Node**)(independent_calloc(n, sizeof(struct Node), 0);
|
| + if (pool == 0) die();
|
| + // organize into a linked list...
|
| + struct Node* first = pool[0];
|
| + for (i = 0; i < n-1; ++i)
|
| + pool[i]->next = pool[i+1];
|
| + free(pool); // Can now free the array (or not, if it is needed later)
|
| + return first;
|
| + }
|
| +*/
|
| +void** dlindependent_calloc(size_t, size_t, void**);
|
| +
|
| +/*
|
| + independent_comalloc(size_t n_elements, size_t sizes[], void* chunks[]);
|
| +
|
| + independent_comalloc allocates, all at once, a set of n_elements
|
| + chunks with sizes indicated in the "sizes" array. It returns
|
| + an array of pointers to these elements, each of which can be
|
| + independently freed, realloc'ed etc. The elements are guaranteed to
|
| + be adjacently allocated (this is not guaranteed to occur with
|
| + multiple callocs or mallocs), which may also improve cache locality
|
| + in some applications.
|
| +
|
| + The "chunks" argument is optional (i.e., may be null). If it is null
|
| + the returned array is itself dynamically allocated and should also
|
| + be freed when it is no longer needed. Otherwise, the chunks array
|
| + must be of at least n_elements in length. It is filled in with the
|
| + pointers to the chunks.
|
| +
|
| + In either case, independent_comalloc returns this pointer array, or
|
| + null if the allocation failed. If n_elements is zero and chunks is
|
| + null, it returns a chunk representing an array with zero elements
|
| + (which should be freed if not wanted).
|
| +
|
| + Each element must be freed when it is no longer needed. This can be
|
| + done all at once using bulk_free.
|
| +
|
| + independent_comallac differs from independent_calloc in that each
|
| + element may have a different size, and also that it does not
|
| + automatically clear elements.
|
| +
|
| + independent_comalloc can be used to speed up allocation in cases
|
| + where several structs or objects must always be allocated at the
|
| + same time. For example:
|
| +
|
| + struct Head { ... }
|
| + struct Foot { ... }
|
| +
|
| + void send_message(char* msg) {
|
| + int msglen = strlen(msg);
|
| + size_t sizes[3] = { sizeof(struct Head), msglen, sizeof(struct Foot) };
|
| + void* chunks[3];
|
| + if (independent_comalloc(3, sizes, chunks) == 0)
|
| + die();
|
| + struct Head* head = (struct Head*)(chunks[0]);
|
| + char* body = (char*)(chunks[1]);
|
| + struct Foot* foot = (struct Foot*)(chunks[2]);
|
| + // ...
|
| + }
|
| +
|
| + In general though, independent_comalloc is worth using only for
|
| + larger values of n_elements. For small values, you probably won't
|
| + detect enough difference from series of malloc calls to bother.
|
| +
|
| + Overuse of independent_comalloc can increase overall memory usage,
|
| + since it cannot reuse existing noncontiguous small chunks that
|
| + might be available for some of the elements.
|
| +*/
|
| +void** dlindependent_comalloc(size_t, size_t*, void**);
|
| +
|
| +/*
|
| + bulk_free(void* array[], size_t n_elements)
|
| + Frees and clears (sets to null) each non-null pointer in the given
|
| + array. This is likely to be faster than freeing them one-by-one.
|
| + If footers are used, pointers that have been allocated in different
|
| + mspaces are not freed or cleared, and the count of all such pointers
|
| + is returned. For large arrays of pointers with poor locality, it
|
| + may be worthwhile to sort this array before calling bulk_free.
|
| +*/
|
| +size_t dlbulk_free(void**, size_t n_elements);
|
| +
|
| +/*
|
| + pvalloc(size_t n);
|
| + Equivalent to valloc(minimum-page-that-holds(n)), that is,
|
| + round up n to nearest pagesize.
|
| + */
|
| +void* dlpvalloc(size_t);
|
| +
|
| +/*
|
| + malloc_trim(size_t pad);
|
| +
|
| + If possible, gives memory back to the system (via negative arguments
|
| + to sbrk) if there is unused memory at the `high' end of the malloc
|
| + pool or in unused MMAP segments. You can call this after freeing
|
| + large blocks of memory to potentially reduce the system-level memory
|
| + requirements of a program. However, it cannot guarantee to reduce
|
| + memory. Under some allocation patterns, some large free blocks of
|
| + memory will be locked between two used chunks, so they cannot be
|
| + given back to the system.
|
| +
|
| + The `pad' argument to malloc_trim represents the amount of free
|
| + trailing space to leave untrimmed. If this argument is zero, only
|
| + the minimum amount of memory to maintain internal data structures
|
| + will be left. Non-zero arguments can be supplied to maintain enough
|
| + trailing space to service future expected allocations without having
|
| + to re-obtain memory from the system.
|
| +
|
| + Malloc_trim returns 1 if it actually released any memory, else 0.
|
| +*/
|
| +int dlmalloc_trim(size_t);
|
| +
|
| +/*
|
| + malloc_stats();
|
| + Prints on stderr the amount of space obtained from the system (both
|
| + via sbrk and mmap), the maximum amount (which may be more than
|
| + current if malloc_trim and/or munmap got called), and the current
|
| + number of bytes allocated via malloc (or realloc, etc) but not yet
|
| + freed. Note that this is the number of bytes allocated, not the
|
| + number requested. It will be larger than the number requested
|
| + because of alignment and bookkeeping overhead. Because it includes
|
| + alignment wastage as being in use, this figure may be greater than
|
| + zero even when no user-level chunks are allocated.
|
| +
|
| + The reported current and maximum system memory can be inaccurate if
|
| + a program makes other calls to system memory allocation functions
|
| + (normally sbrk) outside of malloc.
|
| +
|
| + malloc_stats prints only the most commonly interesting statistics.
|
| + More information can be obtained by calling mallinfo.
|
| +
|
| + malloc_stats is not compiled if NO_MALLOC_STATS is defined.
|
| +*/
|
| +void dlmalloc_stats(void);
|
| +
|
| +#endif /* !ONLY_MSPACES */
|
| +
|
| +#ifdef __cplusplus
|
| +}; /* end of extern "C" */
|
| +#endif
|
| +
|
| +#endif /* MALLOC_280_H */
|
|
|