Index: third_party/afl/src/libdislocator/libdislocator.so.c |
diff --git a/third_party/afl/src/libdislocator/libdislocator.so.c b/third_party/afl/src/libdislocator/libdislocator.so.c |
new file mode 100644 |
index 0000000000000000000000000000000000000000..1d4648f3ce8ad8b258c7f46e4c8c1413e69347b5 |
--- /dev/null |
+++ b/third_party/afl/src/libdislocator/libdislocator.so.c |
@@ -0,0 +1,257 @@ |
+/* |
+ |
+ american fuzzy lop - dislocator, an abusive allocator |
+ ----------------------------------------------------- |
+ |
+ Written and maintained by Michal Zalewski <lcamtuf@google.com> |
+ |
+ Copyright 2016 Google Inc. All rights reserved. |
+ |
+ Licensed under the Apache License, Version 2.0 (the "License"); |
+ you may not use this file except in compliance with the License. |
+ You may obtain a copy of the License at: |
+ |
+ http://www.apache.org/licenses/LICENSE-2.0 |
+ |
+ This is a companion library that can be used as a drop-in replacement |
+ for the libc allocator in the fuzzed binaries. See README.dislocator for |
+ more info. |
+ |
+ */ |
+ |
+#include <stdio.h> |
+#include <stdlib.h> |
+#include <string.h> |
+#include <limits.h> |
+#include <sys/mman.h> |
+ |
+#include "../config.h" |
+#include "../types.h" |
+ |
+#ifndef PAGE_SIZE |
+# define PAGE_SIZE 4096 |
+#endif /* !PAGE_SIZE */ |
+ |
+#ifndef MAP_ANONYMOUS |
+# define MAP_ANONYMOUS MAP_ANON |
+#endif /* !MAP_ANONYMOUS */ |
+ |
+/* Error / message handling: */ |
+ |
+#define DEBUGF(_x...) do { \ |
+ if (alloc_verbose) { \ |
+ if (++call_depth == 1) { \ |
+ fprintf(stderr, "[AFL] " _x); \ |
+ fprintf(stderr, "\n"); \ |
+ } \ |
+ call_depth--; \ |
+ } \ |
+ } while (0) |
+ |
+#define FATAL(_x...) do { \ |
+ if (++call_depth == 1) { \ |
+ fprintf(stderr, "*** [AFL] " _x); \ |
+ fprintf(stderr, " ***\n"); \ |
+ abort(); \ |
+ } \ |
+ call_depth--; \ |
+ } while (0) |
+ |
+/* Macro to count the number of pages needed to store a buffer: */ |
+ |
+#define PG_COUNT(_l) (((_l) + (PAGE_SIZE - 1)) / PAGE_SIZE) |
+ |
+/* Canary & clobber bytes: */ |
+ |
+#define ALLOC_CANARY 0xAACCAACC |
+#define ALLOC_CLOBBER 0x41 |
+ |
+#define PTR_C(_p) (((u32*)(_p))[-1]) |
+#define PTR_L(_p) (((u32*)(_p))[-2]) |
+ |
+/* Configurable stuff (use AFL_LD_* to set): */ |
+ |
+static u32 max_mem = MAX_ALLOC; /* Max heap usage to permit */ |
+static u8 alloc_verbose, /* Additional debug messages */ |
+ hard_fail; /* abort() when max_mem exceeded? */ |
+ |
+static __thread size_t total_mem; /* Currently allocated mem */ |
+ |
+static __thread u32 call_depth; /* To avoid recursion via fprintf() */ |
+ |
+ |
+/* This is the main alloc function. It allocates one page more than necessary, |
+ sets that tailing page to PROT_NONE, and then increments the return address |
+ so that it is right-aligned to that boundary. Since it always uses mmap(), |
+ the returned memory will be zeroed. */ |
+ |
+static void* __dislocator_alloc(size_t len) { |
+ |
+ void* ret; |
+ |
+ if (total_mem + len > max_mem) { |
+ |
+ if (hard_fail) |
+ FATAL("total allocs exceed %u MB", max_mem / 1024 / 1024); |
+ |
+ DEBUGF("total allocs exceed %u MB, returning NULL", |
+ max_mem / 1024 / 1024); |
+ |
+ return NULL; |
+ |
+ } |
+ |
+ /* We will also store buffer length and a canary below the actual buffer, so |
+ let's add 8 bytes for that. */ |
+ |
+ ret = mmap(NULL, (1 + PG_COUNT(len + 8)) * PAGE_SIZE, PROT_READ | PROT_WRITE, |
+ MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); |
+ |
+ if (ret == (void*)-1) { |
+ |
+ if (hard_fail) FATAL("mmap() failed on alloc (OOM?)"); |
+ |
+ DEBUGF("mmap() failed on alloc (OOM?)"); |
+ |
+ return NULL; |
+ |
+ } |
+ |
+ /* Set PROT_NONE on the last page. */ |
+ |
+ if (mprotect(ret + PG_COUNT(len + 8) * PAGE_SIZE, PAGE_SIZE, PROT_NONE)) |
+ FATAL("mprotect() failed when allocating memory"); |
+ |
+ /* Offset the return pointer so that it's right-aligned to the page |
+ boundary. */ |
+ |
+ ret += PAGE_SIZE * PG_COUNT(len + 8) - len - 8; |
+ |
+ /* Store allocation metadata. */ |
+ |
+ ret += 8; |
+ |
+ PTR_L(ret) = len; |
+ PTR_C(ret) = ALLOC_CANARY; |
+ |
+ total_mem += len; |
+ |
+ return ret; |
+ |
+} |
+ |
+ |
+/* The "user-facing" wrapper for calloc(). This just checks for overflows and |
+ displays debug messages if requested. */ |
+ |
+void* calloc(size_t elem_len, size_t elem_cnt) { |
+ |
+ void* ret; |
+ |
+ size_t len = elem_len * elem_cnt; |
+ |
+ /* Perform some sanity checks to detect obvious issues... */ |
+ |
+ if (elem_cnt && len / elem_cnt != elem_len) |
+ FATAL("calloc(%zu, %zu) would overflow", elem_len, elem_cnt); |
+ |
+ ret = __dislocator_alloc(len); |
+ |
+ DEBUGF("calloc(%zu, %zu) = %p [%zu total]", elem_len, elem_cnt, ret, |
+ total_mem); |
+ |
+ return ret; |
+ |
+} |
+ |
+ |
+/* The wrapper for malloc(). Roughly the same, also clobbers the returned |
+ memory (unlike calloc(), malloc() is not guaranteed to return zeroed |
+ memory). */ |
+ |
+void* malloc(size_t len) { |
+ |
+ void* ret; |
+ |
+ ret = __dislocator_alloc(len); |
+ |
+ DEBUGF("malloc(%zu) = %p [%zu total]", len, ret, total_mem); |
+ |
+ if (ret && len) memset(ret, ALLOC_CLOBBER, len); |
+ |
+ return ret; |
+ |
+} |
+ |
+ |
+/* The wrapper for free(). This simply marks the entire region as PROT_NONE. |
+ If the region is already freed, the code will segfault during the attempt to |
+ read the canary. Not very graceful, but works, right? */ |
+ |
+void free(void* ptr) { |
+ |
+ u32 len; |
+ |
+ DEBUGF("free(%p)", ptr); |
+ |
+ if (!ptr) return; |
+ |
+ if (PTR_C(ptr) != ALLOC_CANARY) FATAL("bad allocator canary on free()"); |
+ |
+ len = PTR_L(ptr); |
+ |
+ total_mem -= len; |
+ |
+ /* Protect everything. Note that the extra page at the end is already |
+ set as PROT_NONE, so we don't need to touch that. */ |
+ |
+ ptr -= PAGE_SIZE * PG_COUNT(len + 8) - len - 8; |
+ |
+ if (mprotect(ptr - 8, PG_COUNT(len + 8) * PAGE_SIZE, PROT_NONE)) |
+ FATAL("mprotect() failed when freeing memory"); |
+ |
+ /* Keep the mapping; this is wasteful, but prevents ptr reuse. */ |
+ |
+} |
+ |
+ |
+/* Realloc is pretty straightforward, too. We forcibly reallocate the buffer, |
+ move data, and then free (aka mprotect()) the original one. */ |
+ |
+void* realloc(void* ptr, size_t len) { |
+ |
+ void* ret; |
+ |
+ ret = malloc(len); |
+ |
+ if (ret && ptr) { |
+ |
+ if (PTR_C(ptr) != ALLOC_CANARY) FATAL("bad allocator canary on realloc()"); |
+ |
+ memcpy(ret, ptr, MIN(len, PTR_L(ptr))); |
+ free(ptr); |
+ |
+ } |
+ |
+ DEBUGF("realloc(%p, %zu) = %p [%zu total]", ptr, len, ret, total_mem); |
+ |
+ return ret; |
+ |
+} |
+ |
+ |
+__attribute__((constructor)) void __dislocator_init(void) { |
+ |
+ u8* tmp = getenv("AFL_LD_LIMIT_MB"); |
+ |
+ if (tmp) { |
+ |
+ max_mem = atoi(tmp) * 1024 * 1024; |
+ if (!max_mem) FATAL("Bad value for AFL_LD_LIMIT_MB"); |
+ |
+ } |
+ |
+ alloc_verbose = !!getenv("AFL_LD_VERBOSE"); |
+ hard_fail = !!getenv("AFL_LD_HARD_FAIL"); |
+ |
+} |