diff options
author | Andrea Fioraldi <andreafioraldi@gmail.com> | 2020-01-30 22:40:09 +0100 |
---|---|---|
committer | Andrea Fioraldi <andreafioraldi@gmail.com> | 2020-01-30 22:40:09 +0100 |
commit | 95a98fb3e805d94a3db82d5da21e9a00a79dbdf2 (patch) | |
tree | 4de4f17f898a4f0632098c9902614217fa2e53eb | |
parent | b8bad5a2273b0cddd0244a7f37c20150a08af475 (diff) | |
download | afl++-95a98fb3e805d94a3db82d5da21e9a00a79dbdf2.tar.gz |
cmplog runtime for LLVM
-rw-r--r-- | include/afl-fuzz.h | 14 | ||||
-rw-r--r-- | include/cmplog.h | 30 | ||||
-rw-r--r-- | include/sharedmem.h | 1 | ||||
-rw-r--r-- | include/types.h | 14 | ||||
-rw-r--r-- | llvm_mode/Makefile | 17 | ||||
-rw-r--r-- | llvm_mode/README.cmplog.md | 42 | ||||
-rw-r--r-- | llvm_mode/afl-clang-fast.c | 44 | ||||
-rw-r--r-- | llvm_mode/afl-llvm-cmplog-rt.o.c | 412 | ||||
-rw-r--r-- | llvm_mode/afl-llvm-cmplog.rt.o.c | 422 | ||||
-rw-r--r-- | llvm_mode/afl-llvm-rt.o.c | 123 | ||||
-rw-r--r-- | src/afl-fuzz-cmplog.c | 385 | ||||
-rw-r--r-- | src/afl-fuzz-globals.c | 3 | ||||
-rw-r--r-- | src/afl-fuzz-redqueen.c | 373 |
13 files changed, 1370 insertions, 510 deletions
diff --git a/include/afl-fuzz.h b/include/afl-fuzz.h index ce418931..cd53c703 100644 --- a/include/afl-fuzz.h +++ b/include/afl-fuzz.h @@ -452,6 +452,11 @@ extern u32 a_extras_cnt; /* Total number of tokens available */ u8* (*post_handler)(u8* buf, u32* len); +/* CmpLog */ + +extern u8* cmplog_binary; +extern s32 cmplog_forksrv_pid; + /* hooks for the custom mutator function */ /** * Perform custom mutations on a given input @@ -647,12 +652,13 @@ char** get_qemu_argv(u8*, char**, int); char** get_wine_argv(u8*, char**, int); void save_cmdline(u32, char**); -/* RedQueen */ - -extern u8* cmplog_binary; -extern s32 cmplog_forksrv_pid; +/* CmpLog */ void init_cmplog_forkserver(char** argv); +u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len); + +/* RedQueen */ + u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len, u32 exec_cksum); diff --git a/include/cmplog.h b/include/cmplog.h index d5947226..c02650ee 100644 --- a/include/cmplog.h +++ b/include/cmplog.h @@ -1,5 +1,31 @@ -#ifndef _AFL_REDQUEEN_H -#define _AFL_REDQUEEN_H +/* + american fuzzy lop++ - cmplog header + ------------------------------------ + + Originally written by Michal Zalewski + + Forkserver design by Jann Horn <jannhorn@googlemail.com> + + Now maintained by by Marc Heuse <mh@mh-sec.de>, + Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and + Andrea Fioraldi <andreafioraldi@gmail.com> + + Copyright 2016, 2017 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: + + http://www.apache.org/licenses/LICENSE-2.0 + + Shared code to handle the shared memory. This is used by the fuzzer + as well the other components like afl-tmin, afl-showmap, etc... + + */ + +#ifndef _AFL_CMPLOG_H +#define _AFL_CMPLOG_H #include "config.h" diff --git a/include/sharedmem.h b/include/sharedmem.h index 25c7336d..3540386d 100644 --- a/include/sharedmem.h +++ b/include/sharedmem.h @@ -31,6 +31,7 @@ void setup_shm(unsigned char dumb_mode); void remove_shm(void); extern int cmplog_mode; +extern struct cmp_map* cmp_map; #endif diff --git a/include/types.h b/include/types.h index eba47be7..d5be5920 100644 --- a/include/types.h +++ b/include/types.h @@ -78,6 +78,20 @@ typedef int64_t s64; \ }) +#define SWAP64(_x) \ + ({ \ + \ + u64 _ret = (_x); \ + _ret = \ + (_ret & 0x00000000FFFFFFFF) << 32 | (_ret & 0xFFFFFFFF00000000) >> 32; \ + _ret = \ + (_ret & 0x0000FFFF0000FFFF) << 16 | (_ret & 0xFFFF0000FFFF0000) >> 16; \ + _ret = \ + (_ret & 0x00FF00FF00FF00FF) << 8 | (_ret & 0xFF00FF00FF00FF00) >> 8; \ + _ret; \ + \ + }) + #ifdef AFL_LLVM_PASS #if defined(__linux__) #define AFL_SR(s) (srandom(s)) diff --git a/llvm_mode/Makefile b/llvm_mode/Makefile index a455dc8a..9ee6fc8b 100644 --- a/llvm_mode/Makefile +++ b/llvm_mode/Makefile @@ -19,8 +19,6 @@ # For Heiko: #TEST_MMAP=1 -AFL_TRACE_PC=1 - PREFIX ?= /usr/local HELPER_PATH = $(PREFIX)/lib/afl BIN_PATH = $(PREFIX)/bin @@ -133,9 +131,9 @@ ifeq "$(TEST_MMAP)" "1" endif ifndef AFL_TRACE_PC - PROGS = ../afl-clang-fast ../afl-llvm-pass.so ../libLLVMInsTrim.so ../afl-llvm-rt.o ../afl-llvm-rt-32.o ../afl-llvm-rt-64.o ../compare-transform-pass.so ../split-compares-pass.so ../split-switches-pass.so + PROGS = ../afl-clang-fast ../afl-llvm-cmplog-rt.o ../afl-llvm-cmplog-rt-32.o ../afl-llvm-cmplog-rt-64.o ../afl-llvm-pass.so ../libLLVMInsTrim.so ../afl-llvm-rt.o ../afl-llvm-rt-32.o ../afl-llvm-rt-64.o ../compare-transform-pass.so ../split-compares-pass.so ../split-switches-pass.so else - PROGS = ../afl-clang-fast ../afl-llvm-rt.o ../afl-llvm-rt-32.o ../afl-llvm-rt-64.o ../compare-transform-pass.so ../split-compares-pass.so ../split-switches-pass.so + PROGS = ../afl-clang-fast ../afl-llvm-cmplog-rt.o ../afl-llvm-cmplog-rt-32.o ../afl-llvm-cmplog-rt-64.o ../afl-llvm-rt.o ../afl-llvm-rt-32.o ../afl-llvm-rt-64.o ../compare-transform-pass.so ../split-compares-pass.so ../split-switches-pass.so endif ifneq "$(CLANGVER)" "$(LLVMVER)" @@ -228,6 +226,17 @@ endif @printf "[*] Building 64-bit variant of the runtime (-m64)... " @$(CC) $(CFLAGS) -m64 -fPIC -c $< -o $@ 2>/dev/null; if [ "$$?" = "0" ]; then echo "success!"; else echo "failed (that's fine)"; fi +../afl-llvm-cmplog-rt.o: afl-llvm-cmplog-rt.o.c | test_deps + $(CC) $(CFLAGS) -fPIC -c $< -o $@ + +../afl-llvm-cmplog-rt-32.o: afl-llvm-cmplog-rt.o.c | test_deps + @printf "[*] Building 32-bit variant of the CmpLog runtime (-m32)... " + @$(CC) $(CFLAGS) -m32 -fPIC -c $< -o $@ 2>/dev/null; if [ "$$?" = "0" ]; then echo "success!"; else echo "failed (that's fine)"; fi + +../afl-llvm-cmplog-rt-64.o: afl-llvm-cmplog-rt.o.c | test_deps + @printf "[*] Building 64-bit variant of the CmpLog runtime (-m64)... " + @$(CC) $(CFLAGS) -m64 -fPIC -c $< -o $@ 2>/dev/null; if [ "$$?" = "0" ]; then echo "success!"; else echo "failed (that's fine)"; fi + test_build: $(PROGS) @echo "[*] Testing the CC wrapper and instrumentation output..." unset AFL_USE_ASAN AFL_USE_MSAN AFL_INST_RATIO; AFL_QUIET=1 AFL_PATH=. AFL_CC=$(CC) AFL_LLVM_LAF_SPLIT_SWITCHES=1 AFL_LLVM_LAF_TRANSFORM_COMPARES=1 AFL_LLVM_LAF_SPLIT_COMPARES=1 ../afl-clang-fast $(CFLAGS) ../test-instr.c -o test-instr $(LDFLAGS) diff --git a/llvm_mode/README.cmplog.md b/llvm_mode/README.cmplog.md new file mode 100644 index 00000000..a0e838ad --- /dev/null +++ b/llvm_mode/README.cmplog.md @@ -0,0 +1,42 @@ +# CmpLog instrumentation + +The CmpLog instrumentation enables the logging of the comparisons operands in a +shared memory. + +These values can be used by variuous mutator built on top of it. +At the moment we support the RedQueen mutator (input-2-state instructions only). + +## Build + +Tou use CmpLog, you have to build two versions of the instrumented target +program. + +The first, using the regular AFL++ instrumentation. + +The second, the CmpLog binary, setting AFL_LLVM_CMPLOG during the compilation. + +For example: + +``` +./configure --cc=~/path/to/afl-clang-fast +make +cp ./program ./program.afl +make clean +export AFL_LLVM_CMPLOG=1 +./configure --cc=~/path/to/afl-clang-fast +make +cp ./program ./program.cmplog +``` + +## Use + +AFL++ have the -c option that can be used to specify a CmpLog binary (the second +built). + +For example: + +``` +afl-fuzz -i input -o output -c ./program.cmplog -m none -- ./program.afl @@ +``` + +Be careful to use -m none because CmpLog maps a lot of pages. diff --git a/llvm_mode/afl-clang-fast.c b/llvm_mode/afl-clang-fast.c index 939546d7..ddaa2388 100644 --- a/llvm_mode/afl-clang-fast.c +++ b/llvm_mode/afl-clang-fast.c @@ -40,6 +40,7 @@ static u8* obj_path; /* Path to runtime libraries */ static u8** cc_params; /* Parameters passed to the real CC */ static u32 cc_par_cnt = 1; /* Param count, including argv0 */ static u8 llvm_fullpath[PATH_MAX]; +static u8 cmplog_mode; /* Try to find the runtime libraries. If that fails, abort. */ @@ -196,16 +197,20 @@ static void edit_params(u32 argc, char** argv) { // /laf -#ifdef USE_TRACE_PC - if (getenv("AFL_CMPLOG")) + if (cmplog_mode) { + cc_params[cc_par_cnt++] = "-fsanitize-coverage=trace-pc-guard,trace-cmp"; - else - cc_params[cc_par_cnt++] = - "-fsanitize-coverage=trace-pc-guard"; // edge coverage by default - // cc_params[cc_par_cnt++] = "-mllvm"; - // cc_params[cc_par_cnt++] = - // "-fsanitize-coverage=trace-cmp,trace-div,trace-gep"; - // cc_params[cc_par_cnt++] = "-sanitizer-coverage-block-threshold=0"; + + } else { + +#ifdef USE_TRACE_PC + + cc_params[cc_par_cnt++] = + "-fsanitize-coverage=trace-pc-guard"; // edge coverage by default + // cc_params[cc_par_cnt++] = "-mllvm"; + // cc_params[cc_par_cnt++] = + // "-fsanitize-coverage=trace-cmp,trace-div,trace-gep"; + // cc_params[cc_par_cnt++] = "-sanitizer-coverage-block-threshold=0"; #else cc_params[cc_par_cnt++] = "-Xclang"; cc_params[cc_par_cnt++] = "-load"; @@ -216,6 +221,8 @@ static void edit_params(u32 argc, char** argv) { cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-pass.so", obj_path); #endif /* ^USE_TRACE_PC */ + } + cc_params[cc_par_cnt++] = "-Qunused-arguments"; /* Detect stray -v calls from ./configure scripts. */ @@ -380,11 +387,17 @@ static void edit_params(u32 argc, char** argv) { switch (bit_mode) { case 0: - cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt.o", obj_path); + if (cmplog_mode) + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-cmplog-rt.o", obj_path); + else + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt.o", obj_path); break; case 32: - cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt-32.o", obj_path); + if (cmplog_mode) + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-cmplog-rt-32.o", obj_path); + else + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt-32.o", obj_path); if (access(cc_params[cc_par_cnt - 1], R_OK)) FATAL("-m32 is not supported by your compiler"); @@ -392,7 +405,10 @@ static void edit_params(u32 argc, char** argv) { break; case 64: - cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt-64.o", obj_path); + if (cmplog_mode) + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-cmplog-rt-64.o", obj_path); + else + cc_params[cc_par_cnt++] = alloc_printf("%s/afl-llvm-rt-64.o", obj_path); if (access(cc_params[cc_par_cnt - 1], R_OK)) FATAL("-m64 is not supported by your compiler"); @@ -463,6 +479,10 @@ int main(int argc, char** argv) { #endif /* ^USE_TRACE_PC */ } + + cmplog_mode = getenv("AFL_CMPLOG") || getenv("AFL_LLVM_CMPLOG"); + if (cmplog_mode) + printf("CmpLog mode by <andreafioraldi@gmail.com>\n"); #ifndef __ANDROID__ find_obj(argv[0]); diff --git a/llvm_mode/afl-llvm-cmplog-rt.o.c b/llvm_mode/afl-llvm-cmplog-rt.o.c new file mode 100644 index 00000000..7a513c0d --- /dev/null +++ b/llvm_mode/afl-llvm-cmplog-rt.o.c @@ -0,0 +1,412 @@ +/* + american fuzzy lop++ - LLVM instrumentation bootstrap + --------------------------------------------------- + + Written by Laszlo Szekeres <lszekeres@google.com> and + Michal Zalewski + + LLVM integration design comes from Laszlo Szekeres. + + Copyright 2015, 2016 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: + + http://www.apache.org/licenses/LICENSE-2.0 + + This code is the rewrite of afl-as.h's main_payload. + +*/ + +#ifdef __ANDROID__ +#include "android-ashmem.h" +#endif +#include "config.h" +#include "types.h" +#include "cmplog.h" + +#include <stdio.h> +#include <stdlib.h> +#include <signal.h> +#include <unistd.h> +#include <string.h> +#include <assert.h> +#include <stdint.h> + +#include <sys/mman.h> +#include <sys/shm.h> +#include <sys/wait.h> +#include <sys/types.h> + +/* This is a somewhat ugly hack for the experimental 'trace-pc-guard' mode. + Basically, we need to make sure that the forkserver is initialized after + the LLVM-generated runtime initialization pass, not before. */ + +#ifdef USE_TRACE_PC +#define CONST_PRIO 5 +#else +#define CONST_PRIO 0 +#endif /* ^USE_TRACE_PC */ + +#include <sys/mman.h> +#include <fcntl.h> + +/* Globals needed by the injected instrumentation. The __afl_area_initial region + is used for instrumentation output before __afl_map_shm() has a chance to + run. It will end up as .comm, so it shouldn't be too wasteful. */ + +// In CmpLog, the only usage of __afl_area_ptr is to report errors +u8* __afl_area_ptr; + +struct cmp_map* __afl_cmp_map; +__thread u32 __afl_cmp_counter; + +/* Running in persistent mode? */ + +static u8 is_persistent; + +/* SHM setup. */ + +static void __afl_map_shm(void) { + + u8* id_str = getenv(SHM_ENV_VAR); + + /* If we're running under AFL, attach to the appropriate region, replacing the + early-stage __afl_area_initial region that is needed to allow some really + hacky .init code to work correctly in projects such as OpenSSL. */ + + if (id_str) { + +#ifdef USEMMAP + const char* shm_file_path = id_str; + int shm_fd = -1; + unsigned char* shm_base = NULL; + + /* create the shared memory segment as if it was a file */ + shm_fd = shm_open(shm_file_path, O_RDWR, 0600); + if (shm_fd == -1) { + + printf("shm_open() failed\n"); + exit(1); + + } + + /* map the shared memory segment to the address space of the process */ + shm_base = mmap(0, MAP_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, shm_fd, 0); + if (shm_base == MAP_FAILED) { + + close(shm_fd); + shm_fd = -1; + + printf("mmap() failed\n"); + exit(2); + + } + + __afl_area_ptr = shm_base; +#else + u32 shm_id = atoi(id_str); + + __afl_area_ptr = shmat(shm_id, NULL, 0); +#endif + + /* Whooooops. */ + + if (__afl_area_ptr == (void*)-1) _exit(1); + + /* Write something into the bitmap so that even with low AFL_INST_RATIO, + our parent doesn't give up on us. */ + + __afl_area_ptr[0] = 1; + + } + + id_str = getenv(CMPLOG_SHM_ENV_VAR); + + if (id_str) { + + u32 shm_id = atoi(id_str); + + __afl_cmp_map = shmat(shm_id, NULL, 0); + + if (__afl_cmp_map == (void*)-1) _exit(1); + + } + +} + +/* Fork server logic. */ + +static void __afl_start_forkserver(void) { + + static u8 tmp[4]; + s32 child_pid; + + u8 child_stopped = 0; + + void (*old_sigchld_handler)(int) = 0; // = signal(SIGCHLD, SIG_DFL); + + /* Phone home and tell the parent that we're OK. If parent isn't there, + assume we're not running in forkserver mode and just execute program. */ + + if (write(FORKSRV_FD + 1, tmp, 4) != 4) return; + + while (1) { + + u32 was_killed; + int status; + + /* Wait for parent by reading from the pipe. Abort if read fails. */ + + if (read(FORKSRV_FD, &was_killed, 4) != 4) _exit(1); + + /* If we stopped the child in persistent mode, but there was a race + condition and afl-fuzz already issued SIGKILL, write off the old + process. */ + + if (child_stopped && was_killed) { + + child_stopped = 0; + if (waitpid(child_pid, &status, 0) < 0) _exit(1); + + } + + if (!child_stopped) { + + /* Once woken up, create a clone of our process. */ + + child_pid = fork(); + if (child_pid < 0) _exit(1); + + /* In child process: close fds, resume execution. */ + + if (!child_pid) { + + signal(SIGCHLD, old_sigchld_handler); + + close(FORKSRV_FD); + close(FORKSRV_FD + 1); + return; + + } + + } else { + + /* Special handling for persistent mode: if the child is alive but + currently stopped, simply restart it with SIGCONT. */ + + kill(child_pid, SIGCONT); + child_stopped = 0; + + } + + /* In parent process: write PID to pipe, then wait for child. */ + + if (write(FORKSRV_FD + 1, &child_pid, 4) != 4) _exit(1); + + if (waitpid(child_pid, &status, is_persistent ? WUNTRACED : 0) < 0) + _exit(1); + + /* In persistent mode, the child stops itself with SIGSTOP to indicate + a successful run. In this case, we want to wake it up without forking + again. */ + + if (WIFSTOPPED(status)) child_stopped = 1; + + /* Relay wait status to pipe, then loop back. */ + + if (write(FORKSRV_FD + 1, &status, 4) != 4) _exit(1); + + } + +} + +/* A simplified persistent mode handler, used as explained in README.llvm. */ + +int __afl_persistent_loop(unsigned int max_cnt) { + + static u8 first_pass = 1; + static u32 cycle_cnt; + + if (first_pass) { + + /* Make sure that every iteration of __AFL_LOOP() starts with a clean slate. + On subsequent calls, the parent will take care of that, but on the first + iteration, it's our job to erase any trace of whatever happened + before the loop. */ + + if (is_persistent) { + + // memset(__afl_area_ptr, 0, MAP_SIZE); + __afl_area_ptr[0] = 1; + + } + + cycle_cnt = max_cnt; + first_pass = 0; + return 1; + + } + + if (is_persistent) { + + if (--cycle_cnt) { + + raise(SIGSTOP); + + __afl_area_ptr[0] = 1; + + return 1; + + } else { + + /* When exiting __AFL_LOOP(), make sure that the subsequent code that + follows the loop is not traced. We do that by pivoting back to the + dummy output region. */ + + // __afl_area_ptr = __afl_area_initial; + + } + + } + + return 0; + +} + +/* This one can be called from user code when deferred forkserver mode + is enabled. */ + +void __afl_manual_init(void) { + + static u8 init_done; + + if (!init_done) { + + __afl_map_shm(); + __afl_start_forkserver(); + init_done = 1; + + } + +} + +/* Proper initialization routine. */ + +__attribute__((constructor(CONST_PRIO))) void __afl_auto_init(void) { + + is_persistent = !!getenv(PERSIST_ENV_VAR); + + if (getenv(DEFER_ENV_VAR)) return; + + __afl_manual_init(); + +} + +///// CmpLog instrumentation + +void __sanitizer_cov_trace_cmp1(uint8_t Arg1, uint8_t Arg2) { + + return; + +} + +void __sanitizer_cov_trace_cmp2(uint16_t Arg1, uint16_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + // if (!__afl_cmp_map->headers[k].cnt) + // __afl_cmp_map->headers[k].cnt = __afl_cmp_counter++; + + __afl_cmp_map->headers[k].shape = 1; + //__afl_cmp_map->headers[k].type = CMP_TYPE_INS; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +void __sanitizer_cov_trace_cmp4(uint32_t Arg1, uint32_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 3; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +void __sanitizer_cov_trace_cmp8(uint64_t Arg1, uint64_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 7; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +#if defined(__APPLE__) +#pragma weak __sanitizer_cov_trace_const_cmp1 = __sanitizer_cov_trace_cmp1 +#pragma weak __sanitizer_cov_trace_const_cmp2 = __sanitizer_cov_trace_cmp2 +#pragma weak __sanitizer_cov_trace_const_cmp4 = __sanitizer_cov_trace_cmp4 +#pragma weak __sanitizer_cov_trace_const_cmp8 = __sanitizer_cov_trace_cmp8 +#else +void __sanitizer_cov_trace_const_cmp1(uint8_t Arg1, uint8_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp1"))); +void __sanitizer_cov_trace_const_cmp2(uint16_t Arg1, uint16_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp2"))); +void __sanitizer_cov_trace_const_cmp4(uint32_t Arg1, uint32_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp4"))); +void __sanitizer_cov_trace_const_cmp8(uint64_t Arg1, uint64_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp8"))); +#endif /* defined(__APPLE__) */ + +void __sanitizer_cov_trace_switch(uint64_t Val, uint64_t* Cases) { + + for (uint64_t i = 0; i < Cases[0]; i++) { + + uintptr_t k = (uintptr_t)__builtin_return_address(0) + i; + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 7; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Val; + __afl_cmp_map->log[k][hits].v1 = Cases[i + 2]; + + } + +} + diff --git a/llvm_mode/afl-llvm-cmplog.rt.o.c b/llvm_mode/afl-llvm-cmplog.rt.o.c new file mode 100644 index 00000000..ef944d4f --- /dev/null +++ b/llvm_mode/afl-llvm-cmplog.rt.o.c @@ -0,0 +1,422 @@ +/* + american fuzzy lop++ - LLVM instrumentation bootstrap + --------------------------------------------------- + + Written by Laszlo Szekeres <lszekeres@google.com> and + Michal Zalewski + + LLVM integration design comes from Laszlo Szekeres. + + Copyright 2015, 2016 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: + + http://www.apache.org/licenses/LICENSE-2.0 + + This code is the rewrite of afl-as.h's main_payload. + +*/ + +#ifdef __ANDROID__ +#include "android-ashmem.h" +#endif +#include "config.h" +#include "types.h" +#include "cmplog.h" + +#include <stdio.h> +#include <stdlib.h> +#include <signal.h> +#include <unistd.h> +#include <string.h> +#include <assert.h> +#include <stdint.h> + +#include <sys/mman.h> +#include <sys/shm.h> +#include <sys/wait.h> +#include <sys/types.h> + +/* This is a somewhat ugly hack for the experimental 'trace-pc-guard' mode. + Basically, we need to make sure that the forkserver is initialized after + the LLVM-generated runtime initialization pass, not before. */ + +#ifdef USE_TRACE_PC +#define CONST_PRIO 5 +#else +#define CONST_PRIO 0 +#endif /* ^USE_TRACE_PC */ + +#include <sys/mman.h> +#include <fcntl.h> + +/* Globals needed by the injected instrumentation. The __afl_area_initial region + is used for instrumentation output before __afl_map_shm() has a chance to + run. It will end up as .comm, so it shouldn't be too wasteful. */ + +// In CmpLog, the only usage of __afl_area_ptr is to report errors +u8* __afl_area_ptr; + +struct cmp_map* __afl_cmp_map; +__thread u32 __afl_cmp_counter; + +/* Running in persistent mode? */ + +static u8 is_persistent; + +/* SHM setup. */ + +static void __afl_map_shm(void) { + + u8* id_str = getenv(SHM_ENV_VAR); + + /* If we're running under AFL, attach to the appropriate region, replacing the + early-stage __afl_area_initial region that is needed to allow some really + hacky .init code to work correctly in projects such as OpenSSL. */ + + if (id_str) { + +#ifdef USEMMAP + const char* shm_file_path = id_str; + int shm_fd = -1; + unsigned char* shm_base = NULL; + + /* create the shared memory segment as if it was a file */ + shm_fd = shm_open(shm_file_path, O_RDWR, 0600); + if (shm_fd == -1) { + + printf("shm_open() failed\n"); + exit(1); + + } + + /* map the shared memory segment to the address space of the process */ + shm_base = mmap(0, MAP_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED, shm_fd, 0); + if (shm_base == MAP_FAILED) { + + close(shm_fd); + shm_fd = -1; + + printf("mmap() failed\n"); + exit(2); + + } + + __afl_area_ptr = shm_base; +#else + u32 shm_id = atoi(id_str); + + __afl_area_ptr = shmat(shm_id, NULL, 0); +#endif + + /* Whooooops. */ + + if (__afl_area_ptr == (void*)-1) _exit(1); + + /* Write something into the bitmap so that even with low AFL_INST_RATIO, + our parent doesn't give up on us. */ + + __afl_area_ptr[0] = 1; + + } + + id_str = getenv(CMPLOG_SHM_ENV_VAR); + + if (id_str) { + + u32 shm_id = atoi(id_str); + + __afl_cmp_map = shmat(shm_id, NULL, 0); + + if (__afl_cmp_map == (void*)-1) _exit(1); + + } + +} + +/* Fork server logic. */ + +static void __afl_start_forkserver(void) { + + static u8 tmp[4]; + s32 child_pid; + + u8 child_stopped = 0; + + void (*old_sigchld_handler)(int) = 0; // = signal(SIGCHLD, SIG_DFL); + + /* Phone home and tell the parent that we're OK. If parent isn't there, + assume we're not running in forkserver mode and just execute program. */ + + if (write(FORKSRV_FD + 1, tmp, 4) != 4) return; + + while (1) { + + u32 was_killed; + int status; + + /* Wait for parent by reading from the pipe. Abort if read fails. */ + + if (read(FORKSRV_FD, &was_killed, 4) != 4) _exit(1); + + /* If we stopped the child in persistent mode, but there was a race + condition and afl-fuzz already issued SIGKILL, write off the old + process. */ + + if (child_stopped && was_killed) { + + child_stopped = 0; + if (waitpid(child_pid, &status, 0) < 0) _exit(1); + + } + + if (!child_stopped) { + + /* Once woken up, create a clone of our process. */ + + child_pid = fork(); + if (child_pid < 0) _exit(1); + + /* In child process: close fds, resume execution. */ + + if (!child_pid) { + + signal(SIGCHLD, old_sigchld_handler); + + close(FORKSRV_FD); + close(FORKSRV_FD + 1); + return; + + } + + } else { + + /* Special handling for persistent mode: if the child is alive but + currently stopped, simply restart it with SIGCONT. */ + + kill(child_pid, SIGCONT); + child_stopped = 0; + + } + + /* In parent process: write PID to pipe, then wait for child. */ + + if (write(FORKSRV_FD + 1, &child_pid, 4) != 4) _exit(1); + + if (waitpid(child_pid, &status, is_persistent ? WUNTRACED : 0) < 0) + _exit(1); + + /* In persistent mode, the child stops itself with SIGSTOP to indicate + a successful run. In this case, we want to wake it up without forking + again. */ + + if (WIFSTOPPED(status)) child_stopped = 1; + + /* Relay wait status to pipe, then loop back. */ + + if (write(FORKSRV_FD + 1, &status, 4) != 4) _exit(1); + + } + +} + +/* A simplified persistent mode handler, used as explained in README.llvm. */ + +int __afl_persistent_loop(unsigned int max_cnt) { + + static u8 first_pass = 1; + static u32 cycle_cnt; + + if (first_pass) { + + /* Make sure that every iteration of __AFL_LOOP() starts with a clean slate. + On subsequent calls, the parent will take care of that, but on the first + iteration, it's our job to erase any trace of whatever happened + before the loop. */ + + if (is_persistent) { + + // memset(__afl_area_ptr, 0, MAP_SIZE); + __afl_area_ptr[0] = 1; + + } + + cycle_cnt = max_cnt; + first_pass = 0; + return 1; + + } + + if (is_persistent) { + + if (--cycle_cnt) { + + raise(SIGSTOP); + + __afl_area_ptr[0] = 1; + + return 1; + + } else { + + /* When exiting __AFL_LOOP(), make sure that the subsequent code that + follows the loop is not traced. We do that by pivoting back to the + dummy output region. */ + + // __afl_area_ptr = __afl_area_initial; + + } + + } + + return 0; + +} + +/* This one can be called from user code when deferred forkserver mode + is enabled. */ + +void __afl_manual_init(void) { + + static u8 init_done; + + if (!init_done) { + + __afl_map_shm(); + __afl_start_forkserver(); + init_done = 1; + + } + +} + +/* Proper initialization routine. */ + +__attribute__((constructor(CONST_PRIO))) void __afl_auto_init(void) { + + is_persistent = !!getenv(PERSIST_ENV_VAR); + + if (getenv(DEFER_ENV_VAR)) return; + + __afl_manual_init(); + +} + +///// CmpLog instrumentation + +void __sanitizer_cov_trace_cmp1(uint8_t Arg1, uint8_t Arg2) { + + return; + +} + +void __sanitizer_cov_trace_cmp2(uint16_t Arg1, uint16_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + // if (!__afl_cmp_map->headers[k].cnt) + // __afl_cmp_map->headers[k].cnt = __afl_cmp_counter++; + + __afl_cmp_map->headers[k].shape = 1; + //__afl_cmp_map->headers[k].type = CMP_TYPE_INS; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +void __sanitizer_cov_trace_cmp4(uint32_t Arg1, uint32_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 3; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +void __sanitizer_cov_trace_cmp8(uint64_t Arg1, uint64_t Arg2) { + + if (!__afl_cmp_map) return; + + uintptr_t k = (uintptr_t)__builtin_return_address(0); + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 7; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Arg1; + __afl_cmp_map->log[k][hits].v1 = Arg2; + +} + +#if defined(__APPLE__) +#pragma weak __sanitizer_cov_trace_const_cmp1 = __sanitizer_cov_trace_cmp1 +#pragma weak __sanitizer_cov_trace_const_cmp2 = __sanitizer_cov_trace_cmp2 +#pragma weak __sanitizer_cov_trace_const_cmp4 = __sanitizer_cov_trace_cmp4 +#pragma weak __sanitizer_cov_trace_const_cmp8 = __sanitizer_cov_trace_cmp8 +#else +void __sanitizer_cov_trace_const_cmp1(uint8_t Arg1, uint8_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp1"))); +void __sanitizer_cov_trace_const_cmp2(uint16_t Arg1, uint16_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp2"))); +void __sanitizer_cov_trace_const_cmp4(uint32_t Arg1, uint32_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp4"))); +void __sanitizer_cov_trace_const_cmp8(uint64_t Arg1, uint64_t Arg2) + __attribute__((alias("__sanitizer_cov_trace_cmp8"))); +#endif /* defined(__APPLE__) */ + +void __sanitizer_cov_trace_switch(uint64_t Val, uint64_t* Cases) { + + for (uint64_t i = 0; i < Cases[0]; i++) { + + uintptr_t k = (uintptr_t)__builtin_return_address(0) + i; + k = (k >> 4) ^ (k << 8); + k &= CMP_MAP_W - 1; + + u32 hits = __afl_cmp_map->headers[k].hits; + __afl_cmp_map->headers[k].hits = hits + 1; + + __afl_cmp_map->headers[k].shape = 7; + + hits &= CMP_MAP_H - 1; + __afl_cmp_map->log[k][hits].v0 = Val; + __afl_cmp_map->log[k][hits].v1 = Cases[i + 2]; + + } + +} + + +void __sanitizer_cov_trace_pc_guard(uint32_t* guard) { +} + +void __sanitizer_cov_trace_pc_guard_init(uint32_t* start, uint32_t* stop) { +} + +//// Library functions hooks + +// TODO diff --git a/llvm_mode/afl-llvm-rt.o.c b/llvm_mode/afl-llvm-rt.o.c index c3d1ba7d..9632844f 100644 --- a/llvm_mode/afl-llvm-rt.o.c +++ b/llvm_mode/afl-llvm-rt.o.c @@ -66,9 +66,6 @@ u32 __afl_prev_loc; __thread u32 __afl_prev_loc; #endif -struct cmp_map* __afl_cmp_map; -__thread u32 __afl_cmp_counter; - /* Running in persistent mode? */ static u8 is_persistent; @@ -129,27 +126,6 @@ static void __afl_map_shm(void) { } - if (getenv("__AFL_CMPLOG_MODE__")) { - - id_str = getenv(CMPLOG_SHM_ENV_VAR); - - if (id_str) { - - u32 shm_id = atoi(id_str); - - __afl_cmp_map = shmat(shm_id, NULL, 0); - - if (__afl_cmp_map == (void*)-1) _exit(1); - - } - - } else if (getenv("AFL_CMPLOG")) { - - // during compilation, do this to avoid segfault - __afl_cmp_map = calloc(sizeof(struct cmp_map), 1); - - } - } /* Fork server logic. */ @@ -322,105 +298,6 @@ __attribute__((constructor(CONST_PRIO))) void __afl_auto_init(void) { } -///// CmpLog instrumentation - -void __sanitizer_cov_trace_cmp1(uint8_t Arg1, uint8_t Arg2) { - - return; - -} - -void __sanitizer_cov_trace_cmp2(uint16_t Arg1, uint16_t Arg2) { - - uintptr_t k = (uintptr_t)__builtin_return_address(0); - k = (k >> 4) ^ (k << 8); - k &= CMP_MAP_W - 1; - - u32 hits = __afl_cmp_map->headers[k].hits; - __afl_cmp_map->headers[k].hits = hits + 1; - // if (!__afl_cmp_map->headers[k].cnt) - // __afl_cmp_map->headers[k].cnt = __afl_cmp_counter++; - - __afl_cmp_map->headers[k].shape = 1; - //__afl_cmp_map->headers[k].type = CMP_TYPE_INS; - - hits &= CMP_MAP_H - 1; - __afl_cmp_map->log[k][hits].v0 = Arg1; - __afl_cmp_map->log[k][hits].v1 = Arg2; - -} - -void __sanitizer_cov_trace_cmp4(uint32_t Arg1, uint32_t Arg2) { - - uintptr_t k = (uintptr_t)__builtin_return_address(0); - k = (k >> 4) ^ (k << 8); - k &= CMP_MAP_W - 1; - - u32 hits = __afl_cmp_map->headers[k].hits; - __afl_cmp_map->headers[k].hits = hits + 1; - - __afl_cmp_map->headers[k].shape = 3; - - hits &= CMP_MAP_H - 1; - __afl_cmp_map->log[k][hits].v0 = Arg1; - __afl_cmp_map->log[k][hits].v1 = Arg2; - -} - -void __sanitizer_cov_trace_cmp8(uint64_t Arg1, uint64_t Arg2) { - - uintptr_t k = (uintptr_t)__builtin_return_address(0); - k = (k >> 4) ^ (k << 8); - k &= CMP_MAP_W - 1; - - u32 hits = __afl_cmp_map->headers[k].hits; - __afl_cmp_map->headers[k].hits = hits + 1; - - __afl_cmp_map->headers[k].shape = 7; - - hits &= CMP_MAP_H - 1; - __afl_cmp_map->log[k][hits].v0 = Arg1; - __afl_cmp_map->log[k][hits].v1 = Arg2; - -} - -#if defined(__APPLE__) -#pragma weak __sanitizer_cov_trace_const_cmp1 = __sanitizer_cov_trace_cmp1 -#pragma weak __sanitizer_cov_trace_const_cmp2 = __sanitizer_cov_trace_cmp2 -#pragma weak __sanitizer_cov_trace_const_cmp4 = __sanitizer_cov_trace_cmp4 -#pragma weak __sanitizer_cov_trace_const_cmp8 = __sanitizer_cov_trace_cmp8 -#else -void __sanitizer_cov_trace_const_cmp1(uint8_t Arg1, uint8_t Arg2) - __attribute__((alias("__sanitizer_cov_trace_cmp1"))); -void __sanitizer_cov_trace_const_cmp2(uint16_t Arg1, uint16_t Arg2) - __attribute__((alias("__sanitizer_cov_trace_cmp2"))); -void __sanitizer_cov_trace_const_cmp4(uint32_t Arg1, uint32_t Arg2) - __attribute__((alias("__sanitizer_cov_trace_cmp4"))); -void __sanitizer_cov_trace_const_cmp8(uint64_t Arg1, uint64_t Arg2) - __attribute__((alias("__sanitizer_cov_trace_cmp8"))); -#endif /* defined(__APPLE__) */ - -void __sanitizer_cov_trace_switch(uint64_t Val, uint64_t* Cases) { - - for (uint64_t i = 0; i < Cases[0]; i++) { - - uintptr_t k = (uintptr_t)__builtin_return_address(0) + i; - k = (k >> 4) ^ (k << 8); - k &= CMP_MAP_W - 1; - - u32 hits = __afl_cmp_map->headers[k].hits; - __afl_cmp_map->headers[k].hits = hits + 1; - - __afl_cmp_map->headers[k].shape = 7; - - hits &= CMP_MAP_H - 1; - __afl_cmp_map->log[k][hits].v0 = Val; - __afl_cmp_map->log[k][hits].v1 = Cases[i + 2]; - - } - -} - /* The following stuff deals with supporting -fsanitize-coverage=trace-pc-guard. It remains non-operational in the traditional, plugin-backed LLVM mode. For more info about 'trace-pc-guard', see README.llvm. diff --git a/src/afl-fuzz-cmplog.c b/src/afl-fuzz-cmplog.c index 924784cc..fcb545e1 100644 --- a/src/afl-fuzz-cmplog.c +++ b/src/afl-fuzz-cmplog.c @@ -1,373 +1,33 @@ -#include "afl-fuzz.h" -#include "cmplog.h" - -#define SWAP64(_x) \ - ({ \ - \ - u64 _ret = (_x); \ - _ret = \ - (_ret & 0x00000000FFFFFFFF) << 32 | (_ret & 0xFFFFFFFF00000000) >> 32; \ - _ret = \ - (_ret & 0x0000FFFF0000FFFF) << 16 | (_ret & 0xFFFF0000FFFF0000) >> 16; \ - _ret = \ - (_ret & 0x00FF00FF00FF00FF) << 8 | (_ret & 0xFF00FF00FF00FF00) >> 8; \ - _ret; \ - \ - }) - -u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len); - -extern struct cmp_map* cmp_map; // defined in afl-sharedmem.c - -u8* cmplog_binary; -char** its_argv; - -///// Colorization - -struct range { - - u32 start; - u32 end; - struct range* next; - -}; - -struct range* add_range(struct range* ranges, u32 start, u32 end) { - - struct range* r = ck_alloc_nozero(sizeof(struct range)); - r->start = start; - r->end = end; - r->next = ranges; - return r; - -} - -struct range* pop_biggest_range(struct range** ranges) { - - struct range* r = *ranges; - struct range* prev = NULL; - struct range* rmax = NULL; - struct range* prev_rmax = NULL; - u32 max_size = 0; - - while (r) { - - u32 s = r->end - r->start; - if (s >= max_size) { - - max_size = s; - prev_rmax = prev; - rmax = r; - - } - - prev = r; - r = r->next; - - } - - if (rmax) { - - if (prev_rmax) - prev_rmax->next = rmax->next; - else - *ranges = rmax->next; - - } - - return rmax; - -} - -u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) { - - if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; - - *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST); - return 0; - -} - -static void rand_replace(u8* buf, u32 len) { - - u32 i; - for (i = 0; i < len; ++i) - buf[i] = UR(256); - -} - -u8 colorization(u8* buf, u32 len, u32 exec_cksum) { - - struct range* ranges = add_range(NULL, 0, len); - u8* backup = ck_alloc_nozero(len); - - u64 orig_hit_cnt, new_hit_cnt; - orig_hit_cnt = queued_paths + unique_crashes; - - stage_name = "colorization"; - stage_short = "colorization"; - stage_max = 1000; - - struct range* rng; - stage_cur = stage_max; - while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) { - - u32 s = rng->end - rng->start; - memcpy(backup, buf + rng->start, s); - rand_replace(buf + rng->start, s); - - u32 cksum; - if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1; - - if (cksum != exec_cksum) { - - ranges = add_range(ranges, rng->start, rng->start + s / 2); - ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end); - memcpy(buf + rng->start, backup, s); - - } - - ck_free(rng); - --stage_cur; - - } - - new_hit_cnt = queued_paths + unique_crashes; - stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt; - stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur; - - while (ranges) { - - rng = ranges; - ranges = ranges->next; - ck_free(rng); - - } - - return 0; - -} - -///// Input to State replacement - -u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) { - - u64 orig_hit_cnt, new_hit_cnt; - - orig_hit_cnt = queued_paths + unique_crashes; - - if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; - - new_hit_cnt = queued_paths + unique_crashes; - - if (unlikely(new_hit_cnt != orig_hit_cnt)) { - - *status = 1; - - } else { - - if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA) - maybe_add_auto(&buf[idx], size); - *status = 2; - - } - - return 0; - -} - -u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx, - u8* orig_buf, u8* buf, u32 len, u8 do_reverse, - u8* status) { - - u64* buf_64 = (u64*)&buf[idx]; - u32* buf_32 = (u32*)&buf[idx]; - u16* buf_16 = (u16*)&buf[idx]; - // u8* buf_8 = &buf[idx]; - u64* o_buf_64 = (u64*)&orig_buf[idx]; - u32* o_buf_32 = (u32*)&orig_buf[idx]; - u16* o_buf_16 = (u16*)&orig_buf[idx]; - // u8* o_buf_8 = &orig_buf[idx]; - - u32 its_len = len - idx; - *status = 0; - - if (SHAPE_BYTES(h->shape) == 8) { - - if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) { - - *buf_64 = repl; - if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1; - *buf_64 = pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - if (SHAPE_BYTES(h->shape) == 4 || *status == 2) { - - if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) { - - *buf_32 = (u32)repl; - if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1; - *buf_32 = pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - if (SHAPE_BYTES(h->shape) == 2 || *status == 2) { - - if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) { - - *buf_16 = (u16)repl; - if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1; - *buf_16 = (u16)pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) { - - if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) { - - *buf_8 = (u8)repl; - if (unlikely(its_fuzz(idx, 1, buf, len, status))) - return 1; - *buf_16 = (u16)pattern; - - } - - }*/ - - return 0; - -} - -u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) { - - struct cmp_header* h = &cmp_map->headers[key]; - u32 i, j, idx; - - u32 loggeds = h->hits; - if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H; - - u8 status; - // opt not in the paper - u32 fails = 0; - - for (i = 0; i < loggeds; ++i) { - - struct cmp_operands* o = &cmp_map->log[key][i]; - - // opt not in the paper - for (j = 0; j < i; ++j) - if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1) - goto cmp_fuzz_next_iter; - - for (idx = 0; idx < len && fails < 8; ++idx) { - - if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len, - 1, &status))) - return 1; - if (status == 2) - ++fails; - else if (status == 1) - break; - - if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len, - 1, &status))) - return 1; - if (status == 2) - ++fails; - else if (status == 1) - break; - - } - - cmp_fuzz_next_iter: - stage_cur++; - - } - - return 0; - -} +/* + american fuzzy lop++ - cmplog execution routines + ------------------------------------------------ -///// Input to State stage + Originally written by Michal Zalewski -// queue_cur->exec_cksum -u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len, - u32 exec_cksum) { + Forkserver design by Jann Horn <jannhorn@googlemail.com> - its_argv = argv; + Now maintained by by Marc Heuse <mh@mh-sec.de>, + Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and + Andrea Fioraldi <andreafioraldi@gmail.com> - if (unlikely(colorization(buf, len, exec_cksum))) return 1; + Copyright 2016, 2017 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. - // do it manually, forkserver clear only trace_bits - memset(cmp_map->headers, 0, sizeof(cmp_map->headers)); + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: - if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1; + http://www.apache.org/licenses/LICENSE-2.0 - u64 orig_hit_cnt, new_hit_cnt; - u64 orig_execs = total_execs; - orig_hit_cnt = queued_paths + unique_crashes; + Shared code to handle the shared memory. This is used by the fuzzer + as well the other components like afl-tmin, afl-showmap, etc... - stage_name = "input-to-state"; - stage_short = "its"; - stage_max = 0; - stage_cur = 0; + */ - u32 k; - for (k = 0; k < CMP_MAP_W; ++k) { - - if (!cmp_map->headers[k].hits) continue; - if (cmp_map->headers[k].hits > CMP_MAP_H) - stage_max += CMP_MAP_H; - else - stage_max += cmp_map->headers[k].hits; - - } - - for (k = 0; k < CMP_MAP_W; ++k) { - - if (!cmp_map->headers[k].hits) continue; - cmp_fuzz(k, orig_buf, buf, len); - - } - - memcpy(buf, orig_buf, len); - - new_hit_cnt = queued_paths + unique_crashes; - stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt; - stage_cycles[STAGE_ITS] += total_execs - orig_execs; - - return 0; - -} - -//// CmpLog forkserver +#include "afl-fuzz.h" +#include "cmplog.h" -s32 cmplog_forksrv_pid, cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd; +static s32 cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd; void init_cmplog_forkserver(char** argv) { @@ -490,8 +150,6 @@ void init_cmplog_forkserver(char** argv) { "msan_track_origins=0", 0); - setenv("__AFL_CMPLOG_MODE__", "1", 1); - argv[0] = cmplog_binary; execv(cmplog_binary, argv); @@ -783,8 +441,6 @@ u8 run_cmplog_target(char** argv, u32 timeout) { "symbolize=0:" "msan_track_origins=0", 0); - setenv("__AFL_CMPLOG_MODE__", "1", 1); - argv[0] = cmplog_binary; execv(cmplog_binary, argv); @@ -977,4 +633,3 @@ u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len) { return 0; } - diff --git a/src/afl-fuzz-globals.c b/src/afl-fuzz-globals.c index b3476778..863ee9ad 100644 --- a/src/afl-fuzz-globals.c +++ b/src/afl-fuzz-globals.c @@ -251,6 +251,9 @@ u32 a_extras_cnt; /* Total number of tokens available */ u8 *(*post_handler)(u8 *buf, u32 *len); +u8* cmplog_binary; +s32 cmplog_forksrv_pid; + /* hooks for the custom mutator function */ size_t (*custom_mutator)(u8 *data, size_t size, u8 *mutated_out, size_t max_size, unsigned int seed); diff --git a/src/afl-fuzz-redqueen.c b/src/afl-fuzz-redqueen.c new file mode 100644 index 00000000..f6f659da --- /dev/null +++ b/src/afl-fuzz-redqueen.c @@ -0,0 +1,373 @@ +/* + american fuzzy lop++ - redqueen implementation on top of cmplog + --------------------------------------------------------------- + + Originally written by Michal Zalewski + + Forkserver design by Jann Horn <jannhorn@googlemail.com> + + Now maintained by by Marc Heuse <mh@mh-sec.de>, + Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and + Andrea Fioraldi <andreafioraldi@gmail.com> + + Copyright 2016, 2017 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: + + http://www.apache.org/licenses/LICENSE-2.0 + + Shared code to handle the shared memory. This is used by the fuzzer + as well the other components like afl-tmin, afl-showmap, etc... + + */ + +#include "afl-fuzz.h" +#include "cmplog.h" + +static char** its_argv; + +///// Colorization + +struct range { + + u32 start; + u32 end; + struct range* next; + +}; + +struct range* add_range(struct range* ranges, u32 start, u32 end) { + + struct range* r = ck_alloc_nozero(sizeof(struct range)); + r->start = start; + r->end = end; + r->next = ranges; + return r; + +} + +struct range* pop_biggest_range(struct range** ranges) { + + struct range* r = *ranges; + struct range* prev = NULL; + struct range* rmax = NULL; + struct range* prev_rmax = NULL; + u32 max_size = 0; + + while (r) { + + u32 s = r->end - r->start; + if (s >= max_size) { + + max_size = s; + prev_rmax = prev; + rmax = r; + + } + + prev = r; + r = r->next; + + } + + if (rmax) { + + if (prev_rmax) + prev_rmax->next = rmax->next; + else + *ranges = rmax->next; + + } + + return rmax; + +} + +u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) { + + if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; + + *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST); + return 0; + +} + +static void rand_replace(u8* buf, u32 len) { + + u32 i; + for (i = 0; i < len; ++i) + buf[i] = UR(256); + +} + +u8 colorization(u8* buf, u32 len, u32 exec_cksum) { + + struct range* ranges = add_range(NULL, 0, len); + u8* backup = ck_alloc_nozero(len); + + u64 orig_hit_cnt, new_hit_cnt; + orig_hit_cnt = queued_paths + unique_crashes; + + stage_name = "colorization"; + stage_short = "colorization"; + stage_max = 1000; + + struct range* rng; + stage_cur = stage_max; + while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) { + + u32 s = rng->end - rng->start; + memcpy(backup, buf + rng->start, s); + rand_replace(buf + rng->start, s); + + u32 cksum; + if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1; + + if (cksum != exec_cksum) { + + ranges = add_range(ranges, rng->start, rng->start + s / 2); + ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end); + memcpy(buf + rng->start, backup, s); + + } + + ck_free(rng); + --stage_cur; + + } + + new_hit_cnt = queued_paths + unique_crashes; + stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt; + stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur; + + while (ranges) { + + rng = ranges; + ranges = ranges->next; + ck_free(rng); + + } + + return 0; + +} + +///// Input to State replacement + +u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) { + + u64 orig_hit_cnt, new_hit_cnt; + + orig_hit_cnt = queued_paths + unique_crashes; + + if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; + + new_hit_cnt = queued_paths + unique_crashes; + + if (unlikely(new_hit_cnt != orig_hit_cnt)) { + + *status = 1; + + } else { + + if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA) + maybe_add_auto(&buf[idx], size); + *status = 2; + + } + + return 0; + +} + +u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx, + u8* orig_buf, u8* buf, u32 len, u8 do_reverse, + u8* status) { + + u64* buf_64 = (u64*)&buf[idx]; + u32* buf_32 = (u32*)&buf[idx]; + u16* buf_16 = (u16*)&buf[idx]; + // u8* buf_8 = &buf[idx]; + u64* o_buf_64 = (u64*)&orig_buf[idx]; + u32* o_buf_32 = (u32*)&orig_buf[idx]; + u16* o_buf_16 = (u16*)&orig_buf[idx]; + // u8* o_buf_8 = &orig_buf[idx]; + + u32 its_len = len - idx; + *status = 0; + + if (SHAPE_BYTES(h->shape) == 8) { + + if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) { + + *buf_64 = repl; + if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1; + *buf_64 = pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + if (SHAPE_BYTES(h->shape) == 4 || *status == 2) { + + if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) { + + *buf_32 = (u32)repl; + if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1; + *buf_32 = pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + if (SHAPE_BYTES(h->shape) == 2 || *status == 2) { + + if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) { + + *buf_16 = (u16)repl; + if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1; + *buf_16 = (u16)pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) { + + if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) { + + *buf_8 = (u8)repl; + if (unlikely(its_fuzz(idx, 1, buf, len, status))) + return 1; + *buf_16 = (u16)pattern; + + } + + }*/ + + return 0; + +} + +u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) { + + struct cmp_header* h = &cmp_map->headers[key]; + u32 i, j, idx; + + u32 loggeds = h->hits; + if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H; + + u8 status; + // opt not in the paper + u32 fails = 0; + + for (i = 0; i < loggeds; ++i) { + + struct cmp_operands* o = &cmp_map->log[key][i]; + + // opt not in the paper + for (j = 0; j < i; ++j) + if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1) + goto cmp_fuzz_next_iter; + + for (idx = 0; idx < len && fails < 8; ++idx) { + + if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len, + 1, &status))) + return 1; + if (status == 2) + ++fails; + else if (status == 1) + break; + + if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len, + 1, &status))) + return 1; + if (status == 2) + ++fails; + else if (status == 1) + break; + + } + + cmp_fuzz_next_iter: + stage_cur++; + + } + + return 0; + +} + +///// Input to State stage + +// queue_cur->exec_cksum +u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len, + u32 exec_cksum) { + + its_argv = argv; + + if (unlikely(colorization(buf, len, exec_cksum))) return 1; + + // do it manually, forkserver clear only trace_bits + memset(cmp_map->headers, 0, sizeof(cmp_map->headers)); + + if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1; + + u64 orig_hit_cnt, new_hit_cnt; + u64 orig_execs = total_execs; + orig_hit_cnt = queued_paths + unique_crashes; + + stage_name = "input-to-state"; + stage_short = "its"; + stage_max = 0; + stage_cur = 0; + + u32 k; + for (k = 0; k < CMP_MAP_W; ++k) { + + if (!cmp_map->headers[k].hits) continue; + if (cmp_map->headers[k].hits > CMP_MAP_H) + stage_max += CMP_MAP_H; + else + stage_max += cmp_map->headers[k].hits; + + } + + for (k = 0; k < CMP_MAP_W; ++k) { + + if (!cmp_map->headers[k].hits) continue; + cmp_fuzz(k, orig_buf, buf, len); + + } + + memcpy(buf, orig_buf, len); + + new_hit_cnt = queued_paths + unique_crashes; + stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt; + stage_cycles[STAGE_ITS] += total_execs - orig_execs; + + return 0; + +} |