diff options
Diffstat (limited to 'src')
-rw-r--r-- | src/afl-fuzz-cmplog.c | 385 | ||||
-rw-r--r-- | src/afl-fuzz-globals.c | 3 | ||||
-rw-r--r-- | src/afl-fuzz-redqueen.c | 373 |
3 files changed, 396 insertions, 365 deletions
diff --git a/src/afl-fuzz-cmplog.c b/src/afl-fuzz-cmplog.c index 924784cc..fcb545e1 100644 --- a/src/afl-fuzz-cmplog.c +++ b/src/afl-fuzz-cmplog.c @@ -1,373 +1,33 @@ -#include "afl-fuzz.h" -#include "cmplog.h" - -#define SWAP64(_x) \ - ({ \ - \ - u64 _ret = (_x); \ - _ret = \ - (_ret & 0x00000000FFFFFFFF) << 32 | (_ret & 0xFFFFFFFF00000000) >> 32; \ - _ret = \ - (_ret & 0x0000FFFF0000FFFF) << 16 | (_ret & 0xFFFF0000FFFF0000) >> 16; \ - _ret = \ - (_ret & 0x00FF00FF00FF00FF) << 8 | (_ret & 0xFF00FF00FF00FF00) >> 8; \ - _ret; \ - \ - }) - -u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len); - -extern struct cmp_map* cmp_map; // defined in afl-sharedmem.c - -u8* cmplog_binary; -char** its_argv; - -///// Colorization - -struct range { - - u32 start; - u32 end; - struct range* next; - -}; - -struct range* add_range(struct range* ranges, u32 start, u32 end) { - - struct range* r = ck_alloc_nozero(sizeof(struct range)); - r->start = start; - r->end = end; - r->next = ranges; - return r; - -} - -struct range* pop_biggest_range(struct range** ranges) { - - struct range* r = *ranges; - struct range* prev = NULL; - struct range* rmax = NULL; - struct range* prev_rmax = NULL; - u32 max_size = 0; - - while (r) { - - u32 s = r->end - r->start; - if (s >= max_size) { - - max_size = s; - prev_rmax = prev; - rmax = r; - - } - - prev = r; - r = r->next; - - } - - if (rmax) { - - if (prev_rmax) - prev_rmax->next = rmax->next; - else - *ranges = rmax->next; - - } - - return rmax; - -} - -u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) { - - if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; - - *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST); - return 0; - -} - -static void rand_replace(u8* buf, u32 len) { - - u32 i; - for (i = 0; i < len; ++i) - buf[i] = UR(256); - -} - -u8 colorization(u8* buf, u32 len, u32 exec_cksum) { - - struct range* ranges = add_range(NULL, 0, len); - u8* backup = ck_alloc_nozero(len); - - u64 orig_hit_cnt, new_hit_cnt; - orig_hit_cnt = queued_paths + unique_crashes; - - stage_name = "colorization"; - stage_short = "colorization"; - stage_max = 1000; - - struct range* rng; - stage_cur = stage_max; - while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) { - - u32 s = rng->end - rng->start; - memcpy(backup, buf + rng->start, s); - rand_replace(buf + rng->start, s); - - u32 cksum; - if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1; - - if (cksum != exec_cksum) { - - ranges = add_range(ranges, rng->start, rng->start + s / 2); - ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end); - memcpy(buf + rng->start, backup, s); - - } - - ck_free(rng); - --stage_cur; - - } - - new_hit_cnt = queued_paths + unique_crashes; - stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt; - stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur; - - while (ranges) { - - rng = ranges; - ranges = ranges->next; - ck_free(rng); - - } - - return 0; - -} - -///// Input to State replacement - -u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) { - - u64 orig_hit_cnt, new_hit_cnt; - - orig_hit_cnt = queued_paths + unique_crashes; - - if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; - - new_hit_cnt = queued_paths + unique_crashes; - - if (unlikely(new_hit_cnt != orig_hit_cnt)) { - - *status = 1; - - } else { - - if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA) - maybe_add_auto(&buf[idx], size); - *status = 2; - - } - - return 0; - -} - -u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx, - u8* orig_buf, u8* buf, u32 len, u8 do_reverse, - u8* status) { - - u64* buf_64 = (u64*)&buf[idx]; - u32* buf_32 = (u32*)&buf[idx]; - u16* buf_16 = (u16*)&buf[idx]; - // u8* buf_8 = &buf[idx]; - u64* o_buf_64 = (u64*)&orig_buf[idx]; - u32* o_buf_32 = (u32*)&orig_buf[idx]; - u16* o_buf_16 = (u16*)&orig_buf[idx]; - // u8* o_buf_8 = &orig_buf[idx]; - - u32 its_len = len - idx; - *status = 0; - - if (SHAPE_BYTES(h->shape) == 8) { - - if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) { - - *buf_64 = repl; - if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1; - *buf_64 = pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - if (SHAPE_BYTES(h->shape) == 4 || *status == 2) { - - if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) { - - *buf_32 = (u32)repl; - if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1; - *buf_32 = pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - if (SHAPE_BYTES(h->shape) == 2 || *status == 2) { - - if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) { - - *buf_16 = (u16)repl; - if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1; - *buf_16 = (u16)pattern; - - } - - // reverse encoding - if (do_reverse) - if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx, - orig_buf, buf, len, 0, status))) - return 1; - - } - - /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) { - - if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) { - - *buf_8 = (u8)repl; - if (unlikely(its_fuzz(idx, 1, buf, len, status))) - return 1; - *buf_16 = (u16)pattern; - - } - - }*/ - - return 0; - -} - -u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) { - - struct cmp_header* h = &cmp_map->headers[key]; - u32 i, j, idx; - - u32 loggeds = h->hits; - if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H; - - u8 status; - // opt not in the paper - u32 fails = 0; - - for (i = 0; i < loggeds; ++i) { - - struct cmp_operands* o = &cmp_map->log[key][i]; - - // opt not in the paper - for (j = 0; j < i; ++j) - if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1) - goto cmp_fuzz_next_iter; - - for (idx = 0; idx < len && fails < 8; ++idx) { - - if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len, - 1, &status))) - return 1; - if (status == 2) - ++fails; - else if (status == 1) - break; - - if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len, - 1, &status))) - return 1; - if (status == 2) - ++fails; - else if (status == 1) - break; - - } - - cmp_fuzz_next_iter: - stage_cur++; - - } - - return 0; - -} +/* + american fuzzy lop++ - cmplog execution routines + ------------------------------------------------ -///// Input to State stage + Originally written by Michal Zalewski -// queue_cur->exec_cksum -u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len, - u32 exec_cksum) { + Forkserver design by Jann Horn <jannhorn@googlemail.com> - its_argv = argv; + Now maintained by by Marc Heuse <mh@mh-sec.de>, + Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and + Andrea Fioraldi <andreafioraldi@gmail.com> - if (unlikely(colorization(buf, len, exec_cksum))) return 1; + Copyright 2016, 2017 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. - // do it manually, forkserver clear only trace_bits - memset(cmp_map->headers, 0, sizeof(cmp_map->headers)); + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: - if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1; + http://www.apache.org/licenses/LICENSE-2.0 - u64 orig_hit_cnt, new_hit_cnt; - u64 orig_execs = total_execs; - orig_hit_cnt = queued_paths + unique_crashes; + Shared code to handle the shared memory. This is used by the fuzzer + as well the other components like afl-tmin, afl-showmap, etc... - stage_name = "input-to-state"; - stage_short = "its"; - stage_max = 0; - stage_cur = 0; + */ - u32 k; - for (k = 0; k < CMP_MAP_W; ++k) { - - if (!cmp_map->headers[k].hits) continue; - if (cmp_map->headers[k].hits > CMP_MAP_H) - stage_max += CMP_MAP_H; - else - stage_max += cmp_map->headers[k].hits; - - } - - for (k = 0; k < CMP_MAP_W; ++k) { - - if (!cmp_map->headers[k].hits) continue; - cmp_fuzz(k, orig_buf, buf, len); - - } - - memcpy(buf, orig_buf, len); - - new_hit_cnt = queued_paths + unique_crashes; - stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt; - stage_cycles[STAGE_ITS] += total_execs - orig_execs; - - return 0; - -} - -//// CmpLog forkserver +#include "afl-fuzz.h" +#include "cmplog.h" -s32 cmplog_forksrv_pid, cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd; +static s32 cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd; void init_cmplog_forkserver(char** argv) { @@ -490,8 +150,6 @@ void init_cmplog_forkserver(char** argv) { "msan_track_origins=0", 0); - setenv("__AFL_CMPLOG_MODE__", "1", 1); - argv[0] = cmplog_binary; execv(cmplog_binary, argv); @@ -783,8 +441,6 @@ u8 run_cmplog_target(char** argv, u32 timeout) { "symbolize=0:" "msan_track_origins=0", 0); - setenv("__AFL_CMPLOG_MODE__", "1", 1); - argv[0] = cmplog_binary; execv(cmplog_binary, argv); @@ -977,4 +633,3 @@ u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len) { return 0; } - diff --git a/src/afl-fuzz-globals.c b/src/afl-fuzz-globals.c index b3476778..863ee9ad 100644 --- a/src/afl-fuzz-globals.c +++ b/src/afl-fuzz-globals.c @@ -251,6 +251,9 @@ u32 a_extras_cnt; /* Total number of tokens available */ u8 *(*post_handler)(u8 *buf, u32 *len); +u8* cmplog_binary; +s32 cmplog_forksrv_pid; + /* hooks for the custom mutator function */ size_t (*custom_mutator)(u8 *data, size_t size, u8 *mutated_out, size_t max_size, unsigned int seed); diff --git a/src/afl-fuzz-redqueen.c b/src/afl-fuzz-redqueen.c new file mode 100644 index 00000000..f6f659da --- /dev/null +++ b/src/afl-fuzz-redqueen.c @@ -0,0 +1,373 @@ +/* + american fuzzy lop++ - redqueen implementation on top of cmplog + --------------------------------------------------------------- + + Originally written by Michal Zalewski + + Forkserver design by Jann Horn <jannhorn@googlemail.com> + + Now maintained by by Marc Heuse <mh@mh-sec.de>, + Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and + Andrea Fioraldi <andreafioraldi@gmail.com> + + Copyright 2016, 2017 Google Inc. All rights reserved. + Copyright 2019-2020 AFLplusplus Project. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at: + + http://www.apache.org/licenses/LICENSE-2.0 + + Shared code to handle the shared memory. This is used by the fuzzer + as well the other components like afl-tmin, afl-showmap, etc... + + */ + +#include "afl-fuzz.h" +#include "cmplog.h" + +static char** its_argv; + +///// Colorization + +struct range { + + u32 start; + u32 end; + struct range* next; + +}; + +struct range* add_range(struct range* ranges, u32 start, u32 end) { + + struct range* r = ck_alloc_nozero(sizeof(struct range)); + r->start = start; + r->end = end; + r->next = ranges; + return r; + +} + +struct range* pop_biggest_range(struct range** ranges) { + + struct range* r = *ranges; + struct range* prev = NULL; + struct range* rmax = NULL; + struct range* prev_rmax = NULL; + u32 max_size = 0; + + while (r) { + + u32 s = r->end - r->start; + if (s >= max_size) { + + max_size = s; + prev_rmax = prev; + rmax = r; + + } + + prev = r; + r = r->next; + + } + + if (rmax) { + + if (prev_rmax) + prev_rmax->next = rmax->next; + else + *ranges = rmax->next; + + } + + return rmax; + +} + +u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) { + + if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; + + *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST); + return 0; + +} + +static void rand_replace(u8* buf, u32 len) { + + u32 i; + for (i = 0; i < len; ++i) + buf[i] = UR(256); + +} + +u8 colorization(u8* buf, u32 len, u32 exec_cksum) { + + struct range* ranges = add_range(NULL, 0, len); + u8* backup = ck_alloc_nozero(len); + + u64 orig_hit_cnt, new_hit_cnt; + orig_hit_cnt = queued_paths + unique_crashes; + + stage_name = "colorization"; + stage_short = "colorization"; + stage_max = 1000; + + struct range* rng; + stage_cur = stage_max; + while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) { + + u32 s = rng->end - rng->start; + memcpy(backup, buf + rng->start, s); + rand_replace(buf + rng->start, s); + + u32 cksum; + if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1; + + if (cksum != exec_cksum) { + + ranges = add_range(ranges, rng->start, rng->start + s / 2); + ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end); + memcpy(buf + rng->start, backup, s); + + } + + ck_free(rng); + --stage_cur; + + } + + new_hit_cnt = queued_paths + unique_crashes; + stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt; + stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur; + + while (ranges) { + + rng = ranges; + ranges = ranges->next; + ck_free(rng); + + } + + return 0; + +} + +///// Input to State replacement + +u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) { + + u64 orig_hit_cnt, new_hit_cnt; + + orig_hit_cnt = queued_paths + unique_crashes; + + if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1; + + new_hit_cnt = queued_paths + unique_crashes; + + if (unlikely(new_hit_cnt != orig_hit_cnt)) { + + *status = 1; + + } else { + + if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA) + maybe_add_auto(&buf[idx], size); + *status = 2; + + } + + return 0; + +} + +u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx, + u8* orig_buf, u8* buf, u32 len, u8 do_reverse, + u8* status) { + + u64* buf_64 = (u64*)&buf[idx]; + u32* buf_32 = (u32*)&buf[idx]; + u16* buf_16 = (u16*)&buf[idx]; + // u8* buf_8 = &buf[idx]; + u64* o_buf_64 = (u64*)&orig_buf[idx]; + u32* o_buf_32 = (u32*)&orig_buf[idx]; + u16* o_buf_16 = (u16*)&orig_buf[idx]; + // u8* o_buf_8 = &orig_buf[idx]; + + u32 its_len = len - idx; + *status = 0; + + if (SHAPE_BYTES(h->shape) == 8) { + + if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) { + + *buf_64 = repl; + if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1; + *buf_64 = pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + if (SHAPE_BYTES(h->shape) == 4 || *status == 2) { + + if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) { + + *buf_32 = (u32)repl; + if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1; + *buf_32 = pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + if (SHAPE_BYTES(h->shape) == 2 || *status == 2) { + + if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) { + + *buf_16 = (u16)repl; + if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1; + *buf_16 = (u16)pattern; + + } + + // reverse encoding + if (do_reverse) + if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx, + orig_buf, buf, len, 0, status))) + return 1; + + } + + /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) { + + if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) { + + *buf_8 = (u8)repl; + if (unlikely(its_fuzz(idx, 1, buf, len, status))) + return 1; + *buf_16 = (u16)pattern; + + } + + }*/ + + return 0; + +} + +u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) { + + struct cmp_header* h = &cmp_map->headers[key]; + u32 i, j, idx; + + u32 loggeds = h->hits; + if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H; + + u8 status; + // opt not in the paper + u32 fails = 0; + + for (i = 0; i < loggeds; ++i) { + + struct cmp_operands* o = &cmp_map->log[key][i]; + + // opt not in the paper + for (j = 0; j < i; ++j) + if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1) + goto cmp_fuzz_next_iter; + + for (idx = 0; idx < len && fails < 8; ++idx) { + + if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len, + 1, &status))) + return 1; + if (status == 2) + ++fails; + else if (status == 1) + break; + + if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len, + 1, &status))) + return 1; + if (status == 2) + ++fails; + else if (status == 1) + break; + + } + + cmp_fuzz_next_iter: + stage_cur++; + + } + + return 0; + +} + +///// Input to State stage + +// queue_cur->exec_cksum +u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len, + u32 exec_cksum) { + + its_argv = argv; + + if (unlikely(colorization(buf, len, exec_cksum))) return 1; + + // do it manually, forkserver clear only trace_bits + memset(cmp_map->headers, 0, sizeof(cmp_map->headers)); + + if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1; + + u64 orig_hit_cnt, new_hit_cnt; + u64 orig_execs = total_execs; + orig_hit_cnt = queued_paths + unique_crashes; + + stage_name = "input-to-state"; + stage_short = "its"; + stage_max = 0; + stage_cur = 0; + + u32 k; + for (k = 0; k < CMP_MAP_W; ++k) { + + if (!cmp_map->headers[k].hits) continue; + if (cmp_map->headers[k].hits > CMP_MAP_H) + stage_max += CMP_MAP_H; + else + stage_max += cmp_map->headers[k].hits; + + } + + for (k = 0; k < CMP_MAP_W; ++k) { + + if (!cmp_map->headers[k].hits) continue; + cmp_fuzz(k, orig_buf, buf, len); + + } + + memcpy(buf, orig_buf, len); + + new_hit_cnt = queued_paths + unique_crashes; + stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt; + stage_cycles[STAGE_ITS] += total_execs - orig_execs; + + return 0; + +} |