aboutsummaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorAndrea Fioraldi <andreafioraldi@gmail.com>2020-01-30 22:40:09 +0100
committerAndrea Fioraldi <andreafioraldi@gmail.com>2020-01-30 22:40:09 +0100
commit95a98fb3e805d94a3db82d5da21e9a00a79dbdf2 (patch)
tree4de4f17f898a4f0632098c9902614217fa2e53eb /src
parentb8bad5a2273b0cddd0244a7f37c20150a08af475 (diff)
downloadafl++-95a98fb3e805d94a3db82d5da21e9a00a79dbdf2.tar.gz
cmplog runtime for LLVM
Diffstat (limited to 'src')
-rw-r--r--src/afl-fuzz-cmplog.c385
-rw-r--r--src/afl-fuzz-globals.c3
-rw-r--r--src/afl-fuzz-redqueen.c373
3 files changed, 396 insertions, 365 deletions
diff --git a/src/afl-fuzz-cmplog.c b/src/afl-fuzz-cmplog.c
index 924784cc..fcb545e1 100644
--- a/src/afl-fuzz-cmplog.c
+++ b/src/afl-fuzz-cmplog.c
@@ -1,373 +1,33 @@
-#include "afl-fuzz.h"
-#include "cmplog.h"
-
-#define SWAP64(_x) \
- ({ \
- \
- u64 _ret = (_x); \
- _ret = \
- (_ret & 0x00000000FFFFFFFF) << 32 | (_ret & 0xFFFFFFFF00000000) >> 32; \
- _ret = \
- (_ret & 0x0000FFFF0000FFFF) << 16 | (_ret & 0xFFFF0000FFFF0000) >> 16; \
- _ret = \
- (_ret & 0x00FF00FF00FF00FF) << 8 | (_ret & 0xFF00FF00FF00FF00) >> 8; \
- _ret; \
- \
- })
-
-u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len);
-
-extern struct cmp_map* cmp_map; // defined in afl-sharedmem.c
-
-u8* cmplog_binary;
-char** its_argv;
-
-///// Colorization
-
-struct range {
-
- u32 start;
- u32 end;
- struct range* next;
-
-};
-
-struct range* add_range(struct range* ranges, u32 start, u32 end) {
-
- struct range* r = ck_alloc_nozero(sizeof(struct range));
- r->start = start;
- r->end = end;
- r->next = ranges;
- return r;
-
-}
-
-struct range* pop_biggest_range(struct range** ranges) {
-
- struct range* r = *ranges;
- struct range* prev = NULL;
- struct range* rmax = NULL;
- struct range* prev_rmax = NULL;
- u32 max_size = 0;
-
- while (r) {
-
- u32 s = r->end - r->start;
- if (s >= max_size) {
-
- max_size = s;
- prev_rmax = prev;
- rmax = r;
-
- }
-
- prev = r;
- r = r->next;
-
- }
-
- if (rmax) {
-
- if (prev_rmax)
- prev_rmax->next = rmax->next;
- else
- *ranges = rmax->next;
-
- }
-
- return rmax;
-
-}
-
-u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) {
-
- if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1;
-
- *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST);
- return 0;
-
-}
-
-static void rand_replace(u8* buf, u32 len) {
-
- u32 i;
- for (i = 0; i < len; ++i)
- buf[i] = UR(256);
-
-}
-
-u8 colorization(u8* buf, u32 len, u32 exec_cksum) {
-
- struct range* ranges = add_range(NULL, 0, len);
- u8* backup = ck_alloc_nozero(len);
-
- u64 orig_hit_cnt, new_hit_cnt;
- orig_hit_cnt = queued_paths + unique_crashes;
-
- stage_name = "colorization";
- stage_short = "colorization";
- stage_max = 1000;
-
- struct range* rng;
- stage_cur = stage_max;
- while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) {
-
- u32 s = rng->end - rng->start;
- memcpy(backup, buf + rng->start, s);
- rand_replace(buf + rng->start, s);
-
- u32 cksum;
- if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1;
-
- if (cksum != exec_cksum) {
-
- ranges = add_range(ranges, rng->start, rng->start + s / 2);
- ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end);
- memcpy(buf + rng->start, backup, s);
-
- }
-
- ck_free(rng);
- --stage_cur;
-
- }
-
- new_hit_cnt = queued_paths + unique_crashes;
- stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt;
- stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur;
-
- while (ranges) {
-
- rng = ranges;
- ranges = ranges->next;
- ck_free(rng);
-
- }
-
- return 0;
-
-}
-
-///// Input to State replacement
-
-u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) {
-
- u64 orig_hit_cnt, new_hit_cnt;
-
- orig_hit_cnt = queued_paths + unique_crashes;
-
- if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1;
-
- new_hit_cnt = queued_paths + unique_crashes;
-
- if (unlikely(new_hit_cnt != orig_hit_cnt)) {
-
- *status = 1;
-
- } else {
-
- if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA)
- maybe_add_auto(&buf[idx], size);
- *status = 2;
-
- }
-
- return 0;
-
-}
-
-u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx,
- u8* orig_buf, u8* buf, u32 len, u8 do_reverse,
- u8* status) {
-
- u64* buf_64 = (u64*)&buf[idx];
- u32* buf_32 = (u32*)&buf[idx];
- u16* buf_16 = (u16*)&buf[idx];
- // u8* buf_8 = &buf[idx];
- u64* o_buf_64 = (u64*)&orig_buf[idx];
- u32* o_buf_32 = (u32*)&orig_buf[idx];
- u16* o_buf_16 = (u16*)&orig_buf[idx];
- // u8* o_buf_8 = &orig_buf[idx];
-
- u32 its_len = len - idx;
- *status = 0;
-
- if (SHAPE_BYTES(h->shape) == 8) {
-
- if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) {
-
- *buf_64 = repl;
- if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1;
- *buf_64 = pattern;
-
- }
-
- // reverse encoding
- if (do_reverse)
- if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx,
- orig_buf, buf, len, 0, status)))
- return 1;
-
- }
-
- if (SHAPE_BYTES(h->shape) == 4 || *status == 2) {
-
- if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) {
-
- *buf_32 = (u32)repl;
- if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1;
- *buf_32 = pattern;
-
- }
-
- // reverse encoding
- if (do_reverse)
- if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx,
- orig_buf, buf, len, 0, status)))
- return 1;
-
- }
-
- if (SHAPE_BYTES(h->shape) == 2 || *status == 2) {
-
- if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) {
-
- *buf_16 = (u16)repl;
- if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1;
- *buf_16 = (u16)pattern;
-
- }
-
- // reverse encoding
- if (do_reverse)
- if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx,
- orig_buf, buf, len, 0, status)))
- return 1;
-
- }
-
- /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) {
-
- if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) {
-
- *buf_8 = (u8)repl;
- if (unlikely(its_fuzz(idx, 1, buf, len, status)))
- return 1;
- *buf_16 = (u16)pattern;
-
- }
-
- }*/
-
- return 0;
-
-}
-
-u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) {
-
- struct cmp_header* h = &cmp_map->headers[key];
- u32 i, j, idx;
-
- u32 loggeds = h->hits;
- if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H;
-
- u8 status;
- // opt not in the paper
- u32 fails = 0;
-
- for (i = 0; i < loggeds; ++i) {
-
- struct cmp_operands* o = &cmp_map->log[key][i];
-
- // opt not in the paper
- for (j = 0; j < i; ++j)
- if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1)
- goto cmp_fuzz_next_iter;
-
- for (idx = 0; idx < len && fails < 8; ++idx) {
-
- if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len,
- 1, &status)))
- return 1;
- if (status == 2)
- ++fails;
- else if (status == 1)
- break;
-
- if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len,
- 1, &status)))
- return 1;
- if (status == 2)
- ++fails;
- else if (status == 1)
- break;
-
- }
-
- cmp_fuzz_next_iter:
- stage_cur++;
-
- }
-
- return 0;
-
-}
+/*
+ american fuzzy lop++ - cmplog execution routines
+ ------------------------------------------------
-///// Input to State stage
+ Originally written by Michal Zalewski
-// queue_cur->exec_cksum
-u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len,
- u32 exec_cksum) {
+ Forkserver design by Jann Horn <jannhorn@googlemail.com>
- its_argv = argv;
+ Now maintained by by Marc Heuse <mh@mh-sec.de>,
+ Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and
+ Andrea Fioraldi <andreafioraldi@gmail.com>
- if (unlikely(colorization(buf, len, exec_cksum))) return 1;
+ Copyright 2016, 2017 Google Inc. All rights reserved.
+ Copyright 2019-2020 AFLplusplus Project. All rights reserved.
- // do it manually, forkserver clear only trace_bits
- memset(cmp_map->headers, 0, sizeof(cmp_map->headers));
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at:
- if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1;
+ http://www.apache.org/licenses/LICENSE-2.0
- u64 orig_hit_cnt, new_hit_cnt;
- u64 orig_execs = total_execs;
- orig_hit_cnt = queued_paths + unique_crashes;
+ Shared code to handle the shared memory. This is used by the fuzzer
+ as well the other components like afl-tmin, afl-showmap, etc...
- stage_name = "input-to-state";
- stage_short = "its";
- stage_max = 0;
- stage_cur = 0;
+ */
- u32 k;
- for (k = 0; k < CMP_MAP_W; ++k) {
-
- if (!cmp_map->headers[k].hits) continue;
- if (cmp_map->headers[k].hits > CMP_MAP_H)
- stage_max += CMP_MAP_H;
- else
- stage_max += cmp_map->headers[k].hits;
-
- }
-
- for (k = 0; k < CMP_MAP_W; ++k) {
-
- if (!cmp_map->headers[k].hits) continue;
- cmp_fuzz(k, orig_buf, buf, len);
-
- }
-
- memcpy(buf, orig_buf, len);
-
- new_hit_cnt = queued_paths + unique_crashes;
- stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt;
- stage_cycles[STAGE_ITS] += total_execs - orig_execs;
-
- return 0;
-
-}
-
-//// CmpLog forkserver
+#include "afl-fuzz.h"
+#include "cmplog.h"
-s32 cmplog_forksrv_pid, cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd;
+static s32 cmplog_child_pid, cmplog_fsrv_ctl_fd, cmplog_fsrv_st_fd;
void init_cmplog_forkserver(char** argv) {
@@ -490,8 +150,6 @@ void init_cmplog_forkserver(char** argv) {
"msan_track_origins=0",
0);
- setenv("__AFL_CMPLOG_MODE__", "1", 1);
-
argv[0] = cmplog_binary;
execv(cmplog_binary, argv);
@@ -783,8 +441,6 @@ u8 run_cmplog_target(char** argv, u32 timeout) {
"symbolize=0:"
"msan_track_origins=0", 0);
- setenv("__AFL_CMPLOG_MODE__", "1", 1);
-
argv[0] = cmplog_binary;
execv(cmplog_binary, argv);
@@ -977,4 +633,3 @@ u8 common_fuzz_cmplog_stuff(char** argv, u8* out_buf, u32 len) {
return 0;
}
-
diff --git a/src/afl-fuzz-globals.c b/src/afl-fuzz-globals.c
index b3476778..863ee9ad 100644
--- a/src/afl-fuzz-globals.c
+++ b/src/afl-fuzz-globals.c
@@ -251,6 +251,9 @@ u32 a_extras_cnt; /* Total number of tokens available */
u8 *(*post_handler)(u8 *buf, u32 *len);
+u8* cmplog_binary;
+s32 cmplog_forksrv_pid;
+
/* hooks for the custom mutator function */
size_t (*custom_mutator)(u8 *data, size_t size, u8 *mutated_out,
size_t max_size, unsigned int seed);
diff --git a/src/afl-fuzz-redqueen.c b/src/afl-fuzz-redqueen.c
new file mode 100644
index 00000000..f6f659da
--- /dev/null
+++ b/src/afl-fuzz-redqueen.c
@@ -0,0 +1,373 @@
+/*
+ american fuzzy lop++ - redqueen implementation on top of cmplog
+ ---------------------------------------------------------------
+
+ Originally written by Michal Zalewski
+
+ Forkserver design by Jann Horn <jannhorn@googlemail.com>
+
+ Now maintained by by Marc Heuse <mh@mh-sec.de>,
+ Heiko Eißfeldt <heiko.eissfeldt@hexco.de> and
+ Andrea Fioraldi <andreafioraldi@gmail.com>
+
+ Copyright 2016, 2017 Google Inc. All rights reserved.
+ Copyright 2019-2020 AFLplusplus Project. All rights reserved.
+
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at:
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Shared code to handle the shared memory. This is used by the fuzzer
+ as well the other components like afl-tmin, afl-showmap, etc...
+
+ */
+
+#include "afl-fuzz.h"
+#include "cmplog.h"
+
+static char** its_argv;
+
+///// Colorization
+
+struct range {
+
+ u32 start;
+ u32 end;
+ struct range* next;
+
+};
+
+struct range* add_range(struct range* ranges, u32 start, u32 end) {
+
+ struct range* r = ck_alloc_nozero(sizeof(struct range));
+ r->start = start;
+ r->end = end;
+ r->next = ranges;
+ return r;
+
+}
+
+struct range* pop_biggest_range(struct range** ranges) {
+
+ struct range* r = *ranges;
+ struct range* prev = NULL;
+ struct range* rmax = NULL;
+ struct range* prev_rmax = NULL;
+ u32 max_size = 0;
+
+ while (r) {
+
+ u32 s = r->end - r->start;
+ if (s >= max_size) {
+
+ max_size = s;
+ prev_rmax = prev;
+ rmax = r;
+
+ }
+
+ prev = r;
+ r = r->next;
+
+ }
+
+ if (rmax) {
+
+ if (prev_rmax)
+ prev_rmax->next = rmax->next;
+ else
+ *ranges = rmax->next;
+
+ }
+
+ return rmax;
+
+}
+
+u8 get_exec_checksum(u8* buf, u32 len, u32* cksum) {
+
+ if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1;
+
+ *cksum = hash32(trace_bits, MAP_SIZE, HASH_CONST);
+ return 0;
+
+}
+
+static void rand_replace(u8* buf, u32 len) {
+
+ u32 i;
+ for (i = 0; i < len; ++i)
+ buf[i] = UR(256);
+
+}
+
+u8 colorization(u8* buf, u32 len, u32 exec_cksum) {
+
+ struct range* ranges = add_range(NULL, 0, len);
+ u8* backup = ck_alloc_nozero(len);
+
+ u64 orig_hit_cnt, new_hit_cnt;
+ orig_hit_cnt = queued_paths + unique_crashes;
+
+ stage_name = "colorization";
+ stage_short = "colorization";
+ stage_max = 1000;
+
+ struct range* rng;
+ stage_cur = stage_max;
+ while ((rng = pop_biggest_range(&ranges)) != NULL && stage_cur) {
+
+ u32 s = rng->end - rng->start;
+ memcpy(backup, buf + rng->start, s);
+ rand_replace(buf + rng->start, s);
+
+ u32 cksum;
+ if (unlikely(get_exec_checksum(buf, len, &cksum))) return 1;
+
+ if (cksum != exec_cksum) {
+
+ ranges = add_range(ranges, rng->start, rng->start + s / 2);
+ ranges = add_range(ranges, rng->start + s / 2 + 1, rng->end);
+ memcpy(buf + rng->start, backup, s);
+
+ }
+
+ ck_free(rng);
+ --stage_cur;
+
+ }
+
+ new_hit_cnt = queued_paths + unique_crashes;
+ stage_finds[STAGE_COLORIZATION] += new_hit_cnt - orig_hit_cnt;
+ stage_cycles[STAGE_COLORIZATION] += stage_max - stage_cur;
+
+ while (ranges) {
+
+ rng = ranges;
+ ranges = ranges->next;
+ ck_free(rng);
+
+ }
+
+ return 0;
+
+}
+
+///// Input to State replacement
+
+u8 its_fuzz(u32 idx, u32 size, u8* buf, u32 len, u8* status) {
+
+ u64 orig_hit_cnt, new_hit_cnt;
+
+ orig_hit_cnt = queued_paths + unique_crashes;
+
+ if (unlikely(common_fuzz_stuff(its_argv, buf, len))) return 1;
+
+ new_hit_cnt = queued_paths + unique_crashes;
+
+ if (unlikely(new_hit_cnt != orig_hit_cnt)) {
+
+ *status = 1;
+
+ } else {
+
+ if (size >= MIN_AUTO_EXTRA && size <= MAX_AUTO_EXTRA)
+ maybe_add_auto(&buf[idx], size);
+ *status = 2;
+
+ }
+
+ return 0;
+
+}
+
+u8 cmp_extend_encoding(struct cmp_header* h, u64 pattern, u64 repl, u32 idx,
+ u8* orig_buf, u8* buf, u32 len, u8 do_reverse,
+ u8* status) {
+
+ u64* buf_64 = (u64*)&buf[idx];
+ u32* buf_32 = (u32*)&buf[idx];
+ u16* buf_16 = (u16*)&buf[idx];
+ // u8* buf_8 = &buf[idx];
+ u64* o_buf_64 = (u64*)&orig_buf[idx];
+ u32* o_buf_32 = (u32*)&orig_buf[idx];
+ u16* o_buf_16 = (u16*)&orig_buf[idx];
+ // u8* o_buf_8 = &orig_buf[idx];
+
+ u32 its_len = len - idx;
+ *status = 0;
+
+ if (SHAPE_BYTES(h->shape) == 8) {
+
+ if (its_len >= 8 && *buf_64 == pattern && *o_buf_64 == pattern) {
+
+ *buf_64 = repl;
+ if (unlikely(its_fuzz(idx, 8, buf, len, status))) return 1;
+ *buf_64 = pattern;
+
+ }
+
+ // reverse encoding
+ if (do_reverse)
+ if (unlikely(cmp_extend_encoding(h, SWAP64(pattern), SWAP64(repl), idx,
+ orig_buf, buf, len, 0, status)))
+ return 1;
+
+ }
+
+ if (SHAPE_BYTES(h->shape) == 4 || *status == 2) {
+
+ if (its_len >= 4 && *buf_32 == (u32)pattern && *o_buf_32 == (u32)pattern) {
+
+ *buf_32 = (u32)repl;
+ if (unlikely(its_fuzz(idx, 4, buf, len, status))) return 1;
+ *buf_32 = pattern;
+
+ }
+
+ // reverse encoding
+ if (do_reverse)
+ if (unlikely(cmp_extend_encoding(h, SWAP32(pattern), SWAP32(repl), idx,
+ orig_buf, buf, len, 0, status)))
+ return 1;
+
+ }
+
+ if (SHAPE_BYTES(h->shape) == 2 || *status == 2) {
+
+ if (its_len >= 2 && *buf_16 == (u16)pattern && *o_buf_16 == (u16)pattern) {
+
+ *buf_16 = (u16)repl;
+ if (unlikely(its_fuzz(idx, 2, buf, len, status))) return 1;
+ *buf_16 = (u16)pattern;
+
+ }
+
+ // reverse encoding
+ if (do_reverse)
+ if (unlikely(cmp_extend_encoding(h, SWAP16(pattern), SWAP16(repl), idx,
+ orig_buf, buf, len, 0, status)))
+ return 1;
+
+ }
+
+ /*if (SHAPE_BYTES(h->shape) == 1 || *status == 2) {
+
+ if (its_len >= 2 && *buf_8 == (u8)pattern && *o_buf_8 == (u8)pattern) {
+
+ *buf_8 = (u8)repl;
+ if (unlikely(its_fuzz(idx, 1, buf, len, status)))
+ return 1;
+ *buf_16 = (u16)pattern;
+
+ }
+
+ }*/
+
+ return 0;
+
+}
+
+u8 cmp_fuzz(u32 key, u8* orig_buf, u8* buf, u32 len) {
+
+ struct cmp_header* h = &cmp_map->headers[key];
+ u32 i, j, idx;
+
+ u32 loggeds = h->hits;
+ if (h->hits > CMP_MAP_H) loggeds = CMP_MAP_H;
+
+ u8 status;
+ // opt not in the paper
+ u32 fails = 0;
+
+ for (i = 0; i < loggeds; ++i) {
+
+ struct cmp_operands* o = &cmp_map->log[key][i];
+
+ // opt not in the paper
+ for (j = 0; j < i; ++j)
+ if (cmp_map->log[key][j].v0 == o->v0 && cmp_map->log[key][i].v1 == o->v1)
+ goto cmp_fuzz_next_iter;
+
+ for (idx = 0; idx < len && fails < 8; ++idx) {
+
+ if (unlikely(cmp_extend_encoding(h, o->v0, o->v1, idx, orig_buf, buf, len,
+ 1, &status)))
+ return 1;
+ if (status == 2)
+ ++fails;
+ else if (status == 1)
+ break;
+
+ if (unlikely(cmp_extend_encoding(h, o->v1, o->v0, idx, orig_buf, buf, len,
+ 1, &status)))
+ return 1;
+ if (status == 2)
+ ++fails;
+ else if (status == 1)
+ break;
+
+ }
+
+ cmp_fuzz_next_iter:
+ stage_cur++;
+
+ }
+
+ return 0;
+
+}
+
+///// Input to State stage
+
+// queue_cur->exec_cksum
+u8 input_to_state_stage(char** argv, u8* orig_buf, u8* buf, u32 len,
+ u32 exec_cksum) {
+
+ its_argv = argv;
+
+ if (unlikely(colorization(buf, len, exec_cksum))) return 1;
+
+ // do it manually, forkserver clear only trace_bits
+ memset(cmp_map->headers, 0, sizeof(cmp_map->headers));
+
+ if (unlikely(common_fuzz_cmplog_stuff(argv, buf, len))) return 1;
+
+ u64 orig_hit_cnt, new_hit_cnt;
+ u64 orig_execs = total_execs;
+ orig_hit_cnt = queued_paths + unique_crashes;
+
+ stage_name = "input-to-state";
+ stage_short = "its";
+ stage_max = 0;
+ stage_cur = 0;
+
+ u32 k;
+ for (k = 0; k < CMP_MAP_W; ++k) {
+
+ if (!cmp_map->headers[k].hits) continue;
+ if (cmp_map->headers[k].hits > CMP_MAP_H)
+ stage_max += CMP_MAP_H;
+ else
+ stage_max += cmp_map->headers[k].hits;
+
+ }
+
+ for (k = 0; k < CMP_MAP_W; ++k) {
+
+ if (!cmp_map->headers[k].hits) continue;
+ cmp_fuzz(k, orig_buf, buf, len);
+
+ }
+
+ memcpy(buf, orig_buf, len);
+
+ new_hit_cnt = queued_paths + unique_crashes;
+ stage_finds[STAGE_ITS] += new_hit_cnt - orig_hit_cnt;
+ stage_cycles[STAGE_ITS] += total_execs - orig_execs;
+
+ return 0;
+
+}