hopeful the final testcache improvement ...

This commit is contained in:
van Hauser
2020-10-26 15:24:33 +01:00
parent 44c65fa0a0
commit 868cb61ea6

View File

@ -979,21 +979,23 @@ inline u8 *queue_testcase_get(afl_state_t *afl, struct queue_entry *q) {
if (unlikely(!q->testcase_buf)) { if (unlikely(!q->testcase_buf)) {
/* Buf not cached, let's load it */ /* Buf not cached, let's load it */
u32 tid = afl->q_testcase_max_cache_count; u32 tid = afl->q_testcase_max_cache_count;
static u32 do_once = 0; // because even threaded we would want this. WIP
while (unlikely( while (unlikely(
afl->q_testcase_cache_size + len >= afl->q_testcase_max_cache_size || afl->q_testcase_cache_size + len >= afl->q_testcase_max_cache_size ||
afl->q_testcase_cache_count >= afl->q_testcase_max_cache_entries - 1)) { afl->q_testcase_cache_count >= afl->q_testcase_max_cache_entries - 1)) {
/* Cache full. We neet to evict one or more to map one. /* We want a max number of entries to the cache that we learn.
Get a random one which is not in use */ Very simple: once the cache is filled by size - that is the max. */
if (unlikely(afl->q_testcase_cache_size + len >= if (unlikely(afl->q_testcase_cache_size + len >=
afl->q_testcase_max_cache_size && afl->q_testcase_max_cache_size &&
(afl->q_testcase_cache_count < (afl->q_testcase_cache_count <
afl->q_testcase_max_cache_entries && afl->q_testcase_max_cache_entries &&
afl->q_testcase_max_cache_count < afl->q_testcase_max_cache_count <
afl->q_testcase_max_cache_entries))) { afl->q_testcase_max_cache_entries) &&
!do_once)) {
if (afl->q_testcase_max_cache_count > afl->q_testcase_cache_count) { if (afl->q_testcase_max_cache_count > afl->q_testcase_cache_count) {
@ -1006,8 +1008,19 @@ inline u8 *queue_testcase_get(afl_state_t *afl, struct queue_entry *q) {
} }
do_once = 1;
// release unneeded memory
u8 *ptr = ck_realloc(
afl->q_testcase_cache,
(afl->q_testcase_max_cache_entries + 1) * sizeof(size_t));
if (ptr) { afl->q_testcase_cache = (struct queue_entry **)ptr; }
} }
/* Cache full. We neet to evict one or more to map one.
Get a random one which is not in use */
do { do {
// if the cache (MB) is not enough for the queue then this gets // if the cache (MB) is not enough for the queue then this gets
@ -1065,11 +1078,16 @@ inline u8 *queue_testcase_get(afl_state_t *afl, struct queue_entry *q) {
afl->q_testcase_cache[tid] = q; afl->q_testcase_cache[tid] = q;
afl->q_testcase_cache_size += len; afl->q_testcase_cache_size += len;
++afl->q_testcase_cache_count; ++afl->q_testcase_cache_count;
if (tid >= afl->q_testcase_max_cache_count) if (likely(tid >= afl->q_testcase_max_cache_count)) {
afl->q_testcase_max_cache_count = tid + 1; afl->q_testcase_max_cache_count = tid + 1;
if (tid == afl->q_testcase_smallest_free)
} else if (unlikely(tid == afl->q_testcase_smallest_free)) {
afl->q_testcase_smallest_free = tid + 1; afl->q_testcase_smallest_free = tid + 1;
}
} }
return q->testcase_buf; return q->testcase_buf;
@ -1093,9 +1111,21 @@ inline void queue_testcase_store_mem(afl_state_t *afl, struct queue_entry *q,
} }
u32 tid = 0; u32 tid;
while (likely(afl->q_testcase_cache[tid] != NULL)) if (unlikely(afl->q_testcase_max_cache_count >=
afl->q_testcase_max_cache_entries)) {
// uh we were full, so now we have to search from start
tid = afl->q_testcase_smallest_free;
} else {
tid = afl->q_testcase_max_cache_count;
}
while (unlikely(afl->q_testcase_cache[tid] != NULL))
++tid; ++tid;
/* Map the test case into memory. */ /* Map the test case into memory. */
@ -1114,8 +1144,16 @@ inline void queue_testcase_store_mem(afl_state_t *afl, struct queue_entry *q,
afl->q_testcase_cache[tid] = q; afl->q_testcase_cache[tid] = q;
afl->q_testcase_cache_size += len; afl->q_testcase_cache_size += len;
++afl->q_testcase_cache_count; ++afl->q_testcase_cache_count;
if (tid >= afl->q_testcase_max_cache_count)
if (likely(tid >= afl->q_testcase_max_cache_count)) {
afl->q_testcase_max_cache_count = tid + 1; afl->q_testcase_max_cache_count = tid + 1;
} else if (unlikely(tid == afl->q_testcase_smallest_free)) {
afl->q_testcase_smallest_free = tid + 1;
}
} }