You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I left honggfuzz running overnight, and it found a few crashes (5 different crashes), but my corpus directory ended up with 73k files in the corpus directory (>500MiB).
Now, I fixed all the crashes, and wanted to resume the fuzzing, but it has to run in dry run mode until it processes the entire corpus. Running the minimizer doesn't help because it runs single-threaded, it would take around a day to process the 72k files on my computer (it is going to take ~2 hours to finish the dry run mode with 6 threads).
One idea would be to fix issue #401 , so that minimization can be performed in a reasonable time.
Some other options could be running minimization during the fuzzing process so we don't end up with corpus that is too big? Or allowing the fuzzer to start fuzzing without finishing the dry run... after it has a couple thousand files, it surely has enough info to start fuzzing.
The text was updated successfully, but these errors were encountered:
honggfuzz --input input_corpus --output output_corpus -- bin
When it switches from state 1/3 and 2/3 to 3/3 - then in output_corpus you'll have somewhat minimized corpus, and all of that has been done in parallel.
It's not as good as -M, because samples are not sorted for minimization, but it might work for your use-case
What are states 1/3, 2/3 and 3/3 ? Is this how the honggfuzz is commonly used? I feel like there's something that could be improved here, as every time I restart the fuzzer, it now takes a couple of hours to start properly fuzzing, as the corpus size keeps increasing.
At some point, maybe it makes sense to start fuzzing and mutating before processing all the input corpus? And maybe minimizing it at the same time?
I left honggfuzz running overnight, and it found a few crashes (5 different crashes), but my corpus directory ended up with 73k files in the corpus directory (>500MiB).
Now, I fixed all the crashes, and wanted to resume the fuzzing, but it has to run in dry run mode until it processes the entire corpus. Running the minimizer doesn't help because it runs single-threaded, it would take around a day to process the 72k files on my computer (it is going to take ~2 hours to finish the dry run mode with 6 threads).
One idea would be to fix issue #401 , so that minimization can be performed in a reasonable time.
Some other options could be running minimization during the fuzzing process so we don't end up with corpus that is too big? Or allowing the fuzzer to start fuzzing without finishing the dry run... after it has a couple thousand files, it surely has enough info to start fuzzing.
The text was updated successfully, but these errors were encountered: