aboutsummaryrefslogtreecommitdiffstats
path: root/executor/executor.cc
diff options
context:
space:
mode:
authorAleksandr Nogikh <nogikh@google.com>2021-12-02 10:39:37 +0000
committerAleksandr Nogikh <wp32pw@gmail.com>2021-12-03 12:32:03 +0100
commitc7c20675f58e3edaa53538928c0963144fd524e5 (patch)
tree0552f0896774be7b0d933e95432e63ff8ef13ba4 /executor/executor.cc
parent6df0f018b545aaf2b9bbcfde8b6e530cf90da9be (diff)
executor: reserve fds that will belong to kcov
As now kcov instances may get set up during fuzzing, performing dup2 in cover_open is no longer safe as it may close some important resource. Prevent that by reserving most of fds that belong to the kcov fds range. Unfortunately we must duplicate the code because of the way kcov implementations are organized.
Diffstat (limited to 'executor/executor.cc')
-rw-r--r--executor/executor.cc11
1 files changed, 8 insertions, 3 deletions
diff --git a/executor/executor.cc b/executor/executor.cc
index ac54ce62e..bfda1cbd6 100644
--- a/executor/executor.cc
+++ b/executor/executor.cc
@@ -68,6 +68,7 @@ typedef unsigned char uint8;
const int kMaxFd = 250;
const int kMaxThreads = 32;
const int kPreallocCoverThreads = 3; // the number of kcov instances to be set up during init
+const int kReserveCoverFds = 16; // a compromise between extra fds and the likely neded kcov instances
const int kInPipeFd = kMaxFd - 1; // remapped from stdin
const int kOutPipeFd = kMaxFd - 2; // remapped from stdout
const int kCoverFd = kOutPipeFd - kMaxThreads;
@@ -470,10 +471,14 @@ int main(int argc, char** argv)
if (flag_coverage) {
for (int i = 0; i < kMaxThreads; i++) {
threads[i].cov.fd = kCoverFd + i;
- // Pre-setup coverage collection for some threads. This should be enough for almost
- // all programs, for the remaning few ones coverage will be set up when it's needed.
- if (i < kPreallocCoverThreads)
+ if (i < kPreallocCoverThreads) {
+ // Pre-setup coverage collection for some threads. This should be enough for almost
+ // all programs, for the remaning few ones coverage will be set up when it's needed.
thread_setup_cover(&threads[i]);
+ } else if (i < kReserveCoverFds) {
+ // Ensure that these fds won't be taken during fuzzing or by init routines.
+ cover_reserve_fd(&threads[i].cov);
+ }
}
cover_open(&extra_cov, true);
cover_protect(&extra_cov);