aboutsummaryrefslogtreecommitdiffstats
path: root/tools
diff options
context:
space:
mode:
authorTaras Madan <tarasmadan@google.com>2025-04-02 20:28:15 +0200
committerTaras Madan <tarasmadan@google.com>2025-04-04 14:52:42 +0000
commitc53ea9c902c6745d750c5700631e1556716f4cd4 (patch)
tree855f9483db0aeedee6fe7dbd34e746356db13085 /tools
parent1c4febdbbe8f34403baef506b22cd689576927f6 (diff)
pkg/covermerger: read gcs files sequentially
Quarter long aggregation means thousands of gzip files. Opening all the files in parallel we struggle from: 1. Memory overhead. 2. GCS API errors. It can't read Attrs for 1500+ files.
Diffstat (limited to 'tools')
-rw-r--r--tools/syz-covermerger/syz_covermerger.go12
1 files changed, 4 insertions, 8 deletions
diff --git a/tools/syz-covermerger/syz_covermerger.go b/tools/syz-covermerger/syz_covermerger.go
index 6cd714421..7b51c9997 100644
--- a/tools/syz-covermerger/syz_covermerger.go
+++ b/tools/syz-covermerger/syz_covermerger.go
@@ -71,21 +71,17 @@ func do() error {
panic(fmt.Sprintf("failed to parse time_to: %s", err.Error()))
}
dateFrom = dateTo.AddDays(-int(*flagDuration))
- dbReader := covermerger.MakeBQCSVReader()
- if err = dbReader.InitNsRecords(context.Background(),
+ csvReader, err := covermerger.InitNsRecords(context.Background(),
*flagNamespace,
*flagFilePathPrefix,
"",
dateFrom,
dateTo,
- ); err != nil {
+ )
+ if err != nil {
panic(fmt.Sprintf("failed to dbReader.InitNsRecords: %v", err.Error()))
}
- defer dbReader.Close()
- csvReader, errReader := dbReader.Reader()
- if errReader != nil {
- panic(fmt.Sprintf("failed to dbReader.Reader: %v", errReader.Error()))
- }
+ defer csvReader.Close()
var wc io.WriteCloser
url := *flagToGCS
if *flagToDashAPI != "" {