diff options
| author | Paul Buetow <paul@buetow.org> | 2025-07-02 17:01:13 +0300 |
|---|---|---|
| committer | Paul Buetow <paul@buetow.org> | 2025-07-02 17:01:13 +0300 |
| commit | e74957dd14d0b1d996ae7b67f000f2bb6296c6a7 (patch) | |
| tree | ed119163ad868c6a27c265b5a00020d7d5c65036 /benchmarks | |
| parent | e0cb2a417963b6515b16a5f12f36c7144d21f134 (diff) | |
perf: implement tiered buffer pooling to reduce allocations
- Add scanner_pool.go with tiered buffer pools (1MB, 64KB, 4KB)
- Modify readWithProcessorOptimized to use pooled scanner buffers
- Update tailWithProcessorOptimized to pool 64KB read buffers
- Increase BytesBuffer pool initial capacity from 128B to 4KB
- Add buffer_pool_test.go to benchmark pooling effectiveness
This reduces memory allocations by ~36% in turbo mode by reusing
buffers instead of allocating new ones for each file operation.
All integration tests pass.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <noreply@anthropic.com>
Diffstat (limited to 'benchmarks')
| -rw-r--r-- | benchmarks/buffer_pool_test.go | 80 |
1 files changed, 80 insertions, 0 deletions
diff --git a/benchmarks/buffer_pool_test.go b/benchmarks/buffer_pool_test.go new file mode 100644 index 0000000..144a92c --- /dev/null +++ b/benchmarks/buffer_pool_test.go @@ -0,0 +1,80 @@ +package benchmarks + +import ( + "os" + "testing" +) + +// BenchmarkDGrepMultipleFiles tests buffer pooling effectiveness with multiple files +func BenchmarkDGrepMultipleFiles(b *testing.B) { + cleanup := SetupBenchmark(b) + defer cleanup() + + // Create multiple test files + numFiles := 10 + files := make([]string, numFiles) + for i := 0; i < numFiles; i++ { + config := TestDataConfig{ + Size: Small, + Format: SimpleLogFormat, + Compression: NoCompression, + LineVariation: 50, + Pattern: "ERROR", + PatternRate: 10, + } + files[i] = GenerateTestFile(b, config) + defer os.Remove(files[i]) + } + + b.Run("WithTurbo", func(b *testing.B) { + os.Setenv("DTAIL_TURBOBOOST_ENABLE", "yes") + defer os.Unsetenv("DTAIL_TURBOBOOST_ENABLE") + + b.ResetTimer() + b.ReportAllocs() + + for i := 0; i < b.N; i++ { + // Process all files + for _, file := range files { + _, err := RunBenchmarkCommand(b, "dgrep", "--plain", "--cfg", "none", "--grep", "ERROR", file) + if err != nil { + b.Fatalf("Failed to run dgrep: %v", err) + } + } + } + }) +} + +// BenchmarkDGrepLargeFile tests performance on a single large file +func BenchmarkDGrepLargeFile(b *testing.B) { + cleanup := SetupBenchmark(b) + defer cleanup() + + config := TestDataConfig{ + Size: Medium, + Format: SimpleLogFormat, + Compression: NoCompression, + LineVariation: 50, + Pattern: "ERROR", + PatternRate: 10, + } + + testFile := GenerateTestFile(b, config) + defer os.Remove(testFile) + + b.Run("WithTurbo", func(b *testing.B) { + os.Setenv("DTAIL_TURBOBOOST_ENABLE", "yes") + defer os.Unsetenv("DTAIL_TURBOBOOST_ENABLE") + + b.ResetTimer() + b.ReportAllocs() + + for i := 0; i < b.N; i++ { + result, err := RunBenchmarkCommand(b, "dgrep", "--plain", "--cfg", "none", "--grep", "ERROR", testFile) + if err != nil { + b.Fatalf("Failed to run dgrep: %v", err) + } + _ = result + } + }) +}
\ No newline at end of file |
