1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
|
# CLAUDE.md
This file provides guidance to Claude Code (claude.ai/code) when working with code in this repository.
## Project Overview
DTail (Distributed Tail) is a DevOps tool written in Go for distributed log operations across multiple servers. It provides secure, concurrent access to logs on many machines using SSH protocol, supporting tail, cat, grep, and MapReduce operations.
## Build Commands
```bash
# Build all binaries
make build
# Build individual components
make dtail # Client for tailing log files
make dserver # Server component (required on target machines)
make dcat # Client for displaying files
make dgrep # Client for searching files
make dmap # Client for MapReduce queries
make dtailhealth # Health check client
# Clean build artifacts
make clean
# Enable ACL support (requires libacl-devel)
DTAIL_USE_ACL=yes make build
# Enable proprietary features
DTAIL_USE_PROPRIETARY=yes make build
# Build PGO-optimized binaries (requires existing profiles)
make build-pgo
# Generate PGO profiles and build optimized binaries
make pgo
```
## Testing & Development
```bash
# Run all tests (unit tests only)
make test
# Run all tests including integration tests
# IMPORTANT: Always rebuild binaries before running integration tests
make clean && make build
DTAIL_INTEGRATION_TEST_RUN_MODE=yes make test
# Quick integration test workflow (recommended)
make build && DTAIL_INTEGRATION_TEST_RUN_MODE=yes make test
# Run linting
make lint
# Run go vet
make vet
# Run integration tests individually (requires binaries built first)
cd integrationtests && go test
```
## Benchmarking
```bash
# Run all benchmarks
make benchmark
# Quick benchmarks (subset of tests)
make benchmark-quick
# Full benchmarks with longer runs
make benchmark-full
# Create a baseline for comparison
make benchmark-baseline
# Compare current performance against a baseline
make benchmark-compare BASELINE=benchmarks/baselines/baseline_TIMESTAMP.txt
```
## Profile-Guided Optimization (PGO)
```bash
# Full PGO workflow: generate profiles and build optimized binaries
make pgo
# Quick PGO with smaller datasets (faster)
make pgo-quick
# PGO for specific commands only
make pgo-commands COMMANDS='dcat dgrep'
# Generate PGO profiles only (without building)
make pgo-generate
# Build PGO-optimized binaries using existing profiles
make build-pgo
# Install PGO-optimized binaries to system
make install-pgo
# Clean PGO artifacts
make pgo-clean
# Show PGO help
make pgo-help
```
### PGO Notes
- PGO provides additional performance improvements on top of turbo mode
- Typical improvements: 5-10% for DCat, up to 19% for DGrep with low hit rates
- Profiles are saved in `pgo-profiles/` directory
- Optimized binaries are built in `pgo-build/` directory
- Use `make build-pgo` to rebuild optimized binaries without regenerating profiles
- PGO profiles are workload-specific; consider custom profiles for your use case
## Profiling
```bash
# Profile all commands (dcat, dgrep, dmap)
make profile-all
# Profile individual commands
make profile-dcat # Profile dcat with test data
make profile-dgrep # Profile dgrep with test data
make profile-dmap # Profile dmap MapReduce queries
# Quick profiling with smaller datasets
make profile-quick
# Full automated profiling (includes larger files)
make profile-auto
# Clean all profile data
make profile-clean
# Analyze a specific profile interactively
make profile-analyze PROFILE=profiles/dcat_cpu_*.prof
# Generate flame graph visualization
make profile-flamegraph PROFILE=profiles/dcat_cpu_*.prof
# Custom profiling options
PROFILE_SIZE=10000000 make profile-all # Profile with 10M lines
PROFILE_DIR=myprofiles make profile-dcat # Custom profile directory
# Show all profiling options
make profile-help
```
### Profiling Notes
- Profiles are saved in the `profiles/` directory by default
- Each command generates CPU, memory, and allocation profiles
- Use `go tool pprof` for detailed analysis of profile files
## Test Execution Details
- Integration tests require binaries to be built before execution
- **IMPORTANT:** Always recompile binaries after code changes before running integration tests:
```bash
make clean && make build
DTAIL_INTEGRATION_TEST_RUN_MODE=yes make test
```
- Integration tests are run by setting DTAIL_INTEGRATION_TEST_RUN_MODE to yes, and by running 'make test'
- Integration tests verify: DCat, DGrep, DMap (MapReduce), DServer, DTail, DTailHealth functionality
- All tests run with race detection enabled (`--race` flag)
## Known Limitations
### Turbo Mode and MapReduce Operations
Turbo boost mode is enabled by default and provides performance optimizations for both direct output operations (cat, grep, tail) and MapReduce operations when running in server mode. It can be explicitly disabled via DTAIL_TURBOBOOST_DISABLE=yes or TurboBoostDisable in the config file.
**Technical Details:**
- For cat/grep/tail: Turbo mode bypasses channels for direct writing
- For MapReduce in server mode: Turbo mode uses direct line processing without channels
- For MapReduce in serverless/client mode: Turbo mode is not applicable (client-side aggregation doesn't use server optimizations)
**Server-Side Turbo MapReduce:**
When turbo mode is enabled and using dserver:
- Lines are processed directly without channel overhead
- Batch processing reduces lock contention
- Memory pooling reduces garbage collection pressure
- Same output format and accuracy as regular MapReduce
**High-Concurrency MapReduce Improvements:**
- Channel recycling ensures proper draining before reuse (regular mode)
- Buffer sizes increased from 1,000 to 10,000 for better concurrency
- Direct processing in turbo mode eliminates channel bottlenecks
- Improved synchronization in both regular and turbo aggregate implementations
**Best Practices for High-Concurrency MapReduce:**
1. Turbo boost is enabled by default. To disable if needed: `export DTAIL_TURBOBOOST_DISABLE=yes`
2. Increase MaxConcurrentCats in the server configuration to match workload
3. Use server mode for large-scale MapReduce operations
4. Monitor logs for performance metrics
## Benchmarking & Profiling
```bash
# Run benchmarks
make benchmark
# Run performance profiling
make profile
# Generate profiling reports
make profile-report
# Run specific benchmark suites
make benchmark-network
make benchmark-mapreduce
make benchmark-ssh
```
## Profile-Guided Optimization (PGO)
```bash
# Run PGO for all commands
make pgo
# Quick PGO with smaller datasets
make pgo-quick
# PGO for specific commands
make pgo-commands COMMANDS='dcat dgrep'
# Clean PGO artifacts
make pgo-clean
# Show PGO help
make pgo-help
# Direct usage with dtail-tools
dtail-tools pgo # Optimize all commands
dtail-tools pgo dcat dgrep # Optimize specific commands
dtail-tools pgo -v -iterations 5 # Verbose with 5 iterations
# After PGO, optimized binaries are in pgo-build/
```
### PGO Notes
- PGO uses profile data from real workloads to optimize binary performance
- The process involves: building baseline → generating profiles → building with PGO
- Typical improvements range from 5-20% depending on the workload
- Optimized binaries are placed in the `pgo-build/` directory
## Architecture & Code Organization
### Binary Entry Points
- `/cmd/dtail/` - Remote log tailing client
- `/cmd/dserver/` - Server daemon
- `/cmd/dcat/` - Remote file reading client
- `/cmd/dgrep/` - Remote file searching client
- `/cmd/dmap/` - MapReduce query client
- `/cmd/dtailhealth/` - Health check client
### Core Implementation
- `/internal/clients/` - Client implementations for each tool
- `/internal/server/` - Server daemon logic
- `/internal/mapr/` - MapReduce engine and query parsing
- `/internal/ssh/` - SSH client/server components
- `/internal/config/` - Configuration management
- `/internal/io/` - File operations, logging, compression handling
### Key Architectural Patterns
1. **Client-Server Communication**: All clients communicate with dserver instances via SSH protocol on port 2222 (configurable)
2. **MapReduce Query Engine**: Located in `/internal/mapr/`, implements SQL-like query language for distributed log aggregation
3. **Configuration System**: JSON-based configuration in `/internal/config/`, supports both client and server settings
4. **SSH Integration**: Custom SSH server implementation in `/internal/ssh/server/` and client in `/internal/ssh/client/`
5. **Compression Support**: Automatic handling of gzip and zstd compressed files in `/internal/io/`
## Important Implementation Details
- **Main Server Loop**: `/internal/server/server.go` - Core server processing logic
- **Client Base**: `/internal/clients/baseClient.go` - Common client functionality
- **MapReduce Parser**: `/internal/mapr/parse/` - SQL-like query language parser
- **Log Format Parsers**: `/internal/mapr/logformat/` - Extensible log parsing system
- **SSH Authorization**: `/internal/server/user/authsshkey.go` - SSH key validation
## Configuration Files
- Server config: `/etc/dserver/dtail.json` or `./dtail.json`
- Example configs: `/examples/`
- Docker configs: `/docker/`
## Common Development Tasks
When modifying client behavior:
1. Check `/internal/clients/` for the specific client implementation
2. Common functionality is in `baseClient.go`
3. Client-specific logic is in respective files (e.g., `tail.go`, `cat.go`)
When modifying server behavior:
1. Core server logic is in `/internal/server/server.go`
2. User authentication in `/internal/server/user/`
3. Handler implementations in `/internal/server/handlers/`
When working with MapReduce:
1. Query parsing in `/internal/mapr/parse/`
2. Aggregation logic in `/internal/mapr/reducer/`
3. Log format parsing in `/internal/mapr/logformat/`
|