feat(benchmarks): auto-generate BENCHMARK_REPORT.md at end of run
- Add benchmarks/reports/REPORT_TEMPLATE.md — template with __MARKER__ placeholders for all auto-populated fields (latency, throughput, percentiles, cache delta, benchstat block, environment, ONIX version) - Add benchmarks/tools/generate_report.go — reads latency_report.csv, throughput_report.csv, benchstat_summary.txt and run1.txt metadata, fills the template, and writes BENCHMARK_REPORT.md to the results dir. ONIX version sourced from the latest git tag (falls back to 'dev'). - Update run_benchmarks.sh to call generate_report.go after parse_results.go; also derive ONIX_VERSION from git tag and pass to generator - Update README and directory layout to reflect new files and workflow
This commit is contained in:
@@ -100,11 +100,14 @@ benchmarks/
|
||||
│ ├── init_request.json
|
||||
│ └── confirm_request.json
|
||||
├── tools/
|
||||
│ └── parse_results.go ← CSV exporter for latency + throughput data
|
||||
├── reports/ ← committed benchmark reports
|
||||
│ ├── parse_results.go ← CSV exporter for latency + throughput data
|
||||
│ └── generate_report.go ← fills REPORT_TEMPLATE.md with run data
|
||||
├── reports/ ← committed benchmark reports and template
|
||||
│ ├── REPORT_TEMPLATE.md ← template used to generate each run's report
|
||||
│ └── REPORT_ONIX_v150.md ← baseline report (Apple M5, Beckn v2.0.0)
|
||||
└── results/ ← gitignored; created by run_benchmarks.sh
|
||||
└── <timestamp>/
|
||||
├── BENCHMARK_REPORT.md — generated human-readable report
|
||||
├── run1.txt, run2.txt, run3.txt — raw go test -bench output
|
||||
├── parallel_cpu*.txt — concurrency sweep
|
||||
├── benchstat_summary.txt — statistical aggregation
|
||||
@@ -122,11 +125,11 @@ Committed reports are stored in `benchmarks/reports/`. Each report documents the
|
||||
|------|----------|-----------------|
|
||||
| `REPORT_ONIX_v150.md` | Apple M5 · darwin/arm64 · GOMAXPROCS=10 | beckn-onix v1.5.0 |
|
||||
|
||||
To add a new report after a benchmark run:
|
||||
1. Run `bash benchmarks/run_benchmarks.sh` — results appear in `benchmarks/results/<timestamp>/`.
|
||||
2. Review `benchstat_summary.txt` and the CSV files.
|
||||
3. Write a report (see the existing report as a template) and save it as `benchmarks/reports/REPORT_<tag>.md`.
|
||||
4. Commit only the report file; `benchmarks/results/` remains gitignored.
|
||||
The script auto-generates `BENCHMARK_REPORT.md` in each results directory using `REPORT_TEMPLATE.md`. To permanently record a run:
|
||||
1. Run `bash benchmarks/run_benchmarks.sh` — `BENCHMARK_REPORT.md` is generated automatically.
|
||||
2. Review it, fill in the B5 bottleneck analysis section.
|
||||
3. Copy it to `benchmarks/reports/REPORT_<tag>.md` and commit.
|
||||
4. `benchmarks/results/` stays gitignored; only the curated report goes in.
|
||||
|
||||
---
|
||||
|
||||
|
||||
148
benchmarks/reports/REPORT_TEMPLATE.md
Normal file
148
benchmarks/reports/REPORT_TEMPLATE.md
Normal file
@@ -0,0 +1,148 @@
|
||||
# beckn-onix Adapter — Benchmark Report
|
||||
|
||||
> **Run:** `__TIMESTAMP__`
|
||||
> **Platform:** __CPU__ · __GOOS__/__GOARCH__ · GOMAXPROCS=__GOMAXPROCS__ (default)
|
||||
> **Adapter version:** __ONIX_VERSION__
|
||||
> **Beckn Protocol:** v2.0.0
|
||||
|
||||
---
|
||||
|
||||
## Part A — Executive Summary
|
||||
|
||||
### What Was Tested
|
||||
|
||||
The beckn-onix ONIX adapter was benchmarked end-to-end using Go's native `testing.B`
|
||||
framework and `net/http/httptest`. Requests flowed through a real compiled adapter —
|
||||
with all production plugins active — against in-process mock servers, isolating
|
||||
adapter-internal latency from network variables.
|
||||
|
||||
**Pipeline tested (bapTxnCaller):** `addRoute → sign → validateSchema`
|
||||
|
||||
**Plugins active:** `router`, `signer`, `simplekeymanager`, `cache` (miniredis), `schemav2validator`
|
||||
|
||||
**Actions benchmarked:** `discover`, `select`, `init`, `confirm`
|
||||
|
||||
### Key Results
|
||||
|
||||
| Metric | Value |
|
||||
|--------|-------|
|
||||
| Serial p50 latency (discover) | **__P50_US__ µs** |
|
||||
| Serial p95 latency (discover) | **__P95_US__ µs** |
|
||||
| Serial p99 latency (discover) | **__P99_US__ µs** |
|
||||
| Serial mean latency (discover) | **__MEAN_DISCOVER_US__ µs** |
|
||||
| Peak parallel throughput | **__PEAK_RPS__ req/s** |
|
||||
| Cache warm vs cold delta | **__CACHE_DELTA__** |
|
||||
| Memory per request (discover) | **~__MEM_DISCOVER_KB__ KB · __ALLOCS_DISCOVER__ allocs** |
|
||||
|
||||
### Interpretation
|
||||
|
||||
_Review the numbers above and add interpretation here._
|
||||
|
||||
### Recommendation
|
||||
|
||||
_Add sizing and tuning recommendations here._
|
||||
|
||||
---
|
||||
|
||||
## Part B — Technical Detail
|
||||
|
||||
### B0 — Test Environment
|
||||
|
||||
| Parameter | Value |
|
||||
|-----------|-------|
|
||||
| CPU | __CPU__ (__GOARCH__) |
|
||||
| OS | __GOOS__/__GOARCH__ |
|
||||
| Go package | `github.com/beckn-one/beckn-onix/benchmarks/e2e` |
|
||||
| Default GOMAXPROCS | __GOMAXPROCS__ |
|
||||
| Benchmark timeout | 30 minutes |
|
||||
| Serial run duration | 10s per benchmark × 3 runs |
|
||||
| Parallel sweep duration | 30s per GOMAXPROCS level |
|
||||
| GOMAXPROCS sweep | 1, 2, 4, 8, 16 |
|
||||
| Redis | miniredis (in-process, no network) |
|
||||
| BPP | httptest mock (instant ACK) |
|
||||
| Registry | httptest mock (dev key pair) |
|
||||
| Schema spec | Beckn v2.0.0 OpenAPI (`beckn.yaml`, local file) |
|
||||
|
||||
**Plugins and steps (bapTxnCaller):**
|
||||
|
||||
| Step | Plugin | Role |
|
||||
|------|--------|------|
|
||||
| 1 | `router` | Resolves BPP URL from routing config |
|
||||
| 2 | `signer` + `simplekeymanager` | Signs request body (Ed25519/BLAKE-512) |
|
||||
| 3 | `schemav2validator` | Validates Beckn v2.0 API schema |
|
||||
|
||||
---
|
||||
|
||||
### B1 — Latency by Action
|
||||
|
||||
Averages from `run1.txt` (10s, GOMAXPROCS=__GOMAXPROCS__). Percentile values from `percentiles.txt`.
|
||||
|
||||
| Action | Mean (µs) | p50 (µs) | p95 (µs) | p99 (µs) | Allocs/req | Bytes/req |
|
||||
|--------|----------:|--------:|--------:|--------:|----------:|----------:|
|
||||
| discover (serial) | __MEAN_DISCOVER_US__ | __P50_US__ | __P95_US__ | __P99_US__ | __ALLOCS_DISCOVER__ | __BYTES_DISCOVER__ (~__MEM_DISCOVER_KB__ KB) |
|
||||
| select | __MEAN_SELECT_US__ | — | — | — | __ALLOCS_SELECT__ | __BYTES_SELECT__ (~__MEM_SELECT_KB__ KB) |
|
||||
| init | __MEAN_INIT_US__ | — | — | — | __ALLOCS_INIT__ | __BYTES_INIT__ (~__MEM_INIT_KB__ KB) |
|
||||
| confirm | __MEAN_CONFIRM_US__ | — | — | — | __ALLOCS_CONFIRM__ | __BYTES_CONFIRM__ (~__MEM_CONFIRM_KB__ KB) |
|
||||
|
||||
---
|
||||
|
||||
### B2 — Throughput vs Concurrency
|
||||
|
||||
Results from the concurrency sweep (`parallel_cpu*.txt`, 30s per level).
|
||||
|
||||
__THROUGHPUT_TABLE__
|
||||
|
||||
---
|
||||
|
||||
### B3 — Cache Impact (Redis warm vs cold)
|
||||
|
||||
Results from `cache_comparison.txt` (10s each, GOMAXPROCS=__GOMAXPROCS__).
|
||||
|
||||
| Scenario | Mean (µs) | Allocs/req | Bytes/req |
|
||||
|----------|----------:|-----------:|----------:|
|
||||
| CacheWarm | __CACHE_WARM_US__ | __CACHE_WARM_ALLOCS__ | __CACHE_WARM_BYTES__ |
|
||||
| CacheCold | __CACHE_COLD_US__ | __CACHE_COLD_ALLOCS__ | __CACHE_COLD_BYTES__ |
|
||||
| **Delta** | **__CACHE_DELTA__** | — | — |
|
||||
|
||||
---
|
||||
|
||||
### B4 — benchstat Statistical Summary (3 Runs)
|
||||
|
||||
```
|
||||
__BENCHSTAT_SUMMARY__
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
### B5 — Bottleneck Analysis
|
||||
|
||||
> Populate after reviewing the numbers above and profiling with `go tool pprof`.
|
||||
|
||||
| Rank | Plugin / Step | Estimated contribution | Evidence |
|
||||
|:----:|---------------|------------------------|---------|
|
||||
| 1 | | | |
|
||||
| 2 | | | |
|
||||
| 3 | | | |
|
||||
|
||||
**Profiling commands:**
|
||||
|
||||
```bash
|
||||
# CPU profile
|
||||
go test ./benchmarks/e2e/... -bench=BenchmarkBAPCaller_Discover \
|
||||
-benchtime=30s -cpuprofile=benchmarks/results/cpu.prof -timeout=5m
|
||||
go tool pprof -http=:6060 benchmarks/results/cpu.prof
|
||||
|
||||
# Memory profile
|
||||
go test ./benchmarks/e2e/... -bench=BenchmarkBAPCaller_Discover \
|
||||
-benchtime=30s -memprofile=benchmarks/results/mem.prof -timeout=5m
|
||||
go tool pprof -http=:6060 benchmarks/results/mem.prof
|
||||
|
||||
# Lock contention (find serialisation under parallel load)
|
||||
go test ./benchmarks/e2e/... -bench=BenchmarkBAPCaller_Discover_Parallel \
|
||||
-benchtime=30s -mutexprofile=benchmarks/results/mutex.prof -timeout=5m
|
||||
go tool pprof -http=:6060 benchmarks/results/mutex.prof
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
*Generated from run `__TIMESTAMP__` · beckn-onix __ONIX_VERSION__ · Beckn Protocol v2.0.0*
|
||||
@@ -23,10 +23,13 @@ REPO_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/.." && pwd)"
|
||||
RESULTS_DIR="$REPO_ROOT/benchmarks/results/$(date +%Y-%m-%d_%H-%M-%S)"
|
||||
BENCH_PKG="./benchmarks/e2e/..."
|
||||
BENCH_TIMEOUT="10m"
|
||||
# ── Smoke-test values (small): swap to the "full run" values below once stable ─
|
||||
BENCH_TIME_SERIAL="10s" # full run: 10s
|
||||
BENCH_TIME_PARALLEL="30s" # full run: 30s
|
||||
BENCH_COUNT=1 # full run: keep at 1; benchstat uses the 3 serial files
|
||||
BENCH_TIME_SERIAL="10s"
|
||||
BENCH_TIME_PARALLEL="30s"
|
||||
BENCH_COUNT=1 # benchstat uses the 3 serial files for stability
|
||||
|
||||
# Adapter version — reads from git tag, falls back to "dev"
|
||||
ONIX_VERSION="$(git -C "$REPO_ROOT" describe --tags --abbrev=0 2>/dev/null || echo "dev")"
|
||||
REPORT_TEMPLATE="$REPO_ROOT/benchmarks/reports/REPORT_TEMPLATE.md"
|
||||
|
||||
cd "$REPO_ROOT"
|
||||
|
||||
@@ -120,11 +123,21 @@ echo " Saved → $RESULTS_DIR/benchstat_summary.txt"
|
||||
echo ""
|
||||
|
||||
# ── Parse results to CSV ──────────────────────────────────────────────────────
|
||||
if command -v go &>/dev/null; then
|
||||
echo "Parsing results to CSV..."
|
||||
go run benchmarks/tools/parse_results.go \
|
||||
echo "Parsing results to CSV..."
|
||||
go run "$REPO_ROOT/benchmarks/tools/parse_results.go" \
|
||||
-dir="$RESULTS_DIR" \
|
||||
-out="$RESULTS_DIR" 2>&1 || echo " (parse_results.go: skipping on error)"
|
||||
echo ""
|
||||
|
||||
# ── Generate human-readable report ───────────────────────────────────────────
|
||||
echo "Generating benchmark report..."
|
||||
if [[ -f "$REPORT_TEMPLATE" ]]; then
|
||||
go run "$REPO_ROOT/benchmarks/tools/generate_report.go" \
|
||||
-dir="$RESULTS_DIR" \
|
||||
-out="$RESULTS_DIR" 2>&1 || echo " (parse_results.go: optional step, skipping if errors)"
|
||||
-template="$REPORT_TEMPLATE" \
|
||||
-version="$ONIX_VERSION" 2>&1 || echo " (generate_report.go: skipping on error)"
|
||||
else
|
||||
echo " WARNING: template not found at $REPORT_TEMPLATE — skipping report generation"
|
||||
fi
|
||||
|
||||
# ── Summary ───────────────────────────────────────────────────────────────────
|
||||
@@ -143,11 +156,14 @@ echo "Results written to:"
|
||||
echo " $RESULTS_DIR"
|
||||
echo ""
|
||||
echo "Key files:"
|
||||
echo " BENCHMARK_REPORT.md — generated human-readable report"
|
||||
echo " benchstat_summary.txt — statistical analysis of 3 serial runs"
|
||||
echo " parallel_cpu*.txt — concurrency sweep results"
|
||||
echo " latency_report.csv — per-benchmark latency and allocation data"
|
||||
echo " throughput_report.csv — RPS and latency by GOMAXPROCS level"
|
||||
echo " parallel_cpu*.txt — concurrency sweep raw output"
|
||||
echo " percentiles.txt — p50/p95/p99 latency data"
|
||||
echo " cache_comparison.txt — warm vs cold Redis cache comparison"
|
||||
echo ""
|
||||
echo "To view benchstat summary:"
|
||||
echo " cat $RESULTS_DIR/benchstat_summary.txt"
|
||||
echo "To review the report:"
|
||||
echo " open $RESULTS_DIR/BENCHMARK_REPORT.md"
|
||||
echo "========================================"
|
||||
|
||||
401
benchmarks/tools/generate_report.go
Normal file
401
benchmarks/tools/generate_report.go
Normal file
@@ -0,0 +1,401 @@
|
||||
// generate_report.go — Fills REPORT_TEMPLATE.md with data from a completed
|
||||
// benchmark run and writes BENCHMARK_REPORT.md to the results directory.
|
||||
//
|
||||
// Usage:
|
||||
//
|
||||
// go run benchmarks/tools/generate_report.go \
|
||||
// -dir=benchmarks/results/<timestamp>/ \
|
||||
// -template=benchmarks/reports/REPORT_TEMPLATE.md \
|
||||
// -version=<onix-version>
|
||||
//
|
||||
// The generator reads:
|
||||
// - latency_report.csv — per-benchmark latency and allocation data
|
||||
// - throughput_report.csv — RPS and latency by GOMAXPROCS level
|
||||
// - benchstat_summary.txt — raw benchstat output block
|
||||
// - run1.txt — goos / goarch / cpu metadata
|
||||
//
|
||||
// Placeholders filled in the template:
|
||||
//
|
||||
// __TIMESTAMP__ results dir basename (YYYY-MM-DD_HH-MM-SS)
|
||||
// __ONIX_VERSION__ -version flag value
|
||||
// __GOOS__ from run1.txt header
|
||||
// __GOARCH__ from run1.txt header
|
||||
// __CPU__ from run1.txt header
|
||||
// __GOMAXPROCS__ derived from the benchmark name suffix in run1.txt
|
||||
// __P50_US__ p50 latency in µs (from Discover_Percentiles row)
|
||||
// __P95_US__ p95 latency in µs
|
||||
// __P99_US__ p99 latency in µs
|
||||
// __MEAN_DISCOVER_US__ mean latency in µs for discover
|
||||
// __MEAN_SELECT_US__ mean latency in µs for select
|
||||
// __MEAN_INIT_US__ mean latency in µs for init
|
||||
// __MEAN_CONFIRM_US__ mean latency in µs for confirm
|
||||
// __ALLOCS_DISCOVER__ allocs/req for discover
|
||||
// __ALLOCS_SELECT__ allocs/req for select
|
||||
// __ALLOCS_INIT__ allocs/req for init
|
||||
// __ALLOCS_CONFIRM__ allocs/req for confirm
|
||||
// __BYTES_DISCOVER__ bytes/req for discover
|
||||
// __BYTES_SELECT__ bytes/req for select
|
||||
// __BYTES_INIT__ bytes/req for init
|
||||
// __BYTES_CONFIRM__ bytes/req for confirm
|
||||
// __MEM_DISCOVER_KB__ bytes/req converted to KB for discover
|
||||
// __MEM_SELECT_KB__ bytes/req converted to KB for select
|
||||
// __MEM_INIT_KB__ bytes/req converted to KB for init
|
||||
// __MEM_CONFIRM_KB__ bytes/req converted to KB for confirm
|
||||
// __PEAK_RPS__ highest RPS across all GOMAXPROCS levels
|
||||
// __CACHE_WARM_US__ mean latency in µs for CacheWarm
|
||||
// __CACHE_COLD_US__ mean latency in µs for CacheCold
|
||||
// __CACHE_WARM_ALLOCS__ allocs/req for CacheWarm
|
||||
// __CACHE_COLD_ALLOCS__ allocs/req for CacheCold
|
||||
// __CACHE_WARM_BYTES__ bytes/req for CacheWarm
|
||||
// __CACHE_COLD_BYTES__ bytes/req for CacheCold
|
||||
// __CACHE_DELTA__ formatted warm-vs-cold delta string
|
||||
// __THROUGHPUT_TABLE__ generated markdown table from throughput_report.csv
|
||||
// __BENCHSTAT_SUMMARY__ raw contents of benchstat_summary.txt
|
||||
package main
|
||||
|
||||
import (
|
||||
"bufio"
|
||||
"encoding/csv"
|
||||
"flag"
|
||||
"fmt"
|
||||
"io"
|
||||
"math"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"regexp"
|
||||
"strconv"
|
||||
"strings"
|
||||
)
|
||||
|
||||
func main() {
|
||||
dir := flag.String("dir", "", "Results directory (required)")
|
||||
tmplPath := flag.String("template", "benchmarks/reports/REPORT_TEMPLATE.md", "Path to report template")
|
||||
version := flag.String("version", "unknown", "Adapter version (e.g. v1.5.0)")
|
||||
flag.Parse()
|
||||
|
||||
if *dir == "" {
|
||||
fmt.Fprintln(os.Stderr, "ERROR: -dir is required")
|
||||
os.Exit(1)
|
||||
}
|
||||
|
||||
// Derive timestamp from the directory basename.
|
||||
timestamp := filepath.Base(*dir)
|
||||
|
||||
// ── Read template ──────────────────────────────────────────────────────────
|
||||
tmplBytes, err := os.ReadFile(*tmplPath)
|
||||
if err != nil {
|
||||
fmt.Fprintf(os.Stderr, "ERROR: reading template %s: %v\n", *tmplPath, err)
|
||||
os.Exit(1)
|
||||
}
|
||||
report := string(tmplBytes)
|
||||
|
||||
// ── Parse run1.txt for environment metadata ────────────────────────────────
|
||||
env := parseEnv(filepath.Join(*dir, "run1.txt"))
|
||||
|
||||
// ── Parse latency_report.csv ──────────────────────────────────────────────
|
||||
latency, err := parseLatencyCSV(filepath.Join(*dir, "latency_report.csv"))
|
||||
if err != nil {
|
||||
fmt.Fprintf(os.Stderr, "WARNING: could not parse latency_report.csv: %v\n", err)
|
||||
}
|
||||
|
||||
// ── Parse throughput_report.csv ───────────────────────────────────────────
|
||||
throughput, err := parseThroughputCSV(filepath.Join(*dir, "throughput_report.csv"))
|
||||
if err != nil {
|
||||
fmt.Fprintf(os.Stderr, "WARNING: could not parse throughput_report.csv: %v\n", err)
|
||||
}
|
||||
|
||||
// ── Read benchstat_summary.txt ────────────────────────────────────────────
|
||||
benchstat := readFileOrDefault(filepath.Join(*dir, "benchstat_summary.txt"),
|
||||
"(benchstat output not available)")
|
||||
|
||||
// ── Compute derived values ─────────────────────────────────────────────────
|
||||
|
||||
// Mean latency: convert ms → µs, round to integer.
|
||||
meanDiscoverUS := msToUS(latency["BenchmarkBAPCaller_Discover"]["mean_ms"])
|
||||
meanSelectUS := msToUS(latency["BenchmarkBAPCaller_AllActions/select"]["mean_ms"])
|
||||
meanInitUS := msToUS(latency["BenchmarkBAPCaller_AllActions/init"]["mean_ms"])
|
||||
meanConfirmUS := msToUS(latency["BenchmarkBAPCaller_AllActions/confirm"]["mean_ms"])
|
||||
|
||||
// Percentiles come from the Discover_Percentiles row.
|
||||
perc := latency["BenchmarkBAPCaller_Discover_Percentiles"]
|
||||
p50 := fmtMetric(perc["p50_µs"], "µs")
|
||||
p95 := fmtMetric(perc["p95_µs"], "µs")
|
||||
p99 := fmtMetric(perc["p99_µs"], "µs")
|
||||
|
||||
// Memory: bytes → KB (1 decimal place).
|
||||
memDiscoverKB := bytesToKB(latency["BenchmarkBAPCaller_Discover"]["bytes_op"])
|
||||
memSelectKB := bytesToKB(latency["BenchmarkBAPCaller_AllActions/select"]["bytes_op"])
|
||||
memInitKB := bytesToKB(latency["BenchmarkBAPCaller_AllActions/init"]["bytes_op"])
|
||||
memConfirmKB := bytesToKB(latency["BenchmarkBAPCaller_AllActions/confirm"]["bytes_op"])
|
||||
|
||||
// Cache delta.
|
||||
warmUS := msToUS(latency["BenchmarkBAPCaller_CacheWarm"]["mean_ms"])
|
||||
coldUS := msToUS(latency["BenchmarkBAPCaller_CacheCold"]["mean_ms"])
|
||||
cacheDelta := formatCacheDelta(warmUS, coldUS)
|
||||
|
||||
// Peak RPS across all concurrency levels.
|
||||
peakRPS := "—"
|
||||
var peakRPSVal float64
|
||||
for _, row := range throughput {
|
||||
if v := parseFloatOrZero(row["rps"]); v > peakRPSVal {
|
||||
peakRPSVal = v
|
||||
peakRPS = fmt.Sprintf("%.0f", peakRPSVal)
|
||||
}
|
||||
}
|
||||
|
||||
// ── Build throughput table ─────────────────────────────────────────────────
|
||||
throughputTable := buildThroughputTable(throughput)
|
||||
|
||||
// ── Apply substitutions ────────────────────────────────────────────────────
|
||||
replacements := map[string]string{
|
||||
"__TIMESTAMP__": timestamp,
|
||||
"__ONIX_VERSION__": *version,
|
||||
"__GOOS__": env["goos"],
|
||||
"__GOARCH__": env["goarch"],
|
||||
"__CPU__": env["cpu"],
|
||||
"__GOMAXPROCS__": env["gomaxprocs"],
|
||||
"__P50_US__": p50,
|
||||
"__P95_US__": p95,
|
||||
"__P99_US__": p99,
|
||||
"__MEAN_DISCOVER_US__": meanDiscoverUS,
|
||||
"__MEAN_SELECT_US__": meanSelectUS,
|
||||
"__MEAN_INIT_US__": meanInitUS,
|
||||
"__MEAN_CONFIRM_US__": meanConfirmUS,
|
||||
"__ALLOCS_DISCOVER__": fmtInt(latency["BenchmarkBAPCaller_Discover"]["allocs_op"]),
|
||||
"__ALLOCS_SELECT__": fmtInt(latency["BenchmarkBAPCaller_AllActions/select"]["allocs_op"]),
|
||||
"__ALLOCS_INIT__": fmtInt(latency["BenchmarkBAPCaller_AllActions/init"]["allocs_op"]),
|
||||
"__ALLOCS_CONFIRM__": fmtInt(latency["BenchmarkBAPCaller_AllActions/confirm"]["allocs_op"]),
|
||||
"__BYTES_DISCOVER__": fmtInt(latency["BenchmarkBAPCaller_Discover"]["bytes_op"]),
|
||||
"__BYTES_SELECT__": fmtInt(latency["BenchmarkBAPCaller_AllActions/select"]["bytes_op"]),
|
||||
"__BYTES_INIT__": fmtInt(latency["BenchmarkBAPCaller_AllActions/init"]["bytes_op"]),
|
||||
"__BYTES_CONFIRM__": fmtInt(latency["BenchmarkBAPCaller_AllActions/confirm"]["bytes_op"]),
|
||||
"__MEM_DISCOVER_KB__": memDiscoverKB,
|
||||
"__MEM_SELECT_KB__": memSelectKB,
|
||||
"__MEM_INIT_KB__": memInitKB,
|
||||
"__MEM_CONFIRM_KB__": memConfirmKB,
|
||||
"__PEAK_RPS__": peakRPS,
|
||||
"__CACHE_WARM_US__": warmUS,
|
||||
"__CACHE_COLD_US__": coldUS,
|
||||
"__CACHE_WARM_ALLOCS__": fmtInt(latency["BenchmarkBAPCaller_CacheWarm"]["allocs_op"]),
|
||||
"__CACHE_COLD_ALLOCS__": fmtInt(latency["BenchmarkBAPCaller_CacheCold"]["allocs_op"]),
|
||||
"__CACHE_WARM_BYTES__": fmtInt(latency["BenchmarkBAPCaller_CacheWarm"]["bytes_op"]),
|
||||
"__CACHE_COLD_BYTES__": fmtInt(latency["BenchmarkBAPCaller_CacheCold"]["bytes_op"]),
|
||||
"__CACHE_DELTA__": cacheDelta,
|
||||
"__THROUGHPUT_TABLE__": throughputTable,
|
||||
"__BENCHSTAT_SUMMARY__": benchstat,
|
||||
}
|
||||
|
||||
for placeholder, value := range replacements {
|
||||
report = strings.ReplaceAll(report, placeholder, value)
|
||||
}
|
||||
|
||||
// ── Write output ───────────────────────────────────────────────────────────
|
||||
outPath := filepath.Join(*dir, "BENCHMARK_REPORT.md")
|
||||
if err := os.WriteFile(outPath, []byte(report), 0o644); err != nil {
|
||||
fmt.Fprintf(os.Stderr, "ERROR: writing report: %v\n", err)
|
||||
os.Exit(1)
|
||||
}
|
||||
fmt.Printf(" Written → %s\n", outPath)
|
||||
}
|
||||
|
||||
// ── Parsers ────────────────────────────────────────────────────────────────────
|
||||
|
||||
var gomaxprocsRe = regexp.MustCompile(`-(\d+)$`)
|
||||
|
||||
// parseEnv reads goos, goarch, cpu, and GOMAXPROCS from a run*.txt file header.
|
||||
func parseEnv(path string) map[string]string {
|
||||
env := map[string]string{
|
||||
"goos": "unknown", "goarch": "unknown",
|
||||
"cpu": "unknown", "gomaxprocs": "unknown",
|
||||
}
|
||||
f, err := os.Open(path)
|
||||
if err != nil {
|
||||
return env
|
||||
}
|
||||
defer f.Close()
|
||||
|
||||
scanner := bufio.NewScanner(f)
|
||||
for scanner.Scan() {
|
||||
line := strings.TrimSpace(scanner.Text())
|
||||
switch {
|
||||
case strings.HasPrefix(line, "goos:"):
|
||||
env["goos"] = strings.TrimSpace(strings.TrimPrefix(line, "goos:"))
|
||||
case strings.HasPrefix(line, "goarch:"):
|
||||
env["goarch"] = strings.TrimSpace(strings.TrimPrefix(line, "goarch:"))
|
||||
case strings.HasPrefix(line, "cpu:"):
|
||||
env["cpu"] = strings.TrimSpace(strings.TrimPrefix(line, "cpu:"))
|
||||
case strings.HasPrefix(line, "Benchmark"):
|
||||
// Extract GOMAXPROCS from first benchmark line suffix (e.g. "-10").
|
||||
if m := gomaxprocsRe.FindStringSubmatch(strings.Fields(line)[0]); m != nil {
|
||||
env["gomaxprocs"] = m[1]
|
||||
}
|
||||
}
|
||||
}
|
||||
return env
|
||||
}
|
||||
|
||||
// parseLatencyCSV returns a map of benchmark name → field name → raw string value.
|
||||
// When multiple rows exist for the same benchmark (3 serial runs), values from
|
||||
// the first non-empty occurrence are used.
|
||||
func parseLatencyCSV(path string) (map[string]map[string]string, error) {
|
||||
f, err := os.Open(path)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
defer f.Close()
|
||||
|
||||
r := csv.NewReader(f)
|
||||
header, err := r.Read()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
result := map[string]map[string]string{}
|
||||
for {
|
||||
row, err := r.Read()
|
||||
if err == io.EOF {
|
||||
break
|
||||
}
|
||||
if err != nil || len(row) == 0 {
|
||||
continue
|
||||
}
|
||||
name := row[0]
|
||||
if _, exists := result[name]; !exists {
|
||||
result[name] = map[string]string{}
|
||||
}
|
||||
for i, col := range header[1:] {
|
||||
idx := i + 1
|
||||
if idx < len(row) && row[idx] != "" && result[name][col] == "" {
|
||||
result[name][col] = row[idx]
|
||||
}
|
||||
}
|
||||
}
|
||||
return result, nil
|
||||
}
|
||||
|
||||
// parseThroughputCSV returns rows as a slice of field maps.
|
||||
func parseThroughputCSV(path string) ([]map[string]string, error) {
|
||||
f, err := os.Open(path)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
defer f.Close()
|
||||
|
||||
r := csv.NewReader(f)
|
||||
header, err := r.Read()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
var rows []map[string]string
|
||||
for {
|
||||
row, err := r.Read()
|
||||
if err == io.EOF {
|
||||
break
|
||||
}
|
||||
if err != nil || len(row) == 0 {
|
||||
continue
|
||||
}
|
||||
m := map[string]string{}
|
||||
for i, col := range header {
|
||||
if i < len(row) {
|
||||
m[col] = row[i]
|
||||
}
|
||||
}
|
||||
rows = append(rows, m)
|
||||
}
|
||||
return rows, nil
|
||||
}
|
||||
|
||||
// buildThroughputTable renders the throughput CSV as a markdown table.
|
||||
func buildThroughputTable(rows []map[string]string) string {
|
||||
if len(rows) == 0 {
|
||||
return "_No concurrency sweep data available._"
|
||||
}
|
||||
var sb strings.Builder
|
||||
sb.WriteString("| GOMAXPROCS | Mean Latency (µs) | RPS |\n")
|
||||
sb.WriteString("|:----------:|------------------:|----:|\n")
|
||||
for _, row := range rows {
|
||||
cpu := orDash(row["gomaxprocs"])
|
||||
latUS := "—"
|
||||
if v := parseFloatOrZero(row["mean_latency_ms"]); v > 0 {
|
||||
latUS = fmt.Sprintf("%.0f", v*1000)
|
||||
}
|
||||
rps := orDash(row["rps"])
|
||||
sb.WriteString(fmt.Sprintf("| %s | %s | %s |\n", cpu, latUS, rps))
|
||||
}
|
||||
return sb.String()
|
||||
}
|
||||
|
||||
// ── Formatters ─────────────────────────────────────────────────────────────────
|
||||
|
||||
// msToUS converts a ms string to a rounded µs string.
|
||||
func msToUS(ms string) string {
|
||||
v := parseFloatOrZero(ms)
|
||||
if v == 0 {
|
||||
return "—"
|
||||
}
|
||||
return fmt.Sprintf("%.0f", v*1000)
|
||||
}
|
||||
|
||||
// bytesToKB converts a bytes string to a KB string with 1 decimal place.
|
||||
func bytesToKB(bytes string) string {
|
||||
v := parseFloatOrZero(bytes)
|
||||
if v == 0 {
|
||||
return "—"
|
||||
}
|
||||
return fmt.Sprintf("%.1f", v/1024)
|
||||
}
|
||||
|
||||
// fmtInt formats a float string as a rounded integer string.
|
||||
func fmtInt(s string) string {
|
||||
v := parseFloatOrZero(s)
|
||||
if v == 0 {
|
||||
return "—"
|
||||
}
|
||||
return fmt.Sprintf("%.0f", math.Round(v))
|
||||
}
|
||||
|
||||
// fmtMetric formats a metric value with the given unit, or returns "—".
|
||||
func fmtMetric(s, unit string) string {
|
||||
v := parseFloatOrZero(s)
|
||||
if v == 0 {
|
||||
return "—"
|
||||
}
|
||||
return fmt.Sprintf("%.0f %s", v, unit)
|
||||
}
|
||||
|
||||
// formatCacheDelta produces a human-readable warm-vs-cold delta string.
|
||||
func formatCacheDelta(warmUS, coldUS string) string {
|
||||
w := parseFloatOrZero(warmUS)
|
||||
c := parseFloatOrZero(coldUS)
|
||||
if w == 0 || c == 0 {
|
||||
return "—"
|
||||
}
|
||||
delta := w - c
|
||||
sign := "+"
|
||||
if delta < 0 {
|
||||
sign = ""
|
||||
}
|
||||
return fmt.Sprintf("%s%.0f µs (warm vs cold)", sign, delta)
|
||||
}
|
||||
|
||||
func orDash(s string) string {
|
||||
if s == "" {
|
||||
return "—"
|
||||
}
|
||||
return s
|
||||
}
|
||||
|
||||
func parseFloatOrZero(s string) float64 {
|
||||
v, _ := strconv.ParseFloat(strings.TrimSpace(s), 64)
|
||||
return v
|
||||
}
|
||||
|
||||
func readFileOrDefault(path, def string) string {
|
||||
b, err := os.ReadFile(path)
|
||||
if err != nil {
|
||||
return def
|
||||
}
|
||||
return strings.TrimRight(string(b), "\n")
|
||||
}
|
||||
Reference in New Issue
Block a user