Files
scylladb/test/perf/perf.cc
Marcin Maliszkiewicz 619944555f test: perf: make aggregated_perf_results formatting more human readable
Before:
throughput: mean=170728.58 standard-deviation=1921.76 median=171084.16 median-absolute-deviation=1501.58 maximum=172913.36 minimum=167288.97
instructions_per_op: mean=4685.89 standard-deviation=12.46 median=4683.92 median-absolute-deviation=9.68 maximum=4706.53 minimum=4666.70
cpu_cycles_per_op: mean=3090.94 standard-deviation=52.69 median=3103.43 median-absolute-deviation=24.55 maximum=3192.99 minimum=3003.00

After:
throughput:
	mean=   168224.81 standard-deviation=854.48
	median= 168829.02 median-absolute-deviation=604.21
	maximum=168829.02 minimum=167620.60
instructions_per_op:
	mean=   4837.02 standard-deviation=20.89
	median= 4851.79 median-absolute-deviation=14.77
	maximum=4851.79 minimum=4822.24
cpu_cycles_per_op:
	mean=   3271.42 standard-deviation=46.29
	median= 3304.16 median-absolute-deviation=32.73
	maximum=3304.16 minimum=3238.69
2025-04-09 10:49:20 +02:00

145 lines
5.8 KiB
C++

/*
* Copyright (C) 2021-present ScyllaDB
*/
/*
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
*/
#include "perf.hh"
#include <seastar/core/reactor.hh>
#include <seastar/core/memory.hh>
#include "seastarx.hh"
#include "reader_concurrency_semaphore.hh"
#include "schema/schema.hh"
#include "utils/logalloc.hh"
uint64_t perf_mallocs() {
return memory::stats().mallocs();
}
uint64_t perf_logallocs() {
return logalloc::shard_tracker().statistics().num_allocations;
}
uint64_t perf_tasks_processed() {
return engine().get_sched_stats().tasks_processed;
}
void scheduling_latency_measurer::schedule_tick() {
seastar::schedule(make_task(default_scheduling_group(), [self = weak_from_this()] () mutable {
if (self) {
self->tick();
}
}));
}
auto fmt::formatter<scheduling_latency_measurer>::format(const scheduling_latency_measurer& slm, fmt::format_context& ctx) const
-> decltype(ctx.out()) {
auto to_ms = [] (int64_t nanos) {
return float(nanos) / 1e6;
};
return fmt::format_to(ctx.out(), "{{count: {}, "
//"min: {:.6f} [ms], "
//"50%: {:.6f} [ms], "
//"90%: {:.6f} [ms], "
"99%: {:.6f} [ms], "
"max: {:.6f} [ms]}}",
slm.histogram().count(),
//to_ms(slm.min().count()),
//to_ms(slm.histogram().percentile(0.5)),
//to_ms(slm.histogram().percentile(0.9)),
to_ms(slm.histogram().percentile(0.99)),
to_ms(slm.max().count()));
}
auto fmt::formatter<perf_result>::format(const perf_result& result, fmt::format_context& ctx) const
-> decltype(ctx.out()) {
return fmt::format_to(ctx.out(), "{:.2f} tps ({:5.1f} allocs/op, {:5.1f} logallocs/op, {:5.1f} tasks/op, {:7.0f} insns/op, {:7.0f} cycles/op, {:8} errors)",
result.throughput, result.mallocs_per_op, result.logallocs_per_op, result.tasks_per_op, result.instructions_per_op, result.cpu_cycles_per_op, result.errors);
}
aggregated_perf_results::aggregated_perf_results(std::vector<perf_result>& results) {
stats["throughput"] = calculate_stats(results, std::mem_fn(&perf_result::throughput));
median_by_throughput = results[results.size() / 2];
stats["instructions_per_op"] = calculate_stats(results, std::mem_fn(&perf_result::instructions_per_op));
stats["cpu_cycles_per_op"] = calculate_stats(results, std::mem_fn(&perf_result::cpu_cycles_per_op));
}
aggregated_perf_results::stats_t aggregated_perf_results::calculate_stats(std::vector<perf_result>& results, std::function<double(const perf_result&)> get_stat) const {
stats_t ret;
std::sort(results.begin(), results.end(), [&] (const perf_result& a, const perf_result& b) {
return get_stat(a) < get_stat(b);
});
ret.min = get_stat(results[0]);
ret.median = get_stat(results[results.size() / 2]);
ret.max = get_stat(results[results.size() - 1]);
ret.mean = 0;
for (const auto& pr : results) {
ret.mean += get_stat(pr);
}
ret.mean /= results.size();
double var = 0;
std::vector<double> abs_deviations;
abs_deviations.reserve(results.size());
for (const auto& pr : results) {
auto dev = get_stat(pr) - ret.mean;
abs_deviations.emplace_back(std::abs(dev));
var += dev * dev;
}
// Since the results are samples of the total population
// devide the sum of squared deviations by (n - 1) rather than by n
ret.stdev = std::sqrt(results.size() > 1 ? var / (results.size() - 1) : var);
std::sort(abs_deviations.begin(), abs_deviations.end());
ret.median_absolute_deviation = abs_deviations[abs_deviations.size() / 2];
return ret;
}
std::ostream&
operator<<(std::ostream& os, const aggregated_perf_results& result) {
for (const auto& s : {"throughput", "instructions_per_op", "cpu_cycles_per_op"}) {
auto& t = result.stats.at(s);
fmt::print(os, "{}:\n\tmean= {:.2f} standard-deviation={:.2f}\n\tmedian= {:.2f} median-absolute-deviation={:.2f}\n\tmaximum={:.2f} minimum={:.2f}\n",
s, t.mean, t.stdev, t.median, t.median_absolute_deviation, t.max, t.min);
}
return os;
}
aio_writes_result_mixin::aio_writes_result_mixin()
: aio_writes(engine().get_io_stats().aio_writes)
, aio_write_bytes(engine().get_io_stats().aio_write_bytes)
{}
void aio_writes_result_mixin::update(aio_writes_result_mixin& result, const executor_shard_stats& stats) {
result.aio_writes = (engine().get_io_stats().aio_writes - result.aio_writes) / stats.invocations;
result.aio_write_bytes = (engine().get_io_stats().aio_write_bytes - result.aio_write_bytes) / stats.invocations;
}
auto fmt::formatter<perf_result_with_aio_writes>::format(const perf_result_with_aio_writes& result, fmt::format_context& ctx) const
-> decltype(ctx.out()) {
return fmt::format_to(ctx.out(), "{:.2f} tps ({:5.1f} allocs/op, {:5.1f} logallocs/op, {:5.1f} tasks/op, {:7.0f} insns/op, {:7.0f} cycles/op, {:8} errors, {:7.2f} bytes/op, {:5.1f} writes/op)",
result.throughput, result.mallocs_per_op, result.logallocs_per_op, result.tasks_per_op, result.instructions_per_op, result.cpu_cycles_per_op, result.errors, result.aio_write_bytes, result.aio_writes);
}
namespace perf {
reader_concurrency_semaphore_wrapper::reader_concurrency_semaphore_wrapper(sstring name)
: _semaphore(std::make_unique<reader_concurrency_semaphore>(reader_concurrency_semaphore::no_limits{}, std::move(name),
reader_concurrency_semaphore::register_metrics::no))
{
}
reader_concurrency_semaphore_wrapper::~reader_concurrency_semaphore_wrapper() {
(void)_semaphore->stop().finally([sem = std::move(_semaphore)] { });
}
reader_permit reader_concurrency_semaphore_wrapper::make_permit() {
return _semaphore->make_tracking_only_permit(nullptr, "perf", db::no_timeout, {});
}
} // namespace perf