There are 3 metrics (that goes in every compaction_history entry): total_tombstone_purge_attempt total_tombstone_purge_failure_due_to_overlapping_with_memtable total_tombstone_purge_failure_due_to_overlapping_with_uncompacting_sstable When a tombstone is not expired (e.g. doesn't satisfy "gc_before" or grace period), it can be currently accounted as failure due to overlapping with either memtable or uncompacting sstable. So those 2 last metrics have noise of *unexpired* tombstones. What we should do is to only account for expired tombstones in all those 3 metrics. We lose the info of knowing the amount of tombstones processed by compaction, now we'll only know about the expired ones. But those metrics were primarily added for explaining why expired tombstones cannot be removed. We could have alternatively added a new field purge_failure_due_to_being_unexpired or something, but it requires adding a new field to compaction_history. Fixes https://scylladb.atlassian.net/browse/SCYLLADB-737. Signed-off-by: Raphael S. Carvalho <raphaelsc@scylladb.com> Closes scylladb/scylladb#28669
786 lines
33 KiB
C++
786 lines
33 KiB
C++
/*
|
|
* Copyright (C) 2016-present ScyllaDB
|
|
*/
|
|
|
|
/*
|
|
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include "compaction/compaction_garbage_collector.hh"
|
|
#include "mutation_fragment.hh"
|
|
#include "mutation_fragment_stream_validator.hh"
|
|
#include "mutation_tombstone_stats.hh"
|
|
#include "tombstone_gc.hh"
|
|
#include "keys/full_position.hh"
|
|
#include <type_traits>
|
|
#include "utils/log.hh"
|
|
|
|
extern logging::logger mclog;
|
|
|
|
inline bool has_ck_selector(const query::clustering_row_ranges& ranges) {
|
|
// Like PK range, an empty row range, should be considered an "exclude all" restriction
|
|
return ranges.empty() || std::any_of(ranges.begin(), ranges.end(), [](auto& r) {
|
|
return !r.is_full();
|
|
});
|
|
}
|
|
|
|
enum class compact_for_sstables {
|
|
no,
|
|
yes,
|
|
};
|
|
|
|
template<typename T>
|
|
concept CompactedFragmentsConsumer = requires(T obj, tombstone t, const dht::decorated_key& dk, static_row sr,
|
|
clustering_row cr, range_tombstone_change rtc, tombstone current_tombstone, row_tombstone current_row_tombstone, bool is_alive) {
|
|
obj.consume_new_partition(dk);
|
|
obj.consume(t);
|
|
{ obj.consume(std::move(sr), current_tombstone, is_alive) } -> std::same_as<stop_iteration>;
|
|
{ obj.consume(std::move(cr), current_row_tombstone, is_alive) } -> std::same_as<stop_iteration>;
|
|
{ obj.consume(std::move(rtc)) } -> std::same_as<stop_iteration>;
|
|
{ obj.consume_end_of_partition() } -> std::same_as<stop_iteration>;
|
|
obj.consume_end_of_stream();
|
|
};
|
|
|
|
struct detached_compaction_state {
|
|
::partition_start partition_start;
|
|
std::optional<::static_row> static_row;
|
|
std::optional<range_tombstone_change> current_tombstone;
|
|
};
|
|
|
|
class noop_compacted_fragments_consumer {
|
|
public:
|
|
void consume_new_partition(const dht::decorated_key& dk) {}
|
|
void consume(tombstone t) {}
|
|
stop_iteration consume(static_row&& sr, tombstone, bool) { return stop_iteration::no; }
|
|
stop_iteration consume(clustering_row&& cr, row_tombstone, bool) { return stop_iteration::no; }
|
|
stop_iteration consume(range_tombstone_change&& rtc) { return stop_iteration::no; }
|
|
stop_iteration consume_end_of_partition() { return stop_iteration::no; }
|
|
void consume_end_of_stream() {}
|
|
};
|
|
|
|
class mutation_compactor_garbage_collector : public compaction_garbage_collector {
|
|
const schema& _schema;
|
|
column_kind _kind;
|
|
std::optional<clustering_key> _ckey;
|
|
row_tombstone _tomb;
|
|
row_marker _marker;
|
|
row _row;
|
|
|
|
public:
|
|
explicit mutation_compactor_garbage_collector(const schema& schema)
|
|
: _schema(schema) {
|
|
}
|
|
void start_collecting_static_row() {
|
|
_kind = column_kind::static_column;
|
|
}
|
|
void start_collecting_clustering_row(clustering_key ckey) {
|
|
_kind = column_kind::regular_column;
|
|
_ckey = std::move(ckey);
|
|
}
|
|
void collect(row_tombstone tomb) {
|
|
_tomb = tomb;
|
|
}
|
|
virtual void collect(column_id id, atomic_cell cell) override {
|
|
_row.apply(_schema.column_at(_kind, id), std::move(cell));
|
|
}
|
|
virtual void collect(column_id id, collection_mutation_description mut) override {
|
|
if (mut.tomb || !mut.cells.empty()) {
|
|
const auto& cdef = _schema.column_at(_kind, id);
|
|
_row.apply(cdef, mut.serialize(*cdef.type));
|
|
}
|
|
}
|
|
virtual void collect(row_marker marker) override {
|
|
_marker = marker;
|
|
}
|
|
template <typename Consumer>
|
|
void consume_static_row(Consumer&& consumer) {
|
|
if (!_row.empty()) {
|
|
consumer(static_row(std::move(_row)));
|
|
_row = {};
|
|
}
|
|
}
|
|
template <typename Consumer>
|
|
void consume_clustering_row(Consumer&& consumer) {
|
|
if (_tomb || !_marker.is_missing() || !_row.empty()) {
|
|
consumer(clustering_row(std::move(*_ckey), _tomb, _marker, std::move(_row)));
|
|
_ckey.reset();
|
|
_tomb = {};
|
|
_marker = {};
|
|
_row = {};
|
|
}
|
|
}
|
|
};
|
|
|
|
struct compaction_stats {
|
|
struct row_stats {
|
|
uint64_t live = 0;
|
|
uint64_t dead = 0;
|
|
compact_and_expire_result cell_stats;
|
|
|
|
void add_row(const compact_and_expire_result& new_cell_stats, bool force_live = false) {
|
|
const auto is_live = new_cell_stats.is_live() || force_live;
|
|
live += is_live;
|
|
dead += !is_live;
|
|
cell_stats += new_cell_stats;
|
|
}
|
|
uint64_t total() const {
|
|
return live + dead;
|
|
}
|
|
|
|
row_stats& operator+=(const row_stats& other) {
|
|
live += other.live;
|
|
dead += other.dead;
|
|
cell_stats += other.cell_stats;
|
|
return *this;
|
|
}
|
|
};
|
|
|
|
uint64_t live_cells() const {
|
|
return static_rows.cell_stats.live_cells + clustering_rows.cell_stats.live_cells;
|
|
}
|
|
uint64_t dead_cells() const {
|
|
return static_rows.cell_stats.dead_cells + clustering_rows.cell_stats.dead_cells +
|
|
static_rows.cell_stats.collection_tombstones + clustering_rows.cell_stats.collection_tombstones;
|
|
}
|
|
uint64_t dead_partitions() const {
|
|
return total_partitions - live_partitions;
|
|
}
|
|
|
|
uint64_t total_partitions = 0;
|
|
uint64_t live_partitions = 0;
|
|
row_stats static_rows;
|
|
row_stats clustering_rows;
|
|
uint64_t range_tombstones = 0;
|
|
|
|
compaction_stats& operator+=(const compaction_stats& other) {
|
|
total_partitions += other.total_partitions;
|
|
live_partitions += other.live_partitions;
|
|
static_rows += other.static_rows;
|
|
clustering_rows += other.clustering_rows;
|
|
range_tombstones += other.range_tombstones;
|
|
return *this;
|
|
}
|
|
};
|
|
|
|
template<compact_for_sstables SSTableCompaction>
|
|
class compact_mutation_state {
|
|
const schema& _schema;
|
|
gc_clock::time_point _query_time;
|
|
max_purgeable_fn _get_max_purgeable;
|
|
max_purgeable _max_purgeable_regular;
|
|
max_purgeable _max_purgeable_shadowable;
|
|
std::optional<gc_clock::time_point> _gc_before;
|
|
const query::partition_slice& _slice;
|
|
uint64_t _row_limit{};
|
|
uint32_t _partition_limit{};
|
|
uint64_t _partition_row_limit{};
|
|
tombstone_gc_state _tombstone_gc_state;
|
|
|
|
tombstone _partition_tombstone;
|
|
|
|
bool _static_row_live{};
|
|
uint64_t _rows_in_current_partition;
|
|
uint32_t _current_partition_limit;
|
|
bool _empty_partition{};
|
|
bool _empty_partition_in_gc_consumer{};
|
|
std::optional<dht::decorated_key> _dk;
|
|
bool _return_static_content_on_partition_with_no_rows{};
|
|
|
|
std::optional<static_row> _last_static_row;
|
|
position_in_partition _last_pos;
|
|
// Currently active tombstone, can be different than the tombstone emitted to
|
|
// the regular consumer (_current_emitted_tombstone) because even purged
|
|
// tombstone that are not emitted are still applied to data when compacting.
|
|
tombstone _effective_tombstone;
|
|
// Track last emitted tombstone to regular and gc consumers respectively.
|
|
// Used to determine whether any active tombstones need closing at EOS.
|
|
tombstone _current_emitted_tombstone;
|
|
tombstone _current_emitted_gc_tombstone;
|
|
|
|
std::unique_ptr<mutation_compactor_garbage_collector> _collector;
|
|
|
|
::compaction_stats _stats;
|
|
tombstone_purge_stats* _tombstone_stats = nullptr;
|
|
|
|
mutation_fragment_stream_validating_filter _validator;
|
|
|
|
// Remember if we requested to stop mid-partition.
|
|
stop_iteration _stop = stop_iteration::no;
|
|
private:
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
stop_iteration do_consume(range_tombstone_change&& rtc, Consumer& consumer, GCConsumer& gc_consumer) {
|
|
_validator(mutation_fragment_v2::kind::range_tombstone_change, rtc.position(), rtc.tombstone());
|
|
stop_iteration gc_consumer_stop = stop_iteration::no;
|
|
stop_iteration consumer_stop = stop_iteration::no;
|
|
if (rtc.tombstone() <= _partition_tombstone) {
|
|
rtc.set_tombstone({});
|
|
}
|
|
_effective_tombstone = rtc.tombstone();
|
|
const auto can_purge = rtc.tombstone() && can_purge_tombstone(rtc.tombstone());
|
|
if (can_purge || _current_emitted_gc_tombstone) {
|
|
partition_is_not_empty_for_gc_consumer(gc_consumer);
|
|
auto tomb = can_purge ? rtc.tombstone() : tombstone{};
|
|
_current_emitted_gc_tombstone = tomb;
|
|
gc_consumer_stop = gc_consumer.consume(range_tombstone_change(rtc.position(), tomb));
|
|
if (can_purge) {
|
|
rtc.set_tombstone({});
|
|
}
|
|
}
|
|
// If we have a previous active tombstone we emit the current one even if it is purged.
|
|
if (_current_emitted_tombstone || (rtc.tombstone() && !can_purge)) {
|
|
partition_is_not_empty(consumer);
|
|
_current_emitted_tombstone = rtc.tombstone();
|
|
consumer_stop = consumer.consume(std::move(rtc));
|
|
}
|
|
return gc_consumer_stop || consumer_stop;
|
|
}
|
|
static constexpr bool sstable_compaction() {
|
|
return SSTableCompaction == compact_for_sstables::yes;
|
|
}
|
|
|
|
template <typename GCConsumer>
|
|
void partition_is_not_empty_for_gc_consumer(GCConsumer& gc_consumer) {
|
|
if (_empty_partition_in_gc_consumer) {
|
|
_empty_partition_in_gc_consumer = false;
|
|
gc_consumer.consume_new_partition(*_dk);
|
|
auto pt = _partition_tombstone;
|
|
if (pt && can_purge_tombstone(pt)) {
|
|
gc_consumer.consume(pt);
|
|
}
|
|
}
|
|
}
|
|
|
|
template <typename Consumer>
|
|
void partition_is_not_empty(Consumer& consumer) {
|
|
if (_empty_partition) {
|
|
_empty_partition = false;
|
|
++_stats.live_partitions;
|
|
consumer.consume_new_partition(*_dk);
|
|
auto pt = _partition_tombstone;
|
|
if (pt && !can_purge_tombstone(pt)) {
|
|
consumer.consume(pt);
|
|
}
|
|
}
|
|
}
|
|
|
|
bool satisfy_grace_period(const gc_clock::time_point& deletion_time) {
|
|
return deletion_time < get_gc_before();
|
|
}
|
|
|
|
bool can_purge_tombstone(const tombstone& t, is_shadowable is_shadowable, const gc_clock::time_point deletion_time) {
|
|
max_purgeable::can_purge_result purge_res { };
|
|
std::optional<bool> expired;
|
|
|
|
if (_tombstone_gc_state.cheap_to_get_gc_before(_schema)) {
|
|
// if retrieval of grace period is cheap, can_gc() will only be
|
|
// called for tombstones that are older than grace period, in
|
|
// order to avoid unnecessary bloom filter checks when calculating
|
|
// max purgeable timestamp.
|
|
expired = purge_res.can_purge = satisfy_grace_period(deletion_time);
|
|
if (purge_res.can_purge) {
|
|
purge_res = can_gc(t, is_shadowable);
|
|
}
|
|
} else {
|
|
purge_res = can_gc(t, is_shadowable);
|
|
if (purge_res.can_purge) {
|
|
expired = purge_res.can_purge = satisfy_grace_period(deletion_time);
|
|
}
|
|
}
|
|
|
|
if constexpr (sstable_compaction()) {
|
|
// Tombstone GC stats only account for expired tombstones (those eligible for GC).
|
|
if (!_tombstone_stats || !t || !expired.value_or(satisfy_grace_period(deletion_time))) {
|
|
return purge_res.can_purge;
|
|
}
|
|
|
|
++_tombstone_stats->attempts;
|
|
if (!purge_res.can_purge) {
|
|
static int64_t tombstone_purge_stats::*stats_table[] = {
|
|
&tombstone_purge_stats::failures_other,
|
|
&tombstone_purge_stats::failures_due_to_overlapping_with_memtable,
|
|
&tombstone_purge_stats::failures_due_to_overlapping_with_uncompacting_sstable
|
|
};
|
|
++(_tombstone_stats->*stats_table[static_cast<int>(purge_res.timestamp_source)]);
|
|
}
|
|
}
|
|
|
|
return purge_res.can_purge;
|
|
}
|
|
|
|
bool can_purge_tombstone(const tombstone& t) {
|
|
// Only row tombstones can be shadowable, regular tombstones aren't
|
|
return can_purge_tombstone(t, is_shadowable::no, t.deletion_time);
|
|
};
|
|
|
|
bool can_purge_tombstone(const row_tombstone& t) {
|
|
return can_purge_tombstone(t.tomb(), t.is_shadowable(), t.max_deletion_time());
|
|
};
|
|
|
|
gc_clock::time_point get_gc_before() {
|
|
if (_gc_before) {
|
|
return _gc_before.value();
|
|
} else {
|
|
if (_dk) {
|
|
_gc_before = _tombstone_gc_state.get_gc_before_for_key(_schema.shared_from_this(), *_dk, _query_time);
|
|
return _gc_before.value();
|
|
} else {
|
|
return gc_clock::time_point::min();
|
|
}
|
|
}
|
|
}
|
|
|
|
max_purgeable::can_purge_result can_gc(tombstone t, is_shadowable is_shadowable) {
|
|
if (!sstable_compaction()) {
|
|
return { .can_purge = _tombstone_gc_state.is_gc_enabled(), .timestamp_source = max_purgeable::timestamp_source::none };
|
|
}
|
|
if (!t) {
|
|
return { .can_purge = false };
|
|
}
|
|
auto& max_purgeable = is_shadowable ? _max_purgeable_shadowable : _max_purgeable_regular;
|
|
if (!max_purgeable) {
|
|
max_purgeable = _get_max_purgeable(*_dk, is_shadowable);
|
|
}
|
|
auto ret = max_purgeable.can_purge(t);
|
|
mclog.debug("can_gc: t={} is_shadowable={} max_purgeable={}: can_purge={}, timestamp_source={}", t, is_shadowable, max_purgeable, ret.can_purge, ret.timestamp_source);
|
|
return ret;
|
|
};
|
|
|
|
can_gc_fn make_can_gc_fn() {
|
|
return [this] (tombstone t, is_shadowable is_shadowable) { return can_gc(t, is_shadowable).can_purge; };
|
|
}
|
|
|
|
public:
|
|
compact_mutation_state(compact_mutation_state&&) = delete; // Because 'this' is captured
|
|
|
|
compact_mutation_state(
|
|
const schema& s,
|
|
gc_clock::time_point query_time,
|
|
const query::partition_slice& slice,
|
|
uint64_t limit,
|
|
uint32_t partition_limit,
|
|
const tombstone_gc_state& gc_state,
|
|
mutation_fragment_stream_validation_level validation_level = mutation_fragment_stream_validation_level::token)
|
|
: _schema(s)
|
|
, _query_time(query_time)
|
|
, _slice(slice)
|
|
, _row_limit(limit)
|
|
, _partition_limit(partition_limit)
|
|
, _partition_row_limit(_slice.options.contains(query::partition_slice::option::distinct) ? 1 : slice.partition_row_limit())
|
|
, _tombstone_gc_state(gc_state)
|
|
, _last_pos(position_in_partition::for_partition_end())
|
|
, _validator("mutation_compactor for read", _schema, validation_level)
|
|
{
|
|
static_assert(!sstable_compaction(), "This constructor cannot be used for sstable compaction.");
|
|
}
|
|
|
|
compact_mutation_state(const schema& s, gc_clock::time_point compaction_time,
|
|
max_purgeable_fn get_max_purgeable,
|
|
const tombstone_gc_state& gc_state,
|
|
tombstone_purge_stats* tombstone_stats = nullptr)
|
|
: _schema(s)
|
|
, _query_time(compaction_time)
|
|
, _get_max_purgeable(std::move(get_max_purgeable))
|
|
, _slice(s.full_slice())
|
|
, _tombstone_gc_state(gc_state)
|
|
, _last_pos(position_in_partition::for_partition_end())
|
|
, _collector(std::make_unique<mutation_compactor_garbage_collector>(_schema))
|
|
, _tombstone_stats(tombstone_stats)
|
|
// We already have a validator for compaction in the sstable writer, no need to validate twice
|
|
, _validator("mutation_compactor for compaction", _schema, mutation_fragment_stream_validation_level::none)
|
|
{
|
|
static_assert(sstable_compaction(), "This constructor can only be used for sstable compaction.");
|
|
}
|
|
|
|
void consume_new_partition(const dht::decorated_key& dk) {
|
|
_validator(mutation_fragment_v2::kind::partition_start, position_in_partition_view::for_partition_start(), {});
|
|
_stop = stop_iteration::no;
|
|
_dk = dk;
|
|
auto& pk = _dk->key();
|
|
_validator(*_dk);
|
|
_return_static_content_on_partition_with_no_rows =
|
|
_slice.options.contains(query::partition_slice::option::always_return_static_content) ||
|
|
!has_ck_selector(_slice.row_ranges(_schema, pk));
|
|
_empty_partition = true;
|
|
_empty_partition_in_gc_consumer = true;
|
|
_rows_in_current_partition = 0;
|
|
_static_row_live = false;
|
|
_partition_tombstone = {};
|
|
_current_partition_limit = std::min(_row_limit, _partition_row_limit);
|
|
_max_purgeable_regular = {};
|
|
_max_purgeable_shadowable = {};
|
|
_gc_before = std::nullopt;
|
|
_last_static_row.reset();
|
|
_last_pos = position_in_partition::for_partition_start();
|
|
_effective_tombstone = {};
|
|
_current_emitted_tombstone = {};
|
|
_current_emitted_gc_tombstone = {};
|
|
|
|
++_stats.total_partitions;
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
void consume(tombstone t, Consumer& consumer, GCConsumer& gc_consumer) {
|
|
_partition_tombstone = t;
|
|
if (can_purge_tombstone(t)) {
|
|
partition_is_not_empty_for_gc_consumer(gc_consumer);
|
|
} else {
|
|
partition_is_not_empty(consumer);
|
|
}
|
|
}
|
|
|
|
template <typename Consumer>
|
|
requires CompactedFragmentsConsumer<Consumer>
|
|
void force_partition_not_empty(Consumer& consumer) {
|
|
partition_is_not_empty(consumer);
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
stop_iteration consume(static_row&& sr, Consumer& consumer, GCConsumer& gc_consumer) {
|
|
_validator(mutation_fragment_v2::kind::static_row, sr.position(), {});
|
|
_last_static_row = static_row(_schema, sr);
|
|
_last_pos = position_in_partition(position_in_partition::static_row_tag_t());
|
|
auto current_tombstone = _partition_tombstone;
|
|
if constexpr (sstable_compaction()) {
|
|
_collector->start_collecting_static_row();
|
|
}
|
|
auto can_gc = make_can_gc_fn();
|
|
auto gc_before = get_gc_before();
|
|
auto res = sr.cells().compact_and_expire(_schema, column_kind::static_column, row_tombstone(current_tombstone),
|
|
_query_time, can_gc, gc_before, _collector.get());
|
|
_stats.static_rows.add_row(res);
|
|
const auto is_live = res.is_live();
|
|
if constexpr (sstable_compaction()) {
|
|
_collector->consume_static_row([this, &gc_consumer, current_tombstone] (static_row&& sr_garbage) {
|
|
partition_is_not_empty_for_gc_consumer(gc_consumer);
|
|
// We are passing only dead (purged) data so pass is_live=false.
|
|
gc_consumer.consume(std::move(sr_garbage), current_tombstone, false);
|
|
});
|
|
} else {
|
|
if (can_purge_tombstone(current_tombstone)) {
|
|
current_tombstone = {};
|
|
}
|
|
}
|
|
_static_row_live = is_live;
|
|
if (is_live || !sr.empty()) {
|
|
partition_is_not_empty(consumer);
|
|
_stop = consumer.consume(std::move(sr), current_tombstone, is_live);
|
|
}
|
|
return _stop;
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
stop_iteration consume(clustering_row&& cr, Consumer& consumer, GCConsumer& gc_consumer) {
|
|
_validator(mutation_fragment_v2::kind::clustering_row, cr.position(), {});
|
|
if (!sstable_compaction()) {
|
|
_last_pos = cr.position();
|
|
}
|
|
auto current_tombstone = std::max(_partition_tombstone, _effective_tombstone);
|
|
auto t = cr.tomb();
|
|
t.apply(current_tombstone);
|
|
|
|
if constexpr (sstable_compaction()) {
|
|
_collector->start_collecting_clustering_row(cr.key());
|
|
}
|
|
|
|
{
|
|
const auto rt = cr.tomb();
|
|
if (rt.tomb() <= current_tombstone) {
|
|
cr.remove_tombstone();
|
|
} else if (can_purge_tombstone(rt)) {
|
|
if constexpr (sstable_compaction()) {
|
|
_collector->collect(rt);
|
|
}
|
|
cr.remove_tombstone();
|
|
}
|
|
}
|
|
auto can_gc = make_can_gc_fn();
|
|
auto gc_before = get_gc_before();
|
|
const bool marker_is_live = cr.marker().compact_and_expire(t.tomb(), _query_time, can_gc, gc_before, _collector.get());
|
|
const auto res = cr.cells().compact_and_expire(_schema, column_kind::regular_column, t, _query_time, can_gc, gc_before, cr.marker(),
|
|
_collector.get());
|
|
_stats.clustering_rows.add_row(res, marker_is_live);
|
|
const auto is_live = res.is_live() || marker_is_live;
|
|
|
|
if constexpr (sstable_compaction()) {
|
|
_collector->consume_clustering_row([this, &gc_consumer, t] (clustering_row&& cr_garbage) {
|
|
partition_is_not_empty_for_gc_consumer(gc_consumer);
|
|
// We are passing only dead (purged) data so pass is_live=false.
|
|
gc_consumer.consume(std::move(cr_garbage), t, false);
|
|
});
|
|
} else {
|
|
if (can_purge_tombstone(t)) {
|
|
t = {};
|
|
}
|
|
}
|
|
|
|
if (!cr.empty()) {
|
|
partition_is_not_empty(consumer);
|
|
_stop = consumer.consume(std::move(cr), t, is_live);
|
|
}
|
|
if (!sstable_compaction() && is_live && ++_rows_in_current_partition == _current_partition_limit) {
|
|
_stop = stop_iteration::yes;
|
|
}
|
|
return _stop;
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
stop_iteration consume(range_tombstone_change&& rtc, Consumer& consumer, GCConsumer& gc_consumer) {
|
|
if (!sstable_compaction()) {
|
|
_last_pos = rtc.position();
|
|
}
|
|
++_stats.range_tombstones;
|
|
_stop = do_consume(std::move(rtc), consumer, gc_consumer);
|
|
return _stop;
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
stop_iteration consume_end_of_partition(Consumer& consumer, GCConsumer& gc_consumer) {
|
|
// Only check if the active tombstone has to be closed, if the partition
|
|
// was cut by the consumer. Otherwise, leave the stream as-is.
|
|
if (_stop) {
|
|
if (_effective_tombstone) {
|
|
auto rtc = range_tombstone_change(position_in_partition::after_key(_schema, _last_pos), tombstone{});
|
|
// do_consume() overwrites _effective_tombstone with {}, so save and restore it.
|
|
auto prev_tombstone = _effective_tombstone;
|
|
do_consume(std::move(rtc), consumer, gc_consumer);
|
|
_effective_tombstone = prev_tombstone;
|
|
} else if (_validator.validator().current_tombstone()) {
|
|
// It is possible that the range-tombstone providing the active
|
|
// tombstone was purged and never got to the consumer and therefore
|
|
// didn't set `_effective_tombstone`. In this case we generate a
|
|
// closing tombstone just for the validator.
|
|
_validator(mutation_fragment_v2::kind::range_tombstone_change, position_in_partition::after_key(_schema, _last_pos), tombstone{});
|
|
}
|
|
}
|
|
_validator.on_end_of_partition();
|
|
if (!_empty_partition_in_gc_consumer) {
|
|
gc_consumer.consume_end_of_partition();
|
|
}
|
|
if (!_empty_partition) {
|
|
// #589 - Do not add extra row for static content unless we did a CK range-less query.
|
|
// See comment in query
|
|
if (_rows_in_current_partition == 0 && _static_row_live &&
|
|
_return_static_content_on_partition_with_no_rows) {
|
|
++_rows_in_current_partition;
|
|
}
|
|
|
|
_row_limit -= _rows_in_current_partition;
|
|
_partition_limit -= _rows_in_current_partition > 0;
|
|
auto stop = consumer.consume_end_of_partition();
|
|
if (!sstable_compaction()) {
|
|
stop = _row_limit && _partition_limit && stop != stop_iteration::yes
|
|
? stop_iteration::no : stop_iteration::yes;
|
|
// If we decided to stop earlier but decide to continue now, we
|
|
// are in effect skipping the partition. Do not leave `_stop` at
|
|
// `stop_iteration::yes` in this case, reset it back to
|
|
// `stop_iteration::no` as if we exhausted the partition.
|
|
if (_stop && !stop) {
|
|
_stop = stop_iteration::no;
|
|
}
|
|
return stop;
|
|
}
|
|
}
|
|
return stop_iteration::no;
|
|
}
|
|
|
|
template <typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
auto consume_end_of_stream(Consumer& consumer, GCConsumer& gc_consumer) {
|
|
_validator.on_end_of_stream();
|
|
if constexpr (std::is_void_v<std::invoke_result_t<decltype(&GCConsumer::consume_end_of_stream), GCConsumer&>>) {
|
|
gc_consumer.consume_end_of_stream();
|
|
return consumer.consume_end_of_stream();
|
|
} else {
|
|
return std::pair(consumer.consume_end_of_stream(), gc_consumer.consume_end_of_stream());
|
|
}
|
|
}
|
|
|
|
/// The decorated key of the partition the compaction is positioned in.
|
|
/// Can be null if the compaction wasn't started yet.
|
|
const dht::decorated_key* current_partition() const {
|
|
return _dk ? &*_dk : nullptr;
|
|
}
|
|
|
|
// Only updated when SSTableCompaction == compact_for_sstables::no.
|
|
// Only meaningful if compaction has started already (current_partition() != nullptr).
|
|
position_in_partition_view current_position() const {
|
|
return _last_pos;
|
|
}
|
|
|
|
std::optional<full_position> current_full_position() const {
|
|
if (!_dk) {
|
|
return {};
|
|
}
|
|
return full_position(_dk->key(), _last_pos);
|
|
}
|
|
|
|
/// Reset limits and query-time to the new page's ones and re-emit the
|
|
/// partition-header and static row if there are clustering rows or range
|
|
/// tombstones left in the partition.
|
|
template <typename Consumer>
|
|
requires CompactedFragmentsConsumer<Consumer>
|
|
void start_new_page(uint64_t row_limit,
|
|
uint32_t partition_limit,
|
|
gc_clock::time_point query_time,
|
|
partition_region next_fragment_region,
|
|
Consumer& consumer) {
|
|
_empty_partition = true;
|
|
_static_row_live = false;
|
|
_row_limit = row_limit;
|
|
_partition_limit = partition_limit;
|
|
_rows_in_current_partition = 0;
|
|
_current_partition_limit = std::min(_row_limit, _partition_row_limit);
|
|
_query_time = query_time;
|
|
_stats = {};
|
|
_stop = stop_iteration::no;
|
|
|
|
noop_compacted_fragments_consumer nc;
|
|
|
|
if (next_fragment_region != partition_region::partition_start) {
|
|
_validator.reset(mutation_fragment_v2::kind::partition_start, position_in_partition_view::for_partition_start(), {});
|
|
}
|
|
if (next_fragment_region == partition_region::clustered && _last_static_row) {
|
|
// Stopping here would cause an infinite loop so ignore return value.
|
|
consume(*std::exchange(_last_static_row, {}), consumer, nc);
|
|
}
|
|
if (_effective_tombstone) {
|
|
auto rtc = range_tombstone_change(position_in_partition::after_key(_schema, _last_pos), _effective_tombstone);
|
|
do_consume(std::move(rtc), consumer, nc);
|
|
}
|
|
}
|
|
|
|
/// Signal to the compactor that the current partition will not be finished.
|
|
void abandon_current_partition() {
|
|
_validator.reset(mutation_fragment_v2::kind::partition_end, position_in_partition_view::for_partition_end(), {});
|
|
}
|
|
|
|
bool are_limits_reached() const {
|
|
return _row_limit == 0 || _partition_limit == 0;
|
|
}
|
|
|
|
/// Detach the internal state of the compactor
|
|
///
|
|
/// The state is represented by the last seen partition header, static row
|
|
/// and active range tombstones. Replaying these fragments through a new
|
|
/// compactor will result in the new compactor being in the same state *this
|
|
/// is (given the same outside parameters of course). Practically this
|
|
/// allows the compaction state to be stored in the compacted reader.
|
|
/// If the currently compacted partition is exhausted a disengaged optional
|
|
/// is returned -- in this case there is no state to detach.
|
|
std::optional<detached_compaction_state> detach_state() && {
|
|
// If we exhausted the partition, there is no need to detach-restore the
|
|
// compaction state.
|
|
// We exhausted the partition if `consume_partition_end()` was called
|
|
// without us requesting the consumption to stop (remembered in _stop)
|
|
// from one of the consume() overloads.
|
|
// The consume algorithm calls `consume_partition_end()` in two cases:
|
|
// * on a partition-end fragment
|
|
// * consume() requested to stop
|
|
// In the latter case, the partition is not exhausted. Even if the next
|
|
// fragment to process is a partition-end, it will not be consumed.
|
|
if (!_stop) {
|
|
return {};
|
|
}
|
|
partition_start ps(*std::exchange(_dk, std::nullopt), _partition_tombstone);
|
|
if (_effective_tombstone) {
|
|
return detached_compaction_state{std::move(ps), std::move(_last_static_row),
|
|
range_tombstone_change(position_in_partition::after_key(_schema, _last_pos), _effective_tombstone)};
|
|
} else {
|
|
return detached_compaction_state{std::move(ps), std::move(_last_static_row), std::optional<range_tombstone_change>{}};
|
|
}
|
|
}
|
|
|
|
const ::compaction_stats& stats() const { return _stats; }
|
|
};
|
|
|
|
template<compact_for_sstables SSTableCompaction, typename Consumer, typename GCConsumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
class compact_mutation {
|
|
lw_shared_ptr<compact_mutation_state<SSTableCompaction>> _state;
|
|
Consumer _consumer;
|
|
// Garbage Collected Consumer
|
|
GCConsumer _gc_consumer;
|
|
|
|
public:
|
|
// Can only be used for compact_for_sstables::no
|
|
compact_mutation(const schema& s, gc_clock::time_point query_time, const query::partition_slice& slice, uint64_t limit,
|
|
uint32_t partition_limit, const tombstone_gc_state& gc_state,
|
|
Consumer consumer, GCConsumer gc_consumer = GCConsumer())
|
|
: _state(make_lw_shared<compact_mutation_state<SSTableCompaction>>(s, query_time, slice, limit, partition_limit, gc_state))
|
|
, _consumer(std::move(consumer))
|
|
, _gc_consumer(std::move(gc_consumer)) {
|
|
}
|
|
|
|
// Can only be used for compact_for_sstables::yes
|
|
compact_mutation(const schema& s, gc_clock::time_point compaction_time,
|
|
max_purgeable_fn get_max_purgeable,
|
|
|
|
const tombstone_gc_state& gc_state,
|
|
Consumer consumer, GCConsumer gc_consumer = GCConsumer(), tombstone_purge_stats* tombstone_stats = nullptr)
|
|
: _state(make_lw_shared<compact_mutation_state<SSTableCompaction>>(s, compaction_time, get_max_purgeable, gc_state, tombstone_stats))
|
|
, _consumer(std::move(consumer))
|
|
, _gc_consumer(std::move(gc_consumer)) {
|
|
}
|
|
|
|
compact_mutation(lw_shared_ptr<compact_mutation_state<SSTableCompaction>> state, Consumer consumer,
|
|
GCConsumer gc_consumer = GCConsumer())
|
|
: _state(std::move(state))
|
|
, _consumer(std::move(consumer))
|
|
, _gc_consumer(std::move(gc_consumer)) {
|
|
}
|
|
|
|
void consume_new_partition(const dht::decorated_key& dk) {
|
|
_state->consume_new_partition(dk);
|
|
}
|
|
|
|
void consume(tombstone t) {
|
|
_state->consume(std::move(t), _consumer, _gc_consumer);
|
|
}
|
|
|
|
stop_iteration consume(static_row&& sr) {
|
|
return _state->consume(std::move(sr), _consumer, _gc_consumer);
|
|
}
|
|
|
|
stop_iteration consume(clustering_row&& cr) {
|
|
return _state->consume(std::move(cr), _consumer, _gc_consumer);
|
|
}
|
|
|
|
stop_iteration consume(range_tombstone_change&& rtc) {
|
|
return _state->consume(std::move(rtc), _consumer, _gc_consumer);
|
|
}
|
|
|
|
stop_iteration consume_end_of_partition() {
|
|
return _state->consume_end_of_partition(_consumer, _gc_consumer);
|
|
}
|
|
|
|
auto consume_end_of_stream() {
|
|
return _state->consume_end_of_stream(_consumer, _gc_consumer);
|
|
}
|
|
|
|
lw_shared_ptr<compact_mutation_state<SSTableCompaction>> get_state() {
|
|
return _state;
|
|
}
|
|
};
|
|
|
|
template<typename Consumer>
|
|
requires CompactedFragmentsConsumer<Consumer>
|
|
struct compact_for_query : compact_mutation<compact_for_sstables::no, Consumer, noop_compacted_fragments_consumer> {
|
|
using compact_mutation<compact_for_sstables::no, Consumer, noop_compacted_fragments_consumer>::compact_mutation;
|
|
};
|
|
|
|
using compact_for_query_state = compact_mutation_state<compact_for_sstables::no>;
|
|
|
|
template<typename Consumer, typename GCConsumer = noop_compacted_fragments_consumer>
|
|
requires CompactedFragmentsConsumer<Consumer> && CompactedFragmentsConsumer<GCConsumer>
|
|
struct compact_for_compaction : compact_mutation<compact_for_sstables::yes, Consumer, GCConsumer> {
|
|
using compact_mutation<compact_for_sstables::yes, Consumer, GCConsumer>::compact_mutation;
|
|
};
|